# Pearson's Correlation

- Mar 5 • 16 min read
- Key Terms: correlation, scatter plot, python, pandas

In statistics, the **Pearson correlation coefficient** is a measure of the *linear* relationship between two continuous numerical variables of data.

In business context, the term *correlation* is used to to describe a relationship between two or more variables. However, this article will just cover the Pearson's correlation coefficient. This tutorial will illustrate the concept through simple examples and many visualizations.

### Intro to Pearson Correlation Coefficient

The greek letter \(\rho\) stands for the **Pearson's r**, otherwise known as the **correlation coefficient** (\(\rho\)).

The equation for `r`

is below. The numerator is the covariance of `x`

and `y`

- essentially how much they vary together. The denominator is the standard deviation of `x`

multipled by the standard deviation of `y`

which explains how the two variables vary apart from each other rather than with each other.

\(\rho\) is a proportion. \(\rho\) measures the strength of a relationship by looking at how closely the observations fall along a straight line. If the data points resemble a perfectly straight line with a positive slope, then \(\rho\) is \(1\). If the data resembles a perfectly negatively sloped line, then then \(\rho\) is \(-1\). If the data points are randomly scattered with no resemblence of a perfect line, then \(\rho\) is close to \(0\). In this instace, there's no linear relationship between the two variables.

Any \(\rho\) value closer to the absolute value of \(1\) indicates a *strong* linear relationship between the two variables.

### Import Modules

```
import numpy as np
from scipy import stats
import pandas as pd
import seaborn as sns
import matplotlib.pyplot as plt
% matplotlib inline
```

```
sns.set_context('talk')
sns.set_style("darkgrid")
```

### Example to Illustrate Strength of Pearson's \(\rho\)

Below are two visualizations, labeled *Graph 1* and *Graph 2*. Each shows a scatter plot of the relation of `x`

and `y`

values.

I'm curious - which two variables have a stronger *positive* linear relationship - *Graph 1* or *Graph 2*?

```
x = [-5, -4, -3, -2, -1, 0, 1, 2, 3, 4, 5]
y = [-3.2, -2.2, -5.5, -3.5, -2, -1, 2.3, 1.2, 4.5, 3.1, 5.5]
plt.figure(figsize=(8, 6))
plt.scatter(x, y, color='red')
plt.xlabel("x-values", labelpad=13)
plt.ylabel("y-values", labelpad=13)
plt.title("Graph 1", y=1.015);
```

```
y2 = [-5, -4.1, -3, -2, -1.1, 0, 1.3, 2.2, 3, 4.1, 5]
plt.figure(figsize=(8, 6))
plt.scatter(x, y2, color='dodgerblue')
plt.xlabel("x-values", labelpad=13)
plt.ylabel("y-values", labelpad=13)
plt.title("Graph 2", y=1.015);
```

The scatter points in *Graph 2* seem to have a stronger positive linear relationship because the scatter points look to resemble a near-perfect line with a positive slope of \(1\). In fact, we can figure out the exact Pearson's r of each graph and *Graph 1* indicates a larger value of \(0.999\) while *Graph 2* indicates \(0.909\).

```
correlation_x_y, p_value = stats.pearsonr(x, y)
correlation_x_y
```

```
0.9087195765173497
```

```
correlation_x_y2, p_value2 = stats.pearsonr(x, y2)
correlation_x_y2
```

```
0.9994608748837965
```

### Explanation of Coefficient Values

If the Pearson's r is *positive*, as values from one variable increase, so does the other.

If the coefficient is *negative*, as values from one variable decrease, the values from the other variable increase.

The table below provides some examples of coefficient values and explanations.

Pearson's correlation coefficient | Interpretation |
---|---|

0 | There's no correlation between the two variables |

-0.25 | There's a small negative correlation between the two variables |

-0.75 | There's a large negative correlation between the two variables |

-1.0 | There's a perfect negative correlation between the two variables |

0.25 | There's a small positive correlation between the two variables |

0.75 | There's a large positive correlation between the two variables |

1.0 | There's a perfect positive correlation between the two variables |

### Assumption to Prove Correlation

Generally, for both variables compared, we want them to both be approximately normal. The reason for this is that it's unfair to interpret a correlation pattern if we're just looking at extremes of values.

For example, let's say we want to understand the relationship between income and net worth. We'd initially *assume* there's a positive linear relationship between the two variables: as income increases, net worth increases. However, hypothetically, let's say our dataset just had net income values around 0-20,00 U.S. dollars and 150,000-500,000 U.S. dollars. In this instance, our income values wouldn't represent a normal distribution. We're missing lots of income values between 20,000 and 150,000 U.S. dollars - which I think is a fairly common range for U.S. workers.

We could still compare income and net worth and we'd likely see the people in that small income range have low net worth values while the people in that much higher net income range have larger net worth values. Yet, it's unfair to interpret correlation between these two variables since we're neglecting a huge portion of people with a certain income and it's biased to connect a relationship between *just* the extremes of income data.

Let's move onto some visual examples now to explain correlation.

### Example 1: Tips Dataset - Total Bill and Tip

#### Get the Data

Let's get the `tips`

dataset from the `seaborn`

library and assign it to the DataFrame `df_tips`

.

```
df_tips = sns.load_dataset('tips')
```

Each row represents a unique meal at a restaurant for a party of people; the dataset contains the following fields:

column name | column description |
---|---|

`total_bill` |
financial amount of meal in U.S. dollars |

`tip` |
financial amount of the meal's tip in U.S. dollars |

`sex` |
gender of server |

`smoker` |
boolean to represent if server smokes or not |

`day` |
day of week |

`time` |
meal name (Lunch or Dinner) |

`size` |
count of people eating meal |

Preview the first 5 rows of `df_tips`

.

```
df_tips.head()
```

total_bill | tip | sex | smoker | day | time | size | |
---|---|---|---|---|---|---|---|

0 | 16.99 | 1.01 | Female | No | Sun | Dinner | 2 |

1 | 10.34 | 1.66 | Male | No | Sun | Dinner | 3 |

2 | 21.01 | 3.50 | Male | No | Sun | Dinner | 3 |

3 | 23.68 | 3.31 | Male | No | Sun | Dinner | 2 |

4 | 24.59 | 3.61 | Female | No | Sun | Dinner | 4 |

#### Plot the Relationship Between Total Bill and Tip Amounts

The plot below has total bill values on the x-axis and tips on the y-axis.

```
df_tips.plot.scatter(x='total_bill', y='tip', s=28, alpha=0.6, c='darkgreen', figsize=(9, 7))
plt.xlabel("total bill [$]", labelpad=15)
plt.ylabel("tip [$]", labelpad=15)
plt.title("Total Bill vs. Tips", y=1.015);
```

There seems to be a large positive correlation between the two variables. As total bill sizes increase, tip values increase too. This is what we'd expect since tips given at restaurants are typically 20% of total bill size, so there's mutual dependence between these two variables.

#### Evaluate the Pearson's Correlation Coefficient Between Tips and Bill Amounts

We can use the `scipy`

packages's `stats`

module and `pearsonr()`

method to compute the Pearson correlation coefficient.

This method returns a list in which the first value is the correlation coefficient.

Our returned value is 0.68 which confirms our belief that there's a large positive correlation between the two variables.

```
round(stats.pearsonr(df_tips['tip'], df_tips['total_bill'])[0], 2)
```

```
0.68
```

We can also the use DataFrame method `corr()`

from the pandas package to compute a pairwise correlation of columns.

There'll always be a perfect positive correlation, a coefficient of 1, for each variable with itself. The 0.68 value below is the same as the one from our calculation with scipy above.

```
df_tips[['tip', 'total_bill']].corr()
```

tip | total_bill | |
---|---|---|

tip | 1.000000 | 0.675734 |

total_bill | 0.675734 | 1.000000 |

### Example 2: Diamonds Dataset - Carat and Price

#### Get Diamonds Dataset

Import `diamonds`

dataset from Seaborn library and assign to DataFrame `df_diamonds`

.

Each row of `df_diamonds`

contains details about a specific diamond purchased. We'll just utilize the `price`

and `carat`

columns below.

```
df_diamonds = sns.load_dataset('diamonds')
```

Preview the first few rows of `df_diamonds`

.

```
df_diamonds.head()
```

carat | cut | color | clarity | depth | table | price | x | y | z | |
---|---|---|---|---|---|---|---|---|---|---|

0 | 0.23 | Ideal | E | SI2 | 61.5 | 55.0 | 326 | 3.95 | 3.98 | 2.43 |

1 | 0.21 | Premium | E | SI1 | 59.8 | 61.0 | 326 | 3.89 | 3.84 | 2.31 |

2 | 0.23 | Good | E | VS1 | 56.9 | 65.0 | 327 | 4.05 | 4.07 | 2.31 |

3 | 0.29 | Premium | I | VS2 | 62.4 | 58.0 | 334 | 4.20 | 4.23 | 2.63 |

4 | 0.31 | Good | J | SI2 | 63.3 | 58.0 | 335 | 4.34 | 4.35 | 2.75 |

#### Explanation of carats

`carat`

is a unit of measurement in diamonds to characterize its mass. Typically, bigger/heavier diamonds are known to be more expensive.

Here's a helpful image from womengettingmaried.com.

#### View Distribution of Carats versus Diamond Prices

The plot below has carats on the x-axis and price on the y-axis.

```
df_diamonds.plot.scatter(x='carat', y='price', s=22, alpha=0.25, c='plum', figsize=(11, 9))
plt.xlabel("carats", labelpad=15)
plt.ylabel("price [$]", labelpad=15)
plt.title("Carats versus Price of Diamonds", y=1.015);
```

There seems to be a large positive correlation between the two variables. As carats increases, price increases too. This is expected since we assume larger/heavier diamonds are priced higher than smaller ones.

#### Evaluate the Pearson's Correlation Coefficient Between Tips and Bill Amounts

```
round(stats.pearsonr(df_diamonds['carat'], df_diamonds['price'])[0], 2)
```

```
0.92
```

Our returned value is 0.92 means there's a very strong (nearly perfect) correlation between the two variables.

### Example 3: Titanic Dataset - Age and Fare Price

#### Get Titanic Dataset

Import `titanic`

dataset from Seaborn library and assign to DataFrame `df_titanic`

.

Each row of `df_titanic`

represents details of a passenger of the ship. We'll just utilize the `age`

and `fare`

columns below. The `fare`

column represents the price a passenger paid to board the ship.

```
df_titanic = sns.load_dataset('titanic')
```

Preview the first few rows of `df_titanic`

.

```
df_titanic.head()
```

survived | pclass | sex | age | sibsp | parch | fare | embarked | class | who | adult_male | deck | embark_town | alive | alone | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|

0 | 0 | 3 | male | 22.0 | 1 | 0 | 7.2500 | S | Third | man | True | NaN | Southampton | no | False |

1 | 1 | 1 | female | 38.0 | 1 | 0 | 71.2833 | C | First | woman | False | C | Cherbourg | yes | False |

2 | 1 | 3 | female | 26.0 | 0 | 0 | 7.9250 | S | Third | woman | False | NaN | Southampton | yes | True |

3 | 1 | 1 | female | 35.0 | 1 | 0 | 53.1000 | S | First | woman | False | C | Southampton | yes | False |

4 | 0 | 3 | male | 35.0 | 0 | 0 | 8.0500 | S | Third | man | True | NaN | Southampton | no | True |

Let's create a new DataFrame assigned to the variable `df_titanic_non_nulls`

that filters out the rows in the `age`

column that were nulls.

```
df_titanic_non_nulls = df_titanic[df_titanic['age'].notnull()==True]
```

#### View Distribution of Age vs. Fare Prices

The plot below has age in years on the x-axis and price in U.S. dollars on the y-axis.

```
df_titanic_non_nulls.plot.scatter(x='age', y='fare', s=22, alpha=0.4, c='darkcyan', figsize=(9, 7))
plt.xlabel("age [years]", labelpad=15)
plt.ylabel("fare price [$]", labelpad=15)
plt.title("Titanic Passengers Age [Years] vs. Fare Price [$]", y=1.015);
```

There seems to be no clear correlation between the two variables. The visualization just illustrates just a bunch of random dots.

#### Evaluate the Pearson's Correlation Coefficient Between Age and Fare Price

```
round(stats.pearsonr(df_titanic_non_nulls['age'], df_titanic_non_nulls['fare'])[0], 1)
```

```
0.1
```

Our returned value is 0.1 means there's essentially no correlation between the two variables. A change in the value of one variable has no expected change in the value of the other variable.

### Correlation Does Not Imply Causation

One of the most popular terms in the field of statistics is "correlation does not imply causation". A correlation means there's a linear relationship between two variables. A causation connects one process, the *cause*, with another process or state, the *effect*.

However, this does not mean a change in one variable *causes* a change in the value of the other variable.

For example, there are studies that have likely proved that increased exercise correlates with a decrease in a person's mass. This is considered to be a strong *negative* correlation between the two variables. This relationship has also likely been proven to be cause-and-effect in which once the activity of exercising takes place, this *causes* the decrease in a person's mass.

On the other hand, here's a different example mentioned in this Wikipedia article: as ice cream sales increase, the rate of drowning deaths increases sharply. These two variables can be correlated. So as ice cream sales have increased over time, so have drowning deaths so there's a positive linear relationship between the two variables. Yet, it's incredibly unlikely there there's *any causuality relationship* between those two variables since they're so distinct from one another. Rather, people eat more ice cream in warm-weather months and also likely swim more in these months. If more people are swimming, there's likely a greater chance of drowning deaths. The fact is: ice cream consumption *does not cause* drowning. There's lots of other factors to consider in this relationship such as weather.

```
```