*This notebook contains an excerpt from the Python Data Science Handbook by Jake VanderPlas; the content is available on GitHub.*

*The text is released under the CC-BY-NC-ND license, and code is released under the MIT license. If you find this content useful, please consider supporting the work by buying the book!*

Matplotlib has proven to be an incredibly useful and popular visualization tool, but even avid users will admit it often leaves much to be desired. There are several valid complaints about Matplotlib that often come up:

- Prior to version 2.0, Matplotlib's defaults are not exactly the best choices. It was based off of MATLAB circa 1999, and this often shows.
- Matplotlib's API is relatively low level. Doing sophisticated statistical visualization is possible, but often requires a
*lot*of boilerplate code. - Matplotlib predated Pandas by more than a decade, and thus is not designed for use with Pandas
`DataFrame`

s. In order to visualize data from a Pandas`DataFrame`

, you must extract each`Series`

and often concatenate them together into the right format. It would be nicer to have a plotting library that can intelligently use the`DataFrame`

labels in a plot.

An answer to these problems is Seaborn. Seaborn provides an API on top of Matplotlib that offers sane choices for plot style and color defaults, defines simple high-level functions for common statistical plot types, and integrates with the functionality provided by Pandas `DataFrame`

s.

To be fair, the Matplotlib team is addressing this: it has recently added the `plt.style`

tools discussed in Customizing Matplotlib: Configurations and Style Sheets, and is starting to handle Pandas data more seamlessly.
The 2.0 release of the library will include a new default stylesheet that will improve on the current status quo.
But for all the reasons just discussed, Seaborn remains an extremely useful addon.

Here is an example of a simple random-walk plot in Matplotlib, using its classic plot formatting and colors. We start with the typical imports:

In [1]:

```
import matplotlib.pyplot as plt
plt.style.use('classic')
%matplotlib inline
import numpy as np
import pandas as pd
```

Now we create some random walk data:

In [2]:

```
# Create some data
rng = np.random.RandomState(0)
x = np.linspace(0, 10, 500)
y = np.cumsum(rng.randn(500, 6), 0)
```

And do a simple plot:

In [3]:

```
# Plot the data with Matplotlib defaults
plt.plot(x, y)
plt.legend('ABCDEF', ncol=2, loc='upper left');
```

Although the result contains all the information we'd like it to convey, it does so in a way that is not all that aesthetically pleasing, and even looks a bit old-fashioned in the context of 21st-century data visualization.

Now let's take a look at how it works with Seaborn.
As we will see, Seaborn has many of its own high-level plotting routines, but it can also overwrite Matplotlib's default parameters and in turn get even simple Matplotlib scripts to produce vastly superior output.
We can set the style by calling Seaborn's `set()`

method.
By convention, Seaborn is imported as `sns`

:

In [4]:

```
import seaborn as sns
sns.set()
```

Now let's rerun the same two lines as before:

In [5]:

```
# same plotting code as above!
plt.plot(x, y)
plt.legend('ABCDEF', ncol=2, loc='upper left');
```

Ah, much better!

The main idea of Seaborn is that it provides high-level commands to create a variety of plot types useful for statistical data exploration, and even some statistical model fitting.

Let's take a look at a few of the datasets and plot types available in Seaborn. Note that all of the following *could* be done using raw Matplotlib commands (this is, in fact, what Seaborn does under the hood) but the Seaborn API is much more convenient.

Often in statistical data visualization, all you want is to plot histograms and joint distributions of variables. We have seen that this is relatively straightforward in Matplotlib:

In [6]:

```
data = np.random.multivariate_normal([0, 0], [[5, 2], [2, 2]], size=2000)
data = pd.DataFrame(data, columns=['x', 'y'])
for col in 'xy':
plt.hist(data[col], normed=True, alpha=0.5)
```

`sns.kdeplot`

:

In [7]:

```
for col in 'xy':
sns.kdeplot(data[col], shade=True)
```

Histograms and KDE can be combined using `distplot`

:

In [8]:

```
sns.distplot(data['x'])
sns.distplot(data['y']);
```

`kdeplot`

, we will get a two-dimensional visualization of the data:

In [9]:

```
sns.kdeplot(data);
```

`sns.jointplot`

.
For this plot, we'll set the style to a white background:

In [10]:

```
with sns.axes_style('white'):
sns.jointplot("x", "y", data, kind='kde');
```

`jointplot`

â€”for example, we can use a hexagonally based histogram instead:

In [11]:

```
with sns.axes_style('white'):
sns.jointplot("x", "y", data, kind='hex')
```

When you generalize joint plots to datasets of larger dimensions, you end up with *pair plots*. This is very useful for exploring correlations between multidimensional data, when you'd like to plot all pairs of values against each other.

We'll demo this with the well-known Iris dataset, which lists measurements of petals and sepals of three iris species:

In [12]:

```
iris = sns.load_dataset("iris")
iris.head()
```

Out[12]:

`sns.pairplot`

:

In [13]:

```
sns.pairplot(iris, hue='species', size=2.5);
```

Sometimes the best way to view data is via histograms of subsets. Seaborn's `FacetGrid`

makes this extremely simple.
We'll take a look at some data that shows the amount that restaurant staff receive in tips based on various indicator data:

In [14]:

```
tips = sns.load_dataset('tips')
tips.head()
```

Out[14]:

In [15]:

```
tips['tip_pct'] = 100 * tips['tip'] / tips['total_bill']
grid = sns.FacetGrid(tips, row="sex", col="time", margin_titles=True)
grid.map(plt.hist, "tip_pct", bins=np.linspace(0, 40, 15));
```

Factor plots can be useful for this kind of visualization as well. This allows you to view the distribution of a parameter within bins defined by any other parameter:

In [16]:

```
with sns.axes_style(style='ticks'):
g = sns.factorplot("day", "total_bill", "sex", data=tips, kind="box")
g.set_axis_labels("Day", "Total Bill");
```

Similar to the pairplot we saw earlier, we can use `sns.jointplot`

to show the joint distribution between different datasets, along with the associated marginal distributions:

In [17]:

```
with sns.axes_style('white'):
sns.jointplot("total_bill", "tip", data=tips, kind='hex')
```

The joint plot can even do some automatic kernel density estimation and regression:

In [18]:

```
sns.jointplot("total_bill", "tip", data=tips, kind='reg');
```

Time series can be plotted using `sns.factorplot`

. In the following example, we'll use the Planets data that we first saw in Aggregation and Grouping:

In [19]:

```
planets = sns.load_dataset('planets')
planets.head()
```

Out[19]:

In [20]:

```
with sns.axes_style('white'):
g = sns.factorplot("year", data=planets, aspect=2,
kind="count", color='steelblue')
g.set_xticklabels(step=5)
```

We can learn more by looking at the *method* of discovery of each of these planets:

In [21]:

```
with sns.axes_style('white'):
g = sns.factorplot("year", data=planets, aspect=4.0, kind='count',
hue='method', order=range(2001, 2015))
g.set_ylabels('Number of Planets Discovered')
```

Here we'll look at using Seaborn to help visualize and understand finishing results from a marathon.
I've scraped the data from sources on the Web, aggregated it and removed any identifying information, and put it on GitHub where it can be downloaded
(if you are interested in using Python for web scraping, I would recommend *Web Scraping with Python* by Ryan Mitchell).
We will start by downloading the data from
the Web, and loading it into Pandas:

In [22]:

```
# !curl -O https://raw.githubusercontent.com/jakevdp/marathon-data/master/marathon-data.csv
```

In [23]:

```
data = pd.read_csv('marathon-data.csv')
data.head()
```

Out[23]:

`object`

); we can see this by looking at the `dtypes`

attribute of the DataFrame:

In [24]:

```
data.dtypes
```

Out[24]:

Let's fix this by providing a converter for the times:

In [25]:

```
def convert_time(s):
h, m, s = map(int, s.split(':'))
return pd.datetools.timedelta(hours=h, minutes=m, seconds=s)
data = pd.read_csv('marathon-data.csv',
converters={'split':convert_time, 'final':convert_time})
data.head()
```

Out[25]:

In [26]:

```
data.dtypes
```

Out[26]:

In [27]:

```
data['split_sec'] = data['split'].astype(int) / 1E9
data['final_sec'] = data['final'].astype(int) / 1E9
data.head()
```

Out[27]:

To get an idea of what the data looks like, we can plot a `jointplot`

over the data:

In [28]:

```
with sns.axes_style('white'):
g = sns.jointplot("split_sec", "final_sec", data, kind='hex')
g.ax_joint.plot(np.linspace(4000, 16000),
np.linspace(8000, 32000), ':k')
```

The dotted line shows where someone's time would lie if they ran the marathon at a perfectly steady pace. The fact that the distribution lies above this indicates (as you might expect) that most people slow down over the course of the marathon. If you have run competitively, you'll know that those who do the oppositeâ€”run faster during the second half of the raceâ€”are said to have "negative-split" the race.

Let's create another column in the data, the split fraction, which measures the degree to which each runner negative-splits or positive-splits the race:

In [29]:

```
data['split_frac'] = 1 - 2 * data['split_sec'] / data['final_sec']
data.head()
```

Out[29]:

In [30]:

```
sns.distplot(data['split_frac'], kde=False);
plt.axvline(0, color="k", linestyle="--");
```

In [31]:

```
sum(data.split_frac < 0)
```

Out[31]:

Out of nearly 40,000 participants, there were only 250 people who negative-split their marathon.

Let's see whether there is any correlation between this split fraction and other variables. We'll do this using a `pairgrid`

, which draws plots of all these correlations:

In [32]:

```
g = sns.PairGrid(data, vars=['age', 'split_sec', 'final_sec', 'split_frac'],
hue='gender', palette='RdBu_r')
g.map(plt.scatter, alpha=0.8)
g.add_legend();
```

It looks like the split fraction does not correlate particularly with age, but does correlate with the final time: faster runners tend to have closer to even splits on their marathon time. (We see here that Seaborn is no panacea for Matplotlib's ills when it comes to plot styles: in particular, the x-axis labels overlap. Because the output is a simple Matplotlib plot, however, the methods in Customizing Ticks can be used to adjust such things if desired.)

The difference between men and women here is interesting. Let's look at the histogram of split fractions for these two groups:

In [33]:

```
sns.kdeplot(data.split_frac[data.gender=='M'], label='men', shade=True)
sns.kdeplot(data.split_frac[data.gender=='W'], label='women', shade=True)
plt.xlabel('split_frac');
```

The interesting thing here is that there are many more men than women who are running close to an even split! This almost looks like some kind of bimodal distribution among the men and women. Let's see if we can suss-out what's going on by looking at the distributions as a function of age.

A nice way to compare distributions is to use a *violin plot*

In [34]:

```
sns.violinplot("gender", "split_frac", data=data,
palette=["lightblue", "lightpink"]);
```

This is yet another way to compare the distributions between men and women.

Let's look a little deeper, and compare these violin plots as a function of age. We'll start by creating a new column in the array that specifies the decade of age that each person is in:

In [35]:

```
data['age_dec'] = data.age.map(lambda age: 10 * (age // 10))
data.head()
```

Out[35]:

In [36]:

```
men = (data.gender == 'M')
women = (data.gender == 'W')
with sns.axes_style(style=None):
sns.violinplot("age_dec", "split_frac", hue="gender", data=data,
split=True, inner="quartile",
palette=["lightblue", "lightpink"]);
```

Looking at this, we can see where the distributions of men and women differ: the split distributions of men in their 20s to 50s show a pronounced over-density toward lower splits when compared to women of the same age (or of any age, for that matter).

Also surprisingly, the 80-year-old women seem to outperform *everyone* in terms of their split time. This is probably due to the fact that we're estimating the distribution from small numbers, as there are only a handful of runners in that range:

In [38]:

```
(data.age > 80).sum()
```

Out[38]:

`regplot`

, which will automatically fit a linear regression to the data:

In [37]:

```
g = sns.lmplot('final_sec', 'split_frac', col='gender', data=data,
markers=".", scatter_kws=dict(color='c'))
g.map(plt.axhline, y=0.1, color="k", ls=":");
```