The adult census dataset

The adult census dataset#

This dataset is a collection of demographic information for the adult population as of 1994 in the USA. The prediction task is to predict whether a person is earning a high or low revenue in USD/year.

The column named class is the target variable (i.e., the variable which we want to predict). The two possible classes are " <=50K" (low-revenue) and " >50K" (high-revenue).

Before drawing any conclusions based on its statistics or the predictions of models trained on it, remember that this dataset is not only outdated, but is also not representative of the US population. In fact, the original data contains a feature named fnlwgt that encodes the number of units in the target population that the responding unit represents.

First we load the dataset. We keep only some columns of interest to ease the plotting.

import pandas as pd

adult_census = pd.read_csv("../datasets/adult-census.csv")
columns_to_plot = [
    "age",
    "education-num",
    "capital-loss",
    "capital-gain",
    "hours-per-week",
    "relationship",
    "class",
]
target_name = "class"
target = adult_census[target_name]

We explore this dataset in the first module’s notebook β€œFirst look at our dataset”, where we provide a first intuition on how the data is structured. There, we use a seaborn pairplot to visualize pairwise relationships between the numerical variables in the dataset. This tool aligns scatter plots for every pair of variables and histograms for the plots in the diagonal of the array.

This approach is limited:

  • Pair plots can only deal with numerical features and;

  • by observing pairwise interactions we end up with a two-dimensional projection of a multi-dimensional feature space, which can lead to a wrong interpretation of the individual impact of a feature.

Here we explore with some more detail the relation between features using plotly Parcoords.

import plotly.graph_objects as go
from sklearn.preprocessing import LabelEncoder

le = LabelEncoder()


def generate_dict(col):
    """Check if column is categorical and generate the appropriate dict"""
    if adult_census[col].dtype == "object":  # Categorical column
        encoded = le.fit_transform(adult_census[col])
        return {
            "tickvals": list(range(len(le.classes_))),
            "ticktext": list(le.classes_),
            "label": col,
            "values": encoded,
        }
    else:  # Numerical column
        return {"label": col, "values": adult_census[col]}


plot_list = [generate_dict(col) for col in columns_to_plot]

fig = go.Figure(
    data=go.Parcoords(
        line=dict(
            color=le.fit_transform(target),
            colorscale="Viridis",
        ),
        dimensions=plot_list,
    )
)
fig.show()

The Parcoords plot is quite similar to the parallel coordinates plot that we present in the module on hyperparameters tuning in this mooc. It display the values of the features on different columns while the target class is color coded. Thus, we are able to quickly inspect if there is a range of values for a certain feature which is leading to a particular result.

As in the parallel coordinates plot, it is possible to select one or more ranges of values by clicking and holding on any axis of the plot. You can then slide (move) the range selection and cross two selections to see the intersections. You can undo a selection by clicking once again on the same axis.

In particular for this dataset we observe that values of "age" lower to 20 years are quite predictive of low-income, regardless of the value of other features. Similarly, a "capital-loss" above 4000 seems to lead to low-income.

Even if it is beyond the scope of the present MOOC, one can additionally explore correlations between features, for example, using Spearman’s rank correlation, as the more popular Pearson’s correlation is only appropriate for continuous data that is normally distributed and linearly related. Spearman’s correlation is more versatile in dealing with non-linear relationships and ordinal data, but it is not meant for nominal categorical data.

import matplotlib.pyplot as plt
import numpy as np

from scipy.cluster import hierarchy
from scipy.spatial.distance import squareform
from scipy.stats import spearmanr

# Keep numerical features only
X = adult_census[columns_to_plot].drop(columns=["class", "relationship"])
fig, (ax1, ax2) = plt.subplots(1, 2, figsize=(12, 8))
corr = spearmanr(X).correlation

# Ensure the correlation matrix is symmetric
corr = (corr + corr.T) / 2
np.fill_diagonal(corr, 1)

# We convert the correlation matrix to a distance matrix before performing
# hierarchical clustering using Ward's linkage.
distance_matrix = 1 - np.abs(corr)
dist_linkage = hierarchy.ward(squareform(distance_matrix))
dendro = hierarchy.dendrogram(
    dist_linkage, labels=X.columns.to_list(), ax=ax1, leaf_rotation=90
)
dendro_idx = np.arange(0, len(dendro["ivl"]))

ax2.imshow(corr[dendro["leaves"], :][:, dendro["leaves"]], cmap="coolwarm")
ax2.set_xticks(dendro_idx)
ax2.set_yticks(dendro_idx)
ax2.set_xticklabels(dendro["ivl"], rotation="vertical")
ax2.set_yticklabels(dendro["ivl"])
_ = fig.tight_layout()
../_images/24f1f8568236d75d7e091a00be66061d82c50714ef375692a9d8ba6ff3c43743.png

Using a diverging colormap such as β€œcoolwarm”, the softer the color, the less (anti)correlation between features (no correlation is mapped to white color). In this case dark blue represents strong negative correlations and dark red means strong positive correlations. Indeed, any feature is perfectly correlated to itself.