Datasets#

Raw data often requires significant processing before it can be used effectively in machine learning models.

Data processing#

The most common stages of data processing are:

  • data collection

  • data cleaning

    • handling missing values

    • remove duplicates

    • outlier detection

    • data type conversion and standardization

  • data exploration and visualization

  • feature engineering

    • encoding categorical variables (e.g., One-hot encoding)

    • converting continuous variables into categorical bins (binning/discretization)

    • \(\log\) transformations

  • data annotation (if needed)

  • data augmentation (optional)

The result of these manipulations is what is usually called a dataset: a specific collection of data that is organized and structured in a way that makes it suitable for analysis, processing, or machine learning tasks.

Note

Most examples of datasets in this book have been already carefully prepared and usually do not require much preprocessing. In real life, however, the data preprocessing stage is typically the most crucial and time-consuming stage of ML pipeline.

Feature matrix#

A tabular numerical dataset can be represented as a feature matrix (or design matrix) \(\boldsymbol X\) of shape \(n\times d\) where

  • \(n\) — number of samples (rows)

  • \(d\) — number of features (columns)

Each sample \(\boldsymbol x_i\) is therefore represented by \(i\)-th row of the feature matrix \(\boldsymbol X\):

\[\begin{split} \boldsymbol X = \begin{pmatrix} \boldsymbol x_1^{\mathsf T} \\ \boldsymbol x_2^{\mathsf T} \\ \vdots \\ \boldsymbol x_n^{\mathsf T} \end{pmatrix} = \begin{pmatrix} x_{11} & x_{12} & \cdots & x_{1d} \\ x_{21} & x_{22} & \cdots & x_{2d}\\ \vdots & \vdots & \ddots & \vdots \\ x_{n1} & x_{n2} & \cdots & x_{nd} \end{pmatrix} \end{split}\]

(here \(x_{ij}\) is \(j\)-th feature of \(i\)-th object).

Important

A sample \(\boldsymbol x_i\) is a row vector with \(d\) coordinates. However, in linear algebra a vector is by default a column vector. That’s why in vector-matrix operations a training sample is often denoted as \(\boldsymbol x_i^{\mathsf T}\) to emphasize that it is a row.

Example: Iris dataset#

from sklearn.datasets import load_iris
iris_data = load_iris(as_frame=True)
iris_data['data']
sepal length (cm) sepal width (cm) petal length (cm) petal width (cm)
0 5.1 3.5 1.4 0.2
1 4.9 3.0 1.4 0.2
2 4.7 3.2 1.3 0.2
3 4.6 3.1 1.5 0.2
4 5.0 3.6 1.4 0.2
... ... ... ... ...
145 6.7 3.0 5.2 2.3
146 6.3 2.5 5.0 1.9
147 6.5 3.0 5.2 2.0
148 6.2 3.4 5.4 2.3
149 5.9 3.0 5.1 1.8

150 rows × 4 columns

In this dataset there are \(n=150\) samples and \(d=4\) features; each flower is represented by a \(4\)-dimensional vector, e.g.

\[ \boldsymbol x_1^{\mathsf T} = (5.1, 3.5, 1.4, 0.3). \]

Sources of datasets#

sklearn.datasets#

Scikit-learn provides access to several built-in datasets. This is how Iris dataset was obtained. sklearn.datasets also allows to create synthetic datasets:

from sklearn.datasets import make_moons
import matplotlib.pyplot as plt
%config InlineBackend.figure_formats = ['svg']

X, y = make_moons(n_samples=500, noise=0.2, random_state=42)
colors = ['r','b']
plt.scatter(X[:, 0], X[:,1], c=y)
plt.title("Moons");
../_images/45fdc2aedb6a8efcf1c7ec652a36bd90f983550ce507bb8f24e96fa6ee6d14f4.svg

Another example - classical dataset of handwritten digits:

from sklearn.datasets import fetch_openml

X, Y = fetch_openml('mnist_784', return_X_y=True, parser='auto')
X.shape, Y.shape
((70000, 784), (70000,))
plot_digits(X, Y, random_state=12)
../_images/6d6cc4d8cab8aa9327ff5b9e3a10a2e84804832a46492f4f01d317a108b22b59.svg

Kaggle datasets#

Kaggle is an online platform that serves as a hub for data science and machine learning practitioners. It provides a range of resources, competitions, and community engagement opportunities. Kaggle hosts a large repository of datasets on a wide range of topics, freely available for download.

Note

To fetch datasets from kaggle one needs to obtain an API token from Kaggle and place it in the .kaggle/kaggle.json file.

For example, download MATH dataset:

import kaggle
!kaggle datasets download -d mathurinache/math-dataset -p ../datasets
Dataset URL: https://www.kaggle.com/datasets/mathurinache/math-dataset
License(s): CC-BY-NC-SA-4.0
math-dataset.zip: Skipping, found more recently modified local copy (use --force to force download)

This is a text dataset, it is stored in json files:

import json
from pathlib import Path

problems = []
for p in Path("../datasets/MATH/train/algebra/").rglob("*.json"):
    with open(p) as f:
        problems.append(json.load(f))

print("Total problems:", len(problems))
problems[2]
Total problems: 1744
{'problem': 'If $\\sqrt{2\\sqrt{t-2}} = \\sqrt[4]{7 - t}$, then find $t$.',
 'level': 'Level 4',
 'type': 'Algebra',
 'solution': 'We raise both sides to the fourth power, which is equivalent to squaring twice, in order to get rid of the radicals. The left-hand side becomes $$\\left(\\sqrt{2\\sqrt{t-2}}\\right)^4 = \\left(2\\sqrt{t-2}\\right)^2 = 4 \\cdot (t-2) = 4t-8.$$The right-hand side becomes $\\left(\\sqrt[4]{7-t}\\right)^4 = 7-t$. Setting them equal, $$4t-8 = 7-t \\quad\\Longrightarrow\\quad 5t = 15,$$and $t = \\boxed{3}$.  Checking, we find that this value does indeed satisfy the original equation.'}

Other resources#

Exercises#

  1. What is the shape of the feature matrix of MNIST dataset?

  2. A dataset contains \(10\) features. Six of them are numeric, the rest four are categorical with \(2\), \(3\), \(5\) and 9 categories respectively. How many features will be in this dataset after one-hot encoding with/without dropping?