Skip to content


Switch branches/tags

Name already in use

A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?

Latest commit

* Added two datasets from the Nikuradse 'Flow in Rought Pipes' paper.

* update dataset files

Created by\nfrom 5becab4 on 2023-02-23

* Update metadata.yaml files.


Co-authored-by: github-actions[bot] <github-actions[bot]>

Git stats


Failed to load latest commit information.
Latest commit message
Commit time
February 21, 2023 08:53
February 23, 2023 19:28
August 20, 2021 12:19
October 15, 2020 14:16
February 23, 2023 19:28
October 9, 2020 17:20
August 18, 2020 17:31
October 9, 2020 17:20
February 23, 2023 19:28
November 11, 2016 14:16
September 9, 2020 15:13
August 20, 2021 12:19
October 9, 2020 17:20

Penn Machine Learning Benchmarks

This repository contains the code and data for a large, curated set of benchmark datasets for evaluating and comparing supervised machine learning algorithms. These data sets cover a broad range of applications, and include binary/multi-class classification problems and regression problems, as well as combinations of categorical, ordinal, and continuous features.

Please go to our home page to interactively browse the datasets, vignette, and contribution guide!

Breaking changes in PMLB 1.0

This repository has been restructured, and several dataset names have been changed!

If you have an older version of PMLB, we highly recommend you upgrade it to v1.0 for updated URLs and names of datasets:

pip install pmlb --upgrade


Datasets are tracked with Git Large File Storage (LFS). If you would like to clone the entire repository, please install and set up Git LFS for your user account. Alternatively, you can download the .zip file from GitHub.

All data sets are stored in a common format:

  • First row is the column names
  • Each following row corresponds to one row of the data
  • The target column is named target
  • All columns are tab (\t) separated
  • All files are compressed with gzip to conserve space


The complete table of dataset characteristics is also available for download. Please note, in our documentation, a feature is considered:

  • "binary" if it is of type integer and has 2 unique values (equivalent to pandas profiling's "boolean")
  • "categorical" if it is of type integer and has more than 2 unique values (equivalent to pandas profiling's "categorical")
  • "continuous" if it is of type float (equivalent to pandas profiling's "numeric").

Python wrapper

For easy access to the benchmark data sets, we have provided a Python wrapper named pmlb. The wrapper can be installed on Python via pip:

pip install pmlb

and used in Python scripts as follows:

from pmlb import fetch_data

# Returns a pandas DataFrame
adult_data = fetch_data('adult')

The fetch_data function has two additional parameters:

  • return_X_y (True/False): Whether to return the data in scikit-learn format, with the features and labels stored in separate NumPy arrays.
  • local_cache_dir (string): The directory on your local machine to store the data files so you don't have to fetch them over the web again. By default, the wrapper does not use a local cache directory.

For example:

from pmlb import fetch_data

# Returns NumPy arrays
adult_X, adult_y = fetch_data('adult', return_X_y=True, local_cache_dir='./')

You can also list all of the available data sets as follows:

from pmlb import dataset_names


Or if you only want a list of available classification or regression datasets:

from pmlb import classification_dataset_names, regression_dataset_names


Example usage: Compare two classification algorithms with PMLB

PMLB is designed to make it easy to benchmark machine learning algorithms against each other. Below is a Python code snippet showing the most basic way to use PMLB to compare two algorithms.

from sklearn.linear_model import LogisticRegression
from sklearn.naive_bayes import GaussianNB
from sklearn.model_selection import train_test_split

import matplotlib.pyplot as plt
import seaborn as sb

from pmlb import fetch_data, classification_dataset_names

logit_test_scores = []
gnb_test_scores = []

for classification_dataset in classification_dataset_names:
    X, y = fetch_data(classification_dataset, return_X_y=True)
    train_X, test_X, train_y, test_y = train_test_split(X, y)

    logit = LogisticRegression()
    gnb = GaussianNB(), train_y), train_y)

    logit_test_scores.append(logit.score(test_X, test_y))
    gnb_test_scores.append(gnb.score(test_X, test_y))

sb.boxplot(data=[logit_test_scores, gnb_test_scores], notch=True)
plt.xticks([0, 1], ['LogisticRegression', 'GaussianNB'])
plt.ylabel('Test Accuracy')


See our Contributing Guide. We're looking for help with documentation, and also appreciate new dataset and functionality contributions.

Citing PMLB

If you use PMLB in a scientific publication, please consider citing one of the following papers:

Joseph D. Romano, Le, Trang T., William La Cava, John T. Gregg, Daniel J. Goldberg, Praneel Chakraborty, Natasha L. Ray, Daniel Himmelstein, Weixuan Fu, and Jason H. Moore. PMLB v1.0: an open source dataset collection for benchmarking machine learning methods. arXiv preprint arXiv:2012.00058 (2020).

  title={PMLB v1.0: an open source dataset collection for benchmarking machine learning methods},
  author={Romano, Joseph D and Le, Trang T and La Cava, William and Gregg, John T and Goldberg, Daniel J and Chakraborty, Praneel and Ray, Natasha L and Himmelstein, Daniel and Fu, Weixuan and Moore, Jason H},
  journal={arXiv preprint arXiv:2012.00058v2},

Randal S. Olson, William La Cava, Patryk Orzechowski, Ryan J. Urbanowicz, and Jason H. Moore (2017). PMLB: a large benchmark suite for machine learning evaluation and comparison. BioData Mining 10, page 36.

BibTeX entry:

    author="Olson, Randal S. and La Cava, William and Orzechowski, Patryk and Urbanowicz, Ryan J. and Moore, Jason H.",
    title="PMLB: a large benchmark suite for machine learning evaluation and comparison",
    journal="BioData Mining",

Support for PMLB

PMLB was developed in the Computational Genetics Lab at the University of Pennsylvania with funding from the NIH under grant AI117694, LM010098 and LM012601. We are incredibly grateful for the support of the NIH and the University of Pennsylvania during the development of this project.


PMLB: A large, curated repository of benchmark datasets for evaluating supervised machine learning algorithms.







No packages published