Skip to content

Fast Hyperparameter Bayesian Optimization for Large Dataset

Notifications You must be signed in to change notification settings

rickie95/fabolas

Repository files navigation

FABOLAS: Fast Bayesian Optimization of Hyperparameters on Large Datasets

This repository replicates experiments from Klein, Falkner, Bartels, Henning, Hutter's "Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets" (arXiv:1605.07079v2).

Experiments:

  • Showcase: grid search on SVM
  • SMV on
    • MNIST,
    • Vehicle Registration,
    • Forest Cover Types
  • CNN on:
    • CIFAR-10,
    • SVHN
  • Deep Residual Network on CIFAR-10

Showcase: Grid Search on SVM

File: svm-grid-search.py

Just a playground, shows the impact of the dataset size on the hyperparameters search. The script uses methods and API of ScikitLearn library, which provides a handy way to execute a grid search with cross validation.

Grid search is run on SVMs equipped with RBF kernel, searching for the best C and gamma couple that fit best the data.

SVM on MNIST

File: svm-mnist.py

Searches for the best couple of C and gamma, benchmarking three methods: Expected Improvement, Entropy Search and FABOLAS

CNN on CIFAR10

File: cnn_cifar10.py

Tries to find the best configuration choosing:

  • # of filters for convolutional layer L1, L2, L3, mapped in log_2 space and bounded in [4, 9]
  • batch normalization
  • leanring rate, mapped in log_10, bounded in [-6, 0]

Methods tested: EI, ES, FABOLAS

CNN on Street View House Numbers

File: cnn_svhn.py

Tries to find the best configuration choosing:

  • # of filters for convolutional layer L1, L2, L3, mapped in log_2 space and bounded in [4, 9]
  • batch normalization
  • leanring rate, mapped in log_10, bounded in [-6, 0]

Methods tested: EI, ES, FABOLAS

About

Fast Hyperparameter Bayesian Optimization for Large Dataset

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published