Milano is a tool for automating hyper-parameters search for your models on a backend of your choice.
Clone or download
Latest commit bb1e268 Nov 7, 2018
Type Name Latest commit message Commit time
Failed to load latest commit information.
benchmarking initial code commit Sep 5, 2018
docs Update Quick_start md Oct 31, 2018
examples visualize Nov 7, 2018
milano initial code commit Sep 5, 2018
.gitignore initial code commit Sep 5, 2018 initial code commit Sep 5, 2018
LICENSE initial code commit Sep 5, 2018 Merge branch 'master' into visualize Nov 7, 2018
iconMilano.png initial code commit Sep 5, 2018
requirements.txt initial code commit Sep 5, 2018 initial code commit Sep 5, 2018 :q:xvisualize Nov 7, 2018

License Documentation


(This is a research project, not an official NVIDIA product.)



Milano (Machine learning autotuner and network optimizer) is a tool for enabling machine learning researchers and practitioners to perform massive hyperparameters and architecture searches.

You can use it to:

Your script can use any framework of your choice, for example, TensorFlow, PyTorch, Microsoft Cognitive Toolkit etc. or no framework at all. Milano only requires minimal changes to what your script accepts via command line and what it returns to stdout.

Currently supported backends:

  • Azkaban - on a single multi-GPU machine or server with Azkaban installed
  • AWS - Amazon cloud using GPU instances
  • SLURM - any cluster which is running SLURM


  • Linux
  • Python 3
  • Ensure you have Python version 3.5 or later with packages listed in the requirements.txt file.
  • Backend with NVIDIA GPU

How to Get Started

  1. Install all dependencies with the following command pip install -r requirements.txt.
  2. Follow this mini-tutorial for local machine or this mini-tutorial for AWS


We provide a script to convert the csv file output into two kinds of graphs:

  • Graphs of each hyperparameter with the benchmark (e.g. valid perplexity)
  • Color graphs that show the relationship between any two hyperparameters and the benchmark

To run the script, use:

python3 --file [the name of the results csv file] 
                     --n [the number of samples to visualize]
                     --subplots [the number of subplots to show in a plot]
                     --max [the max value of benchmark you care about]