Skip to content

tanglang96/MDENAS

master
Switch branches/tags
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Multinomial Distribution Learning for Effective Neural Architecture Search

Here we propose a method to extremely accelerate NAS, without reinforcement learning or gradient, just by sampling architectures from a distribution and comparing these architectures, estimating their relative performance rather than absolute performance, iteratively updating parameters of the distribution while training.

Our paper is available at 《Multinomial Distribution Learning for Effective Neural Architecture Search》

Here we provide our test codes and pretrained models, our code is based on DARTS and ProxylessNAS, pretrained models can be downloaded here

Requirements

  • PyTorch 1.0
  • DALI

Evaluate

You need to modified your path to dataset in data_providers/cifar10.py and data_providers/imagenet.pyconfig.sh is used to prepare your environment, you should write this file by yourself and here we use it to prepare dataset and packages

To evaluate the model in DARTS setting, just run

for cifar10 :

chmod +x run_darts_cifar.sh
./run_darts_cifar.sh

for imagenet :

chmod +x run_darts_imagenet.sh
./run_darts_imagenet.sh

To evaluate the model in Mobile setting, just run

for gpu :

chmod +x run_gpu_imagenet.sh
./run_gpu_imagenet.sh

for cpu :

chmod +x run_cpu_imagenet.sh
./run_cpu_imagenet.sh

Performance

About

Multinomial Distribution Learning for Effective Neural Architecture Search

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published