Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Replicating results #10

Closed
ilia10000 opened this issue Mar 31, 2019 · 3 comments
Closed

Replicating results #10

ilia10000 opened this issue Mar 31, 2019 · 3 comments

Comments

@ilia10000
Copy link

I think your paper is fascinating so I have been experimenting with it for a few weeks now.
I was wondering what hyperparams you used to get 10 images that achieve almost 94% accuracy on MNIST after 1 GD step and 3 epochs. I can't seem to hit this when I run the suggested code for 200 epochs. At most I managed to get around 91%.

python3 main.py --mode distill_basic --dataset MNIST --arch LeNet --distill_steps 1 --train_nets_type known_init --n_nets 1 --test_nets_type same_as_train

@ssnl
Copy link
Owner

ssnl commented Mar 31, 2019

Thanks for your interest in our work! It turned out indeed that I set the default training epochs to be too low for some of the settings. In 3d07dd5 I changed the default values to ones that should cover all settings. I also specified the initial distill_lr for some experiments, although in my experience it doesn't matter much.

Thanks again. Let me know if you still have problems using the updated settings.

@ssnl ssnl closed this as completed Mar 31, 2019
@ilia10000
Copy link
Author

Thanks, already getting better results with the new defaults!
Is the reported result in the paper based on the final distilled images (ie images after epoch 350) or is it based on the best intermediate result (e.g. one of the checkpoints at earlier epoch)?

@ssnl
Copy link
Owner

ssnl commented Mar 31, 2019

@ilia10000 The results in the paper are based on the final images. It is possible that some intermediate results are better.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants