Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch size for all the experiments in the papaer #8

Closed
timqqt opened this issue Oct 2, 2019 · 2 comments
Closed

Batch size for all the experiments in the papaer #8

timqqt opened this issue Oct 2, 2019 · 2 comments

Comments

@timqqt
Copy link

timqqt commented Oct 2, 2019

Hey, I just want to repeat your work. I saw that your paper implied that the batch size for all experiments should be 1. However, I find that if I set the batch size to 1, I cannot get the same error as your experiments (about 10 times difference). But if I set the batch size to 32, I got decent result.

I would very appreciate your help to explain details in your experiments!

@angeloskath
Copy link
Collaborator

Hi, batch size of 1 is used to report memory requirements for all methods. The larger the batch size the better, since you want to minimize the variance of the gradients.

In the example scripts you can see the default batch size is 32 for the speed limits and 128 for the mnist experiment.

Let me know if I can help with anything more.

Cheers,
Angelos

@angeloskath
Copy link
Collaborator

I am closing the issue but feel free to reopen it (or another one) if needed.

Angelos

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants