Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About learning rate setting #22

Closed
bityangke opened this issue Jun 16, 2018 · 5 comments
Closed

About learning rate setting #22

bityangke opened this issue Jun 16, 2018 · 5 comments

Comments

@bityangke
Copy link

Hi Yi,
How did you decide the lr step ?
Did you follow somewhere else or experiment it youself ?
Thanks in advance!

@bryanyzhu
Copy link
Owner

Hi, I decided the lr step according to this paper.
But usually in my experiments, I just see when the loss/accuracy saturate, and then decay the lr. I find it more effective.

@bityangke
Copy link
Author

Thanks very much.

@bityangke bityangke reopened this Jun 17, 2018
@bityangke
Copy link
Author

bityangke commented Jun 17, 2018

Hi Yi
I noted that Yuanjun used a batchsize(16x4,16 samples on each card) and iter size 4, so the “batch size” is equal to 256. They used step size 4000, 8000, 10000 (iters) , the corresponding batch steps should be 1000,2000,2500, and epoch number should be about 27,54, 67.
Am I right?
Thanks!

@bryanyzhu
Copy link
Owner

Hi, I am so sorry for the slow response, I am at CVPR this week.

I am also using similar strategy iter_size at here. So my corresponding batch steps are still 4000, 8000 and 10000, which is about 100, 200 and 250. Hope this is clear.

@bityangke
Copy link
Author

Thank you!
Hope you have happy CVPR days!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants