Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss stuck at ~6.90 #12

Closed
bobi461993 opened this issue Apr 9, 2020 · 3 comments
Closed

Loss stuck at ~6.90 #12

bobi461993 opened this issue Apr 9, 2020 · 3 comments

Comments

@bobi461993
Copy link

I am trying to train MoCo V2 on a machine with 2 GPUs using the hyperparameters recommended in this repo. However, the loss function gets stuck at value 6.90-ish. Is this behaviour normal or should I try with a different set of hyperparameters? I see that you have used a machine with 8 GPUs. Could this explain the difference? Thanks!

@amsword
Copy link

amsword commented Apr 9, 2020

i had similar loss values, and the fine-tuned performance is 66.9 (67.5 as shown in README). Thus, 6.9 loss value looks not that bad.

@bobi461993
Copy link
Author

Thank you for the prompt response!

@KaimingHe
Copy link
Contributor

I suggest you finish training and check the final result. The loss should be decreasing if you monitor for a longer time. 6.9 is not any special number here, as the random guess is log(65536), not log(1000).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants