Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I use SMU instead of SILU in YoloV5, but loss shows up as nan #8

Open
mzzjuve opened this issue Apr 10, 2022 · 2 comments
Open

I use SMU instead of SILU in YoloV5, but loss shows up as nan #8

mzzjuve opened this issue Apr 10, 2022 · 2 comments

Comments

@mzzjuve
Copy link

mzzjuve commented Apr 10, 2022

I use SMU instead of SILU in YoloV5, but loss shows up as nan.

Could you please tell me the possible reason?Or maybe it's normal that this happened in previous epochs?

image

@koushik313
Copy link

@mzzjuve
Thanks for the information you shared. I suppose you use consider alpha=0.25 and mu=100000. Instead, I will recommend you try to initialize alpha at 0.01 and mu at 2.0 or 2.5 (use mu as a trainable parameter) for SMU and then run your experiments. From, my experience, these initializations provide better results. Loss should not be nan with these parameter values. Please let me know if you still got nan.

I use SMU instead of SILU in YoloV5, but loss shows up as nan.

Could you please tell me the possible reason?Or maybe it's normal that this happened in previous epochs?

image

@mzzjuve
Copy link
Author

mzzjuve commented Apr 11, 2022

@mzzjuve Thanks for the information you shared. I suppose you use consider alpha=0.25 and mu=100000. Instead, I will recommend you try to initialize alpha at 0.01 and mu at 2.0 or 2.5 (use mu as a trainable parameter) for SMU and then run your experiments. From, my experience, these initializations provide better results. Loss should not be nan with these parameter values. Please let me know if you still got nan.

I use SMU instead of SILU in YoloV5, but loss shows up as nan.
Could you please tell me the possible reason?Or maybe it's normal that this happened in previous epochs?
image

Thank you for your timely reply. The problem was solved after I modified the parameters. I'll keep training. Thank you for your excellent work

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants