Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A question about the computation of loss #3

Closed
liuyuan-pal opened this issue Dec 24, 2019 · 3 comments
Closed

A question about the computation of loss #3

liuyuan-pal opened this issue Dec 24, 2019 · 3 comments

Comments

@liuyuan-pal
Copy link

Hi Jiahui, I'm trying to reproduce the results on YFCC.
I have a question about the computation of loss.
I find that the essential loss is used after 20k steps and only those essential loss less than 0.1 is used in the backward. (https://github.com/zjhthu/OANet/blob/master/core/loss.py#L49 and https://github.com/zjhthu/OANet/blob/master/core/loss.py#L84) I am wondering what is the motivation behind this implementation and what will happen if we use all essential losses all the time.
Thank you for the excellent work.

@zjhthu
Copy link
Owner

zjhthu commented Dec 24, 2019

It is a common practice in the robust loss function. As said in the DFE, Clamping the residuals ensures that hard problem instances in the training set do not dominate the training loss. I remember that we have tested different thresholds, 0.1 works best. Larger or smaller thresholds gives a little worse results.

@liuyuan-pal
Copy link
Author

Thanks for your answer!

@zjhthu
Copy link
Owner

zjhthu commented Dec 24, 2019

As for using essential loss only after 20k steps, this is inherited from learning-corr.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants