Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss differs slightly from paper #8

Open
poier opened this issue Jun 29, 2022 · 0 comments
Open

Loss differs slightly from paper #8

poier opened this issue Jun 29, 2022 · 0 comments

Comments

@poier
Copy link

poier commented Jun 29, 2022

I just saw that the final loss in the implementation differs from what was described in the paper (Eqn. (5)). In the implementation (https://github.com/JizhiziLi/P3M/blob/master/core/train.py#L85), loss_fusion_alpha is used in two terms, giving it a higher weight than described in the paper. The second term can simply be skipped to make it equivalent to the paper, i.e., change it to:

loss = loss_global/6+loss_local*2+loss_fusion_alpha*2+loss_fusion_comp

instead of:

loss = loss_global/6+loss_local*2+loss_fusion_alpha*2+loss_fusion_alpha+loss_fusion_comp
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant