New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implementation of Consistency Loss #13
Comments
Hi, as shown in Eq. 12, the loss is divided by the number of decoder layers, which equals As for the summation in Eq. 13, we are sorry for the mistake, it should be an average over |
Thanks for pointing it out, we will correct it and upload a new file to Arxiv. |
Each element in the list |
I think you are right. Sorry I have not run the code yet myself so was just reading along. Thanks. |
Hello, so I noticed that there is a slight difference between the paper's fornulation of consistency loss and the actual implementation. Please correct me if I am wrong. As can be seen, the actual formulation are as follows:
However, I noticed that for the final consistency loss, the actual implementation uses a sum instead of an average across layers as indicated in equation 12 (weight_dict for this type of loss is 1 by defualt, therefore notion of average is perhaps not incorporated). As for the per layer consistency loss, the actual implementation uses an average over all M object queries instead of just the sum, as indicated in equation 13.
I am wondering what are the right formulations to follow? Thanks.
The text was updated successfully, but these errors were encountered: