New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error finetuning from pretrained checkpoint #30
Comments
Thank you so much for the fast reply. With the full checkpoint I see two messages about incompatible keys and the training time is 0:00:00. However, once I specify more epochs than what was used for the pretrained model everything looks good. |
I just finished finetuning for 50 epochs but the self-attention maps look worse than those I get from the pretrained checkpoint. Not sure if that's due to a poor choice of args or related to the messages about incompatible keys.
|
Maybe specify args, such as: https://dl.fbaipublicfiles.com/dino/dino_deitsmall16_pretrain/args.txt |
Thanks for the suggestion @woctezuma. I'll definitely need to spend some time to find args that work well for my single GPU setup. The main reason I re-opened the issue is to double-check that the messages about incompatible keys can indeed be safely ignored. Maybe @mathildecaron31 could comment on that. |
It looks like the checkpoints were trained on a slightly different version of the released code. Luckily it's not difficult to change the names of the affected keys.
Now training starts at a much smaller loss and I see the following message.
|
Hi @yadamonk Thanks for flagging this. I indeed need to change the keys from the released checkpoints to match this codebase. I changed the keys from the projection head during refactoring to prepare the code release. |
Hi @ymathildecaron31 Thank you so much for your wonderful work and all the time you're putting into helping others build on it. |
Thanks a lot for your kind words @yadamonk :) :). |
@mathildecaron31 Any Updates on this? I tried the method suggested by @yadamonk but it doesn't seem to work for me. |
Hi @tejassp2002 The checkpoint seems to have been updated since I posted my workaround. If you replace the corresponding sections with the snippets below things should work again. Hope that helps.
|
Hey @yadamonk, that really helped! |
Hi all, I'm running into an error when trying to fine-tune from one of the pretrained checkpoints.
Code
Error
Any suggestions would be very much appreciated.
The text was updated successfully, but these errors were encountered: