Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine-tuning based on the DETR architecture code, but the verification indicators are all 0 #69

Closed
Flyooofly opened this issue Nov 18, 2022 · 1 comment

Comments

@Flyooofly
Copy link

Thanks for your work. I noticed that you open-sourced the detreg of the DETR architecture, and then I tried to use the pre-trained model on the imagenet dataset you provided to fine-tune training for my custom dataset. But I found that all the indicators are still 0 after more than fifty batches of pre-training. I have followed the tips in the related issues of DETR (https://github.com/facebookresearch/detr/issues?page=1&q=zero) , the num_calss was modified. Many people mentioned that DETR requires a large amount of training data, or fine-tuning. But I am currently using fine-tuning, and the number of fine-tuning datasets is about one thousand. But the effect is still very poor, may I ask why. It's normal for me to use deformable-detr architecture.
image

Repository owner deleted a comment from luoyq6 Jul 18, 2023
@amirbar
Copy link
Owner

amirbar commented Jul 18, 2023

Apologies for the delay in response. What is the size of your dataset? For example, ImageNet is 1m images, so 50 epochs might be enough to train over it. If your dataset is significantly slower, you would need to increase the number of epochs proportionally. Also, make sure you don't drop the learning rate early (this should be mostly relevant for Deformable-DETR).

@amirbar amirbar closed this as completed Jul 18, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants