New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New SOTA reported by original TF repo, any plans to sync-up the changes? #70
Comments
Just weight changes from a new set of runs, also not clear on what they changed for the better result. I have D0-D5 converted and evaluated, haven't had a moment to validate D6. I'd already done the D3, D7, D7X updates which were trained with longer epochs. |
Longer epochs? How many epochs? More than 300? On D0, I have played lots with your code on the hyper-parameters, I don't think I can achieve 34.3 to be honest, just with the hyper-params changes |
@factplay1 Reviewing the changes between version 6 and 7 of the paper, it seems that they now use soft-NMS for all models instead of NMS. They didn't train during more epochs. Here is an extract of the paragraph (section 5.1, first paragraph, at the end):
While the old was:
|
softnms is the default for the eval results on latest models, but most of the gains aren't due to that, but due to the retrainining. So still not clear what they changed for D0-D6 that improved the accuracy with the same number of epochs |
It'd be very interesting to find that out. I'll try to also investigate the difference, will report if I find anything useful. |
Hi @rwightman How you trained D7, as far as I know it is not possible to fit in 16GB GPUs with even batch size of one. |
Its the same problem, I am facing. |
@afaq-ahmad @Naivepro1990 I have never tried to train a D7, with the resolution in the paper you'd need 4-8 32-GB V100 cards or A100 cards (if that even works) and many weeks. The originals were trained with (a lot) of TPUs. I have no plans to spend $100K+ on a system or spend $20-50K in cloud compute so no, there is no point. I think a few people have fine-tuned a D7 at lower resolutions for some Kaggle challenges. |
Thanks for your good work.
https://github.com/google/automl/tree/master/efficientdet
They now report 34.3 val mAP for EffDet-D0. Not sure what exactly they did, but seems like they made some changes in their model definition etc. It'd be great if you could have a look at their changes, and potentially sync-up your repo so it reproduces those numbers.
Thanks a million :)
The text was updated successfully, but these errors were encountered: