New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
(oicrbaseline+mist wi reg) setup map50 result 41.05 #15
Comments
have you tried the ITER_SIZE flag? Also, in addition to decrease the LR, you should also train longer (4x). Unfortunately we never test our model using 1 or 2 GPUs, so I'm not sure about the best parameters. |
Do the all the related parameters WARMUP_ITERS, STEPS, MAX_ITER (4x)? |
You should only have to set ITER_SIZE, the rest should be automatically scaled for you, it'll throw up a warning to let you know. I was able to reproduce with these settings |
ok, thanks. I will try. |
Setup: |
Setup: |
I got ~50 at last epoch with BS=1, ITER=8. ITER_SIZE is an approximation and a somewhat shaky one at that, it isn't surprising for it to perform differently. If we want to reproduce their results then using their original hyperparameters is the only way to be sure |
Thanks @bradezard131 @bradezard131 @bityangke @liz6688 for sharing. @bradezard131 contributed to this feature "ITER_SIZE". I believe he knows better than me how to use it. Unfortunately we didn't test under 1/2 GPUs setting. We report the best performance as we found it's usually not achieved at the last iteration, especially for small dataset like VOC. The things I don't understand is why the variance is so big between @bityangke @bradezard131 reported. Seems strange. I would still recommend using 8GPUs and I'll look into this. Thanks again. |
Hopefully I will be able to clean up and release my re-implementation soon, which works with 1 GPU and is somewhat simplified |
Thanks. Looking forward to your re-implementation. |
Thanks for your reply.
As our computing resources are limited, there is no way to use 8 gpus. In order to get the performance of the paper, I setup my experiments as follows (oicrbaseline+mist wi reg):
removed the cbd module
BATCH_SIZE=2
base lr = 0.0025 (0.01/4)
torch = 1.5.0
cuda=9.2
The result of voc_2007_test data is map50 = 41.5. There is still a big gap from the results 51.4 in the paper. Do you have any suggestions for further improvement?
The text was updated successfully, but these errors were encountered: