Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about experiments #2

Closed
Hongbin98 opened this issue Jul 27, 2021 · 6 comments
Closed

Questions about experiments #2

Hongbin98 opened this issue Jul 27, 2021 · 6 comments

Comments

@Hongbin98
Copy link

Thanks for your ICCV work. However, I find you directly use the test set of ImageNet-LT to store the best models, which may lead to overfitting in practice and seems to be unfair to other compared methods. Could you please provide the results on ImageNet-LT using the validation set to store models? It would be easier for us to compare with PaCo in our work. Thanks very much.

@jiequancui
Copy link
Collaborator

Hi,

Very thanks for your suggestion! In fact, we observe there is no difference for selecting models by validation set or test set in our experiments. Further, for all re-implemented baselines, we use the same setting for comparisons.

@xxxzhi
Copy link

xxxzhi commented Aug 2, 2021

hi, do you run experiment on 4 gpus with total batch size 128? what's the running time? I find I can not reproduce the result with batch size 128 (single gpu). Much worse than the reported.

@jiequancui
Copy link
Collaborator

Hi,
Thank you for your question.
We run all experiments on 4gpus with a batch size of 128 in 400 epochs, that's 32 images on a single GPU.
what network is used in your experiment? Can you provide me a training log to check the problem?

@xxxzhi
Copy link

xxxzhi commented Aug 2, 2021

Thanks for your response. 400 epochs, 128 batch size. That might be very time-consuming. My experiment has not been finished. I use 4 gpus with a batch size of 128*4 in 400 epochs. here is the log.
R50_mocot0.2_augrandcls_sim_400epochs_lr0.02_t1_gpu4.log

here is the log of 2 gpu
R50_mocot0.2_augrandcls_sim_400epochs_lr0.02_t1_gpu2.log

how many gpu hours your code run with 400 epochs on Imagenet-LT?

@jiequancui
Copy link
Collaborator

Hi,
You may misunderstand it.
We use 4gpus with a batch size of 128 (32 images per gpu) and train 400 epochs. It takes about 2 days with 4gpus for ResNet-50 on ImageNet-LT.

I think that, with 4gpus, running with our scripts in this repo can reproduce our results reported in the paper.

@xxxzhi
Copy link

xxxzhi commented Aug 2, 2021

Thanks, I will check your released log. I want to use larger batch size to reduce the training time.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants