Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The model is very sensitive to the batch size #15

Open
nebuladream opened this issue Apr 21, 2020 · 2 comments
Open

The model is very sensitive to the batch size #15

nebuladream opened this issue Apr 21, 2020 · 2 comments

Comments

@nebuladream
Copy link

We train the model on 4-GPUs with different batch size. The results on validation change largely with batch size, for batch=128 we get all_recall=286; batch = 256 we get all_recall=268; batch=1280 all_recall=240;
We train the model on 1-GPU with different batch size: batch=128, all_recall=295; batch=256, all_recall=285;
We tried different learning rate, but it seems has no affect for the decreasing result.
Do you have the similar result?

@danieljf24
Copy link
Owner

danieljf24 commented Apr 23, 2020

Sorry, I just trained the model on 1 GPU. The results you posted are interesting. I think it may be caused by the triplet loss with hard example mining.
Additionally, I am wondering why you posted all_recall is so high, I only obtained all_recall about 150.

@nebuladream
Copy link
Author

Sorry, I just trained the model on 1 GPU. The results you posted are interesting. I think it may be caused by the triplet loss with hard exampling mining.
Additionally, I am wondering why you posted all_recall is so high, I only obtained all_recall about 150.

it may because we report all recall on all direction, more details as follow:
Text to video:
r_1_5_10: [20.433, 47.042, 57.455]
medr, meanr: [7.0, 37.884]
Video to text:
r_1_5_10: [32.998, 62.777, 74.245]
medr, meanr: [3.0, 18.048]
best sum recall: 294.9496981891348

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants