New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Which results in paper correspond to the finetune command? #41
Comments
This is finetuning from imagenet initialisation, to finetune from the pretrained model download it |
Thanks! I will try to implement the pretaining with this checkpoint. BTW, I have finetuned the checkpoint WebVid2M+CC3M+COCO, 4-frames, base_patch_16_224. I achieve:
Are these results correct? The R1 is slightly lower than reported values (31.0). Is this regular fluctuation? |
I got the similar results as in paper with batch_size = 64. |
run the test script the sliding window stride argument adds temporal averaging over multiple frame samples :) |
Exactly! I got further improvements with this testing script. R1 comes to 33.7. |
I experiment the finetune procedure and run the command
python train.py --config configs/msrvtt_4f_i21k.json
.I got:
There are two R1 resutls. Which results corresponding to the results in paper.
I found the R1 in Table5 is 31.0. It seems far from these implementation.
The text was updated successfully, but these errors were encountered: