-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How many gpu days does the training procedure of AVQA take? #8
Comments
Thanks for your kind response. I can see your result is 77 in the end of epoch 13, but I can only reach 76.0 there. I have some idea about that. Could you please check something for me:
|
Oh, another, have you pretrained the grounding module as the original code? I haven't done it and I thought that for the backbone have been changed, the pretrained parameter may be useless, so I commented out the loading code in line 227 in main_avst.py. I think this may be the main cause of this difference! |
@Rainlt Thanks for pointing that out. I found we did use the pretrained grounding module. Gonna fix this bug soon. |
Hello, I'm intrested in your pretty work and trying to reproduce the result. But I found that I have to spend nearly 5 gpu days to train for the AVQA task on 1 3090 gpu. Is this normal?
This is the recorded time during training:
It can be seen that encoding one audio and positive visual sample using swin transformer with adapter spend 0.2s. So it will take 2 gpu days just to encode the positive feature for 30 epochs.
The text was updated successfully, but these errors were encountered: