New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Time required to train one epoch. #5
Comments
Thanks for your suggestion. I will answer your question in a few days. |
hardware setup : 8 * A6000 with per GPU holding 128 images MAE ConvMAE ConvMAE by skipping masked region computation in stage 1 and stage 2 Thank you so much for reminding us about training speed comparison. We will include speed/GPU memory/FLOPs comparison in updated version. |
Thanks for your detailed reply! |
By default, ConvMAE stands for ConvMAE + Multi-scale Decoder proposed in our paper. |
We are going to release Fast ConvMAE which can significantly accelerate the pretraining of ConvMAE in a few days. |
Fast ConvMAE have been released which accelerates the pretraining time by half. |
Dear author:
Thank you for sharing the excellent work! May I ask how the time overhead of ConvMAE pre-training compares to MAE? Can you provide the time required to train an epoch for these two methods on the same type of GPU?
The text was updated successfully, but these errors were encountered: