Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Time required to train one epoch. #5

Closed
charlesCXK opened this issue May 23, 2022 · 6 comments
Closed

Time required to train one epoch. #5

charlesCXK opened this issue May 23, 2022 · 6 comments

Comments

@charlesCXK
Copy link

Dear author:
Thank you for sharing the excellent work! May I ask how the time overhead of ConvMAE pre-training compares to MAE? Can you provide the time required to train an epoch for these two methods on the same type of GPU?

@gaopengpjlab
Copy link

Thanks for your suggestion. I will answer your question in a few days.

@gaopengpjlab
Copy link

hardware setup : 8 * A6000 with per GPU holding 128 images

MAE
0.4084s per iteration, GPU memory printed by nvidia-smi command : 17926MB

ConvMAE
0.8306s per iteration, GPU memory printed by nvidia-smi command : 27049MB

ConvMAE by skipping masked region computation in stage 1 and stage 2
0.5480s per iteration, GPU memory printed by nvidia-smi command : 21250MB

Thank you so much for reminding us about training speed comparison. We will include speed/GPU memory/FLOPs comparison in updated version.

@charlesCXK
Copy link
Author

Thanks for your detailed reply!

@gaopengpjlab
Copy link

By default, ConvMAE stands for ConvMAE + Multi-scale Decoder proposed in our paper.

@gaopengpjlab
Copy link

We are going to release Fast ConvMAE which can significantly accelerate the pretraining of ConvMAE in a few days.
https://github.com/Alpha-VL/FastConvMAE

@gaopengpjlab
Copy link

Fast ConvMAE have been released which accelerates the pretraining time by half.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants