Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

When will release mae pretrained weights? #14

Closed
shi-lt opened this issue Jun 16, 2023 · 8 comments
Closed

When will release mae pretrained weights? #14

shi-lt opened this issue Jun 16, 2023 · 8 comments

Comments

@shi-lt
Copy link

shi-lt commented Jun 16, 2023

No description provided.

@UmaisZahid
Copy link

Any update on on the pretrained MAE release? Even just having the pretrained decoder weights would be quite useful.

@dbolya
Copy link
Contributor

dbolya commented Jul 10, 2023

@UmaisZahid We are currently going back and forth about how to release these weights. May I ask, what do you plan to use the decoder for?

The issue is the models were trained with a normalized pixel loss (i.e., each patch that the decoder had to predict was normalized within that patch). Thus, using the pretrained decoder for reconstructing a masked image wouldn't look correct. Hence, we were planning on removing the decoder when releasing the weights (I believe MAE does the same thing).

Would the decoder still be useful in spite of that?

@UmaisZahid
Copy link

Thanks for the response @dbolya.

I believe so yes. One use case for the decoder would be to further pretrain on another dataset. The ImageNet pretraining would act as a warm start or inductive bias of sorts in this case - which I imagine is particularly useful in the cases where the second pretraining dataset is smaller.

You could release the decoder and perhaps remind people explicitly in the GitHub Readme what you just mentioned about the reconstructive loss - that would be very useful!

@dbolya
Copy link
Contributor

dbolya commented Jul 21, 2023

Thanks for the context. We've decided to release all the pretrained weights with decoders and all!

The pretrained models are available in v0.1.2 (see the readme on how to use them). Closing the issue now, but feel free to reopen if you run into any trouble.

@dbolya dbolya closed this as completed Jul 21, 2023
@abhishekaich27
Copy link

@dbolya Are these weights for MViT2 or modified MViT2 as proposed in the paper?

@dbolya
Copy link
Contributor

dbolya commented Sep 7, 2023

@abhishekaich27 These weights are for Hiera, the final model that we end up with after making all modifications to MViTv2.

@abhishekaich27
Copy link

Thanks @dbolya! Do you plan to release or provide the MAE pre-trained MViT2 model?

@dbolya
Copy link
Contributor

dbolya commented Sep 7, 2023

@abhishekaich27 No plans for that afaik.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants