Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pretraining segformer on ImageNet-22K #32

Closed
htzheng opened this issue Jul 1, 2021 · 4 comments
Closed

Pretraining segformer on ImageNet-22K #32

htzheng opened this issue Jul 1, 2021 · 4 comments

Comments

@htzheng
Copy link

htzheng commented Jul 1, 2021

The Swin transformer release a large model pretrained on ImageNet-22K for semantic segmentation and achieved a good result. I wonder if you are interested in improving segformer in a similar way? Thanks!

@xieenze
Copy link
Contributor

xieenze commented Jul 2, 2021

Hi,
Good question!
Yes, we are planning to pre-train on Imagenet-22k. But actually, it is not easy for us.
Firstly before we did not have enough GPU resources, but recently we have sufficient resources now.
Then we haven't master multi-node training on our GPU clusters yet, we are still trying to figure it out.
We believe that recently we can pre-train it on Imagenet-22k and release stronger models.

Thanks for the attention!

@htzheng
Copy link
Author

htzheng commented Jul 2, 2021

That would be really great! Thanks for the great work and good luck to you!

@htzheng htzheng closed this as completed Jul 2, 2021
@armored-guitar
Copy link

@xieenze how many epochs have you trained your model on imagenet? cannot find this info

@pablogiaccaglia
Copy link

Hi, any news on this?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants