Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add PoolFormer Model #14584

Closed
3 tasks done
tanaymeh opened this issue Dec 1, 2021 · 0 comments
Closed
3 tasks done

Add PoolFormer Model #14584

tanaymeh opened this issue Dec 1, 2021 · 0 comments

Comments

@tanaymeh
Copy link
Contributor

tanaymeh commented Dec 1, 2021

🌟 New model addition

I would like to add the recently announced PoolFormer model to the Transformers library.

Model description

PoolFormer model was proposed in the paper, “MetaFormer is Actually What You Need for Vision” by Sea AI Lab and the main argument behind this is that performance of transformer/MLP-like models primarily comes from the general architecture instead of the specific token mixers they are using (like Attention for example).

To show this, they have added a basic non-parametric Pooling operator which does basic token mixing. This suggested model outperforms DeiT and ResMLP.

Link to the Model Repo: PoolFormer
Link to the Paper: MetaFormer is Actually What You Need for Vision

Open-source status

  • the model implementation is available: In PyTorch on PoolFormer repository
  • the model weights are available: In the same repository, link
  • who are the authors: Sea AI Lab
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant