Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Whats the point of the class token here? #20

Closed
Manojbhat09 opened this issue Oct 21, 2020 · 3 comments
Closed

Whats the point of the class token here? #20

Manojbhat09 opened this issue Oct 21, 2020 · 3 comments

Comments

@Manojbhat09
Copy link

Really appreciate your work.
Question : As the topic.

@lucidrains
Copy link
Owner

lucidrains commented Oct 21, 2020

@Manojbhat09 it's a common practice in NLP where you have one token that pools information from the rest of the tokens through rounds of attention, usually to classify the sentence at the end. whether it is completely necessary for ViT to work is up to debate. my take is it isn't that important. you could pool all the embeddings from the last layer and probably still get great results at scale

@Manojbhat09
Copy link
Author

Thank you for your clear explanation. Hope others too find it helpful.

@zhongyy
Copy link

zhongyy commented Jan 15, 2021

Thank you for the explaination !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants