Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[feat] Compositional Attention #41

Closed
blefaudeux opened this issue Oct 26, 2021 · 1 comment
Closed

[feat] Compositional Attention #41

blefaudeux opened this issue Oct 26, 2021 · 1 comment
Assignees
Labels
brainstorm dropping an idea, may or may not be implemented in the end. RFC enhancement New feature or request

Comments

@blefaudeux
Copy link
Contributor

🚀 Feature

Intriguing paper, keep the softmax(QKt) and V untangled, in that retrievals (*V_i in the vanilla attention) can have a look at all the searchs, that is it can be evaluated against all the softmax(QKt)_j, on a per head basis ("heads" become how many searchs and and many retrieval you support, possibly different)

Motivation

Interesting take for some tasks, does not seem life changing for classical MLM but seems very relevant to reasoning or vision related tasks

Pitch

Implement this, see how it goes in something like Dino ?

Alternatives

Not doing it

Additional context

Paper
Reference implementation

@blefaudeux blefaudeux added the enhancement New feature or request label Oct 26, 2021
@blefaudeux blefaudeux self-assigned this Oct 26, 2021
@blefaudeux blefaudeux added the brainstorm dropping an idea, may or may not be implemented in the end. RFC label Oct 28, 2021
@blefaudeux
Copy link
Contributor Author

done

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
brainstorm dropping an idea, may or may not be implemented in the end. RFC enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant