Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ALiBi in self-Attention #4

Closed
Ldoun opened this issue Nov 15, 2021 · 2 comments
Closed

ALiBi in self-Attention #4

Ldoun opened this issue Nov 15, 2021 · 2 comments
Labels
question Further information is requested

Comments

@Ldoun
Copy link

Ldoun commented Nov 15, 2021

What is your question?

first of all nice work! and thank you for sharing the code.
I noticed that the code use ALiBi in encoder-decoder attention but not in the transformer's self-Attention. Have you tried ALiBi in transformer self-attention? And Is there a reason you didn't use it for the self-attention layer?

@Ldoun Ldoun added the question Further information is requested label Nov 15, 2021
@ofirpress
Copy link
Owner

Thanks!
We've actually currently implemented ALiBi for self-attention (casually masked) only. We have not implemented it for encoder-decoder attention or non-masked encoder attention yet.

@Ldoun
Copy link
Author

Ldoun commented Nov 15, 2021

oh, I missed that. thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants