You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
in the blog you says there is a more efficient way of implementation? see lecture at the top. Do you mean the youtube vide at the top?
but there is no code explaination in the video , do i have to watch the video and implement myself or any blogs about the more efficient way of self attention?
thanks a lot!
The text was updated successfully, but these errors were encountered:
This refers to slide 25-26 in this lectures: https://dlvu.github.io/slides/dlvu.lecture12.pdf Slide 25 shows the basic idea of multi-head self attention, and 26 shows how to implement it efficiently.
This is implemented in the default self-attention here:
in the blog you says there is a more efficient way of implementation? see lecture at the top. Do you mean the youtube vide at the top?
but there is no code explaination in the video , do i have to watch the video and implement myself or any blogs about the more efficient way of self attention?
thanks a lot!
The text was updated successfully, but these errors were encountered: