You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It seems that you atrank is different from the one described in paper. For example, bilinear attention is used in paper, but scale-dot attention here. Vanilla attention in paper, but multi-head attention here.
The text was updated successfully, but these errors were encountered:
Query comes from the decoding layer, key and value comes from the encoding layer and is called vanilla attention (which the paper didn't say), the most basic attention. Query, key and value are all from the encoding layer called self attention.
It seems that you atrank is different from the one described in paper. For example, bilinear attention is used in paper, but scale-dot attention here. Vanilla attention in paper, but multi-head attention here.
The text was updated successfully, but these errors were encountered: