You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Do you realized key_padding_mask (a parameter in torch.nn.multiheadattention) in the long-term attention and short-term Attention? Or you just do not need this in your network?
The text was updated successfully, but these errors were encountered:
zhouweii234
changed the title
Do you realized key_padding_mask in you long-term attention and short-term Attention
Do you realized key_padding_mask in you long-term attention and short-term Attention?
Aug 27, 2021
Long-term attention doesn't need any mask. The implementation of short-term attention is complicated. Since short-term attention has some processes related to the relative positions between pixels. Masking out the correlations out of feature boundary is necessary.
Do you realized
key_padding_mask
(a parameter intorch.nn.multiheadattention
) in the long-term attention and short-term Attention? Or you just do not need this in your network?The text was updated successfully, but these errors were encountered: