-
Notifications
You must be signed in to change notification settings - Fork 134
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Percevier_IO mask #49
Comments
@xesdiny the latents never need any masking at all https://github.com/lucidrains/perceiver-pytorch/blob/main/perceiver_pytorch/perceiver_io.py#L168 |
Hi, not sure if I'm missing a key detail here but from what I can see the mask in this implementation would not work like a normal transformer. The mask as applied here allows you to control which latents get information from which part of the input sequence (i.e the mask is To match existing transformers concept of an attention mask (or at least the one used in autoregressive LM's) the mask would need to be something like
Does this makes sense, am I missing something? |
Hi,guy.I learing you code to attention mask code in perceiver io.
If it is the step of encode, it can be well understood as whether the input information in the mask is mapped to the hidden space, but in the decode part, the logic of this code is not explained.
Mask in lantent space array mapping to Output array in lantent dimension, What's means?
The text was updated successfully, but these errors were encountered: