New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
A question about input dimensions #3
Comments
For pytorch built-in transformer module, you can input with the dimension order of Here is the doc for torch.nn.Tranformer |
oh, I got it. Thanks a lot for your reply! |
And I also wander that why seq_len you chose is 512 instead of 1024 or 256. I would appreciate it if you could reply. |
@DaiZhenrong So, in general, the model performs better with a longer sequence length as long as your GPU memory permits. |
Ok, I see. Thank you for your prompt reply, help and work! |
Hi Ian Shih,
Thank for your amazing job.
In my memories,the input dimesions of Transformer is [batch_size, seq_len,d_model]. So I wonder that why your input dimesions of model is seq_len, batch_size, d_model?
I will appreciate that if I can your reply.
Thanks.
The text was updated successfully, but these errors were encountered: