New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
regarding hierarchical representation pattern #2
Comments
In libs/models/tcn.py(L472-L473),
where num_layers=10, which means that the last attention module has the windows of 512. If you want to reproduce the ablation study, just replace the 2**i with 512, so that each layer will have windows of 512. |
@ChinaYi thank you for your answer,
|
|
thanks for the hints! appreciate it! |
In the paper it mentioned that
To demonstrate its effectiveness, we conduct a non-hierarchical version by setting the size of the local window in all attention layers to 512, which is the largest window size in the last encoder/decoder block (out of memory if we directly set the size of the local window in all attention layers to the video length).
I was not able to figure out where in the code the attention layer has windows of
512
. can you please point me to the right direction?The text was updated successfully, but these errors were encountered: