Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Max pooling and layer dimensions inside attention layer #30

Closed
valillon opened this issue May 16, 2021 · 1 comment
Closed

Max pooling and layer dimensions inside attention layer #30

valillon opened this issue May 16, 2021 · 1 comment

Comments

@valillon
Copy link

valillon commented May 16, 2021

  • There are a couple of max_pooling2d() layers inside the attention layer sn_non_local_block_sim() which reduce the number of local features by 4 as such downsampled_num = location_num // 4. However, no downsampling step is reported in the original paper.

  • Also, the first two sn_conv1x1() layers, which stand for Wg and Wf in the paper, have equal sizes C/8 x C, but the third one standing for Wh has C/2 x C shape, while should be also C/8 x C. Similarly the last conv layer.

Is there a reason for such discrepancies?

Related #8

@valillon
Copy link
Author

valillon commented May 26, 2021

Apparently, as mentioned here, max pooling inside the attention layer is just motivated by design-wise to save computation/memory overhead.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant