Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Asking for code component #20

Closed
trqminh opened this issue May 18, 2022 · 5 comments
Closed

Asking for code component #20

trqminh opened this issue May 18, 2022 · 5 comments

Comments

@trqminh
Copy link

trqminh commented May 18, 2022

Hi @lkeab,
Thank you for your consideration.

I see that the proposed method is mainly in mask_head.py, however, I find the code is hard to keep track. Would you mind showing me where is the code for this component below?

image

@lkeab
Copy link
Collaborator

lkeab commented May 18, 2022

Hi, thanks for watching our work Transfiner.
The sequence encoder is: class TransformerEncoder, L928 to L957.
The pixel decoder is conv_r1, L935.
The incoherent query sequence sent into the transformer for refinement is here during training, while here during inference.

@trqminh
Copy link
Author

trqminh commented May 19, 2022

Thanks, @lkeab for your answer.
Could I ask about the size of the incoherent query sequence according to those pyramid features maps?
Thank you.

image

@lkeab
Copy link
Collaborator

lkeab commented May 19, 2022

N x C, N is the number of the detected incoherent query points. The size of N depends on different objects but much smaller than total points in the RoI region.

@trqminh
Copy link
Author

trqminh commented May 19, 2022

Thank you @lkeab .

@128ve900
Copy link

May I ask whether different samples have different values ​​of N? Is the input size of the incoherent query sequence indeterminate?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants