Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is BertEncoder fine-tuned on the visual grounding tasks? #4

Closed
PlumedSerpent opened this issue Aug 15, 2021 · 2 comments
Closed

Is BertEncoder fine-tuned on the visual grounding tasks? #4

PlumedSerpent opened this issue Aug 15, 2021 · 2 comments

Comments

@PlumedSerpent
Copy link

It seems that BertEncoder will be finetuned and updated during training, which is unfair compared to "Improving one-stage visual grounding by recursive sub-query construction" and "A fast and accurate one- stage approach to visual grounding".

@hbb1
Copy link
Collaborator

hbb1 commented Aug 26, 2021

@PlumedSerpent
Thanks for you attention. We fine-tuned Bert for comparing to SQC-large (see their codebase) in refcocog.
Actually, we achieve comparable or even better results with LSTM to SQC-large (BertEncoder) in refcoco, referit, refcoco+.
Completely fair comparison to their algorithms (SQC) is not practical, e.g. they use more parameters, large resolution input, etc. Furthermore, our contribution does not lie in the language side so our paper seems orthogonal to SQC.

@hbb1
Copy link
Collaborator

hbb1 commented Sep 12, 2021

I am going to close this issue. Please open more if there is any other comment.

@hbb1 hbb1 closed this as completed Sep 12, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants