You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It seems that BertEncoder will be finetuned and updated during training, which is unfair compared to "Improving one-stage visual grounding by recursive sub-query construction" and "A fast and accurate one- stage approach to visual grounding".
The text was updated successfully, but these errors were encountered:
@PlumedSerpent
Thanks for you attention. We fine-tuned Bert for comparing to SQC-large (see their codebase) in refcocog.
Actually, we achieve comparable or even better results with LSTM to SQC-large (BertEncoder) in refcoco, referit, refcoco+.
Completely fair comparison to their algorithms (SQC) is not practical, e.g. they use more parameters, large resolution input, etc. Furthermore, our contribution does not lie in the language side so our paper seems orthogonal to SQC.
It seems that BertEncoder will be finetuned and updated during training, which is unfair compared to "Improving one-stage visual grounding by recursive sub-query construction" and "A fast and accurate one- stage approach to visual grounding".
The text was updated successfully, but these errors were encountered: