Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About total data performance #2

Closed
bugczw opened this issue Apr 29, 2021 · 4 comments
Closed

About total data performance #2

bugczw opened this issue Apr 29, 2021 · 4 comments

Comments

@bugczw
Copy link

bugczw commented Apr 29, 2021

Hi, I am glad to read this article. This essay is the first work that focuses on efficiently adapting large pretrained language
models for image captioning, which inspires me a lot!
In the result display section, it mainly shows the results of training using some data sets at different sampling rates. Therefore, I would like to ask, have you tested the results on all data sets without sampling? How does the performance compare to M2Transformer?

@junchen14
Copy link
Collaborator

Hi, I am glad to read this article. This essay is the first work that focuses on efficiently adapting large pretrained language
models for image captioning, which inspires me a lot!
In the result display section, it mainly shows the results of training using some data sets at different sampling rates. Therefore, I would like to ask, have you tested the results on all data sets without sampling? How does the performance compare to M2Transformer?

hi thanks for the interest of our work.
we also tested on training with full COCO training data, but it shows our model's performance is slightly less than M2Transformer, and we have already put this result in the supplementary. We hypotheses the performance drop is due to the domain gap of GPT-2 knowledge and COCO captions.

@qizhust
Copy link

qizhust commented May 1, 2021

Hi, do you have the performance under cross-entropy training only? @junchen14

@junchen14
Copy link
Collaborator

Hi, do you have the performance under cross-entropy training only? @junchen14

We did not report the performance under cross-entropy training only, because all the SOTA results are gained under reinforcement setting, we therefore also report the performance after reinforcement.
but through our observation, our visualgpt is always better than the baselines with or without reinforcement.

@qizhust
Copy link

qizhust commented May 1, 2021

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants