Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Qwen-VL-Chat doesn't follow prompt #7

Open
RifleZhang opened this issue Apr 18, 2024 · 1 comment
Open

Qwen-VL-Chat doesn't follow prompt #7

RifleZhang opened this issue Apr 18, 2024 · 1 comment

Comments

@RifleZhang
Copy link

Following set up in https://github.com/open-compass/VLMEvalKit/tree/main for Qwen set up.

Qwen-VL-Chat directly outputs answer instead of a letter choice. Did you use any customized prompt or did post processing of model responses?

@RifleZhang
Copy link
Author

Using the default prompt in https://github.com/open-compass/VLMEvalKit/tree/main , I got 22.26 for Qwen-VL-Chat. Using different prompt or evaluation post-processing method can lead to large variance. Similar for Deepseek_vl_7b, I got 26.86 with the LLava-Next prompt provided, and 32.6 with the default prompt in VLMEvalKit.

Is there an evaluation pipeline for other models reported in the paper. I found it hard to replicate the exact number without the exact prompt.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant