New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model] How to evaluate Idefics Model's ability with in context examples? #25803
Comments
Hi @Luodian ,
To perfectly match the format used during the training of the instructed versions, you should modify slightly the template you are showing: beyond the additional
We use the <end_of_utterance> in the dialogue setup to have an easier exit condition. it marks both the end of a user and assistant turn. We found that not having this token makes it harder in a dialogue setup to stop the generation. |
Thanks! Then for in context examples, should it be like?
|
No need for double line breaks but otherwise, it is correct, that is the most straightforward way to do in-context evaluation |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
Hi the recent release of Idefics-9/80B-Instruct model is superbly promising!
We would like to evaluate them on a customized benchmarks with in context examples. May I ask how should I arrange the prompt template, especially for
instruct
version?We had some problems previously when evaluating the model on single images, the model will ramble and wont stop, but managed to resolve them somehow.
For single image we use the template to evaluate instruct version model.
Would it be perfectly correct (matching your training template?) or do you have better recommendation. Sorry we have a customized pipeline so it's not easy to adopt your designed
IdeficsProcessor
. 😭Also we migrate the code on
image_attention_mask
withI have read all related blogs and docs but still got confused about the usage of
<end_of_utterance>
. Is it used to break the in context examples with query example?My guess is
Besides, very curious that the model would generate the normal
<end_of_utterance>
at the last of sentence instead of normal llama's<|endofchunk|>
?The text was updated successfully, but these errors were encountered: