You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, thanks for your interest. I think that's due to different versions of Llava - the older version I used when developing this repo seems to need this slicing (I'll need to double check it). But if decoding the whole generated sentence works for you, that's totally fine to just go without truncation.
Thanks for the awesome paper and repo!
I was trying out LLaVA and noticed that the model predictions were always empty strings, and was able to narrow it down to this line
VL-ICL/utils/model_inference.py
Line 76 in 6ad043d
It seems LLaVA already outputs only the tokens it generated, and not the whole context tokens + generated tokens.
The fix is quite easy, just decode the whole generated sentence, without truncating it first with the input_token_len.
Is my thinking correct or am I missing something?
The text was updated successfully, but these errors were encountered: