-
Notifications
You must be signed in to change notification settings - Fork 732
Jni layer use multimodal runner #13825
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/13825
Note: Links to docs will display an error until the docs builds have been completed. ❌ 3 New Failures, 62 PendingAs of commit bb2bfe0 with merge base b759ae8 ( NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
jackzhxng
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice, how did you test?
Honestly not sure how to test multimodal. I can just check for llama regression for now. Maybe llava? Do you have idea? |
|
Yeah, I think just checking Llama for regression would be good for now |
Validated |
|
@kirklandsign Is there a test for jni layer? |
No. We use E2E CI. However, I have trouble with validating this. Multimodal runner requires method |
|
I have some difficulty validating this. Getting error executorch/extension/llm/runner/multimodal_prefiller.cpp Lines 110 to 111 in 41c299f
|
|
@kirklandsign is this for text prefill or image prefill? If it's text prefill, then you should check if the encoder embeddings has a batch dimension |
text. |
|
Now the part I'm not sure is
which is different from llava runner
Otherwise we can merge this PR |
No description provided.