-
Notifications
You must be signed in to change notification settings - Fork 248
support text-only input with llama3.2-11b #1216
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchchat/1216
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 New Failure, 1 Cancelled JobAs of commit c0c8033 with merge base ec7b510 ( NEW FAILURE - The following job has failed:
CANCELLED JOB - The following job was cancelled. Please retry:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
| is_multimodal = False | ||
|
|
||
| seq_len = tokens.size(1) | ||
| seq_len = x.size(1) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What's the difference between tokens and x?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No difference, all equal to text input
Previously we pull the text input out of batch inside prefill and call it tokens
Now we pull out text input first then forward it into prefill function and call it x
The reason we change the name is other single modality models usually use x to represent the text token inputs.
|
You can also check for multimodal content like this so you don't have to check the model type explicitly: |
|
@joecummings thanks for your suggestions! |
|
The failing test is not relevant. Pushing through |
This PR enables llama3.2-11b model with text-only input.
Note that this PR is only for CLI pipeline. Will have a separate PR for openai-api update