-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added prompt_template
preprocessing param for text features
#3298
Conversation
for more information, see https://pre-commit.ci
Unit Test Results 6 files ± 0 6 suites ±0 31m 8s ⏱️ - 1h 21m 22s Results for commit 491ab56. ± Comparison against base commit 87a56fa. This pull request removes 139 and adds 52 tests. Note that renamed tests count towards both.
This pull request removes 11 skipped tests and adds 1 skipped test. Note that renamed tests count towards both.
♻️ This comment has been updated with latest results. |
inverse_vocabulary=metadata[f"{prefix}str2idx"], | ||
tokenizer_type=preprocessing_parameters[f"{prefix}tokenizer"], |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Would you also be able to remove the rest of the backwards compatibility workaround (introduced in #1859)?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not too familiar with this code tbh, maybe you can take it in a follow-up?
Useful when fine-tuning large language models (LLMs) that benefit from providing additional context to improve the quality of the embeddings they generate. Particularly valuable when using fixed weights (
trainable=false
).