-
Notifications
You must be signed in to change notification settings - Fork 418
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] Special tokens are still mismatched. #715
Comments
Please install the correct transformers version as described in the readme and requirements file: Line 3 in bd57ff3
Lower versions of transformers cannot correctly identify the id set in |
yes, the final reason is
Thanks for reply! |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Describe the bug
This commit added additional_special_tokens, which seems result in mismatch of tokenizer length and vocablary size in my
transformers==4.31.0
version (although < 4.34).It seems that the additional special tokens are made new ids, which is mismatched with the input_embeddings. But this pr seems to resolve the bug in 4.33.2 as described in this issue.
Environment
I'm still not sure. It seems that
transformer==4.31.0
requiresrevision="f7dc28191037a297c086b5b70c6a226e2134e46d"
for from_pretrained.Other information
No response
The text was updated successfully, but these errors were encountered: