-
Notifications
You must be signed in to change notification settings - Fork 25.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.
Already on GitHub? Sign in to your account
AutoTokenizer not able to load saved Roberta Tokenizer #4197
Comments
It looks like when you load a tokenizer from a dir it's also looking for files to load it's related model config via
Either:
|
I had the same issue loading a saved tokenizer. Separately loading and saving the config file to the tokenizer directory worked.
|
+1 to @jaymody. It seems as though that For those that happen to be loading a tokenizer and model using the
|
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
馃悰 Bug
Information
Model I am using (Bert, XLNet ...): Roberta
Language I am using the model on (English, Chinese ...): English
The problem arises when using:
I'm trying to run
run_language_modelling.py
with my own tokenizerThe tasks I am working on is:
To reproduce
Steps to reproduce the behavior:
Stack Trace:
Expected behavior
AutoTokenizer should be able to load the tokenizer from the file.
Possible duplicate of #1063 #3838
Environment info
transformers
version:The text was updated successfully, but these errors were encountered: