-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Does fine-tuned model use the same vocabulary as pre-trained model? #5
Comments
yes |
appreciate |
If I want to fine-tune this model to support new languages, new tokens should be added to existing file? |
Yes, you are right, if you want to expand the supported languages, you need to merge the newly added tokens into the existing vocabulary, then randomly initialize the embedding vectors of the new tokens. |
appreciate * 2 |
The pre-trained model use learned joint vocabulary with 64867 tokens, merged by 32k operations. That is the file vocab.bpe.32000.
Does fine-tuned model (e.g. en2de) released in icloud use the same vocabulary file as pre-trained model?
The text was updated successfully, but these errors were encountered: