Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tokenizer #64

Closed
pepi99 opened this issue Apr 30, 2022 · 2 comments
Closed

Tokenizer #64

pepi99 opened this issue Apr 30, 2022 · 2 comments

Comments

@pepi99
Copy link

pepi99 commented Apr 30, 2022

Hello!
Since FinBERT is further trained on BERT, and the original BERT has its own tokenizer, I suppose you are not training your own tokenizer to train the model, right? Otherwise, the model might get confused since the original model was trained with a different tokenizer.

@doguaraci
Copy link
Member

Yes, we’re using BERT tokenizer.

@pepi99
Copy link
Author

pepi99 commented May 4, 2022

This is very useful to know. Thank you and have a good day.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants