Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Trainer trains the Model in-place #526

Closed
n1t0 opened this issue Nov 13, 2020 · 0 comments · Fixed by #519
Closed

Trainer trains the Model in-place #526

n1t0 opened this issue Nov 13, 2020 · 0 comments · Fixed by #519
Labels
enhancement New feature or request

Comments

@n1t0
Copy link
Member

n1t0 commented Nov 13, 2020

Current state

When training a Tokenizer, its Model gets replaced after training since the Trainer generates a new Model. This has several limitations:

  • When the Model being replaced has been customized (dropout, unk_token, ...), we lose all of this when we replace it (cf BPE dropout not working as expected #201)
  • In Python, if we keep a reference to the model added to the Tokenizer, this reference does not point to the actual model used by the Tokenizer after training.

Goal

Change the Trainer to actually train to Model in-place.

This was referenced Nov 13, 2020
@n1t0 n1t0 added the enhancement New feature or request label Nov 13, 2020
@n1t0 n1t0 closed this as completed in #519 Nov 20, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant