Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MaskedLM for longer texts #8

Closed
AsmaBaccouche opened this issue Jun 23, 2022 · 1 comment
Closed

MaskedLM for longer texts #8

AsmaBaccouche opened this issue Jun 23, 2022 · 1 comment

Comments

@AsmaBaccouche
Copy link

AsmaBaccouche commented Jun 23, 2022

Is it possible to apply the same logic on Roberta for the MaskedLM model? I need it for pretraining on a custom dataset that has long texts - Thanks

@MichalBrzozowski91
Copy link
Collaborator

Hi, the method we used here is applied on already pre-trained model during fine-tuning stage. I am not sure if it can be applied during pre-training stage. Maybe it could be better to train the models with architecture modified for longer texts like BigBird or Longformer if you want to pre-train it from scratch.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants