Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

masked_lm_accuracy is low at 0.51, but next_sentence_accuracy is high at 0.93 #557

Closed
SeekPoint opened this issue Apr 6, 2019 · 4 comments

Comments

@SeekPoint
Copy link

how to explain that,

my training set about 1M line, runing 50000 steps, batchsize is 32

@KavyaGujjala
Copy link

Hi, Did you solve this issue?

I am pretraining bert on domain specific dataset ( 1 million sentences, all as one document ) masked lm accuracy doesnt go beyond 70%

Tried running for 100000 steps also, didnt help much.

What to do about this?

I am pretraining to get sentence level embeddings and compare the similarities between them.

@SeekPoint
Copy link
Author

acturally, after 300000 steps, masked_lm_accuracy is low at 0.88
so, there is no problemn, just not enough training steps

@maggieezzat
Copy link

Hello, my data isn't that large: 720 Million words only, so 300,000 steps with sequence length 128 would mean 50 epochs for me, would that lead to overfitting? Also what learning rate and warmup steps did you use? @lovejasmine

@linWujl
Copy link

linWujl commented Oct 24, 2019

Hi, I follow the guide for pre-trained, which use sample_text.txt, but got low mlm accuracy and high nsp accuracy, do you know why?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants