Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Will the Pre-trained ELECTRA-1.75M be released? #17

Closed
xf05888 opened this issue Mar 15, 2020 · 3 comments
Closed

Will the Pre-trained ELECTRA-1.75M be released? #17

xf05888 opened this issue Mar 15, 2020 · 3 comments

Comments

@xf05888
Copy link

xf05888 commented Mar 15, 2020

In the paper, it also described a model called ELECTRA-1.75M and has a better performance than ELECTRA-1.45M ( ELECTRA-large ).
So will the pre-trained ELECTRA-1.75M be released?

@vprelovac
Copy link

It looks like a typo, 1.45M is never mentioned in the paper.

@xf05888
Copy link
Author

xf05888 commented Mar 15, 2020

@vprelovac According to README.md,

The models were trained on uncased English text. They correspond to ELECTRA-Small++, ELECTRA-Base++, ELECTRA-1.45M in our paper.

And in the paper it mentioned ELECTRA-Large which is same as ELECTRA-1.45M

Also,
We therefore also report resultsfor ELECTRA-1.75M with the only trick being dev-set model

@clarkkev
Copy link
Collaborator

clarkkev commented Apr 8, 2020

Yes, that is a typo!

@clarkkev clarkkev closed this as completed Apr 8, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants