-
Notifications
You must be signed in to change notification settings - Fork 25.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
if crf needed when do ner? #444
Comments
A CRF gives better NER F1 scores in some cases, but not necessarily in all cases. In the BERT paper, no CRF is used and hence also no CRF in this repository. I'd presume the BERT authors tested both with and without CRF and found that a CRF layer gives no improvement, since using a CRF is kind of the default setting nowadays. |
Issue #64 is a good reference for discussion on NER. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
This repository have showed how to add a CRF layer on transformers to get a better performance on token classification task. |
your code does not work |
* Add dinov2 models * Add `BitImageProcessor` * Update list of supported models
If crf needed when do ner? In BertForTokenClassification, just Linear is used to predict tag. If not, why?
The text was updated successfully, but these errors were encountered: