Skip to content

Commit

Permalink
doc typos/minor details
Browse files Browse the repository at this point in the history
  • Loading branch information
kermitt2 committed Aug 31, 2016
1 parent 2f06b0d commit cc07cd0
Show file tree
Hide file tree
Showing 3 changed files with 4 additions and 4 deletions.
2 changes: 1 addition & 1 deletion grobid-ner/doc/Licence.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ Grobid and Grobid NER are distributed under [Apache 2.0 license](http://www.apac

Author and contact: Patrice Lopez (<patrice.lopez@inria.fr>)

For citing the tool, please refer to the github project: <https://github.com/grobid/grobid> (2014)
For citing the tool, please refer to the github project: <https://github.com/grobid/grobid> (2016)

### Datasets

Expand Down
2 changes: 1 addition & 1 deletion grobid-ner/doc/annotation-guidelines.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Annotation guidelines for GROBID Named Entities Recognition
# Annotation guidelines for GROBID NER

### Principle

Expand Down
4 changes: 2 additions & 2 deletions grobid-ner/doc/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,8 @@ GROBID NER is a Named-Entity Recogniser module for [GROBID](https://raw.github.c
GROBID NER has been developed more specifically for the purpose of further supporting post disambiguation and resolution of entities against knowledge bases such as Wikipedia.

The current models shipped with the source uses 26 Named Entity [classes](classes-ane-senses.md) and have been trained using the following dataset:
- Reuters NER [CONLL 2003](http://www.cnts.ua.ac.be/conll2003/ner/) manually annotated training data (10k words). This dataset is not public, ence not shipped with the code. In order to obtain it,
- Manually annotated extract from the Wikipedia article on World War 1 (approximately 10k words)
- Reuters NER [CONLL 2003](http://www.cnts.ua.ac.be/conll2003/ner/) manually annotated training data (10k words, 26 classes). This dataset is not public, so not shipped with the code. In order to obtain it,
- Manually annotated extract from the Wikipedia article on World War 1 (approximately 10k words, 26 classes)

The training has been completed with a very large semi-supervised training based on the Wikipedia Idilia data set.

Expand Down

0 comments on commit cc07cd0

Please sign in to comment.