Skip to content

cltl/clin26-eval

Repository files navigation

CLIN-26: CONLL NERC, NED, COREF, and Factuality evaluation

Terminology

Key and response:

  1. key: expected output
  2. response: system output

The folder dev_corpora contains six folders:

  1. factuality: factuality data
  2. ne: named entity data
  3. e: entity data
  4. coref_ne: named entity coreference
  5. coref_event: event coreference
  6. coref: named entity and event coreference

Requirements

The scripts scoring NED, NERC and Factuality make use of numpy and pytest. They are tested on Python 2.7.

General instructions

  1. Put your response files in one flat directory (dev corpus and test corpus will be provided in this repository)
  2. Make sure the files in the response directory have the same name as those in key directory (otherwise, they will be ignored)
  3. Response and key files need to be in the CONLL 2011 format
  4. A response file needs to be provided for each key file.
  5. Run the score-XXX.py or score-XXX.sh script from the command line. Run each script without commands to see information about how to run the script.

Reports are written to default stdout and stderr.

Task-specific instructions

Entity coreference

bash score_coref.sh coref_ne key response measurement

The development key folder can be found at: /dev_corpora/coref_ne.

The measurement is always: blanc.

Event coreference

bash score_coref.sh coref_event key response measurement

The development key folder can be found at: /dev_corpora/coref_event.

The measurement is always: blanc.

Entity and event coreference

bash score_coref.sh coref key response measurement

The development key folder can be found at: /dev_corpora/coref.

The measurement is always: blanc.

Named-Entity Recognition and Classification (NERC)

python score_nerc.py key response

Named-Entity Disambiguation (NED)

python score_ned.py key response

Factuality

python score_factuality.py key response