Skip to content

ufal/MTEQA

Repository files navigation

MTEQA

This repository is the official implementation of Just Ask! Evaluating Machine Translation by Asking and Answering Questions.

MTEQA-basic

Requirements

Our implementation is based on the code and models provided in: question_generation.

To clone this repo, run:

git clone https://github.com/ufal/MTEQA
cd MTEQA
git submodule init
git submodule update

or just:

git clone --recurse-submodules https://github.com/ufal/MTEQA
cd MTEQA

and then install required dependencies and download required models with:

pip install -r requirements.txt
./mteqa/download_models.py

Code was tested with python 3.8.

Scoring

To score the MT output, run

python mteqa_score.py --reference ref --hypothesis out --lang en > mteqa_score.tsv

Essential arguments are:

  • reference: Path to the file with reference translations, one segment per line.
  • hypothesis: Path to the file with MT system output, one segment per line.
  • lang: ISO code of the target language (23.11.2021 - only English [en] is supported).

Additional flags are:

  • --cpu: Force interference on CPU, by default GPU is used if detected.
  • --baseline_qe: Use the baseline system for Answer Extraction, see Section 3.3 - 2). By default keyword extraction based on POS pattern matching/NER is used.
  • --gen_from_out: Extract Question/Answer pairs from both the reference and the MT output, see Section 3.3 - 1).
  • --verbose: Outputs per-segment score. By default, only the single system-level score is reported.

Output is tab delimited, with a single column for each string-comparison metric that we used.

To reproduce our results from the WMT 2021 Metrics Shared Task you should use the default parameters (i.e. POS/NER based keyword extraction, questions generated only from reference) and consider the chrf metric for answer comparison.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages