Skip to content

Latest commit

 

History

History
53 lines (38 loc) · 1.85 KB

README.md

File metadata and controls

53 lines (38 loc) · 1.85 KB

3AM: An Ambiguity-Aware Multi-Modal Machine Translation Dataset

We introduce 3AM: an ambiguity-aware multimodal machine translation dataset with ~26K image-text pairs for multimodal machine translation. Compared with previous MMT datasets, our dataset encompasses a greater diversity of caption styles and a wider range of visual concepts. Please check out our paper for more details.

Download

Please download the dataset at here. The text data is also available at the data folder.

Training

Selective Attention

The code for training the selective attention model is available here, which is base on fairseq-mmt.

# train
bash train_mmt.sh
# test
bash translate_mmt.sh

VL-Bart, VL-T5

The code for training the VL-Bart and VL-T5 model is available here, which is based on VL-T5.

# VL-Bart
bash scripts/MMT_VLBart.sh
# VL-T5
bash scripts/MMT_VLT5.sh

Contact

If you have any questions, please email yc27434@umac.mo.

Citation

If you use this dataset in your research, please cite:

@inproceedings{ma-etal-2024-3am-ambiguity,
    title = "3{AM}: An Ambiguity-Aware Multi-Modal Machine Translation Dataset",
    author = {Ma, Xinyu and Liu, Xuebo and Wong, Derek F and Rao, Jun and Li, Bei and Ding, Liang and Chao, Lidia S and Tao, Dacheng and Zhang, Min},
    booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
    month = may,
    year = "2024",
    address = "Torino, Italia",
    publisher = "ELRA and ICCL",
    url = "https://aclanthology.org/2024.lrec-main.1",
    pages = "1--13",
}