Skip to content
forked from e-bug/iglue

Code and data for our paper "IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages".

License

Notifications You must be signed in to change notification settings

danoneata/iglue

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

IGLUE: The Image-Grounded Language Understanding Evaluation Benchmark

This is the implementation of the approaches described in the paper:

Emanuele Bugliarello, Fangyu Liu, Jonas Pfeiffer, Siva Reddy, Desmond Elliott, Edoardo Maria Ponti, Ivan Vulić. IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages. arXiv 2022; abs/2201.11732.

We provide the code for reproducing our results, preprocessed data and pretrained models.

IGLUE models and tasks will also be integrated into VOLTA, upon which our repository was origally built.

Repository Setup

To set the environment to reproduce our results, see "Repository Setup" in the VOLTA's README.

Data

datasets/ contains the textual data for each dataset.

Check out its README for links to preprocessed data

Features extraction steps for each of dataset and backbone can be found under features_extraction/.

Models

The checkpoints of all the pretrained V&L models can be downloaded from ERDA.

For more details on defining new models in VOLTA, see volta/MODELS.md.

Model configuration files are stored in volta/config/.

Training and Evaluation

We provide the scripts we used to train and evaluate models in experiments/:

  • zero_shot/: English fine-tuning and zero-shot/`translate test' evaluation
  • few_shot/: Few-shot experiments for each dataset-language-shots triplet
  • few_shot.dev-mt/: Few-shot experiments when using dev sets in the target languages (MT)
  • translate_train.de/: `Translate train' experiments on xFLickr&CO in German
  • translate_train.ja/: `Translate train' experiments on xFLickr&CO in Japanese

Task configuration files are stored in config_tasks/.

License

This work is licensed under the MIT license. See LICENSE for details. Third-party software and data are subject to their respective licenses.
If you find our code/data/models or ideas useful in your research, please consider citing the paper:

@article{bugliarello-etal-2022-iglue,
    title = "{IGLUE}: {A} Benchmark for Transfer Learning across Modalities, Tasks, and Languages",
    author="Bugliarello, Emanuele and 
        Liu, Fangyu and 
        Pfeiffer, Jonas and 
        Reddy, Siva and 
        Elliott, Desmond and 
        Ponti, Edoardo Maria and 
        Vuli{\'c}, Ivan",
    journal = "arXiv preprint arXiv:2201.11732"
    year = "2022",
    url = "https://arxiv.org/abs/2201.11732",
}

About

Code and data for our paper "IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages".

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Shell 32.4%
  • Jupyter Notebook 32.3%
  • Python 29.3%
  • C++ 2.5%
  • Cuda 2.5%
  • MATLAB 0.4%
  • Other 0.6%