This repository contains code to reproduce the results from:
- Chrupała, G., Gelderloos, L., & Alishahi, A. (2017). Representations of language in a model of visually grounded speech signal. ACL. arXiv preprint: https://arxiv.org/abs/1702.01991
First, download and install funktional version 0.6: https://github.com/gchrupala/funktional/releases/tag/0.6
Second, install the code in the current repo:
python setup.py develop
You also need to download and unpack the files
models.tgz from http://doi.org/10.5281/zenodo.804392.
The files in
models.tgz contain the pre-trained models used for the analyses in the paper.
After unpacking these files you should have the directories
For the scripts in analysis, you also need to install:
In order to re-train one of the models, change to the corresponding directory in experiments, and execute:
python2.7 run.py > log.txt