This repository contains code to reproduce the results from:
- Chrupała, G., Gelderloos, L., & Alishahi, A. (2017). Representations of language in a model of visually grounded speech signal. ACL. arXiv preprint: https://arxiv.org/abs/1702.01991
First, download and install funktional version 0.6: https://github.com/gchrupala/funktional/releases/tag/0.6
Second, install the code in the current repo:
python setup.py develop
You also need to download and unpack the files data.tgz
and models.tgz
from http://doi.org/10.5281/zenodo.804392.
The files in models.tgz
contain the pre-trained models used for the analyses in the paper.
After unpacking these files you should have the directories data
and models
.
For the scripts in analysis, you also need to install:
- sklearn
- pandas
- matplotlib
- keras
In order to re-train one of the models, change to the corresponding directory in experiments, and execute:
python2.7 run.py > log.txt