Skip to content

gchrupala/visually-grounded-speech

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

70 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

visually-grounded-speech

This repository contains code to reproduce the results from:

  • Chrupała, G., Gelderloos, L., & Alishahi, A. (2017). Representations of language in a model of visually grounded speech signal. ACL. arXiv preprint: https://arxiv.org/abs/1702.01991

Installation

First, download and install funktional version 0.6: https://github.com/gchrupala/funktional/releases/tag/0.6

Second, install the code in the current repo:

python setup.py develop

You also need to download and unpack the files data.tgz and models.tgz from http://doi.org/10.5281/zenodo.804392. The files in models.tgz contain the pre-trained models used for the analyses in the paper.

After unpacking these files you should have the directories data and models.

For the scripts in analysis, you also need to install:

  • sklearn
  • pandas
  • matplotlib
  • keras

Usage

Training models

In order to re-train one of the models, change to the corresponding directory in experiments, and execute:

python2.7 run.py > log.txt

Analysis

See analysis/README.md

About

Representations of language in a model of visually grounded speech signal.

Resources

License

Stars

Watchers

Forks

Packages

No packages published