Skip to content
Multilingual acoustic word embedding approaches applied and evaluated on GlobalPhone data.
Jupyter Notebook Python Other
Branch: master
Clone or download

Latest commit

Fetching latest commit…
Cannot retrieve the latest commit at this time.


Type Name Latest commit message Commit time
Failed to load latest commit information.
embeddings Added utility script Mar 30, 2020

Multilingual Acoustic Word Embeddings on GlobalPhone


Multilingual acoustic word embedding approaches are implemented and evaluated on the GlobalPhone corpus. The experiments are described in:

  • H. Kamper, Y. Matusevych, and S.J. Goldwater "Multilingual acoustic word embedding models for processing zero-resource languages," in Proc. ICASSP, 2020. [arXiv]

Please cite this paper if you use the code.


The code provided here is not pretty. But I believe that research should be reproducible. I provide no guarantees with the code, but please let me know if you have any problems, find bugs or have general comments.

Download datasets

The GlobalPhone corpus and forced alignments of the data needs to be obtained. GlobalPhone needs to be paid for. If you have proof of payment, we can give you access to the forced alignments. Save the data and forced alignments in a separate directory and update the file to point to the data directories.

Install dependencies

You will require the following:

To install speech_dtw (required for same-different evaluation) and shorten (required for processing audio), run ./

Extract speech features

Update the paths in to point to the data directories. If you are using docker, will already point to the mounted directories. Extract MFCC features in the features/ directory as follows:

cd features
./ SP

You need to run for all languages; run it without any arguments to see all 16 language codes.

Evaluate frame-level features using the same-different task

This is optional. To perform frame-level same-different evaluation based on dynamic time warping (DTW), follow samediff/

Obtain downsampled acoustic word embeddings

Extract and evaluate downsampled acoustic word embeddings by running the steps in downsample/

Train neural acoustic word embeddings

Train and evaluate neural network acoustic word embedding models by running the steps in embeddings/

Analyse embedding models

Analyse different properties/aspects of the acoustic word embedding models by running the steps in blackbox/

Query-by-example search

Perform query-by-example search experiments by running the steps in qbe/

Unit tests

In the root project directory, run make test to run unit tests.




The code is distributed under the Creative Commons Attribution-ShareAlike license (CC BY-SA 4.0).

You can’t perform that action at this time.