Skip to content
master
Switch branches/tags
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 

This is the implementation of the Temporal Attention-Gated Model, which is proposed in the following paper: Wenjie Pei, Tadas Baltrušaitis, David M.J. Tax and Louis-Philippe Morency. "Temporal Attention-Gated Model for Robust Sequence Classification", https://arxiv.org/pdf/1612.00385.pdf, which is accepted by CVPR 2017.

The code is implemented in Lua and Torch. It contains mainly the following parts:

  • main.lua: the starting point of the entire code.
  • train_process.lua: the training process.
  • evaluate_process.lua: the evaluation process.
  • package 'model' contains the required models including attention model, TAGM, LSTM, GRU and plain-RNN.
  • package 'util' contains the required small utilities such as data loader.

Data
The clean arabic data used in the paper is uploaded named 'data/arabic/window_2.t7'. You can generate the noised version by running 'util/arabic_preprocess.lua'. It will generate a noised version named 'arabic_append_noise.t7', which can be used in the main.lua for conducting the experiments for Speech Recognition Experiments in the paper.
The datasets for the Sentiment analysis and Event recognition are too large to upload to github. They can be obtained from the original data source online.

Feel free to contact Wenjie Pei (wenjiecoder@gmail.com).

About

The source code for Temporal Attention-Gated Model.

Resources

Releases

No releases published

Packages

No packages published

Languages