Plan, Attend, Generate: Planning for Sequence-to-Sequence Models
This is a theano implementation of Plan, Attend, Generate, described in our paper:
Francis Dutil, Caglar Gulcehre, Adam Trischler, Yoshua Bengio, Plan, Attend, Generate: Planning for Sequence-to-Sequence Models (NIPS 2017)
The code for character-level translation is based on the code provided by:
Junyoung Chung, Kyunghyun Cho, Yoshua Bengio, A Character-Level Decoder without Explicit Segmentation for Neural Machine Translation: https://github.com/nyu-dl/dl4mt-cdec
To reproduce the results, theano version 0.9.0 needs to be used.