Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hybrid Self-Attention Network for Machine Translation #1

Open
zomux opened this issue Jan 28, 2019 · 0 comments
Open

Hybrid Self-Attention Network for Machine Translation #1

zomux opened this issue Jan 28, 2019 · 0 comments

Comments

@zomux
Copy link
Contributor

zomux commented Jan 28, 2019

Main Authors / Organization

Kaitao Song, Xu Tan, Furong Peng, Jianfeng Lu

Nanjing University of Science and Technology
Microsoft Research
Institute of Big Data Science and Industry, Shanxi University

PDF link

https://arxiv.org/pdf/1811.00253.pdf

Hypothesis

Adding relative positional information to self-attention improves Transformer.

Approach

Using directional mask and local mask then fuse the representations together in the self-attention mechanism.

Main Experimental Result

WMT14 De-En Translation task: + 0.4 BLEU
WMT17 Ch-En Translation task: +1.0 BLEU

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant