Skip to content

A pytorch implementation of The Transformer from the paper "Attention Is All You Need".

Notifications You must be signed in to change notification settings

yonghee12/transformer_torch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Transformer Implementation

A Pytorch Implementation of the paper "Attention is All You Need".
I checked out several popular implementations and I have found a few points which was quite different from the original paper.
This repository is the result of fixing errors and cleaning codes in pytorch-OOP manner.

Examples

  • Trained on 20k Korean-English parallel corpus for two hours with general GPU.
  • These test sentences are not from the train corpus.
  • Hyperparameters
    • d_model = 32
    • d_ff = 128
    • n_layers = 3
    • n_heads = 2
우리 내일 어디로 갈까?
<sos> where should we go tomorrow ? <eos>
너 나 좋아하니?
<sos> do you like to go ? <eos>
이번 시험에서 내가 잘할 수 있을까요?
<sos> can i get a good job this exam ? <eos>
정말 이번에는 졸업하고 싶은데 잘 될지 걱정이야.
<sos> i want to graduate this time , but i 'm worried about you . <eos>

References

Author

About

A pytorch implementation of The Transformer from the paper "Attention Is All You Need".

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages