Skip to content

Seq2Seq networks (Encoder-Decoder architecture) to solve the task of Machine Translation.

License

Notifications You must be signed in to change notification settings

dddat1017/Machine-Translation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Machine-Translation

Solving the task of machine translation (from English to Vietnamese) with a regular Seq2Seq network and a global attention-based, dot product Seq2Seq network. Both following the Encoder-Decoder architecture.

Preprocessed data courtesy of the Stanford NLP group.

TODO:

  • [] Attention-based network: Compute exact loss for each sequence (i.e., don't compute loss for sequences with padding). Look into utilizing torch.nn.utils.rnn.pack_padded_sequence
  • [] Regular network: Get on track with the attention network, files are in .src/TODO/
  • [] Training: Run for more epochs
  • [] Training: Implement K-fold CV
  • [] Training: Hyperparameter tuning
  • [] Testing: Implement model evaluation for user inputs and test data
  • [] Miscellaneous: More documentation + better organization

Latest Results

Attention-based network [01/01/2021]:

Standard Seq2Seq (Encoder-Decoder architecture) network

Global attention-based, dot product Seq2Seq (Encoder-Decoder architecture) network

About

Seq2Seq networks (Encoder-Decoder architecture) to solve the task of Machine Translation.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published