Skip to content

MokkeMeguru/transformer

Repository files navigation

roadmap

  • [X] dataset
  • [X] metrics
  • [X] losses
  • [X] optimizers
  • [ ] conf
  • [8/8] models
    • [X] Encoder
    • [X] Decoder
    • [X] MultiheadAttention
    • [X] PositionwiseFeedForward
    • [X] ScaledDotProdAttention
    • [X] PositionalEncoding
    • [X] Embedding
    • [X] share proj weight
  • [X] task
  • [X] data
  • [X] hydra conf
  • [X] optuna settings
  • [ ] mlflow settings

dataset form

  • train/dev/test.en
    this is a example .
    hello !
        
  • train/dev.ja
    これ は 例 です 。
    こんにち は 。
        

developper’s guide

unittest

python -m unittest discover

single run to test one model

python task.py --help
python task.py [some argument for update parameters]

multirun to change model

ref. hydra (facebook)

python task.py --help
python task.py [some argument for update parameters with comma] -m

tune with optuna

ref. optuna (PFN)

python gen_conf.py [some argument for update parameters] > optuna_args.yaml
# (edit some line in tuning.py for your tuning parameter)
python tuninig.py

Thanks

https://github.com/jadore801120/attention-is-all-you-need-pytorch

About

transformer for jap <-> en with hydra + optuna + tensorboardx (pytorch)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages