Skip to content

Commit

Permalink
Update wmt.md (#972)
Browse files Browse the repository at this point in the history
  • Loading branch information
pappagari committed Nov 7, 2021
1 parent 35dd717 commit ec63500
Showing 1 changed file with 5 additions and 2 deletions.
7 changes: 5 additions & 2 deletions docs/tutorials/wmt.md
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,8 @@ Before we start training we will prepare the training data by splitting it into
python -m sockeye.prepare_data \
-s corpus.tc.BPE.de \
-t corpus.tc.BPE.en \
-o train_data
-o train_data \
--shared-vocab
```
While this is an optional step it has the advantage of considerably lowering the time needed before training starts and also limiting the memory usage as only one shard is loaded into memory at a time.

Expand All @@ -98,7 +99,9 @@ python -m sockeye.train -d train_data \
--max-seq-len 60 \
--decode-and-evaluate 500 \
--use-cpu \
-o wmt_model
-o wmt_model \
--shared-vocab \
--max-num-epochs 3
```

This will train a "base" [Transformer](https://arxiv.org/abs/1706.03762) model.
Expand Down

0 comments on commit ec63500

Please sign in to comment.