Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproduce the results of the paper, but there may be something wrong with my pre-processing steps. #1

Open
lingxiaoxue opened this issue Oct 13, 2023 · 0 comments

Comments

@lingxiaoxue
Copy link

lingxiaoxue commented Oct 13, 2023

I used L=3, K=6 WMT2015 de-en Transformer-Base to reproduce the results of the paper, but the bleu value was nearly 1.64 lower than the original paper.
image

As shown in the figure above, the result in the paper is bleu 29.29, but the reproduced result is bleu 27.65.
The training and inference steps are the same as those on github, so I think there may be something wrong with my pre-processing steps.

Refer to prepare-wmt14en2de.shhttps://github.com/ictnlp/HMT/blob/main/examples/translation/prepare-wmt14en2de.sh), change it to wmt15, and delete lines 114-118; As the paper, I use newstest2013 (3000 pairs) as the validation set and newstest2015 (2169 pairs) as the test set.
image

and BPE_TOKENS =32000;The rest of the steps remain the same.
python $BPEROOT/learn_bpe.py -s $BPE_TOKENS < $TRAIN > $BPE_CODE

image

Finally, perform length filtering:
perl $CLEAN -ratio 1.5 $tmp/bpe.train $src $tgt $prep/train 1 250
perl $CLEAN -ratio 1.5 $tmp/bpe.valid $src $tgt $prep/valid 1 250

Before pre-processing:
image
After pre-processing:
image

Can you help analyze the problem, or provide a pre-processing script?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant