Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question on vocabulary size of the Chinese poem dataset #18

Open
zl1zl opened this issue Aug 4, 2018 · 1 comment
Open

Question on vocabulary size of the Chinese poem dataset #18

zl1zl opened this issue Aug 4, 2018 · 1 comment

Comments

@zl1zl
Copy link

zl1zl commented Aug 4, 2018

I'm trying to reproduce the Poem BLEU-2 result in the SeqGan paper, but I couldn't find out the vocabulary size used in the paper. In the RankGan paper, it uses a different dataset with size of 13,123 poems and filters out the words that occurs less than 5 times. Do you know the vocabulary size used in the SeqGan paper? Thanks a lot!

@YongfeiYan
Copy link

Hi, have you reproduced the results? I tried use all words in training data, but got BLEU-2 ~0.394 for MLE, which is lower than reported. Besides, what is the configuration of your SeqGAN model? lstm_hidden_size 32, emb_dim 32?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants