Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

Pretrained word embeddings #5

Closed
YunwonTae opened this issue Feb 14, 2019 · 1 comment
Closed

Pretrained word embeddings #5

YunwonTae opened this issue Feb 14, 2019 · 1 comment

Comments

@YunwonTae
Copy link

YunwonTae commented Feb 14, 2019

First, thanks for sharing your code!

I really appreciate it.

I have a question about pre-trained word embeddings for unsupervised NMT task.

While reviewing code, I could find out that you guys never used pre-trained word embeddings.
(since --reload_emb is empty)

If this is true that pre-trained word embeddings has not beed used, is there a specific reason for not using pre-trained word embeddings?

Thank You!

@glample
Copy link
Contributor

glample commented Feb 14, 2019

Hi,

In the paper we compared pretraining the full model (--reload_model) vs pretraining embeddings only (--reload_emb) and we found that pretraining the embeddings doesn't work as well, so we didn't put any examples of this in the repo. You could pretrain the embeddings before you pretrain the language model, but this wouldn't be useful since training the cross-lingual language model already provides high quality cross-lingual embeddings.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants