Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

translation with language models #495

Open
VictorChen2012 opened this issue Sep 20, 2019 · 4 comments
Open

translation with language models #495

VictorChen2012 opened this issue Sep 20, 2019 · 4 comments

Comments

@VictorChen2012
Copy link

I'm wondering if it is possible to combine an LM with a seq2seq model under OpenNMT-tf, e.g. shallow fusion, deep fusion or cold fusion.

Currently, vars and ops of LM decoder and seq2seq decoder are in different name scope. It's too complicated to directly load and merge two pretrained models, i.e. LM and the seq2seq model under the same name scope.

Any suggestions to the goal above?

@guillaumekln
Copy link
Contributor

Shallow fusion should be the most accessible but it may not be easy to integrate at this time. However, there are some incoming changes that should facilitate such combinations.

I'm interested in supporting shallow fusion in the near future.

@VictorChen2012
Copy link
Author

Shallow fusion should be the most accessible but it may not be easy to integrate at this time. However, there are some incoming changes that should facilitate such combinations.

I'm interested in supporting shallow fusion in the near future.

Thanks for your quick reply! I'll see if I can contribute then.

@lkluo
Copy link

lkluo commented Nov 5, 2020

Any update?

@guillaumekln
Copy link
Contributor

No one is currently working on this as far as I know.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants