Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can I use another language without re-training? #14

Closed
kikirizki opened this issue Oct 7, 2021 · 2 comments
Closed

Can I use another language without re-training? #14

kikirizki opened this issue Oct 7, 2021 · 2 comments

Comments

@kikirizki
Copy link

Hi I wonder if this method can be used for any language, without retraining ? thank you

@Wendison
Copy link
Owner

Wendison commented Oct 7, 2021

Hi, that's a good question :) According to my testing, if the model is trained on one language, e.g., English, it can be also applied to Chinese but the performance is not as decent as English, e.g., the original content of Chinese utterances may not be well preserved. The key of successful conversion for unseen languages is that the VQCPC-based content encoder can accurately discover acoustic units related to underlying linguistic content or pronunciations of speech. Different languages may share some similar pronunciations, but there're still many different articulation units among different languages. Therefore, if the model is trained on multiple languages (i.e., letting the content encoder to discover more articulation units), it should have better generalization ability to unseen languages.

@kikirizki
Copy link
Author

Thank you for your brief response and suggestion, I will try it

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants