Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Problems in tokenizing LibriTTS #5

Closed
jry-king opened this issue Feb 7, 2023 · 10 comments
Closed

Problems in tokenizing LibriTTS #5

jry-king opened this issue Feb 7, 2023 · 10 comments

Comments

@jry-king
Copy link

jry-king commented Feb 7, 2023

Thanks for your reproduction of the VALL-E paper! When I tried to prepare LibriTTS data with prepare.sh I encountered this problem:
image
I'm only using 4 tar files (train-clean-100, train-clean-360, test-clean and dev-clean) out of 7 in LibriTTS. Could you give me some suggestions about what's going on? Thanks!

@lifeiteng
Copy link
Owner

datasets fix a23f7c9

@jry-king
Copy link
Author

jry-king commented Feb 8, 2023

Looks like the command is fixed but the "words count mismatch" warning still exists. I only found that it might be caused by some inconsistency between text files and phonemes. Will it actually influence the result?

@lifeiteng
Copy link
Owner

I haven't figured out these warnings yet, I checked the results and there is nothing wrong.

@pandamq
Copy link

pandamq commented Feb 23, 2023

@jry-king Hi, i meet the same warning info, i assume the reason is espeak tokenizer frequently take many words into one phone, e.g. ' of the' to 'ʌvðə'.
Do you have some solution to fix it, or we can igonre the warning to train the model correctly? Thanks!

@lifeiteng
Copy link
Owner

@jry-king Hi, i meet the same warning info, i assume the reason is espeak tokenizer frequently take many words into one phone, e.g. ' of the' to 'ʌvðə'. Do you have some solution to fix it, or we can igonre the warning to train the model correctly? Thanks!

can you provide more cases?

we need dig into https://github.com/bootphon/phonemizer.

@pandamq
Copy link

pandamq commented Feb 26, 2023

@jry-king Hi, i meet the same warning info, i assume the reason is espeak tokenizer frequently take many words into one phone, e.g. ' of the' to 'ʌvðə'. Do you have some solution to fix it, or we can igonre the warning to train the model correctly? Thanks!

it's hard to count them all. i can give you some cases.
"on the floor" to "ɔnðə floːɹ"
"had been" to "hɐdbɪn"
"of a" to "əvə"
"would have" to "wʊdhɐv"
"there is" to "ðɛɹɪz"

@lifeiteng
Copy link
Owner

errors shoud be fixed

ɜː -> ɜ ː

@lifeiteng
Copy link
Owner

@jry-king @mingqilearning Thank you for report these issues. It would be great if you guys could fix these issues.

Training looks good now #37

@pandamq
Copy link

pandamq commented Mar 3, 2023

@jry-king @mingqilearning Thank you for report these issues. It would be great if you guys could fix these issues.

Training looks good now #37

i tried to fix it, but failed. training good means the warning doesn't influent the training and inference?
it would be great if you share the phomezier and espeak-ng version.

@lifeiteng
Copy link
Owner

espeak-ng/espeak-ng#1703

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants