Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TTS recipe #67

Closed
jimbozhang opened this issue Sep 1, 2020 · 8 comments
Closed

TTS recipe #67

jimbozhang opened this issue Sep 1, 2020 · 8 comments

Comments

@jimbozhang
Copy link
Contributor

I'm not sure if it useful to let lhotse support TTS training. If needed, I can make a TTS recipe using the LJSpeech corpus.
@danpovey @pzelasko

@danpovey
Copy link
Collaborator

danpovey commented Sep 1, 2020 via email

@pzelasko
Copy link
Collaborator

pzelasko commented Sep 1, 2020

Yes, it will be most useful. Thanks!

@jimbozhang
Copy link
Contributor Author

Just started working on this. A new recipe ljspeech will be added for end-to-end (e.g. Tacotron2) training. I'll make a PR next week.

@pzelasko
Copy link
Collaborator

pzelasko commented Sep 9, 2020 via email

@jimbozhang
Copy link
Contributor Author

Sounds great! A heads up - I made a couple of general changes in the meanwhile, e.g. the preferred manifest format is now JSON because it’s much faster to read/write.

I had a little worry: if the dataset is extremely huge, the JSON would be very large. To avoid out-of-memory, the manifest has to be splitted. I think splitting YAML is easier than JSON, because if you want to split a JSON, you have to load it into memory first.

@danpovey
Copy link
Collaborator

danpovey commented Sep 9, 2020 via email

@pzelasko
Copy link
Collaborator

pzelasko commented Sep 9, 2020

Hmm, you're raising a good point. We can add JSONL quite easily - we're already supporting two formats anyway...

I wonder if it makes sense to use a binary format for such super-large datasets. E.g. in https://github.com/huggingface/nlp/ they use Apache Arrow to traverse mmap-ed files, allowing them to iterate very large datasets with almost zero memory footprint (likely at the cost of storage size).

@danpovey
Copy link
Collaborator

danpovey commented Sep 9, 2020 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants