Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

dataset size / audio generation #1

Open
markhanslip opened this issue Jun 4, 2020 · 0 comments
Open

dataset size / audio generation #1

markhanslip opened this issue Jun 4, 2020 · 0 comments

Comments

@markhanslip
Copy link

Hi, thanks for the implementation :)
Do you have suggestions for how long the input audio chunks should be and what size the dataset should be?
At the moment I'm getting weird negative discriminator loss readings, which I wondered might be because of not enough / inappropriately segmented data. I was working with ~50mins of 4 second wav chunks. Also should the input audio data be downsampled to 16000?
I've only trained as far as ~5000 epochs and at the moment there are only glitches at the very beginning and end of the generated waveform with flat lines in between - is this a normal thing for early iterations that improves in later epochs, or is something wrong?
Sorry for multiple questions!
thanks again,
Mark

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant