Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Jump in sub_loss/train_dur_loss_step #11

Open
vn09 opened this issue Dec 4, 2023 · 6 comments
Open

Jump in sub_loss/train_dur_loss_step #11

vn09 opened this issue Dec 4, 2023 · 6 comments

Comments

@vn09
Copy link

vn09 commented Dec 4, 2023

Hi @p0p4k ,

I hope this message finds you well. I am currently working on training the pflowtts model with my own dataset and have encountered an unexpected behavior that I'm hoping to get some assistance with.

During training, I've observed significant jumps in the sub_loss/train_dur_loss_step metric, as illustrated in the screenshot below:

Screenshot 2023-12-04 at 17 54 38

I have followed the recommended setup and training guidelines, but I am unsure what might be causing these fluctuations. Here are some details about my training configuration and dataset:

   batch_size: 64
   n_spks: 1
   ...
  data_statistics:
    mel_mean: -6.489412784576416
    mel_std: 2.281172275543213

I would greatly appreciate it if you could provide any insights or suggestions that might help resolve this issue. Perhaps there are known factors that could lead to such behavior or additional steps I could take to stabilize the training loss?

@p0p4k
Copy link
Owner

p0p4k commented Dec 5, 2023

Could be some anomaly in the dataset. Don't worry about it, let the model train and check inference, that is when you start debugging

@rafaelvalle
Copy link

rafaelvalle commented Dec 5, 2023

To find problematic samples, one can generate transcriptions with whisper v3 and compare them with the transcriptions in the data by looking for samples with high edit distance for example.

@p0p4k
Copy link
Owner

p0p4k commented Dec 5, 2023

@rafaelvalle [unrelated question] given any sliced prompt from the target mel, since they all are supposed to give out the same target mel; is there a way to add some loss for this in one forward pass while using multiple slices for the same target mel? Thanks.

@vuong-ts
Copy link

vuong-ts commented Dec 8, 2023

Thanks @rafaelvalle @p0p4k . I use the trained model to run again on training data to filter out outlier samples (dur_loss > 5) and it helps. The loss of new train is smooth now.

@rafaelvalle
Copy link

rafaelvalle commented Dec 8, 2023 via email

@vn09
Copy link
Author

vn09 commented Dec 8, 2023

The issues mostly are:

  • incorrect transcription like wrong text inserted at the beginning
  • long pause at the end.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants