-
Notifications
You must be signed in to change notification settings - Fork 167
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Baker Loss在18左右,声音没有预训练模型的干净 #36
Comments
非常抱歉给你带来了不好的体验,我没有足够的时间去做充分的验证,预训练模型我是训练好没有loss_kl_r的模型后,然后使用loss_kl_r继续训练模型的;发音问题是pypinyin的错误,不是模型的问题。sorry!!! |
@startreker-shzy 我想起来了,这个噪声我遇到过,就是打开fp16会这样 |
感谢大佬的回复。发音问题是pinyin问题,不是大问题。 |
2,保持false就好 |
好的,谢谢分享。我先试一下分开训练。 |
不用改了 |
好的,感谢大佬~ |
@startreker-shzy 感谢分享填坑经验 |
@MaxMax2016 |
楼主你好,请问一下怎么去掉loss_kl_r呀?另外,我在微调时,显示no teacher model. 是什么原因呢?这是不是表示迁移失败? |
感谢大佬提供的训练代码及预训练模型;
我用Baker的数据跑了一遍pipeline,batch_size设定为32,目前训练到了200k step:
基本上Loss在18左右,很难再下降了;
bert_vits INFO loss_disc=2.473, loss_gen=2.393, loss_fm=5.935
bert_vits INFO loss_mel=19.083, loss_dur=0.133, loss_kl=0.908
bert_vits INFO loss_kl_r=1.710
bert_vits INFO Train Epoch: 659 [55%]
bert_vits INFO [200200, 9.205765022545685e-05]
bert_vits INFO loss_disc=2.600, loss_gen=2.050, loss_fm=5.560
bert_vits INFO loss_mel=18.343, loss_dur=0.120, loss_kl=0.851
bert_vits INFO loss_kl_r=1.387
结果链接: https://pan.baidu.com/s/11_qTi-ubfLoGOjZu565ymQ 提取码: 1sg8
除了偶尔的发音问题,感觉音质不错。但是对比预训练的模型,声音感觉不够干净,高频内容有点多的样子。请问这个gap是什么原因呢?
The text was updated successfully, but these errors were encountered: