-
Notifications
You must be signed in to change notification settings - Fork 260
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how about generalization of sincNet ? #13
Comments
In the studies that we have conducted so far (for both speaker and speech recognition) we have done training and test on the same dataset (as often done in the speech/speaker recognition community). We didn't check so far cross-dataset performance. I think the behavior is similar to that of standard neural networks: if the datasets A and B are similar you might have good performance, otherwise, you likely have a performance drop. We are currently working on some adaptation strategies (that will probably be an object of a future paper) in order to perform a quick unsupervised adaptation of SincNet to conditions very different from that seen during training. The big advantage of SincNet if that the first convolutional layers is based on a few hundreds of learnable parameters only, while standard CNNs are based on thousands of them. This feature could make adaptation much easier! |
alright, hope you successfully. |
I appreciated your sinc_conv. |
nevertheless, sentence embedding a little more in SincNet.forward in dnn_models.py |
First, thanks for your contribution.
In your expriment, Classification Error Rates - CER% for speaker-id task and Equal Error Rate - EER% for speaker verification used, howeve, at present, deep feature representition and similarity scores were used to speaker recognition, so whether if should explain generalization of sincNet from A dataset training to B dataset testing.
The text was updated successfully, but these errors were encountered: