-
Notifications
You must be signed in to change notification settings - Fork 414
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
question, performance not match with the paper and feature generation #6
Comments
meixitu
changed the title
performance not match with the paper
question, performance not match with the paper and feature generation
Jan 4, 2018
Hi @zhangjinhong17
|
Hi @navsuda ,
Thanks for your other reply. Thanks |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hi,
Thanks for your wonderful work. It really help me much.
I have several question about this project
1 , I run the code with your train_commands.txt, and I found the performance is a little worse than the result in the paper, Table 7. For DS-CNN, small model, the highest validation accuracy is 92.98% in codee, and it is 93.6% in the paper.
My question is, do you get the Table 7 performance with the same code setting?
It does not use the model when validation dataset accuracy is maximum.
do you calculate the test accuracy in the paper with the same method?
3, did you compare the performance of LBFE vs MFCC? in google's paper, it use LFBE. But MFCC can use small feature, you only use 10 MFCC features. If we use more MFCC features, can we get the higher performance?
4, do you consider the feature normalization to compatible with the different signal power range?
5, if some frame, the signal power is zero, how do you calculate the log(LBFE)? I can't see it in the code. In general, it will use log(LBFE+delta), delta is a constant small value. what is the delta value?
but in DS-CNN, you use window_size=40ms and window_stride=20ms
big window_stride can reduce the OPERATIONS, I can understand.
But I don't know why use 40ms window_size, for 16k input sample rate, it should use 1024 FFT, it is power consuming.
7, I run the simulation, almost need 4 hours. I use GPU Geforce 1080 TI and CPU E5-2650. But I saw in your other reply that you only need 1 hours to complete the simulation. Is there any way to speed up? I found the feature generation use most of the time.
Thanks
Jinhong
The text was updated successfully, but these errors were encountered: