-
Notifications
You must be signed in to change notification settings - Fork 228
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Precise Custom Wake Word mic to be used and additional data set training #53
Comments
I have indeed noticed a dependency on the mic used. This could be both because of the different noise frequencies it picks up, and because of the different volumes. Apart from recording samples on a few different mics, you can try recording a long audio of silence on a few other mics (almost like getting the "noise profile") and using However, this might not solve the volume issue. Thinking about it, I should probably add a feature to Edit: And you can train the same model on a new dataset just like you would expect (passing the name of the model in the command, but passing in the folder with the new dataset). Let me know if this helps. |
Thanks Matthew, i will try first with collecting noise from different mic and also plan to collect data from different mic sources. |
Have one question on precise-add-noise, will the syntax be After this I assume I should use precise-train <model.net> path to output from above |
You can always run the command with
As you can see, the 3 required arguments, in order, are:
As you may now guess, this will create a duplicated dataset with the noise added in. You would then point the data folder in |
Hi Matthew, |
The problem might be that there's been a change with the audio processing library Precise uses. Mycroft Core I think is still using the old one. Sorry I just realized this, but I think your problem should be fixed if you modify vectorizer= in ListenerParams within precise/params.py to be |
Thanks Matthew, I will try as suggested and share results here. Edit: |
I retrained my models with this setting as well. They definitely feel more correctly responsive. The training took a few more steps to get where I liked it. |
Hi Matthew, |
Awesome to hear! Closing, but let me know if you have any other issues. |
@MatthewScholefield I dont find any usage of "precise-add-noise" in training tutorial |
@EuphoriaCelestial Not removed, it's just I don't think it's ever been documented. However, we'd love contributions. Feel free to create a new page on the wiki about it. |
yeah I've never known about that command until I reached this issue. Can you provide more information on what it does and how to use in training progress? |
@EuphoriaCelestial Is there any part you'd like me to expand on? What I explained from before covers most of it:
This would be most useful to do in cases where you don't have a lot of wakewords and want to generate more variations of the data. |
@MatthewScholefield should I do this? |
@EuphoriaCelestial First, just to clarify, this only pertains to Mycroft Core. Now, in order to make it work with Mycroft Core, I think it's actually more realiable (but still a bit hacky) to basically do a source install of precise on the same platform you use mycroft core on. Then you can just link the source install engine script to where mycroft core expects it: default_engine_path=~/.mycroft/precise/precise-engine/precise-engine
# Back up default precise-engine
mv "$default_engine_path" "$default_engine_path.bak"
# Link source install to Mycroft Core
cd mycroft-precise/ # Source install location
ln -s "$(pwd)/.venv/bin/precise-engine" "$default_engine_path" If you do this you would definitely not need to modify the vectorizer. |
@MatthewScholefield I encountered this error when run precise-add-noise : I dont know what happened, just yesterday it still working fine, I generated hundreds of file using this command |
I trained model with 200 + dataset (about 15 sample from different individuals) usind method 2 to take out false negative. Now the model when tested works good with the mic i used to collect the dataset. When i use other mics then the result is different, when i use laptop mic there is lot of false positive when i use another mic then to get activation is very hard. I also exported the model to tensorflow and used it with 6 array mic and couldnt good activation.
There are 2 questions here,
The text was updated successfully, but these errors were encountered: