New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add wake word support #5
Comments
+1 on this @AlexxIT any plans to add support for this? Seems like the missing piece |
Without having looked at the code, you can probably test it out by changing this line locally to
|
See the updated Assist docs here for wake word detection: https://developers.home-assistant.io/docs/voice/pipelines/#wake-word-detection You will also have the option of passing in parameters for audio enhancement, so HA can clean up noise and boost the volume if needed. |
So I made a fork here, https://github.com/starsoccer/StreamAssist, and tried to get this working but did not have much luck. I will prefix all of the following by saying while I am a developer I dont really know python but anyways theory/details below. So I made the change mentioned by @balloob to use
@synesthesiam As long as your on this thread, are there any plans to support something like this natively in HA? I think a lot of people likely have cameras/microphones and just having a simple UI that lets you make an |
I guess something like these changes need to be done: home-assistant/core@7856189#diff-4ff817d7964242e3c079f2f2799985713b8a4983de705b6fcf620542fe5897ff |
Oh good find, yeah that shows it was previously taking a chunk of aaudio but now it wants a float for the chunk seconds. Honestly I am not sure if this is even the right function to use anymore. I dont really understand the order things should happen in but my thinking is that maybe I should instead be calling Whats not really clear to me is if I continue to use this process function and instead pass in the chunk time, how its going to actually get the audio chunk as I dont see it being passed in anywhere else. Even in the example test files it seems to just call it without any audio which I dont really get, https://github.com/home-assistant/core/blob/22c21fdc180fec24e3a45e038aba6fb685acd776/tests/components/assist_pipeline/test_vad.py#L33C48-L33C48 |
@starsoccer The Let me know if you have any more questions, since I wrote the code 😄 |
Got it, do you have an example of how the audio is passed to the function that I can maybe work from? Also any plans to build this functionality into HA directly rather then need this custom integration? |
Any updates on this so far? It would be REALLY great to have a possibility to get Wall-Panels which already have an internal mic to work as an Assist-Device with Wake-Word 😍 |
+1 would love to keep working on this but its a bit outside my skill set honestly. Hopefully HA adds/builds this feature in natively letting users specify a input device either video or microphone, and then allowing any speaker to function as an output |
Asked about this in the year of the voice chapter 5 live chat but no response (or I missed it). I was thinking this could also be integrated into frigate maybe? Seeing there's vad going on in frigate 0.13 |
Yeah I made an issue about putting it into frigate but was told this isnt planned, blakeblackshear/frigate#8644 Hopefully someone more technical then me will get this working. Ive asked in the discord and seen someone else ask about it, but so far I dont know anyone who has this working. |
Seems this can be a starting point! https://github.com/asmsaifs/StreamAssist |
Cool, I tried using it but getting this vague error, Not sure if I am missing something or not. But any value I seem to put in the url for the stream seems to give this error |
In testing in the latest main/master version |
@AlexxIT Do you have any info on how we can use the new version and test/debug it? I gave it a try but cant seem to get anything to happen. Ive tried using a RTSP url as well as a camera entity. They both seem to just get stuck in the start phase for wake and then never change. |
Reinstall via HACS with manual selecting main version tag |
I already did that |
It works flawlessly! Awesome! My tablets around the house just got superpowers! |
Thanks. Unfortunately I don't have time to do a complete test. Also, it all works just horribly in my language. |
Can you add some more troubleshooting info? For instance how to ensure its detecting voice and it actually gets past the wakeword stage/step as that is where mine seems to be stuck |
You can enable debug logs |
This should get a shout out on year of the voice part 6! I'm using this on Amazon Fire HD 10+ tablets running IP Webcam for rtsp stream, and Fully Kiosk for the media_player. And it works out of the box! (Mostly using Extended OpenAI Conversation agent, while not local the results are so impressive, having to work with intents seems puny) Small FR/Q: Should these show up as assist devices? |
I have never saw what is assist device. |
Now that core has added the wake word integration (home-assistant/core#96380) this component could make use of it after running the Voice Activity Detector.
It could be integrated using wyoming protocol, since that core integration also provides a Wyoming implementation that can be used with an openWakeWord container.
The text was updated successfully, but these errors were encountered: