Skip to content

Through this project I wanted to add to the voice assistant made in Home Assistant a visual image and random personalized responses that are displayed on an android tablet.

License

Notifications You must be signed in to change notification settings

relust/VisualStreamAssist

Repository files navigation

VisualStreamAssist

Through this project I wanted to add to AlexxIT integration StreamAssist personalized and visual responses to be played on an android tablet using the browser mod.

Pre-requisites

  • Home Assistant 2023.11.3 or newer
  • A voice assistant configured in HA with STT and TTS in a language of your choice
  • Install Browser Mod integration with HACS. The browser mod media player of android tablet will be used to stream gif files with browser_mod.popup service and audio responses with browser mod media player.
  • Install Rtpmic app on android tablet or another application to stream mic or camera with sound on tablet. If you use Rtpmic app, on default settings check auto start streaming and start at boot, target adress 255.255.255.255, port 5555and audio codec G.711a.
  • Optionally install Fully kiosk browser on android tablet and Fully Kiosk Browser integration on Home Assistant.

Installation

HACS > Integrations > 3 dots (upper top corner) > Custom repositories > URL: https://github.com/relust/VisualStreamAssist, Category: Integration > Add > wait > Stream Assist > Install

Config Stream Assist

  1. Add Stream Assist Integration
    Settings > Integrations > Add Integration > Stream Assist
  2. Config Stream Assist Integration
    Settings > Integrations > Stream Assist > Configure
  • If you use Rtpmic app, Stream URL is rtp://192.168.0.xxx:5555
  • On Player Entity copy exact name of your BROWSER MODE PLAYER of tablet browser (media_player.xxx_xxx).
  • On Browser ID copy exact name of your BROWSER MODE BROWSER (from tablet Browser Mod tab/Browser ID field).
  • For complete TTS service for wake word detection and TTS language for wake word detection fields you can simulate a new automation, add action Media Player, select Play media, select a media player, and from Pick media select Text to speech, select your language and write a message. Then go to yaml mode and copy the tts service and tts language. IMPORTANT IT MUST BE THE SAME AS IN THE SELECTED PIPELINE
    • Example: - From media-source://tts/edge_tts?message=how can I help you&language=en-US-MichelleNeural - copy edge_tts to TTT service field and en-US-MichelleNeural to TTS language field
  • On Wake Word detection responses you can put many response swith a comma between them.
    • Example: how can I hel you, how can assist you, yes im listening`
  • Copy speech.gif and listen.gif or, after integration insallation, from Home Assistant /config/custom_components/stream_assist/gifsdirectory on www/gifs directory and on UI Speech Gif and Listen Gif fields write the path:
    • /local/gifs/jarvis_speech.gif
    • /local/gifs/jarvis_listen.gif
  • You can select Voice Assistant Pipeline for recognition process: WAKE => STT => NLP => TTS. By default componen will use default pipeline. You can create several Pipelines with different settings. And several Stream Assist components with different settings.

Using

Component has MIC switch and multiple sensors - WAKE, STT, INTENT, TTS. There may be fewer sensors, depending on the Pipeline settings.

The sensor attributes contain a lot of useful information about the results of each step of the assistant.

You can also view the pipelines running history in the Home Assistant interface:

  • Settings > Voice assistants > Pipeline > 3 dots > Debug

About

Through this project I wanted to add to the voice assistant made in Home Assistant a visual image and random personalized responses that are displayed on an android tablet.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published