Follow these steps to get the add-on installed on your system:
- Navigate in your Home Assistant frontend to Settings -> Add-ons -> Add-on store.
- Add the store https://github.com/rhasspy/hassio-addons
- Find the "vosk" add-on and click it.
- Click on the "INSTALL" button.
After this add-on is installed and running, it will be automatically discovered by the Wyoming integration in Home Assistant. To finish the setup, click the following my button:
Alternatively, you can install the Wyoming integration manually, see the Wyoming integration documentation for more information.
Models are automatically downloaded from HuggingFace, but they are originally from Alpha Cephei. Please review the license of each model that you use (model list).
There are three operating modes:
- Open-ended - any sentence can be spoken, but recognition is very poor compared to Whisper
- Corrected - sentences similar to templates are forced to match
- Limited - only sentences from templates can be spoken
This is the default mode: transcripts from vosk are used directly.
Recognition is very poor compared to Whisper unless you use one of the larger models.
To use a specific model, such as vosk-model-en-us-0.21
(1.6GB):
- Create a directory in
/share/vosk/models
with the name of the model's language (e.g.,en
) - Download and extract the model
- Copy the contents of the directory named after the model into
/share/vosk/models/<LANGUAGE>
In the English example, all of the files inside the extracted vosk-model-en-us-0.21
directory will be put into /share/vosk/models/en
, so you would have a file named /share/vosk/models/en/am/final.mdl
.
By specifying which sentences will be spoken ahead of time, transcripts from vosk can be corrected using rapidfuzz.
Create your sentence templates and save them to a file named /share/vosk/sentences/<LANGUAGE>.yaml
where <LANGUAGE>
is one of the supported language codes. For example, English sentences should be saved in /share/vosk/sentences/en.yaml
.
You may adjust the correct_sentences
config value to:
- 0 - force transcript to be one of the template sentences
- greater than 0 - allow more sentences that are not similar to templates to pass through
When correct_sentences
is large, speech recognition is effectively open-ended again. Experiment with different values to find one that lets you speak sentences outside your templates without sacrificing accuracy too much.
If you have a set of sentences with a specific pattern that you'd like to skip correction, add them to your no-correct patterns.
Follow the instructions for corrected mode to create your sentence templates, then enable the limit_sentences
config option.
This will tell vosk that only the sentences from you templates can ever be spoken. Sentence correction is still needed (due to how vosk works internally), but it will ensure that sentences outside the templates cannot be sent.
This mode will get you the highest possible accuracy, with the trade-off being that you cannot speak sentences outside the templates.
Each language may have a YAML file with sentence templates. Most syntax is supported, including:
- Optional words, surrounded with
[square brackets]
- Alternative words,
(surrounded|with|parens)
- Lists of values, referenced by
{name}
- Expansion rules, inserted by
<name>
The general format of a language's YAML file is:
sentences:
- this is a plain sentence
- this is a sentence with a {list} and a <rule>
lists:
list:
values:
- value 1
- value 2
expansion_rules:
rule: body of the rule
Sentences have a special in/out
form as well, which lets you say one thing (in
) but put something else in the transcript (out
).
For example:
sentences:
- in: lou mo ss # lumos
out: turn on all the lights
- in: knocks # nox
out: turn off all the lights
lets you say "lumos" to send "turn on all the lights", and "nox" to send "turn off all the lights".
Notice that we used words that sound like "lumos" and "nox" because the vocabulary of the default English model is limited (vosk-model-small-en-us-0.15
).
The in
key can also take a list of sentences, all of them outputting the same out
string.
Lists are useful when you many possible words/phrases in a sentence.
For example:
sentences:
- set light to {color}
lists:
color:
values:
- red
- green
- blue
- orange
- yellow
- purple
lets you set a light to one of six colors.
This could also be written as set light to (red|green|blue|orange|yellow|purple)
, but the list is more manageable and can be shared between sentences.
List values have a special in/out
form that lets you say one thing (in
) but put something else in the transcript (out
).
For example:
sentences:
- turn (on|off) {device}
lists:
device:
values:
- in: tv
out: living room tv
- in: light
out: bedroom room light
lets you say "turn on tv" to turn on the living room TV, and "turn off light" to turn off the bedroom light.
Repeated parts of a sentence template can be abstracted into an expansion rule.
For example:
sentences:
- turn on <the> light
- turn off <the> light
expansion_rules:
the: [the|my]
lets you say "turn on light" or "turn off my light" without having to repeat the optional part.
When you correct sentences, you want to keep the score cutoff as low as possible to avoid letting invalid sentences though. But what if you just want some open-ended sentences, such as "draw me a picture of ..." which you can then forward to an image generator?
Add the following to your sentences YAML file:
sentences:
...
no_correct_patterns:
- <regular expression>
- <regular expression>
...
You can add as many regular expressions to no_correct_patterns
as you'd like. If the transcript matches any of these patterns, it will be sent with no further corrections. This effectively lets you "punch holes" in the sentence templates to allow some sentences through.
With --allow-unknown
, you can enable the detection of "unknown" words/phrases outside of the model's vocabulary. Transcripts that are "unknown" will be set to empty strings, indicating that nothing was recognized. When combined with limited sentences, this lets you differentiate between in and out of domain sentences.
Strictness when correcting sentences, where 0 is the most strict and larger values get less strict. This is only used when a YAML file exists for the model's language at /share/vosk/sentences/<LANGUAGE>.yaml
.
When enabled, only sentences from the file /share/vosk/sentences/<LANGUAGE>.yaml
can be spoken.
When enabled with limit_sentences
, sentences that are not part of the templates (or "no correct" patterns) will be returned as empty strings.
Preloads the speech-to-text model for the selected language. Other models are loaded as requested.
Enable debug logging. Useful for seeing satellite connections and each wake word detection in the logs.
Got questions?
You have several options to get them answered:
- The Home Assistant Discord Chat Server.
- The Home Assistant Community Forum.
- Join the Reddit subreddit in /r/homeassistant
In case you've found an bug, please open an issue on our GitHub.