VoCoMi (Voice-Controlled Mixer) is our hack submitted to the WearHacks Montreal 2015 Hackathon. It is a music sequencer for visually impaired people that does not require any visual feedback to be operated.
The system prompts input to the user using Nuance's TTS engine, and it understand voice commands through Nuance's speech recognition and understanding API's. Once a sample bank is selected, the user can browse its contents with gestures using the Myo armband and compose with them using a loop machine.
- Clone or download the repo.
- Make sure you have Python 3. You might have to run
python3all the time (default if you use homebrew and don't touch anything).
- Install Myo Connect and check that it works.
- Get Nuance user and application API keys, and put them in
- Install the
numpyPython packages from
pip. Installl also the
pygame(look for instructions online) and
pyspeex(https://github.com/NuanceDev/pyspeex) packages. You may also have to install the
speexsystem packages beforehand (both available on homebrew).
- Add myo-sdk to the LDPATH:
python vocomi.py, depending on your setup.
If you get an error getting the number of channels, it is because the system is using the wrong device to get audio input. Try changing the constant
You can change the audio samples (just put whatever
wav files inside the
assets sub-folders respecting the existing folder structure), but not the categories (that would require adapting the language model, and some code if you change the top-level categories with sub-groups).