This particular project known as Extension Management Machine Assistant(E.M.M.A.) is mostly a learning experience in using various different libraries to develop a digital assistant that can do various tasks for the user, similar to Alexa, Siri, and Google Home. While this is a very early iteration, it has a wide array of functionalities from doing searches on the internet to playing music for the user. There will be future iterations of this project.
To run this project, simply type in the terminal: python speech.py
and this project is currently running under python 3.7
In the configuration file under config.py
please add in the details of the following:
- WOLFRAM_ID: YOUR WOLFRAM API KEY
- username: your email
- password: your email password
- code: /path/to/your/code_editor
- play_directory: /path/to/your/music_file
- root_user: your name
- chat_log: if you plan on using rainmeter it would be /path/to/your/chat_log/chat.txt
- shutdown: path/to/your/sound_file/shutdown.wav
Additional change to be done is under this function: def on_startup_sound
the variable for the interface sound needs to be changed to match where your path to the sound_file
is similar to shutdown
.
The GUI library is from PySimpleGui
The development of the interface that I found appealing was to use a rainmeter skin interface to display the information of what the digital assistant was presenting whenever a query was answered. Essentially, I combined python with a fairly simple .ini
script and had it be displayed there. Rainmeter needs to be installed to display this type of UI setup.Below is the current interface.
Soon:tm: