ASL Interpretation using Leap Motion
Disabled persons have always been limited in their methods to communicate with the world. Due to the rapidly growing popularity and devleopment towards VR, we wanted to implement a method to allow mute people to communicate in VR using their "natural" language.
Leap motion detects sign language gestures converting it into command line text. Implementation in reality would be to allow mute people to sign in VR and communicate with those people who don't understand sign language.
We used Leap Motion SDK and built it using python 2.7.
Lots of technological difficulties. The most trouble we had was the intimidating factor of trying to understand Leap Motion API in a short amount of time. This was the first time we worked with Leap Motion.
Managing to produce a minimum viable product despite having switched project ideas multiple times in 24 hours.
We learned how to use LeapMotion API as well as python.
Add text to voice feature to allow mute people to sign in VR and having the ASL being interpreted as voice. This will have many applications such as in conference calls, mute people will be able to dynamically voice their opinions.
- Clone Repo
- Install Leap Motion SKD (Orion version is strongly suggested)
- Install Python 2.7.13
- Navigate to the src folder
- Run main.py
- Place you hand over the Leap Motion device and make a sign
- View result in console print