โ Indian Sign Language Recognition to Speech
A real-time sign language recognition system that converts hand gestures into text and speech. This project uses computer vision and machine learning to help bridge the communication gap between sign language users and those who donโt understand sign language.
๐ Overview
Imagine having a translator that can watch your hand movements and instantly convert them into spoken words. Thatโs exactly what this project does:
๐ท Captures video from your computerโs camera.
๐ค Recognizes sign language letters (currently supports A, C, and T).
๐ Displays recognized letters as text in real time.
๐ Converts text to speech using AI voice.
โก Runs live and instantly (low latency, ~30 FPS).
This project demonstrates how AI-powered assistive technology can support deaf and hard-of-hearing communities, while also being an educational tool for learning computer vision and sign language.
๐ก Why This Project Matters
Sign language is used by millions worldwide, but very few non-signers understand it. In India especially, tools for Indian Sign Language (ISL) are scarce.
This project:
Promotes inclusion for deaf and hard-of-hearing individuals.
Makes education and workplaces more accessible.
Helps non-signers communicate with ISL users.
Provides a hands-on learning platform for AI, HCI, and accessibility research.
๐ How It Works
Think of this as teaching a computer to "see" and "understand" hand gestures:
-
๐๏ธ Seeing โ Camera captures hand movements.
-
๐ Understanding โ MediaPipe detects 21 hand landmarks (fingertips, joints, palm).
-
๐ง Learning โ AI model (Random Forest or SVM) recognizes gestures.
-
๐ฃ๏ธ Speaking โ Converts recognized gestures into text and speaks aloud.
๐ Project Structure
sign-language-recognition/ โโโ 1.data.collection.py # Collects training images for gestures โโโ 2.HandLandmarks.py # Extracts hand landmarks from images โโโ 3.TrainModel.py # Trains AI models (Random Forest / SVM) โโโ 4.Deployment.py # Main real-time translator โโโ 4.1Deployment-2-frame.py # Enhanced version with better detection โโโ data/ # Folder for gesture image datasets โโโ model.p # Trained ML model (generated after training) โโโ data.pickle # Processed landmark data (generated)
๐ง Step-by-Step Setup
- Install Python
Download and install Python 3.8+ from python.org. (Ensure you check โAdd Python to PATHโ during installation.)
- Install Dependencies
Run the following in your terminal:
pip install opencv-python mediapipe scikit-learn numpy matplotlib tkinter gtts pickle
- Collect Training Data
Capture your own gesture images:
python 1.data.collection.py
Show gestures for A, C, and T.
System captures 100+ images per gesture.
- Process Data (Extract Hand Landmarks)
python 2.HandLandmarks.py
Converts images into numerical hand landmark data.
Creates data.pickle.
- Train the AI Model
python 3.TrainModel.py
Trains both Random Forest and SVM classifiers.
Saves best model as model.p.
- Run the Translator
python 4.Deployment.py
Opens camera feed.
Recognizes signs โ Converts to text โ Speaks aloud.
For better accuracy, run the enhanced version:
python 4.1Deployment-2-frame.py
๐ฎ How to Use
Show gesture like A, C, or T to the camera.
Recognized letter will appear in text box.
Use GUI buttons:
๐ง Clear Text โ Reset text box.
๐ฉ Speak Text โ Speak letters aloud in English.
๐ช Translate Speech โ Speak in French.
๐ฅ Exit โ Close application.
๐ Performance Metrics
โ Accuracy: 85โ95% (depends on dataset quality).
โก Processing speed: ~30 FPS.
โฑ๏ธ Response time: < 1 second.
๐พ Memory usage: ~200 MB during runtime.
๐ Training time: 2โ5 minutes for 3 gestures.
๐ Applications
This system can be used in:
๐ซ Education โ Teach/learn sign language.
๐ฅ Healthcare โ Doctor-patient communication.
๐ Customer Service โ Accessible counters/kiosks.
๐ Transportation โ Public info kiosks.
๐ป Video Conferencing โ Real-time ISL translation.
๐ฎ Gaming & HCI โ Gesture-controlled interfaces.
๐ฎ Future Enhancements
Full AโZ alphabet support.
Recognition of words and phrases.
Support for Indian Sign Language (ISL) grammar.
Mobile app versions (Android/iOS).
Cloud training & sharing of datasets.
Real-time conversational translation.
๐ Educational Value
This project demonstrates:
Computer Vision โ Hand landmark detection with MediaPipe.
Machine Learning โ Training classifiers for gesture recognition.
Real-Time AI โ Running inference on live video.
Accessibility Technology โ Practical assistive tools for communication.
Python Development โ End-to-end ML + HCI project pipeline.
๐ค Contributing
You can improve this project by:
Adding more gestures (letters/words).
Improving the UI/UX.
Enhancing accuracy with deep learning (CNNs/LSTMs).
Creating mobile/web app versions.
Supporting multi-language speech synthesis.
๐ Privacy & Security
All recognition happens locally on your computer.
Works offline (except text-to-speech).
No images are stored permanently unless you save them.
๐ Learning Opportunities
By working with this project, youโll gain hands-on experience with:
Python programming.
Data collection and preprocessing.
ML model training and evaluation.
Real-time video processing.
GUI development for accessibility.