Skip to content

ajiteshgpt/SignAssist

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

11 Commits
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

โœ‹ Indian Sign Language Recognition to Speech

A real-time sign language recognition system that converts hand gestures into text and speech. This project uses computer vision and machine learning to help bridge the communication gap between sign language users and those who donโ€™t understand sign language.


๐ŸŒŸ Overview

Imagine having a translator that can watch your hand movements and instantly convert them into spoken words. Thatโ€™s exactly what this project does:

๐Ÿ“ท Captures video from your computerโ€™s camera.

๐Ÿค– Recognizes sign language letters (currently supports A, C, and T).

๐Ÿ“ Displays recognized letters as text in real time.

๐Ÿ”Š Converts text to speech using AI voice.

โšก Runs live and instantly (low latency, ~30 FPS).

This project demonstrates how AI-powered assistive technology can support deaf and hard-of-hearing communities, while also being an educational tool for learning computer vision and sign language.


๐Ÿ’ก Why This Project Matters

Sign language is used by millions worldwide, but very few non-signers understand it. In India especially, tools for Indian Sign Language (ISL) are scarce.

This project:

Promotes inclusion for deaf and hard-of-hearing individuals.

Makes education and workplaces more accessible.

Helps non-signers communicate with ISL users.

Provides a hands-on learning platform for AI, HCI, and accessibility research.


๐Ÿš€ How It Works

Think of this as teaching a computer to "see" and "understand" hand gestures:

  1. ๐Ÿ‘๏ธ Seeing โ†’ Camera captures hand movements.

  2. ๐Ÿ” Understanding โ†’ MediaPipe detects 21 hand landmarks (fingertips, joints, palm).

  3. ๐Ÿง  Learning โ†’ AI model (Random Forest or SVM) recognizes gestures.

  4. ๐Ÿ—ฃ๏ธ Speaking โ†’ Converts recognized gestures into text and speaks aloud.


๐Ÿ“‚ Project Structure

sign-language-recognition/ โ”‚โ”€โ”€ 1.data.collection.py # Collects training images for gestures โ”‚โ”€โ”€ 2.HandLandmarks.py # Extracts hand landmarks from images โ”‚โ”€โ”€ 3.TrainModel.py # Trains AI models (Random Forest / SVM) โ”‚โ”€โ”€ 4.Deployment.py # Main real-time translator โ”‚โ”€โ”€ 4.1Deployment-2-frame.py # Enhanced version with better detection โ”‚โ”€โ”€ data/ # Folder for gesture image datasets โ”‚โ”€โ”€ model.p # Trained ML model (generated after training) โ”‚โ”€โ”€ data.pickle # Processed landmark data (generated)


๐Ÿ”ง Step-by-Step Setup

  1. Install Python

Download and install Python 3.8+ from python.org. (Ensure you check โ€œAdd Python to PATHโ€ during installation.)

  1. Install Dependencies

Run the following in your terminal:

pip install opencv-python mediapipe scikit-learn numpy matplotlib tkinter gtts pickle

  1. Collect Training Data

Capture your own gesture images:

python 1.data.collection.py

Show gestures for A, C, and T.

System captures 100+ images per gesture.

  1. Process Data (Extract Hand Landmarks)

python 2.HandLandmarks.py

Converts images into numerical hand landmark data.

Creates data.pickle.

  1. Train the AI Model

python 3.TrainModel.py

Trains both Random Forest and SVM classifiers.

Saves best model as model.p.

  1. Run the Translator

python 4.Deployment.py

Opens camera feed.

Recognizes signs โ†’ Converts to text โ†’ Speaks aloud.

For better accuracy, run the enhanced version:

python 4.1Deployment-2-frame.py


๐ŸŽฎ How to Use

Show gesture like A, C, or T to the camera.

Recognized letter will appear in text box.

Use GUI buttons:

๐ŸŸง Clear Text โ€“ Reset text box.

๐ŸŸฉ Speak Text โ€“ Speak letters aloud in English.

๐ŸŸช Translate Speech โ€“ Speak in French.

๐ŸŸฅ Exit โ€“ Close application.


๐Ÿ“Š Performance Metrics

โœ… Accuracy: 85โ€“95% (depends on dataset quality).

โšก Processing speed: ~30 FPS.

โฑ๏ธ Response time: < 1 second.

๐Ÿ’พ Memory usage: ~200 MB during runtime.

๐Ÿ• Training time: 2โ€“5 minutes for 3 gestures.


๐ŸŒ Applications

This system can be used in:

๐Ÿซ Education โ†’ Teach/learn sign language.

๐Ÿฅ Healthcare โ†’ Doctor-patient communication.

๐Ÿ›’ Customer Service โ†’ Accessible counters/kiosks.

๐Ÿš‰ Transportation โ†’ Public info kiosks.

๐Ÿ’ป Video Conferencing โ†’ Real-time ISL translation.

๐ŸŽฎ Gaming & HCI โ†’ Gesture-controlled interfaces.


๐Ÿ”ฎ Future Enhancements

Full Aโ€“Z alphabet support.

Recognition of words and phrases.

Support for Indian Sign Language (ISL) grammar.

Mobile app versions (Android/iOS).

Cloud training & sharing of datasets.

Real-time conversational translation.


๐Ÿ“– Educational Value

This project demonstrates:

Computer Vision โ†’ Hand landmark detection with MediaPipe.

Machine Learning โ†’ Training classifiers for gesture recognition.

Real-Time AI โ†’ Running inference on live video.

Accessibility Technology โ†’ Practical assistive tools for communication.

Python Development โ†’ End-to-end ML + HCI project pipeline.


๐Ÿค Contributing

You can improve this project by:

Adding more gestures (letters/words).

Improving the UI/UX.

Enhancing accuracy with deep learning (CNNs/LSTMs).

Creating mobile/web app versions.

Supporting multi-language speech synthesis.


๐Ÿ” Privacy & Security

All recognition happens locally on your computer.

Works offline (except text-to-speech).

No images are stored permanently unless you save them.


๐ŸŽ“ Learning Opportunities

By working with this project, youโ€™ll gain hands-on experience with:

Python programming.

Data collection and preprocessing.

ML model training and evaluation.

Real-time video processing.

GUI development for accessibility.


About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •