This project implements a real-time sign language detection system using TensorFlow and OpenCV.
The goal is to bridge the communication gap between the hearing-impaired and others by translating sign language gestures into text or speech.
- Real-time sign detection using webcam.
- Deep learning models (TensorFlow Object Detection API).
- 🗂 Custom dataset support for training your own gestures.
- 🖼 Image preprocessing & augmentation pipeline.
- Evaluation metrics and visualization.
- Easy-to-run scripts for training and inference.
SignLanguageDetection/ │── Tensorflow/ │ ├── workspace/ │ │ ├── annotations/ # Label files (.xml/.csv) │ │ ├── images/ # Train/Test images │ │ ├── models/ # Trained models and checkpoints │ │ └── pre-trained-models/ # Downloaded pre-trained TF models │ └── scripts/ # Utility scripts (training, evaluation, etc.) │── README.md # Project documentation
- Clone the repository
git clone https://github.com/anjaliy11/SignLanguageDetection.git cd SignLanguageDetection ---
conda create -n signlang python=3.10 -y conda activate signlang
Install TensorFlow Object Detection API git clone https://github.com/tensorflow/models Tensorflow/models cd Tensorflow/models/research protoc object_detection/protos/*.proto --python_out=. pip install . cd ../../..
Use your webcam to capture images for different gestures: python Tensorflow/scripts/collect_images.py
Label the collected images using LabelImg.
python Tensorflow/scripts/train.py
Run the detection script: python Tensorflow/scripts/detect.py
Python
TensorFlow
Keras
OpenCV
NumPy
Pandas
Matplotlib
LabelImg (for annotation)