Skip to content

A Sign language to English Conversion AI which works on multiple models like Keypoint Tracking and DeepLearning. A beautiful UI which makes the user feel like translation is realtime

License

Notifications You must be signed in to change notification settings

CodeWithInferno/SignSpeak

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

61 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

🧠 SignSpeak: Real-Time ASL Recognition System

Banner

Release License: MIT


🚀 Overview

SignSpeak is a cross-platform American Sign Language (ASL) recognition platform that:

  • 🕒 Detects ASL signs in real-time via your device camera
  • 📱 Runs on mobile devices (iOS & Android) using React Native + Expo
  • 🤖 Processes frames on a Python Flask backend with MediaPipe, OpenCV, and trained ML models
  • 🏗️ Dockerized for one‑command deployment (development or production)

🎥 Model Demo: Watch proof video


📋 Table of Contents


🔥 Features

  • Real-Time Recognition: 30+ frames per second hand detection with MediaPipe

  • Multiple Strategies:

    • Keypoint-based classification
    • CNN-powered image classification
    • Optional YOLOv5 object detection pipeline
  • Session Recording: Start/stop sessions, save raw frames, export metadata

  • Cross-Platform UI: Shared codebase for iOS, Android, and Web via Expo

  • Scalable Backend: Containerized Flask API, ready for production with Gunicorn + Nginx


🗂 Project Structure

SignSpeak/
├── UI_expo/           # React Native (Expo) mobile app
│   ├── components/    # UI & camera overlay
│   ├── services/      # API wrappers & data handling
│   └── app.json       # Expo config
├── server/            # Flask backend API
│   ├── App.py         # Main application
│   ├── recordings/    # Stored session data
│   └── requirements.txt
├── models/            # Pretrained ML models & label files
│   ├── random_forest_model.pkl
│   └── label_classes.txt
├── deploy/            # Docker & Nginx configs
│   └── docker-compose.yml
├── assets/
│   ├── SignSpeak.png       # README banner
│   ├── ASL Alphabet.jpg    # ASL chart reference
│   └── Model_proof.mp4     # Demo video
├── docs/              # (Optional) Extended documentation
├── Scripts/           # Training & preprocessing scripts
└── README.md          # This file

⚙️ Getting Started

Prerequisites

Local Development

  1. Clone the repository

    git clone https://github.com/CodeWithInferno/SignSpeak.git
    cd SignSpeak
  2. Backend

    cd server
    python3 -m venv venv
    source venv/bin/activate
    pip install -r requirements.txt
    python App.py

    The API will run at: http://localhost:5050

  3. Mobile App

    cd UI_expo
    npm install
    npx expo start
    • Open the Expo Go app on your phone and scan the QR code, or launch on an emulator.

🐳 Docker Deployment

  1. Start with Docker Compose

    docker-compose up --build -d
  2. Verify services

    docker-compose ps
  3. Access UI on your LAN at http://<your-host-ip>:19006


🖼️ Assets

  • Banner: ./SignSpeak.png

  • ASL Alphabet Chart:

    ASL Alphabet

  • Model Proof Video:


🛠️ Tech Stack

Layer Technology
Frontend React Native (Expo)
Backend Python, Flask, Flask-CORS
CV/ML MediaPipe Hands, OpenCV, scikit-learn (RF), YOLOv5
Deployment Docker, Docker Compose, Gunicorn
Platform Android, iOS & Web (Progressive)

🤝 Contributing

Contributions are welcome! Please:

  1. Fork the repo
  2. Create a feature branch (git checkout -b feature/YourFeature)
  3. Commit your changes (git commit -m "Add YourFeature")
  4. Push (git push origin feature/YourFeature)
  5. Open a Pull Request

For major changes, open an issue first to discuss what you’d like to change.


📖 License

This project is licensed under the MIT License. See LICENSE for details.


📞 Contact


Made with ❤️ by Pratham Patel, Raahil Desai & Tashvi Patel

About

A Sign language to English Conversion AI which works on multiple models like Keypoint Tracking and DeepLearning. A beautiful UI which makes the user feel like translation is realtime

Topics

Resources

License

Contributing

Security policy

Stars

Watchers

Forks

Contributors 3

  •  
  •  
  •