Skip to content

Srushti-T18/SignCast

Repository files navigation

SignCast: Real-Time ASL-to-Speech Translation

Developed by Team Anyhow Samagra Agarwal, Srushti Talandage, and Bhoumik Sangle

SignCast is a real-time accessibility tool designed to bridge the communication gap for the speech-impaired. By leveraging computer vision and deep learning, SignCast translates American Sign Language (ASL) gestures into spoken language instantaneously.

The Mission

For millions of individuals who rely on ASL, communicating with non-signers in real-time remains a significant challenge. SignCast provides a software-based bridge, converting complex hand movements into synthesized speech, empowering users to express themselves naturally in any environment.

Technical Architecture

SignCast uses a sophisticated pipeline to ensure that both the shape and the motion of signs are captured accurately.

  1. Input: Real-time video stream processed at 30 FPS.
  2. Landmark Extraction: MediaPipe extracts 21 3D hand landmarks per frame.
  3. Normalization: To ensure the model is invariant to where the user is sitting, all landmarks are normalized relative to the wrist ().
  4. Temporal Processing: A Bi-Directional LSTM (Long Short-Term Memory) network processes a sliding window of 30 frames to understand the motion.
  5. Classification: A Softmax layer chooses the correct gloss from a vocabulary of 2,700 signs.
  6. Output: The predicted text is converted to audio via a Text-to-Speech (TTS) engine.

Dataset & Performance

  • Dataset: ASL Citizen (2,700+ distinct signs).
  • Data Split: 60% Training | 25% Validation | 15% Testing.
  • Accuracy: Achieved a final validation accuracy of 85.74%.

Tech Stack

  1. TypeScript & JavaScript – Frontend logic and real-time interactions
  2. HTML5 & CSS3 – Responsive user interface and layout
  3. MediaPipe Hands – Real-time hand landmark detection
  4. OpenCV – Webcam capture and frame processing
  5. TensorFlow / Keras – Sequence-based sign language classification model
  6. NumPy – Landmark preprocessing and tensor handling
  7. Scikit-learn – Label encoding and dataset utilities
  8. pyttsx3 – Offline text-to-speech conversion
  9. Python – Core ML pipeline and inference logic
  10. Git & GitHub – Version control and collaboration

Installation & Setup

Prerequisites: Python 3.10

# Clone the repository
git clone https://github.com/your-repo/SignCast.git

# Create a virtual environment
python3 -m venv signcast_env
source signcast_env/bin/activate

# Install dependencies
pip install tensorflow mediapipe opencv-python pandas numpy

Project Structure

├── /models
│   ├── SignCast_best_model.h5  # Trained Bi-LSTM model
│   └── label_map.json          # Dictionary for 2,700 signs
├── data_loader.py              # Custom ASLDataGenerator
├── train_model.py              # Model architecture & training script
└── webcam_test.py              # Real-time inference & TTS integration

Future Roadmap

Team Anyhow is committed to expanding SignCast into a full-scale accessibility suite:

  • Custom Word Feature: Allowing users to record and train personal idiosyncratic signs.
  • Subtitles: Integrated overlay for video conferencing platforms.
  • Multi-Hand Support: Expanding landmarks to include facial expressions and body pose for more nuanced translation.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors