Skip to content

HandSign-Translator is a real-time American Sign Language (ASL) alphabet recognition system powered by a Convolutional Neural Network (CNN). It captures hand gestures through your webcam, identifies the corresponding ASL letter, and intelligently builds complete sentences based on confident predictions making silent communication , smart and easy.

Notifications You must be signed in to change notification settings

Venkat-023/HandSignTranslator-CNN

Repository files navigation

Real-Time American Sign Language (ASL) Recognition Using CNN

HandSign-Translator is a real-time ASL alphabet recognition system that detects hand gestures from a webcam feed, classifies them using a custom-built Convolutional Neural Network (CNN), and dynamically constructs sentences based on confident predictions. The result is a smooth and intelligent system that bridges visual language and readable text.

📌 Project Highlights 🧠 Custom CNN architecture built from scratch

🎥Real-time gesture recognition via webcam

✋ MediaPipe used for precise hand detection and dataset cleaning

⚪ Grayscale image processing for faster and more robust model inference

🧪 Trained on a self-curated dataset with over 94% testing accuracy

💬 Confidence-based sentence construction displayed live on-screen

🖼️ Dataset Preparation Source: ASL alphabet dataset from Kaggle

Due to dataset size and noise, it was split and preprocessed in parts

MediaPipe was used to extract and crop only the hand region from images

Cleaned dataset was saved and converted to grayscale for better model performance

Final dataset was significantly smaller, cleaner, and optimized for CNN training

🧠 Model Architecture Designed and trained 15+ custom CNN models

Used TensorFlow/Keras for model building and evaluation

Best-performing model achieved 94%+ accuracy on test data

Final model was exported and integrated into a real-time prediction pipeline

🎥 Real-Time Pipeline Capture live video input using OpenCV

Detect hand region using MediaPipe

Preprocess the frame (resize & grayscale)

Predict gesture using the trained CNN

Validate prediction confidence

Build sentence from validated predictions

Display the result live on the webcam feed

⚙️ Technologies Used Python 3.x

TensorFlow / Keras

OpenCV

MediaPipe

NumPy, Matplotlib

📦 Use Cases ASL learning tools and educational platforms

Assistive communication for individuals with speech or hearing impairments

Research in gesture recognition and HCI (Human-Computer Interaction)

Gesture-based user interfaces and control systems

🎯 Future Enhancements Extend support to ASL words and phrases

Add multi-hand support for more complex gestures

Improve UI/UX with overlay controls and customization

Export as a lightweight desktop app using PyInstaller or Electron

About

HandSign-Translator is a real-time American Sign Language (ASL) alphabet recognition system powered by a Convolutional Neural Network (CNN). It captures hand gestures through your webcam, identifies the corresponding ASL letter, and intelligently builds complete sentences based on confident predictions making silent communication , smart and easy.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published