This project is a real-time Sign Language Translator that uses a webcam to recognize hand gestures and convert them into text and speech. It utilizes Computer Vision, Machine Learning, and Deep Learning techniques to interpret sign language effectively.
- Real-time gesture recognition using webcam input
- Translates hand gestures into corresponding text
- Converts translated text into speech
- Preprocessing and landmark extraction using MediaPipe
- Model training using scikit-learn and Keras
- Modular code for easy extension
image_to_csv.pyโ Capture hand landmarks and export to CSVdata_preprocessing.pyโ Normalize and preprocess datasetmodel_training.pyโ Train gesture recognition modelpredict_gesture.pyโ Load model and predict gesturesreal_time_gesture.pyโ Real-time webcam detection and translation
Ensure the following are installed:
- Python 3.9.12
- Libraries:
opencv-pythonmediapipenumpypandasscikit-learnkeraspyttsx3(for speech output)
You can install dependencies with:
pip install opencv-python mediapipe numpy pandas scikit-learn keras pyttsx3-
Capture landmarks
Runimage_to_csv.pyto collect gesture data (if required) -
Preprocess data
Rundata_preprocessing.pyto prepare data for training -
Train the model
Runmodel_training.pyto train and save your model -
Predict single image
Usepredict_gesture.pyto predict a gesture from static image/landmarks -
Real-time recognition
Launchreal_time_gesture.pyto start webcam detection and translation
- In image_to_csv.py, change the path to point to your dataset image directory before running the script.
- Example:
- DATASET_DIR = 'path//to//your//dataset'