A real-time sign language recognition system built with Python, MediaPipe, TensorFlow, and Streamlit. This application can detect and interpret sign language gestures in real-time using computer vision and machine learning.
- Real-time Hand Detection: Uses MediaPipe for accurate hand landmark detection
- Sign Language Recognition: Supports 29 gestures including:
- Numbers: 1, 2, 4, 5, 6, 7, 8
- Letters: A-X (excluding E)
- Live Webcam Processing: Real-time gesture recognition through webcam
- Interactive Web Interface: Built with Streamlit for easy use
- Adjustable Confidence: Configurable detection confidence threshold
- Optimized Performance: Fast response with temporal prediction smoothing
- Python 3.8+
- Webcam
- Git
- Clone the repository
git clone https://github.com/yourusername/ai-sign-language-interpreter.git
cd ai-sign-language-interpreter- Install dependencies
pip install -r requirements.txt- Run the application
streamlit run simple_sign_interpreter.py- Open your browser and navigate to
http://localhost:8501
- Start the application using the command above
- Allow camera access when prompted by your browser
- Click "START" to begin webcam capture
- Show hand gestures to the camera
- Adjust confidence threshold using the sidebar slider for better detection
- View real-time predictions in the "Current Sign" panel
ai-sign-language-interpreter/
├── simple_sign_interpreter.py # Main Streamlit application
├── sign_utils.py # Core prediction and landmark detection
├── models/
│ ├── sign_model.tflite # Trained TensorFlow Lite model
│ └── labels.txt # Class labels for gestures
├── requirements.txt # Python dependencies
├── .gitignore # Git ignore file
└── README.md # Project documentation
- Frontend: Streamlit web interface
- Computer Vision: MediaPipe for hand landmark detection
- Machine Learning: TensorFlow Lite for gesture classification
- Real-time Processing: Optimized prediction pipeline with temporal smoothing
- Input: 63-dimensional hand landmark features (21 landmarks × 3 coordinates)
- Output: 29 sign language gesture classes
- Format: TensorFlow Lite (.tflite) for optimized inference
- Performance: Real-time processing at 30+ FPS
- MediaPipeLandmarksModel: Handles hand detection and landmark extraction
- SignLanguagePredictor: Manages model inference and prediction smoothing
- Streamlit Interface: Provides user-friendly web interface
- Detection Confidence: Minimum confidence for hand detection (default: 0.5)
- Tracking Confidence: Minimum confidence for hand tracking (default: 0.5)
- Prediction Threshold: Minimum confidence for gesture prediction (default: 0.5)
- Buffer Size: Number of frames for prediction smoothing (default: 8)
- Fork the repository
- Create a feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
This project is licensed under the MIT License - see the LICENSE file for details.
- MediaPipe for hand landmark detection
- TensorFlow for machine learning framework
- Streamlit for the web interface
- OpenCV for computer vision utilities
If you encounter any issues or have questions:
- Check the Issues page
- Create a new issue with detailed description
- Include system information and error messages
- Support for more sign language gestures
- Multi-hand gesture recognition
- Sign language sentence formation
- Mobile app version
- Real-time translation to multiple languages