Sign Language Interpreter is an AI-powered system that translates hand gestures into natural language text and speech. Using MediaPipe for hand detection, KNN for gesture classification, and Gemini AI for structuring sentences, it provides a user-friendly interface with Tkinter. The app supports webcam and file upload inputs, making sign language communication more accessible and interactive.
- Real-time Gesture Detection: Uses MediaPipe to detect hand landmarks live from webcam or uploaded videos/images.
- Accurate Gesture Classification: KNN model trained on hand gesture images to recognize alphabets (A-Z).
- Prediction Buffering: Stabilizes predictions using a 15-frame buffer to avoid jittery results.
- Natural Language Structuring: Sends recognized letters to Gemini AI to generate meaningful sentences.
- GUI Display & Text-to-Speech: Shows recognized sentences on Tkinter window and reads them aloud.
- Multiple Input Modes: Supports both webcam live feed and uploaded media files.
- Frontend: Tkinter (Python GUI)
- Backend: Python, OpenCV
- Hand Detection: MediaPipe
- Gesture Classification: scikit-learn (KNN)
- Language Processing: Gemini AI
- Text-to-Speech: pyttsx3
- Python 3.8 to 3.10(any) installed
- Basic command line knowledge
- Clone the repository:
git clone https://github.com/Eshita-Badhe/Sign-Language-Interpreter.git - Navigate to the project directory:
cd Sign-Language-Interpreter - Install required packages:
pip install -r requirements.txt - Run the application:
python app.py
- Launch the app.
- Select input mode: Webcam or Upload a media file.
- Start making hand gestures representing alphabets and click 'q' when completed.
- Wait as the app detects, predicts letters, buffers results, and forms sentences.
- View the recognized sentence on the GUI and listen to the audio output.
Contributions are welcome! Feel free to fork the repo, create branches for your features, and submit pull requests with clear explanations.
- Replace KNN with CNN or other deep learning models for better accuracy.
- Implement hand cropping and preprocessing to improve gesture focus.
- Expand vocabulary to include numbers, special signs, and phrases.
- Add GUI features like sentence editing and correction.
- Enable multilingual support using Gemini AI capabilities.
This project is licensed under the MIT License - see the LICENSE file for details.
For feedback or collaboration, reach out to:
- Name: Eshita Badhe
- Email: sge.eshita31gb@gmail.com
- GitHub: https://github.com/Eshita-Badhe

