Sign Language Interpreter App Sessions
Below is an overview of the three sessions dedicated to building our Sign Language Interpreter App. Each session spans two weeks, starting August 30.
Session 1 (August 30 - September 12): Theory & Setup
- Topics Covered:
- Project Architecture & Goals
- Real-Time Image Capture from Webcam
- Basics of Sign Language Recognition Approaches
- Introduction to OpenAI GPT API
- Outcome:
- Environment Setup (Python, Flask, OpenAI key)
- Understanding how images will be captured and processed
- High-level plan for GPT integration
Session 2 (September 13 - September 26): Show Examples & Implementation
- Topics Covered:
- Step-by-Step Example of Image Processing
- Recognizing Sign Language Gestures (Placeholder or real models)
- Building Context and Tracking Recognized Words
- Constructing Sentences with GPT
- Outcome:
- A working prototype that captures images and identifies gestures
- Integration with GPT to form partial sentences
- Understanding how to handle multiple recognized words over time
Session 3 (September 27 - October 10): Practice, Review & Finalization
- Topics Covered:
- Debugging and Refining Sign Language Recognition
- Final GPT Integration for Coherent Sentences
- Testing the End-to-End Flow
- Addressing Q&A, Homework, and Future Improvements
- Outcome:
- A polished Sign Language Interpreter App
- Lessons learned on AI integration in web apps
- Ideas for further development (e.g., advanced ML models, UI enhancements)