Lead developer for AMAVA, an accessibility research project translating video input into real-time audio feedback for visually impaired users.
- Engineered a motion-aware audio engine using YOLOv2 to classify environmental context and dynamically throttle audio playback.
- Built a low-latency pipeline with captioning and caching using Gemini, torchaudio, ElevenLabs, and FFmpeg.
- Authored an 8-page research paper detailing system architecture, user trials, and performance benchmarks.