Multimodal AI Training Data Annotation for LLMs
Executed large-scale video annotation and quality assurance projects supporting computer vision and multimodal AI systems. Tasks included object detection, scene classification, emotion recognition, and action tracking across diverse video datasets. Applied bounding boxes, segmentation, and keypoint labeling to improve model recall by 18% and precision in action recognition tasks. Reviewed and labeled audio-visual datasets for speech, background noise, accents, and emotional tone, enhancing ASR (Automatic Speech Recognition) and voice assistant models. Delivered consistent annotation quality with IAA > 0.85, ensuring high inter-annotator agreement and ISO-aligned QA standards. Collaborated with cross-functional QA teams to refine annotation guidelines, reducing systemic errors and boosting annotation consistency by 35%. Demonstrated ability to scale projects from thousands to hundreds of thousands of video frames while maintaining strict accuracy thresholds.