High precison Multimodal data annotation for Computer vision model
Currently working on an audio data annotation project supporting speech recognition and NLP model training. Responsible for high-accuracy transcription of audio recordings, including speaker segmentation, timestamping, and intent classification. Performed emotion recognition labeling, background noise identification, and audio quality assessment to improve model robustness across diverse acoustic environments. Annotated multilingual and accented speech datasets while ensuring clarity and consistency. Maintained over 98% accuracy by following strict annotation guidelines, conducting quality assurance reviews, and correcting inconsistencies. Managed large datasets efficiently while meeting tight deadlines and adhering to data confidentiality standards. Contributed to improving AI model performance by identifying edge cases such as overlapping speech, low-volume recordings, heavy accents, and noisy backgrounds.