Senior Linguistic Quality Analyst
Evaluated and ranked AI-generated text outputs for clarity, fluency, and coherence to enhance large language models. Developed annotation guidelines to standardize assessments across a team of over 15 annotators. Collaborated with machine learning engineers to refine evaluation metrics and improve model quality scores. • Ensured detailed rationale for ranking and assessment decisions • Maintained consistency and high inter-annotator agreement • Focused on outputs for Copilot and Azure AI applications • Enhanced model output quality by 23% through annotation-driven feedback