Conversational AI Trainer & LLM Evaluation Specialist
As a Conversational AI Trainer and Large Language Model (LLM) Evaluation Specialist, I designed and executed advanced conversational simulations to test model reasoning, safety, and consistency. I performed high-accuracy dialogue annotation and adversarial prompt evaluation across diverse user personas and contexts. My responsibilities included analyzing transcripts, documenting model behaviors, and ensuring data quality through structured annotation workflows. • Designed complex multi-turn dialogue simulations to evaluate LLM reasoning, safety alignment, and emotional intelligence. • Performed advanced data annotation, transcript labeling, and adversarial prompt engineering tasks. • Collaborated with distributed remote teams via specialized annotation platforms to maintain evaluation consistency. • Supported AI optimization by documenting model performance insights and behavioral inconsistencies.