AI Trainer & LLM Evaluation Specialist
As an AI Trainer, I built and executed AI training pipelines focused on data annotation and rubric-based scoring for LLM evaluation. I created structured annotation workflows and participated in prompt engineering to boost LLM performance and accuracy. My work directly improved AI datasets and enabled rigorous LLM output quality assessment. • Designed and implemented scoring pipelines for evaluating large language model outputs. • Created and maintained structured annotation guidelines for consistent data labeling. • Engineered 1000+ prompts to assess and align model behavior. • Authored evaluation frameworks and dashboards for AI model performance monitoring.