AI trainer / model evaluator
As an AI trainer and model evaluator, I designed structured training tasks targeting large language models (LLMs) in varied business and operational contexts. I developed workflows and scoring rubrics to systematically assess language model outputs for accuracy, reasoning, and instruction following. My responsibilities included quality assurance reviews, feedback provision, and contributing to iterative model improvement cycles. • Crafted and scored prompt-based LLM tasks to evaluate performance. • Conducted comparative analysis and quality control on AI outputs. • Developed evaluation criteria to ensure dataset consistency and reliability. • Provided structured feedback to improve dataset utility and robustness.