AI Data Trainer / Annotation Quality Lead
Led production of reinforcement learning with human feedback (RLHF) preference data for large language models, ensuring top-tier annotation quality. Designed detailed annotation guidelines and ontologies for instruction-following, toxicity detection, and factual accuracy labeling. Built Python quality assurance scripts to identify errors, lowering labeling error rate by 31% overall. • Authored calibration playbooks adopted by 3 client projects • Collaborated with ML engineers for model evaluation and feedback • Coordinated a team of 18 annotators and optimized workflow • Standardized project-level annotation complexity and agreement (Cohen's κ > 0.88)