Senior LLM Evaluation & RLHF Data Labeling Project
Worked as a Senior AI Trainer & Model Evaluator supporting large language model (LLM) fine-tuning and alignment initiatives. Evaluated AI-generated responses for reasoning quality, factual accuracy, hallucination detection, and safety compliance. Performed comparative ranking of multiple model outputs as part of Reinforcement Learning from Human Feedback (RLHF) workflows. Applied strict annotation rubrics to ensure consistent scoring across large batches of text data. Identified recurring model weaknesses, documented error trends, and flagged policy-sensitive outputs. Maintained a 97%+ average quality score while adhering to detailed evaluation guidelines and QA standards. Contributed directly to improving dataset reliability and enhancing model alignment and performance.