AI Content Evaluator / Data Labeler
I collaborated on reinforcement learning from human feedback (RLHF) projects to support the training and alignment of Large Language Models (LLMs). My responsibilities included evaluating, ranking, and grading AI-generated text responses to ensure accuracy, safety, and adherence to instructions. I was also tasked with identifying edge cases, hallucinations, and potential biases, contributing to the improvement of training datasets. • Conducted factuality, coherence, and instruction-following assessment of model outputs • Maintained high accuracy and quality assurance scores by following strict guidelines • Reported and documented edge cases, hallucinations, and biases in model data • Used bespoke data annotation platforms to deliver timely and high-quality results