AI Model Evaluator
As an AI Model Evaluator, I assessed and refined AI-generated content to ensure factual accuracy and adherence to guidelines. I provided detailed feedback and actionable annotations to enhance the performance and reliability of large language models. My role directly contributed to improving the consistency, accuracy, and human alignment of conversational AI outputs. • Evaluated over 5,000 AI-generated text outputs weekly for accuracy and reasoning • Fact-checked and annotated model responses using internal proprietary tools • Reduced annotation discrepancies by 20% through rubric refinement • Collaborated with teams to optimize evaluation processes