AI Content Evaluator & Data Annotator
Contributed to the improvement of Large Language Models (LLMs) by evaluating and ranking AI-generated responses based on accuracy, tone, and safety guidelines. My work involved detailed fact-checking, identifying linguistic nuances, and providing structured feedback to enhance model performance. I ensured all annotations followed strict project-specific rubrics, focusing on reducing hallucinations and improving the helpfulness of the AI.