AI Rater / Content Evaluation Contributor
I evaluated AI-generated content using structured rubrics and pairwise comparison methodologies. My work involved rating, reviewing, and providing feedback on large language model outputs. This supported model improvement in coherence, factual accuracy, and compliance with task instructions.• Compared model responses and scored outputs for relevance and quality • Provided structured human feedback to enhance LLM performance • Reviewed prompts and responses to flag errors and inconsistencies • Ensured outputs met established rubrics and guideline standards