AI response evaluation and data annotation
Worked as an AI Data Rater on the Aether project, evaluating and annotating large language model outputs for quality, accuracy, safety, and instruction adherence. Compared multiple model responses and selected the best-performing outputs based on structured evaluation guidelines. Rewrote and improved responses where necessary to align with quality standards. Maintained high consistency, attention to detail, and compliance with project-specific scoring framework