AI Response Evaluation & Data Annotation
Reviewed and evaluated AI-generated responses against predefined rubrics and quality standards. Tasks included identifying factual inaccuracies, logical gaps, ambiguous reasoning, and instruction-following failures. Provided structured feedback and scoring to support model refinement and training iterations. Work required strict adherence to confidentiality requirements, consistency across evaluations, and precise documentation of decisions.