AI Response Evaluation and Text Annotation
Reviewed and evaluated AI-generated text responses for accuracy, reasoning quality, tone, completeness, and instruction adherence using structured rubrics and project guidelines. Compared multiple model outputs, selected preferred responses, and produced concise written justifications explaining evaluation decisions. Interpreted complex prompts and constraints to identify logical gaps, factual inaccuracies, and policy misalignment. Maintained consistent quality standards while completing high-volume text annotation and evaluation tasks in a remote, guideline-driven environment.