AI Response Evaluation & Text Annotation Specialist
Worked on AI model improvement projects focused on evaluating and refining AI-generated responses. Reviewed pairs of responses using structured quality guidelines and assessed them for clarity, accuracy, relevance, coherence, instruction-following, and safety compliance. Ranked outputs based on predefined scoring rubrics and provided detailed justifications for evaluation decisions. Identified weaknesses such as incomplete answers, logical inconsistencies, vague explanations, or policy risks, and produced improved rewritten versions to enhance overall response quality. Contributed to improving model reliability, consistency, and alignment with user intent across various subject domains.