Freelance AI Evaluator & Data Annotator
Reviewed AI-generated text responses using structured rubrics to assess accuracy, helpfulness, and safety. Applied annotation guidelines to label and categorize outputs for large language model analysis. Maintained high consistency and accuracy throughout substantial annotation workloads. • Performed pairwise comparison and scoring of model responses. • Identified hallucinations and logical inconsistencies in LLM outputs. • Delivered structured feedback to enhance model performance. • Adapted quickly to evolving project requirements and provided evidence-based feedback.