Freelance AI Data Annotator & Evaluator
Evaluated and ranked large language model (LLM) outputs for accuracy, relevance, safety, and helpfulness using detailed project guidelines. Provided written justifications for decisions and identified policy violations in responses. Worked independently on remote-based AI training assignments, maintaining strict adherence to quality and deadlines. • Assessed LLM responses against established criteria • Generated feedback to improve model performance • Maintained high accuracy in data annotation tasks • Submitted timely evaluations in a remote work environment