AI Data Labeler / AI Trainer (Freelance) - Anthropic (Claude AI)
Contributed to the training and evaluation of large language models (Claude) for Anthropic. Responsibilities included performing response evaluation, ranking, and quality scoring of AI outputs across diverse tasks. Ensured improved model accuracy by providing structured feedback and annotation through detailed analysis of reasoning gaps, hallucinations, and inconsistencies. • Conducted in-depth evaluation of AI-generated responses for logical consistency and factual accuracy. • Identified reasoning gaps, hallucinations, and provided corrective feedback. • Maintained strict adherence to annotation guidelines and workflows. • Enhanced overall model quality through well-structured annotation and review processes.