Remote AI Data Annotation & Evaluation Specialist (Freelance / Project-Based)
As a remote AI Data Annotation & Evaluation Specialist, I annotated and reviewed text data to train and evaluate artificial intelligence models. My tasks involved intent recognition, consistency checking, and classification of prompts and user scenarios according to detailed guidelines. I evaluated large language model (LLM) outputs for accuracy, tone, relevance, and policy compliance while maintaining high productivity and attention to detail. • Conducted intent-checking, classification, and flagging of ambiguous or unsafe responses. • Suggested improvements on low-quality outputs and ensured factual correctness. • Followed complex annotation and evaluation instructions to meet quality benchmarks. • Specialized in text-based data labeling and LLM evaluation projects.