AI Data Annotation & LLM Evaluation Specialist (Freelance)
As an AI Data Annotation & LLM Evaluation Specialist, I performed data labeling and model evaluation for large language models. My work involved annotation, prompt engineering, response ranking, and comprehensive quality assurance for LLM outputs. I applied RLHF principles, logical reasoning, and hallucination detection within strict guidelines to support continual model improvement. • Conducted text data labeling, response ranking, and prompt-response pair generation for AI models. • Evaluated AI-generated outputs for accuracy, logical consistency, and factual correctness. • Applied RLHF, hallucination detection, and content validation within quality protocols. • Used multiple platforms including Remotasks, Toloka, Hivework, OneForma, and Outlier.ai for annotation and evaluation tasks.