AI Agent Assistant / Evaluation & Annotation Contributor (Freelance)
Collaborated with large language models (LLMs) to perform data annotation, prompt evaluation, and output analysis in freelance remote projects. Refined and rated AI-generated outputs for clarity, accuracy, and usability, ensuring alignment with project guidelines. Conducted web-based fact-checking, prompt testing, and ambiguity resolution as part of systematic LLM evaluation workflows. • Carried out annotation and evaluation tasks using AI evaluation and annotation platforms • Reviewed and improved prompts to enhance LLM performance • Identified model errors, hallucinations, and usability edge cases • Maintained high-quality standards and adhered to complex task specifications