AI Model Evaluation & Response Quality Annotation
I have been working on AI training and evaluation tasks with Vetto AI, focusing on improving the performance and reliability of large language models. My responsibilities include evaluating AI-generated text responses for accuracy, coherence, relevance, and safety based on detailed annotation guidelines. The project involves structured evaluation rating tasks where I assess and compare model outputs, identify errors such as hallucinations, bias, or logical inconsistencies, and provide high-quality feedback to improve future model behavior. I consistently apply strict labeling criteria to ensure consistency and reliability across all evaluations. The work requires strong attention to detail, critical thinking, and the ability to interpret nuanced language outputs across different contexts. I maintain high accuracy standards while working in a fast-paced, remote environment, contributing to the improvement of AI systems at scale.