Independent AI Trainer & Data Annotation Specialist
As an independent AI trainer and data annotator, I evaluated AI-generated responses for clarity, accuracy, and completeness. I created structured prompts to test response quality and consistency and applied guideline-based decision-making to assess and rate AI responses. I compared outputs across different scenarios to identify response patterns and inconsistencies. • Evaluated large language model (LLM) outputs using defined metrics and guidelines. • Developed and tested prompts to improve AI response accuracy. • Assessed and categorized responses into distinct quality levels. • Utilized internal/proprietary tooling and mainstream AI tools such as ChatGPT and Claude.