AI Model Evaluation & Response Quality Review
Worked as a Freelance AI Trainer on an LLM evaluation project focused on improving response quality and safety. Reviewed prompts and AI-generated responses to verify factual accuracy, instruction adherence, relevance, tone, and harmful content risks. Rated outputs using defined rubrics, identified issues such as hallucinations and policy violations, and provided corrected/ideal responses and structured feedback. Followed strict quality guidelines and consistency checks to ensure high annotation accuracy and reliable dataset quality.