AI Model Evaluation Specialist – Hallucination Detection & Rating
I reviewed and evaluated AI model outputs to detect hallucinations, biases, and errors. My work focused on assessing the quality of text responses and providing precise ratings based on strict guidelines. This process improved model reliability by highlighting problems and contributing corrections. • Evaluated search relevance through the UHRS platform. • Detected labeling inconsistencies and AI hallucinations. • Delivered feedback to enhance AI system accuracy. • Supported guideline-based evaluation tasks for various AI platforms.