AI Response Evaluation Analyst
As an AI Response Evaluation Analyst, I evaluated AI-generated text responses from large language models for quality, relevance, and safety. My duties included fact-checking outputs, selecting optimal responses, and providing structured feedback to improve LLM accuracy. Consistent application of project guidelines was required to ensure unbiased and reliable ratings. • Detected hallucinations, factual errors, and misleading statements in responses. • Compared and selected the most helpful and correct outputs. • Collaborated in structured dataset review processes for LLM training. • Used proprietary or online platforms for response evaluation and rating.