AI Evaluation Analyst
As an AI Evaluation Analyst, I systematically evaluate AI-generated content and model outputs for logical consistency and accuracy. My primary responsibility is to rate, assess, and compare language model outputs across various subject matters, ensuring clarity and compliance with guidelines. I collaborate internationally to train large language models (LLMs) by selecting the most contextually appropriate responses to enhance model performance. • Reviewed and rated diverse AI-generated text outputs across multiple domains. • Applied critical thinking to ensure neutrality, accuracy, and logical coherence in content assessments. • Participated in ongoing qualitative research and content analysis tasks. • Contributed to iterative improvement of natural language model capabilities through expert feedback.