AI Conversation Evaluator & Prompt Specialist - Project Horizon
As an AI Conversation Evaluator & Prompt Specialist for Project Horizon, I designed multi-turn scenarios to assess and optimize language model outputs. My work included evaluating contextual retention, reasoning consistency, and instruction-following for AI-generated content in various domains. I documented structured reports with actionable improvement recommendations and provided comparative analysis for alternative model responses. • Employed adversarial testing strategies for edge case exposure. • Analyzed model performance in academic, professional, and customer settings. • Identified subtle failure modes like context drift and emotional miscalibration. • Ensured real-world applicability through robust scenario design.