LLM Output Evaluator (Embedded within Software Engineer role)
I evaluated AI-generated and LLM model outputs against specified criteria to ensure accuracy, clarity, and context sensitivity. My work involved rubric-based assessment, prompt analysis, and identifying issues related to bias and cultural sensitivity. Detailed documentation supported improvements in model performance and reliability. • Applied standardized evaluation templates and review processes • Focused on linguistic nuance, tone, and context • Used spreadsheets for consistent recording and tracking • Contributed to ethical and responsible AI system behavior