AI Output Evaluator and Prompt Engineer (Project/Extracurricular)
I tested and evaluated outputs from large language models such as ChatGPT and Claude. My work involved assessing model responses for accuracy, tone, bias, and overall helpfulness. I applied prompt engineering best practices to generate targeted model behaviors and provided constructive written feedback to improve model performance. • Consistently assessed a variety of generated text for compliance with quality standards • Utilized logical reasoning and attention to detail when evaluating outputs • Reported findings and improvement suggestions in clear, concise documentation • Developed a strong understanding of responsible AI use and prompt optimization techniques