RAG Pipeline Output Evaluation (CSA Agent Project)
As part of the CSA Agent AI Customer Support Automation project, I evaluated the contextual relevance and traceability of LLM-generated responses. I participated in the systematic testing of Retrieval Augmented Generation (RAG) pipelines, verifying accuracy and appropriateness of results. My focus included the identification and flagging of mismatches between AI-generated outputs and expected customer support resolutions. • Performed RAG-specific validation of conversational customer support agents. • Assessed LLM response quality for real-world customer support scenarios. • Collaborated on process improvements for RAG pipeline output review. • Implemented structured review and annotation of support dialogues.