AI Agent Response Factuality & Grounding Analysis
This project focused on the critical evaluation of AI agent responses to ensure factual accuracy. For each task, I performed sentence-level analysis, labeling every sentence from the agent's output as 'supported', 'unsupported', 'contradictory', 'disputed', or 'no_rad' based on a strict source-of-truth context. The core of my work involved not only assigning a label but also writing a detailed rationale for each decision and meticulously extracting the exact text or code excerpts from the context that proved or disproved the agent's statement. This required a deep analysis of the agent's reasoning, including verifying the logic of its tool-use code (e.g., Python API calls) and the validity of tool outputs before they could be used as trusted evidence for subsequent sentences.