Independent Applied AI Researcher – Multi-Agent LLM Evaluation/Guardrail Testing
Designed and tested multi-agent LLM evaluation frameworks for AI output reliability. Developed automated adversarial guardrail and prompt-injection stress testing procedures for agentic workflows. Created structured criteria for judging LLM responses in credibility models. • Designed evaluation metrics to assess LLM reliability for complex outputs • Automated adversarial attacks and guardrail tests for AI safety • Utilized tools like LangSmith and Labelbox for structured monitoring • Documented and analyzed results to guide AI security enhancements