Lead Adversarial Tester
I spent this period designing sophisticated ‘jailbreak’ scenarios for LLM adversarial safety tests. I specialized in multi-turn dialogue testing sessions to probe for safety policy violations, notably around PII and dangerous instructions. These sessions often lasted up to 90 minutes per scenario. • Identified and documented model vulnerabilities • Explored multi-dimensional prompt techniques • Pioneered adversarial red teaming best practices • Informed design of model safety guardrails.