Freelance AI Safety Evaluator (Red Teaming)
Conducted red teaming and AI safety evaluation tasks focused on identifying failure modes in generative model behavior. Work involved crafting adversarial text prompts, testing boundary conditions, and documenting model responses using structured feedback fields rather than traditional annotation labels. Outputs included categorization of failure types, severity assessments, and concise written rationales, all completed within client-provided proprietary evaluation platforms and aligned with project quality and safety guidelines.