Establishment of Private AI Red Teaming Lab – LLM Evaluation
I established a private AI Red Teaming lab for adversarial evaluation of local AI models. My work involved developing new strategies for model testing and conducting comparative attacks using the 'Attacker – Victim – Judge' principle. I built a portfolio of practical red teaming tactics for use in AI safety projects. • Constructed local infrastructure for LLM red teaming • Researched and implemented novel adversarial approaches • Conducted model-to-model comparative vulnerability testing • Developed documentation for private use and knowledge sharing