Red Teaming for AI Text-Based Models
Participated in a Red Teaming project to test the ethical and functional robustness of AI models handling text-based inputs. My tasks included crafting adversarial scenarios and prompts to expose vulnerabilities in text generation and classification models. Additionally, I annotated responses to flag inappropriate outputs, biases, and inconsistencies, ensuring the model adhered to ethical standards and improved reliability.