AI Model Evaluator and Data Annotator
Evaluated and red teamed AI models and agents for safety and vulnerabilities, focusing on large language model (LLM) output assessment and prompt evaluation. Built and executed offline, reproducible, and auto-evaluable test cases for robust model evaluation processes. Applied data annotation techniques to classify and rate LLM responses, contributing to the continuous improvement of AI text models. • Labeled and evaluated text-based outputs from AI models. • Applied cybersecurity knowledge to identify prompt injection and LLM risks. • Utilized scripting tools and Jupyter Notebooks for annotation workflow automation. • Documented evaluation findings with precise, reproducible steps.