AI Testing & Data Labeling Specialist
Contributed to AI training and evaluation projects including Shadow Tracker (AI‑powered competitive intelligence monitoring) and RegRadar (AI‑driven compliance monitoring). Responsibilities included designing structured annotation workflows, labeling and validating LLM outputs for factual accuracy, detecting hallucinations and biases, and conducting adversarial prompt testing. Leveraged tools such as Python, SQL, LangChain, Hugging Face, Salesforce, and Jira to ensure high‑quality datasets and reliable AI performance. Achieved measurable improvements in resolution times and reduced repeat escalations through rigorous quality audits and dataset curation.