AI Testing QA & Prompt Engineering Specialist
I evaluated AI-generated outputs as part of LLM quality assurance and prompt engineering efforts. Responsibilities included systematically identifying model errors, biases, and inconsistencies while stress-testing language models with structured prompts. Rigorous evaluation ensured only high-quality outputs and datasets were contributed to model teams. • Rated and reviewed language model responses for quality. • Created diverse prompts for thorough model evaluation. • Reported bugs and inconsistencies in AI outputs. • Contributed to continuous improvement via QA feedback.