AI Trainer & Prompt Engineer
Designed, developed, and evaluated structured instruction datasets and prompts for large language model (LLM) training and evaluation. Assessed AI outputs for factual accuracy, safety alignment, technical correctness, and clarity according to defined rubrics. Performed comprehensive benchmarking and qualitative error analysis to identify hallucinations, bias, and unsafe recommendations. • Developed structured instructions to enhance LLM contextual understanding • Evaluated AI-generated text for multiple quality dimensions • Collaborated with research teams to refine training data and evaluation frameworks • Supported deployment readiness by improving model output reliability