Physics & Mathematics QA and AI Model Evaluation
Led multiple projects focused on evaluating and improving AI model outputs in Physics and Mathematics, ensuring technical correctness, conceptual clarity, and adherence to domain-specific guidelines. Designed and implemented QA rubrics and scoring frameworks for assessing LLM-generated solutions across problem-solving tasks, derivations, and applied reasoning. Developed and refined educational content, mathematical proofs, and physics explanations, embedding references and step-by-step examples to enhance clarity and learning outcomes. Conducted peer audits to maintain quality consistency and compliance with project standards. Delivered training workshops for engineers and analysts, covering QA methodologies, mathematical reasoning in NLP tasks, and model alignment evaluation. Integrated transformer-based models (BERT, GPT, and custom LLMs) into client systems, enabling STEM-focused automation and evaluation workflows. This work resulted in higher model accuracy.