Physics-Based AI Response Evaluation & Annotation
Contributed to AI model improvement initiatives by evaluating and annotating advanced physics-related prompts and responses for scientific accuracy, mathematical correctness, logical consistency, and physical validity. Responsibilities included: Designing graduate-level physics problems across classical mechanics, electromagnetism, quantum mechanics, and statistical mechanics. Reviewing AI-generated solutions for multi-step mathematical derivations and theoretical soundness. Rating responses based on correctness, reasoning quality, clarity, and adherence to physical laws. Identifying logical gaps, flawed assumptions, and computational errors. Providing structured feedback to improve model reasoning and domain reliability. Project scope involved reviewing 500+ complex physics responses with strict quality benchmarks (≥95% evaluation consistency). Adhered to detailed annotation rubrics and maintained high inter-rater reliability standards.