LLM Math Prompt Engineering and Evaluation for AI Training
Contributed to a large-scale LLM training initiative by developing and evaluating 850+ advanced math prompts covering algebra, calculus, geometry, and probability. Tasks included writing domain-specific questions, generating high-quality answers, and rating AI responses based on correctness, clarity, and logical reasoning. Designed detailed rubrics to standardize evaluation and ensure high accuracy. Collaborated with prompt engineers to improve dataset quality, reducing error rates by 35%.