AI Coding Evaluator
In this role, I evaluated and annotated AI-generated Python code to improve large language models. I wrote, edited, and debugged Python code solutions to serve as high-quality references for model training. My tasks included developing test cases, evaluation rubrics, and identifying edge cases through prompt workflows. • Assessed correctness, reasoning, and usability of AI outputs. • Provided structured feedback to enhance model performance. • Developed and maintained coding evaluation procedures. • Participated in ongoing team reviews and process improvements.