Source Code Annotation, Evaluation, and AI Training Data Preparation for Programming Models
Worked on programming-related data labeling and evaluation tasks to support training and improvement of AI code generation models. Responsibilities included reviewing AI-generated code outputs, validating correctness, and annotating responses based on quality, accuracy, and instruction compliance. Evaluated code across multiple programming languages including Java, Python, and JavaScript. Performed tasks such as: Comparing multiple code solutions for correctness and efficiency Identifying logical errors, syntax issues, and edge case failures Writing high-quality reference solutions for model training Creating prompt-response pairs for supervised fine-tuning datasets Ensuring outputs followed formatting, compilation, and execution requirements Worked with structured evaluation rubrics focusing on: Code correctness Readability and maintainability Time and space complexity awareness Instruction-following accuracy Edge case handling Maintained high annotation quality by follo