LLM Coding Task Creation, Golden Answer Development and Evaluation Rubric Design
In this LLM training project, I worked on designing high quality coding tasks to improve the code generation abilities of large language models. My work included creating clear and structured problem statements in Python, Django, JavaScript, and TypeScript, writing complete and production ready golden answers, and building detailed evaluation rubrics to assess correctness, reasoning, edge case handling, and overall code quality. I also reviewed and rated model generated responses, identified hallucinations and logical errors, and provided structured feedback to support fine tuning and RLHF improvements. Most of the tasks focused on backend development, APIs, and algorithmic problem solving.