AI Trainer
Worked on multiple projects evaluating code responses produced by different AI models. For each prompt, I reviewed two candidate code solutions, assessed them against a detailed rubric (including correctness, efficiency, code quality/readability, error handling, security considerations, and adherence to user requirements), and assigned structured scores. I then ranked which model response performed better for each criterion and produced an overall judgment. This work required careful execution of test cases, deep understanding of programming concepts (e.g. algorithms, API usage, concurrency, data structures), and consistent application of labeling guidelines. The resulting annotations were used to train and fine-tune AI models for higher-quality code generation and more reliable developer assistance. Languages and ecosystems frequently involved included Python, Go, JavaScript/TypeScript, and backend API patterns. I also provided written rationales for decisions where required, helping model trainers and researchers understand edge cases and nuanced trade-offs between different code solutions.