AI Trainer and LLM Evaluator (Software Engineering Domain)
I reviewed, evaluated, and rated AI-generated code and technical documentation outputs related to backend systems. My role involved assessing the accuracy, reliability, and quality of model outputs using explicit guidelines and structured annotation processes. I ensured the consistency and fairness of AI model evaluations by following standardized procedures. • Evaluated code generation quality in Node.js and TypeScript outputs • Performed RLHF ranking of LLM-generated technical responses • Conducted correctness and performance reviews for API and backend solutions • Contributed to documentation for annotation and evaluation guidelines