AI Model Evaluator (LLM Evaluation & Annotation)
Worked on AI model evaluation and annotation tasks, focusing on assessing and comparing LLM-generated responses based on predefined rubrics. Evaluated outputs for correctness, relevance, coherence, and factual accuracy. Performed side-by-side comparisons, identified reasoning errors, and ensured adherence to quality guidelines. Contributed to improving model performance by providing structured feedback and maintaining consistency across evaluations.