AI Model Evaluation & Data Annotation Project
Worked on ongoing AI training and evaluation projects focused on improving large language model (LLM) performance. Tasks included evaluating and rating model-generated text, comparing multiple responses, classifying outputs based on detailed guidelines, and providing structured feedback to improve accuracy, reasoning, and instruction following. Maintained high consistency and quality standards while completing high-volume text evaluation tasks within defined accuracy and review thresholds, using internal annotation and QA workflow.