AI Training Data Annotation & Model Evaluation Project
Worked on AI training and data labeling projects focused on improving large language models. Responsibilities included evaluating AI-generated responses for correctness, reasoning quality, and edge-case handling, particularly in coding-related tasks. Identified failure points in model outputs and provided detailed, structured feedback to support model improvement and fine-tuning. Additionally contributed to Arabic language data annotation by recording high-quality audio prompts, performing text and audio classification, and writing complex prompts for supervised fine-tuning. Followed strict quality guidelines to ensure linguistic accuracy, consistency, and reliability across datasets. Maintained high annotation accuracy while handling large volumes of data.