Text Annotation & Model Output Evaluation (LLM Projects)
Worked on text annotation and AI model evaluation projects across multiple platforms. Responsibilities included: • Comparing receipts from different retail stores to ensure data consistency and structured alignment • Rubrics alignment and evaluation based on predefined quality criteria • Comparing outputs from different AI models and selecting the best response according to accuracy, clarity, and instruction adherence • Writing and refining Arabic prompts for LLM training and evaluation • Identifying logical inconsistencies, factual errors, and formatting issues in generated responses Maintained high accuracy standards, followed detailed guidelines, and ensured consistency across large batches of annotated data. Demonstrated strong analytical thinking and attention to detail while working independently in remote environments.