Multilingual AI Response Evaluation & Data Annotation (Arabic–English)
Worked on evaluating and improving AI-generated responses as part of reinforcement learning with human feedback (RLHF) workflows. Assessed outputs for factual accuracy, relevance, coherence, and safety, ensuring alignment with human expectations. Ranked multiple responses to identify the best-performing outputs and contributed to refining model behavior. Identified hallucinations, bias, and inconsistencies, particularly in complex domains such as medical queries, while strictly adhering to annotation guidelines and maintaining high consistency.