AI Training & RLHF Contributor
Participated in reinforcement learning from human feedback (RLHF) to fine-tune AI model responses for improved performance. Provided high-quality justifications for model outputs and audited model outputs for safety, bias, and cultural relevancy. Leveraged subject matter expertise in Swahili and engineering to enhance AI content moderation and safety protocols. • Identified patterns and anomalies in AI-generated text. • Rated and evaluated model outputs for appropriateness and bias. • Applied safety-first criteria to language review tasks. • Informed AI training development with cross-lingual insights and feedback.