AI Trainer & Evaluator (Independent Contributor)
I performed Reinforcement Learning from Human Feedback (RLHF) tasks, evaluating and rating AI model outputs for factual accuracy, logical consistency, and safety. My work included writing detailed justifications for each rating to ensure clarity and transparency in model assessment. I consistently achieved high-quality scores in annotation and search relevance projects across various platforms. • Platforms used included Aurora Studio, Prolific, and Neevo. • Tasks covered text evaluation, annotation, and providing feedback on AI responses. • Emphasis was placed on analytical reasoning and linguistic precision. • Worked remotely with a focus on quality assurance and RLHF best practices.