AI Training Annotator (RLHF, LLM Evaluation)
I collaborated with AI development platforms to refine Large Language Models using Reinforcement Learning from Human Feedback (RLHF). My responsibilities included fact-checking model outputs and evaluating language correctness to improve model accuracy. This work demanded meticulous attention to instruction adherence and consistency. • Reviewed and rated AI-generated responses in various contexts • Provided structured feedback to improve AI reasoning and relevance • Fact-checked answers for factual accuracy and logical coherence • Worked within strict guidelines to maintain annotation quality