AI Quality Assurance, RLHF Rater
I engaged in daily interactions with Large Language Models to provide Reinforcement Learning from Human Feedback (RLHF). My responsibilities included evaluating AI-generated outputs and identifying hallucinations or factual inconsistencies. The goal was to improve the accuracy and integrity of language model responses. • Evaluated and ranked AI-generated text for relevancy and correctness. • Detected and flagged hallucinations in model completions. • Provided detailed human feedback for ongoing LLM improvement. • Leveraged analytical skills to enhance the quality of AI outputs.