AI Training Specialist – RLHF & Response Evaluation
I evaluated and refined large language model outputs for clarity, coherence, and instruction-following using RLHF methodologies on the Aether project. Tasked with reviewing AI-generated text, I flagged errors, rewrote responses, and assessed outputs against multi-criteria rubrics. My role emphasized factual accuracy, logical consistency, and high-quality language use. • Conducted prompt writing and response ranking to enhance model performance • Identified and addressed hallucinations and factual inaccuracies • Followed detailed style guides and rubric-based evaluation protocols • Maintained consistent high quality and pass rates in annotation tasks