AI Trainer & RLHF Specialist
As an AI Trainer & RLHF Specialist at Outlier, I produced preference ranking data for LLM fine-tuning and assessed LLM outputs for helpfulness, safety, and factual accuracy. I crafted adversarial and red-team prompts to uncover model vulnerabilities and generated SFT demonstration data for code, reasoning, and instruction tasks. I maintained exceptional quality scores across all annotation projects. • Produced high-quality preference and ranking data for language model fine-tuning. • Crafted adversarial/red-team prompts for safety evaluation and model robustness. • Generated SFT demonstration datasets, including code generation and reasoning tasks. • Consistently achieved over 95% quality scores, adhering to detailed annotation rubrics.