AI Trainer & Domain Expert — Outlier (Scale AI)
Worked as an AI trainer and domain expert on the Outlier (Scale AI) platform, specializing in the evaluation and improvement of AI-generated content. Assessed and rated the accuracy, quality, and safety of large language model (LLM) outputs, spanning domains such as cybersecurity, code review, mathematics, and general AI response evaluation. Provided structured feedback, comparative output ranking (RLHF), and technical quality reviews for AI-generated text and code. • Conducted detailed LLM response assessments for correctness, reasoning, and instruction relevance. • Evaluated code solutions, cybersecurity recommendations, and mathematical problem-solving. • Reviewed and rated safety and ethical alignment of AI outputs, identifying edge cases and failure modes. • Delivered over 500+ completed labeling and evaluation tasks using Outlier (Scale AI) as proprietary tooling.