AI Trainer, Scale AI - Independent Contractor
This role involves evaluating and refining AI-generated text to ensure its accuracy, coherence, and contextual suitability. It includes developing specific rubric criteria to assess artificial intelligence model outputs against intended user goals and contributing benchmarking tasks for large language models. Feedback is provided on submissions to maintain high quality standards. • Refined AI-generated text for improved accuracy and relevance. • Developed and applied rubric criteria for AI output evaluation. • Authored advanced, peer-reviewed questions to test LLMs. • Reviewed contributor work and delivered quality feedback.