AI Data Evaluator & Content Reviewer, Outlier AI
Served as an AI Data Evaluator & Content Reviewer, focusing on evaluation and annotation tasks to support large language model training. Assessed AI-generated responses across factuality, helpfulness, instruction-following, and harmlessness, utilizing structured rubrics and workflows. Reviewed code, technical documentation, and web content for correctness and adherence to standards. • Conducted high-volume response and content evaluation, maintaining a >95% quality acceptance rate. • Performed comparative A/B testing and multi-metric scoring of model outputs. • Identified edge cases, model limitations, and contributed to quality assurance frameworks. • Evaluated technical content, code generation, and web outputs for multidimensional accuracy.