AI Code Evaluator & LLM Data Labeler (Full-Stack AI Specialist, Micro1/Mercor/Outlier)
Reviewed and evaluated AI-generated code outputs as part of advanced LLM training workflows for production systems. Used structured rubrics to assess code submissions across several complexity tiers, focusing on API design, data structures, and integration patterns. Ensured reliability and scalability by incorporating automated evaluation workflows and validating against multiple testing environments. • Evaluated code for correctness, performance, and production-readiness. • Specialized in assessing RESTful APIs and MongoDB schema integration. • Ran automated end-to-end testing using established tools. • Contributed to improving robustness and quality in LLM pipelines.