Freelance AI Research Evaluator & Data Consultant
Performed factual verification and quality assurance of AI-generated technical, mathematical, and scientific content for dataset integrity. Created and evaluated original problem sets to assess model reasoning and conceptual depth. Developed and applied evaluation guidelines focused on accuracy, factual robustness, and ethical bias detection. • Utilized platforms including Labelbox, Snorkel Flow, and AnnotatePro for annotation and review. • Evaluated and scored LLM outputs on technical correctness and contextual coherence. • Collaborated on creation of research-oriented benchmark datasets. • Applied bias-correction workflows to improve data fairness and reliability.