AI Training Expert & Data Evaluator
I conducted high-quality data labeling and annotation to help train and refine advanced Large Language Models (LLMs). My core responsibilities included applying Reinforcement Learning from Human Feedback (RLHF) by ranking, rewriting, and rating AI-generated responses for accuracy, safety, and helpfulness. I also evaluated multilingual outputs and inspected programmatic reasoning and Python-based logic outputs for correctness and optimization. • Designed and tested complex prompts to identify model vulnerabilities and edge cases. • Performed in-depth evaluation of AI-generated responses with a focus on semantic accuracy and cultural alignment. • Collaborated with global teams to ensure quality and throughput within evolving project guidelines. • Provided feedback on model alignment, fact-checking, and hallucination identification.