AI Training Data Annotation & LLM Evaluation Specialist
Worked on AI training and data annotation projects involving text classification, Named Entity Recognition (NER), sentiment tagging, and LLM response evaluation. The scope included labeling structured and unstructured datasets to improve model accuracy, safety, and contextual understanding. Tasks involved rating AI-generated responses for factual correctness, relevance, bias detection, coherence, and policy compliance under RLHF frameworks. Handled datasets ranging from 5,000+ to 20,000+ entries per project while maintaining 95%+ accuracy standards. Followed strict annotation guidelines and participated in quality review cycles, including cross-validation and feedback-based corrections. Utilized tools like Labelbox, CVAT, and spreadsheet-based tracking systems to ensure consistency, timely delivery, and high-quality outputs suitable for fine-tuning and model optimization.