Data Annotation
LLM project was fine-tuning a customer support chatbot for an e-commerce company to accurately answer product questions, process refunds, and escalate complex issues. The data labeling tasks included annotating user intent, tagging sentiment, ranking multiple AI responses for quality, rewriting ideal responses, and flagging hallucinations or policy violations. The project used approximately 120,000 high-quality annotated conversations drawn from a larger dataset of about 500,000 interactions, completed by a team of around 50 annotators and QA reviewers over three to four months. Strict quality measures were followed, including detailed annotation guidelines, annotator certification tests, inter-annotator agreement targets above 0.75, hidden gold-standard checks, and multi-layer quality audits. These controls ensured improved response accuracy, reduced hallucinations, and safe, policy-compliant model behavior.