LLM Training and Content understanding
Worked on large-scale NLP data labeling projects to support the training and evaluation of large language models. Responsibilities included annotating text for named entity recognition (persons, locations, organizations, products), intent and topic classification, sentiment and emotion recognition, and response quality evaluation. The project involved labeling tens of thousands of text samples, including user queries, chatbot conversations, and generated responses. Strict annotation guidelines were followed, with special attention to edge cases, ambiguity resolution, and contextual accuracy. Quality measures included peer reviews, gold-standard tests, inter-annotator agreement checks, and continuous feedback loops to maintain high accuracy and consistency.