LLM Text Data Evaluation and Annotation for NLP Training
Worked on a large-scale LLM training project involving the annotation and evaluation of English text data to improve model quality and alignment. Responsibilities included classifying prompts, evaluating model-generated responses for factuality, coherence, and helpfulness, and providing written feedback for reinforcement learning with human feedback (RLHF). Also participated in prompt-response generation for supervised fine-tuning (SFT). The project required consistently high levels of attention to detail, understanding of nuanced language, and adherence to strict quality assurance guidelines. Delivered over 15,000+ labeled examples with a 98% reviewer acceptance rate.