LLM Instruction Tuning and Human Feedback Annotation
Worked on large scale instruction tuning and human feedback data generation for a conversational AI model. The project involved writing high quality prompt and response pairs, ranking multiple model outputs based on helpfulness, accuracy, and safety, and providing structured feedback for reinforcement learning from human feedback workflows. I also reviewed annotator work to ensure consistency and reduce bias across datasets. The dataset included over 200,000 prompt response pairs spanning general knowledge, reasoning, coding, and long form generation tasks. Strict quality guidelines were followed, including double review sampling, inter annotator agreement checks, and rubric based scoring to maintain consistency and reliability.