LLM Data Annotation & AI Response Evaluation
Worked as a contributor on AI training data projects involving evaluation and annotation of large language model outputs. Tasks included reviewing conversation transcripts, rating model responses based on relevance, accuracy, and safety, and generating concise summaries of dialogues. Performed text classification and response quality evaluation aligned with Reinforcement Learning from Human Feedback (RLHF) guidelines to help improve model performance. Ensured adherence to annotation instructions and maintained high quality standards while completing micro-tasks across different AI training datasets.