LLM Response Evaluation and Prompt Annotation for RLHF
Participated in Reinforcement Learning from Human Feedback (RLHF) projects by evaluating and ranking AI-generated responses based on multiple quality criteria such as relevance, coherence, and accuracy. Provided corrective feedback and annotations to improve model outputs. Tasks included prompt refinement, response rating, and fine-tuning support. Worked with large datasets in a high-accuracy environment adhering to strict quality guidelines.