Prompt-Response Comparison & Ranking (RLHF)
Ranked multiple model-generated responses to prompts based on accuracy, helpfulness, coherence, and stylistic quality. Applied detailed rubric-based assessments to support reinforcement learning through human feedback (RLHF). Helped fine-tune model behavior by providing comparative judgments and constructive analysis to guide reward modeling.