LLM Response Evaluation & RLHF Support
Evaluated AI-generated responses for correctness, relevance, tone, and safety using detailed project guidelines. Provided structured human feedback to improve LLM performance through RLHF-style workflows, including ranking, scoring, and justification of outputs.