Comparative Evaluation of AI-Generated Responses to Prompts
Contributed to a reinforcement learning from human feedback (RLHF) project aimed at improving conversational AI. The task involved evaluating two different responses generated by an AI model for the same prompt, and ranking or scoring them based on relevance, coherence, helpfulness, and tone. Applied detailed evaluation guidelines to ensure unbiased and consistent judgment. Also provided written feedback or justification in certain cases to help guide model refinement. The project played a key role in training models to better align with human expectations in various conversational contexts, particularly in French.