RLHF Logic Evaluation & Hallucination Detection
Evaluated and improved the reasoning capabilities of LLMs on complex tasks. Identified hallucinations in mathematical and logical workflows, providing step-by-step corrections. Ranked model responses based on truthfulness and instruction adherence. Drafted "Gold Standard" responses to teach the model correct logical deduction paths.