LLM Reasoning and Factuality Evaluation
Performed Reinforcement Learning from Human Feedback (RLHF) to improve Large Language Model (LLM) accuracy. Tasks included evaluating model-generated responses for logical consistency, factual precision, and safety compliance. I identified edge cases in reasoning and refined prompts to test model boundaries. My focus was on providing nuanced feedback to reduce hallucinations and ensure the model followed complex multi-step instructions.