STEM Response Evaluator
I contributed to large language model (LLM) training through reinforcement learning from human feedback (RLHF). My responsibilities included evaluating STEM-related responses and providing feedback to improve model accuracy. I utilized my advanced knowledge in physics and technical expertise to ensure high-quality evaluations. • Conducted RLHF tasks focused on science, technology, engineering, and mathematics content. • Provided detailed evaluations and feedback on model-generated text. • Applied critical reasoning to assess factual and conceptual accuracy in responses. • Ensured alignment with project guidelines and quality standards.