AI Training & Full Stack Developer Contributor
As an AI Training & Full Stack Developer Contributor at Outlier AI, I was responsible for refining and rewriting over 1,000 complex code snippets in the MERN stack to fine-tune large language models (LLMs) for Google DeepMind and OpenAI. My contributions directly impacted the improvement of LLM reasoning and code production capabilities. During this period, I maintained a high quality acceptance rate while performing reinforcement learning from human feedback (RLHF) reviews. • Refined and rewrote computer code as annotated data for AI model training • Identified and corrected logic errors in AI-generated code outputs • Ensured quality through rigorous review cycles using RLHF • Worked on next-generation LLMs for industry leaders such as Google DeepMind and OpenAI.