LLM - Python Engineer
Contributed to enterprise LLM evaluation and training workflows involving reinforcement learning from human feedback (RLHF) and supervised fine-tuning (SFT) of large language models. Developed Python tooling for dataset evaluation and AI model testing to improve model performance. Responsibilities included evaluating model outputs and providing structured feedback for continuous model refinement. • Performed RLHF and SFT workflows to fine-tune language models. • Designed and ran automated evaluation tests on LLM datasets. • Generated structured feedback for model improvement. • Collaborated with teams on enterprise-level LLM projects.