AI Training – RLHF, Data Modeling, Data Quality/Validation
Responsible for AI training by providing Reinforcement Learning from Human Feedback (RLHF) and related data modeling. Tasks included assessing AI outputs, evaluating language models, and ensuring data quality and validation. Developed logical reasoning approaches for optimizing AI behavior in real-world applications. • Consistently performed data quality checks for AI model outputs. • Provided structured feedback for language model improvements. • Applied data modeling techniques to AI training datasets. • Collaborated on process automation for quality validation.