LLM Training and Reinforcement Learning with Human Feedback (RLHF)
Helped develop and evaluate large language models (LLMs) by annotating AI-generated responses for accuracy, coherence, and ethical concerns. Supported RLHF-based AI improvement, wherein I provided response ranking, prompt engineering, and data curation to improve AI-generated text. Task was to identify biases, improve factual grounding, and refine AI responses for real-world usability. Ensured high-quality annotations through cross-validation, inter-annotator agreement, and rigorous AI evaluation processes.