AI/ML Analyst – LLM Evaluation and RLHF Assessment
As an AI/ML Analyst at Innodata (Client: Meta), I evaluated large language model (LLM) responses using reinforcement learning from human feedback (RLHF) methodologies. My responsibilities included daily assessment of over 500 LLM outputs for quality, policy compliance, and adherence to guidelines. I contributed to enhancing AI safety and reliability through comprehensive QA and hallucination detection processes. • Performed RLHF-based evaluation and response ranking of LLM outputs. • Ensured policy compliance, factuality, and helpfulness through justification writing. • Identified and documented model hallucinations and edge cases. • Maintained strict SOP and dataset confidentiality protocols.