AI Engineer – LLM Evaluation & Validation
This role involved evaluating and validating the outputs of large language models to ensure quality and correctness. The process required methodical assessment of model responses to user queries and prompts. Feedback loops and detailed evaluations contributed to continuous improvement of AI-driven learning features. • Conducted validation and assessment of LLM-generated text outputs • Provided detailed feedback and ratings to enhance model accuracy • Supported reliability testing strategies for AI functions • Collaborated on evaluation processes with engineering teams