AI Trainer / LLM Evaluator
Assessed AI-generated responses for correctness, clarity, and relevance across various evaluation scenarios. Conducted rankings and structured feedback to enhance large language model (LLM) performance. Compared and evaluated multiple AI outputs to identify the most suitable and accurate results. • Applied standardized guidelines to ensure consistency in evaluation. • Flagged errors, low-quality outputs, and inconsistencies with user intent. • Optimized prompts to improve AI response quality and relevance. • Utilized ChatGPT and similar AI evaluation workflows throughout projects.