AI Evaluator – Prompt Engineering and RLHF Evaluation
As an AI Evaluator, I assessed the outputs of large language models (LLMs) using reinforcement learning from human feedback (RLHF) techniques. My work involved prompt engineering, AI model evaluation, and structured data annotation to ensure high-quality AI-driven solutions. I applied strong critical thinking and maintained ethical standards in all evaluations. • Evaluated LLM-generated responses for accuracy, coherence, and relevance • Authored structured feedback and quality ratings on diverse language tasks • Applied RLHF and prompt engineering methodologies to improve model performance • Engaged in data annotation to support ongoing AI training and evaluation efforts