LLM Evaluation and AI Model Fine-Tuning
As part of an AI model fine-tuning project, I evaluated responses generated by large language models (LLMs) to ensure accuracy, coherence, and alignment with human intent. The work involved: Ranking model outputs based on factual correctness, fluency, and relevance. Classifying responses for bias detection, sentiment analysis, and content safety. Writing and refining prompt-response pairs to enhance model learning. Contributing to reinforcement learning (RLHF) by training models with human feedback loops.