AI Training & LLM Evaluation Project
Contributed to AI training projects focused on improving large language models (LLMs). Responsibilities included evaluating and ranking AI-generated responses, annotating text data, identifying inaccuracies and inconsistencies, and ensuring outputs met detailed quality and compliance guidelines. Worked with structured instructions, applied consistent evaluation criteria, and maintained high accuracy standards in a remote, task-based environment.