AI Annotator — Deccan AI
In this role, I evaluated large language model (LLM) responses for quality and accuracy. My main responsibility was to assess the relevance, correctness, and completeness of AI-generated answers using structured rubrics. I provided comparative assessments and actionable feedback to improve model outputs. • Evaluated over 150 LLM responses against detailed quality rubrics • Assessed multiple dimensions including accuracy, relevance, and completeness • Conducted comparative analysis to rank and select optimal responses • Provided feedback to inform and enhance model reliability