LLM Prompt–Response Annotation & Evaluation for AI Applications
Worked on prompt–response annotation and evaluation tasks for LLM-based applications. Responsibilities included creating high-quality prompts, reviewing AI-generated responses, rating outputs for accuracy and relevance, and improving response quality through iterative feedback. Handled structured datasets in JSON/CSV format and ensured consistency, clarity, and correctness in labeled data. Applied quality control by cross-checking outputs and refining prompts to reduce hallucinations and improve model reliability. Contributed to AI chatbot and automation projects where labeled data supported fine-tuning and performance optimization.