Outlier AI Prompt Evaluation & Annotation Project
Worked on a high-impact annotation project focused on enhancing the reasoning and ethical performance of AI models. The scope included creating image-based and text-based prompts designed to assess model capabilities across problem-solving, ethical reasoning, and bias detection. Tasks involved annotating model responses, labeling logical inconsistencies, identifying hallucinations, and flagging harmful or inappropriate content. Delivered detailed feedback loops to guide model improvement, involving multiple annotation types like emotion tagging, named entity recognition, and red-teaming edge-case prompts. The project helped shape model training strategies by contributing over 500+ labeled samples, with adherence to strict quality guidelines, benchmark rubrics, and weekly performance audits.