AI Data Annotation & Evaluation Project
I work on AI data annotation and evaluation tasks focused on improving the performance of language models. My responsibilities include labeling and classifying text data, evaluating AI-generated responses based on accuracy, relevance, helpfulness, and safety criteria, and ensuring outputs align with structured guidelines. I also compare multiple AI responses to determine which best follows instructions and provides the most accurate and useful information. The work requires high attention to detail, consistency, and adherence to strict quality standards to ensure reliable training data for model improvement.