LLM Response Evaluation & Quality Annotation
This project involved the systematic evaluation and annotation of AI-generated text responses to assess quality, accuracy, safety, and adherence to user instructions. The goal was to provide high-quality human feedback to train and refine large language models.