Data Annotation Project 2
In this project, I evaluated dialogues between annotators and a language model that generated two possible responses to each prompt. Rather than relying on predefined criteria, I was responsible for developing my own evaluation framework to assess the quality of each response. I then applied these criteria to justify my choice of the better response and documented my reasoning in detail. This process played a key role in refining the model’s output and improving its interpretive accuracy.