AI Text Annotation and Evaluation
I annotated and evaluated AI-generated text outputs using structured guidelines. Some tasks included classification of responses, quality evaluation and rating across multiple dimensions such as factuality, coherence, verbosity, and transcript of audio but mainly text where required. Additionally, this included comparing multiple model outputs or a thread of outputs across dimensions such as following instructions across multiple counts, and identifying quality issues to support improvement of large language model training datasets. This work required careful guideline adherence and attention to detail.