Text Annotation & AI Response Labelling
Performed text annotation and response labeling tasks to help train and evaluate LLMs and coding models. My responsibilities included reviewing model generated responses and categorizing them based on their predefined guidelines and metrics like correctness, instruction following, safety, and overall quality. In some cases, I had to create and rate against a list of rubrics or criterion. The work involved analyzing or testing models and responses to complex prompts and identifying errors, halluciantions and edge cases.