Image-Based QA Annotation & Linguistic Evaluation for AI Training
Evaluated and approved large-scale image-based question-answering datasets for training multimodal AI models. Duties included checking the correctness of labels, evaluating the correctness of answers, and checking for consistency between images and answers. Identified accurate text from images for evidence-based validation and reformatted answers for self-contained clarity, grammatical correctness, and fluent natural U.S. English language usage. Used structured evaluation criteria for classification, question-answering, and text generation tasks, maintaining quality scores above 98% at all times. Identified ambiguous or unanswered questions with sound reasoning and provided actionable feedback to annotators to minimize systematic errors. Performed linguistic evaluation for grammar, punctuation, semantic correctness, tone, and relevance. Completed work on various platforms (Labelbox, Scale AI, Appen Connect, TELUS AI) in a fast-paced remote setting, averaging about 7 minutes per task