RWS
Actively involved in the annotation and quality assurance of large language model (LLM) outputs for English conversational datasets. Tasks included rewriting user prompts, verifying assistant intent alignment, labeling question-answer pairs, and evaluating multi-turn dialog responses. Delivered consistent high-accuracy results (>95%) across QA reviews, NER tagging, and instruction-based prompt refinement. Communicated feedback in structured formats to help optimize LLM behavior for real-world use cases. Operated in a fully remote environment using version-controlled workflows and detailed guidelines.