Content Moderation & Safety Annotation
Performed moderation and labeling of content to identify harmful, misleading, or policy-violating material. Classified data based on predefined categories such as toxicity, spam, and unsafe content. Ensured consistency and compliance with moderation guidelines while maintaining high attention to detail. Flagged ambiguous or borderline cases and applied careful judgment to maintain dataset integrity. Contributed to improving AI safety and reliability by supporting the development of responsible AI systems.