Candidates must have a bachelor’s degree or higher in a relevant field; be near-native or native in Korean; have at least C1 English proficiency; and show experience in Trust & Safety, content moderation, or policy operations at a senior level. Direct experience in LLM red-teaming/adversarial testing, documented edge-case mitigation, and localization/translation is highly preferred. Strong analytical writing and emotional resilience are essential due to the challenging nature of the content. Contributors will review, annotate, and evaluate AI-generated responses, focusing on explicit safety issues such as bias, harassment, and misinformation. Responsibilities include fact-checking, reviewing policy alignment, documenting rationales, and identifying edge cases in both Korean and English. Annotations will inform foundation-model labs, preventing unsafe or adversarial AI outputs through refined safety policy application.
Estimated Total Earnings
$3,200.00
Pay per Hour
$32.00/hr
Time Requirement
20+ hrs/week
Duration
3-6 months
AI-generated safety and policy evaluation in Korean/English
Software
Hiring Type
Required Location
Workload / Schedule
Expected weekly commitment is at least 20 hours. Project duration is expected to run for 3 to 6 months. Labelers should follow milestone deadlines and quality checkpoints.
Software
Data Type
Label Types
Subject Matter / Industry
Language
Proposals: 0
Invites sent: 0
Unanswered invites: 0
Share link