Candidates must have a bachelor’s degree or equivalent professional experience in a relevant field, with near-native or native proficiency in Hebrew and at least C1 proficiency in English. Required qualifications include experience in trust and safety, content moderation, LLM red teaming, and strong familiarity with safety domains such as hate, harassment, violence, sexual content, self-harm, and misinformation. Applicants must demonstrate strong judgment, resilience to explicit content, and prior experience using AI tools such as Perplexity, Gemini, or ChatGPT. Project contributors will annotate, curate, and review model responses for safety, accuracy, and policy compliance. Work includes generating adversarial cases to test model robustness, scoring responses, documenting failure modes, and maintaining annotation standards. This work is critical to preventing unsafe outputs and involves collaboration with a global team to improve leading AI models.
Estimated Total Earnings
$3,200.00
Pay per Hour
$32.00/hr
Time Requirement
20+ hrs/week
Duration
3-6 months
Hebrew & English LLM outputs on safety and adversarial behavior
Software
Hiring Type
Required Location
Workload / Schedule
Expected weekly commitment is at least 20 hours. Project duration is expected to run for 3 to 6 months. Labelers should follow milestone deadlines and quality checkpoints.
Software
Data Type
Label Types
Subject Matter / Industry
Language
Proposals: 0
Invites sent: 0
Unanswered invites: 0
Share link