Contributors will help train and evaluate AI systems by creating, testing, and refining LLM prompts and responses within a variety of real-world automation workflows. Your responsibilities will span from designing complex prompt frameworks and quality evaluation rubrics to structuring inputs for reliable outputs and reviewing AI-generated content for adherence to criteria such as accuracy, structure, and tone. To succeed, you’ll need hands-on experience integrating LLM APIs (such as OpenAI, Claude, or Gemini), advanced prompt engineering skills, and the ability to design and assess AI automation pipelines. Strong English language skills (B2 or higher), as well as experience with content generation automation, data structuring, and workflow integration, are essential.
Estimated Total Earnings
$900.00
Pay per Hour
$45.00/hr
Time Requirement
20+ hrs/week
Duration
3-6 months
LLM workflow prompts and evaluations
Software
Hiring Type
Required Location
Workload / Schedule
Weekly commitment can be adjusted based on throughput targets. Project duration is expected to run for 3 to 6 months. Labelers should follow milestone deadlines and quality checkpoints.
Software
Data Type
Task Types
Subject Matter / Industry
Proposals: 287
Invites sent: 0
Unanswered invites: 0
Share link