AI Training Contractor - LLM Evaluation and Data Annotation
I conducted AI model training involving Reinforcement Learning from Human Feedback (RLHF) by reviewing, ranking, and evaluating AI-generated code and technical outputs. Work included structured prompt engineering, benchmark testing, and extensive annotation of JSON datasets for software engineering, DevOps, and data science tasks. I consistently delivered 95%+ annotation quality and processed a high volume of outputs each month. • Evaluated and ranked AI code and responses across Python, JavaScript, and DevOps scenarios • Annotated over 1,000 structured JSON datasets for fine-tuning and SFT workflows • Created mathematical reasoning benchmarks in JSON for LLM capability assessments • Reviewed code for correctness, security vulnerabilities, and production readiness