For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
R
Rishabh Jha

Rishabh Jha

AI Model Evaluator & Red Teamer

United Kingdom flagGuildford, United Kingdom
$50.00/hrIntermediateOther

Key Skills

Software

Other

Top Subject Matter

LLM Evaluation
Rlhf Domain Expertise
Alignment Domain Expertise

Top Data Types

TextText
ImageImage
DocumentDocument

Top Task Types

RLHFRLHF
Red TeamingRed Teaming
Fine-tuningFine-tuning
Evaluation/RatingEvaluation/Rating
Prompt + Response Writing (SFT)Prompt + Response Writing (SFT)
SegmentationSegmentation
ClassificationClassification

Freelancer Overview

AI Model Evaluator & Red Teamer with experience in AI training data, structured evaluation, and model quality assessment. Brings a strong background in reviewing LLM outputs for mathematical correctness, logical reasoning, consistency, and edge-case behaviour across benchmark tasks. Core strengths include rubric-based evaluation, adversarial prompt design, error analysis, RLHF support workflows, and applying detailed annotation standards with precision. Education includes Doctor of Philosophy, University of Surrey (2029) and Master of Science, University of Surrey (2025). AI-training focus includes data types such as Text and labeling workflows including Evaluation and Rating.

IntermediateEnglishHindi

Labeling Experience

AI Model Evaluator & Red Teamer

Text
Evaluated large language model (LLM) outputs for correctness, logical reasoning, and consistency across structured benchmark tasks. Designed and applied evaluation rubrics to assess model behavior and analyzed error patterns in diverse mathematical and computational problem sets. Performed adversarial prompt engineering to stress-test LLMs for alignment and robustness improvements. • Contributed to reinforcement learning from human feedback (RLHF) and alignment workflows. • Used internal or proprietary tools for annotation and evaluation tasks. • Specialized in evaluating text and mathematical responses. • Collaborated remotely and documented outcomes to inform model improvement.

Evaluated large language model (LLM) outputs for correctness, logical reasoning, and consistency across structured benchmark tasks. Designed and applied evaluation rubrics to assess model behavior and analyzed error patterns in diverse mathematical and computational problem sets. Performed adversarial prompt engineering to stress-test LLMs for alignment and robustness improvements. • Contributed to reinforcement learning from human feedback (RLHF) and alignment workflows. • Used internal or proprietary tools for annotation and evaluation tasks. • Specialized in evaluating text and mathematical responses. • Collaborated remotely and documented outcomes to inform model improvement.

2025 - 2026

Education

U

University of Surrey

Doctor of Philosophy, Artificial Intelligence

Doctor of Philosophy
2025 - 2029
U

University of Surrey

Master of Science, Artificial Intelligence

Master of Science
2024 - 2025

Work History

U

University of Surrey

Teaching Assistant

Guildford
2026 - Present
S

Self-Employed

Freelance Data Analyst

New Delhi
2022 - 2023