For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
Shahnawaz Ahmed

Shahnawaz Ahmed

AI Trainer LLM Evaluation & RLHF Specialist Multimodal Annotation Expert

India flagNEW DELHI, India
$18.00/hrExpertAws SagemakerCVATDoccano

Key Skills

Software

AWS SageMakerAWS SageMaker
CVATCVAT
DoccanoDoccano
Google Cloud Vertex AIGoogle Cloud Vertex AI
LabelboxLabelbox
MindriftMindrift
Scale AIScale AI
Snorkel AISnorkel AI
SuperAnnotateSuperAnnotate
SuperviselySupervisely
Label StudioLabel Studio

Top Subject Matter

No subject matter listed

Top Data Types

Computer Code ProgrammingComputer Code Programming
ImageImage
TextText

Top Task Types

Evaluation Rating
Fine Tuning
Prompt Response Writing SFT
Question Answering
RLHF

Freelancer Overview

AI Trainer and LLM Evaluation Specialist with hands-on experience in Reinforcement Learning from Human Feedback (RLHF), prompt engineering, and multi-modal annotation. Worked on projects for Amazon Phoenix, Google Gemini, and Fika GenAI via Invisible Technologies, Outlier AI, and Mindrift. Skilled in side-by-side evaluation, hallucination detection, and reward-model training data curation that improve factuality, safety, and alignment in LLMs.

ExpertUrduHindiFrenchGermanEnglish

Labeling Experience

Label Studio

Invisible Technologies — Amazon Phoenix (RLHF 3H Framework)

Label StudioTextQuestion AnsweringRLHF
Evaluated large volumes of model-generated responses under the Honest–Helpful–Harmless (3H) framework for Amazon’s Phoenix RLHF pipeline. Applied hallucination detection, bias and safety screening, and factuality verification across text-based dialogue and Q&A tasks. Authored structured preference rationales that clarified trade-offs between helpfulness and safety. Contributed to a measurable ~40% improvement in factual accuracy and reduced unsafe outputs across downstream LLM deployments.

Evaluated large volumes of model-generated responses under the Honest–Helpful–Harmless (3H) framework for Amazon’s Phoenix RLHF pipeline. Applied hallucination detection, bias and safety screening, and factuality verification across text-based dialogue and Q&A tasks. Authored structured preference rationales that clarified trade-offs between helpfulness and safety. Contributed to a measurable ~40% improvement in factual accuracy and reduced unsafe outputs across downstream LLM deployments.

2023 - 2024

Education

J

Jamia Hamdard University

Bachelor of Technology, Computer Science & Engineering

Bachelor of Technology
2017 - 2021

Work History

G

G2i (Databricks, Bosch, IBM)

Java Engineer & Code Reviewer

New Delhi
2021 - 2022