For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
J
John Williams

John Williams

AI Evaluation Engineer

USA flagN/A, Usa
$50.00/hrExpertLabelboxAppenMercor

Key Skills

Software

LabelboxLabelbox
AppenAppen
MercorMercor
OneFormaOneForma
RemotasksRemotasks
Scale AIScale AI
TelusTelus
Internal/Proprietary Tooling

Top Subject Matter

AI Code Evaluation and Model Alignment

Top Data Types

TextText
Computer Code ProgrammingComputer Code Programming
ImageImage

Top Task Types

Text GenerationText Generation
RLHFRLHF
Evaluation/RatingEvaluation/Rating
Computer Programming/CodingComputer Programming/Coding
Red TeamingRed Teaming
Data CollectionData Collection
Function CallingFunction Calling
Prompt + Response Writing (SFT)Prompt + Response Writing (SFT)
TranscriptionTranscription
Fine-tuningFine-tuning
Text SummarizationText Summarization
Question AnsweringQuestion Answering
Object DetectionObject Detection
Entity (NER) ClassificationEntity (NER) Classification
ClassificationClassification
SegmentationSegmentation
Bounding BoxBounding Box

Freelancer Overview

AI Evaluation Engineer. Brings 3+ years of professional experience across complex professional workflows, research, and quality-focused execution. Core strengths include Internal and Proprietary Tooling. Education includes Master of Science, University of Illinois Urbana-Champaign (2025) and Bachelor of Science, Western Governors University (2023). AI-training focus includes data types such as Computer Code and Programming and labeling workflows including Evaluation and Rating.

ExpertEnglish

Labeling Experience

AI Evaluation Engineer

As an AI Evaluation Engineer, I contributed to the evaluation of AI-generated code and alignment tasks. My work involved testing and reviewing outputs of frontier models to ensure their correctness and utility. This process required designing challenging algorithmic tasks as well as providing reference solutions and feedback. • Evaluated code models' outputs for accuracy and adherence to task instructions • Designed adversarial test problems to stress-test AI reasoning and logic • Collaborated on open-source contributions to align AI agents with human software practices • Provided feedback to improve model performance based on test results

As an AI Evaluation Engineer, I contributed to the evaluation of AI-generated code and alignment tasks. My work involved testing and reviewing outputs of frontier models to ensure their correctness and utility. This process required designing challenging algorithmic tasks as well as providing reference solutions and feedback. • Evaluated code models' outputs for accuracy and adherence to task instructions • Designed adversarial test problems to stress-test AI reasoning and logic • Collaborated on open-source contributions to align AI agents with human software practices • Provided feedback to improve model performance based on test results

2025 - Present

Education

U

University of Illinois Urbana-Champaign

Master of Science, Computer Science

Master of Science
2024 - 2025
W

Western Governors University

Bachelor of Science, Computer Science

Bachelor of Science
2020 - 2023

Work History

S

Self

AI Evaluation Engineer

N/A
2025 - Present
H

Heaton Media Group

Software Engineer

N/A
2024 - 2024