For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
B

Bill Malea

Engineering Expert Reviewer (Multi‑Disciplinary)

Kenya flagNairobi, Kenya
$90.00/hrExpertMercorAws SagemakerOther

Key Skills

Software

MercorMercor
AWS SageMakerAWS SageMaker
Other

Top Subject Matter

Engineering (Software, Electrical, Chemical)
Software Engineering (AI/LLM)
AI Systems Evaluation

Top Data Types

TextText
Computer Code ProgrammingComputer Code Programming
DocumentDocument

Top Task Types

Prompt Response Writing SFT
Evaluation Rating
Computer Programming Coding
Function Calling
RLHF

Freelancer Overview

Engineering Expert Reviewer (Multi‑Disciplinary). Brings 6+ years of professional experience across legal operations, contract review, compliance, and structured analysis. Core strengths include Mercor, Internal, and Proprietary Tooling. Education includes Master of Science, University College London (2026) and Bachelor of Science, University of Eldoret (2020). AI-training focus includes data types such as Computer Code, Programming, and Text and labeling workflows including Evaluation and Rating.

ExpertEnglishSwahili

Labeling Experience

Mercor

Code Quality Software Engineer Reviewer

Mercor
Reviewed production-level software code and associated outputs as training data for AI reasoning models. Labeled, corrected, and improved step-by-step outputs to strengthen model reliability and reduce hallucinations. Applied engineering expertise to align training data with authentic real-world system behavior. • Evaluated model-generated reasoning traces for validity • Enhanced data completeness and logical consistency • Improved engineering outcome alignment through targeted edits • Ensured high-quality LLM code training samples

Reviewed production-level software code and associated outputs as training data for AI reasoning models. Labeled, corrected, and improved step-by-step outputs to strengthen model reliability and reduce hallucinations. Applied engineering expertise to align training data with authentic real-world system behavior. • Evaluated model-generated reasoning traces for validity • Enhanced data completeness and logical consistency • Improved engineering outcome alignment through targeted edits • Ensured high-quality LLM code training samples

2026 - 2026
Mercor

Engineering Expert Reviewer (Multi‑Disciplinary)

Mercor
Reviewed and evaluated technical submissions for inclusion in AI training pipelines. Applied standardized criteria to assess the quality and applicability of engineering solutions for large-scale machine learning training. Ensured data met rigorous requirements before incorporation into model development workflows. • Validated logical and methodological soundness across multidisciplinary engineering data • Detected ambiguities and flagged unsafe assumptions impacting model accuracy • Provided detailed judgment beyond mere code review • Contributed to error reduction in AI output through refined labeling practices

Reviewed and evaluated technical submissions for inclusion in AI training pipelines. Applied standardized criteria to assess the quality and applicability of engineering solutions for large-scale machine learning training. Ensured data met rigorous requirements before incorporation into model development workflows. • Validated logical and methodological soundness across multidisciplinary engineering data • Detected ambiguities and flagged unsafe assumptions impacting model accuracy • Provided detailed judgment beyond mere code review • Contributed to error reduction in AI output through refined labeling practices

2026 - 2026

Software Engineer (AI Systems & Evaluation)

Text
Designed tasks and evaluation protocols to benchmark AI model quality. Analyzed and tested model responses for consistency, failure modes, and output reliability. Contributed to systematic process improvements for large-scale AI model evaluation. • Developed structured edge-case scenarios • Ran systematic quality assurance and benchmarking • Provided actionable feedback for reliability improvement • Applied multidomain engineering expertise to output assessment

Designed tasks and evaluation protocols to benchmark AI model quality. Analyzed and tested model responses for consistency, failure modes, and output reliability. Contributed to systematic process improvements for large-scale AI model evaluation. • Developed structured edge-case scenarios • Ran systematic quality assurance and benchmarking • Provided actionable feedback for reliability improvement • Applied multidomain engineering expertise to output assessment

2024 - 2025

Education

U

University of Eldoret

Bachelor of Science, Mathematics and Computer Science

Bachelor of Science
2016 - 2020
U

University College London

Master of Science, Embedded Systems Engineering

Master of Science
2026

Work History

E

Evolve Software Development Agency

Software Engineer / Systems Architect

London
2023 - 2025
F

Flutterwave

Senior Software Engineer

Nairobi
2022 - 2024