For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
T

Tochukwu Ndumnwere

Python & AI Systems Engineer (AI Training/Evaluation)

Nigeria flagAbuja, Nigeria
$25.00/hrExpertOtherScale AISurge AI

Key Skills

Software

Other
Scale AIScale AI
Surge AISurge AI
RemotasksRemotasks
OneFormaOneForma
MercorMercor
Data Annotation TechData Annotation Tech

Top Subject Matter

AI-generated Python code evaluation
LLM-driven applications
AI-generated code/text evaluation

Top Data Types

TextText
DocumentDocument
Computer Code ProgrammingComputer Code Programming

Top Task Types

Entity (NER) ClassificationEntity (NER) Classification
ClassificationClassification
Computer Programming/CodingComputer Programming/Coding
Prompt + Response Writing (SFT)Prompt + Response Writing (SFT)
Evaluation/RatingEvaluation/Rating
Object DetectionObject Detection
Text GenerationText Generation
Question AnsweringQuestion Answering
Function CallingFunction Calling

Freelancer Overview

Python & AI Systems Engineer (AI Training/Evaluation). Brings 8+ years of professional experience across legal operations, contract review, compliance, and structured analysis. Core strengths include Internal, Proprietary Tooling, and Outlier Ai. AI-training focus includes data types such as Computer Code and Programming and labeling workflows including Evaluation, Rating, and Classification.

ExpertEnglish

Labeling Experience

Python & AI Systems Engineer (AI Training/Evaluation)

In this role, I evaluated and validated AI-generated Python code outputs, reviewing responses for correctness, logical consistency, and execution reliability. My work focused on identifying failure patterns in AI reasoning and improving the quality of LLM outputs at scale. I utilized internal Python-based tools for code assessment and automated workflow validation. • Assessed 100+ AI-generated Python code responses weekly • Designed and improved validation workflows for API-integrated AI systems • Contributed structured feedback to boost model reliability • Operated remotely in asynchronous AI training environments

In this role, I evaluated and validated AI-generated Python code outputs, reviewing responses for correctness, logical consistency, and execution reliability. My work focused on identifying failure patterns in AI reasoning and improving the quality of LLM outputs at scale. I utilized internal Python-based tools for code assessment and automated workflow validation. • Assessed 100+ AI-generated Python code responses weekly • Designed and improved validation workflows for API-integrated AI systems • Contributed structured feedback to boost model reliability • Operated remotely in asynchronous AI training environments

2025 - Present

Freelance Data Annotation & QA Specialist

OtherText
I evaluated AI-generated outputs with a focus on reasoning quality and adherence to safety guidelines. I performed audits on datasets and documented edge cases impacting AI training projects. Identified and reported inconsistencies to help improve data reliability. • Conducted systematic evaluations of AI outputs for correctness. • Assessed compliance with specified project standards and safety requirements. • Audited training data for gaps or anomalies. • Provided actionable insights to enhance labeling accuracy.

I evaluated AI-generated outputs with a focus on reasoning quality and adherence to safety guidelines. I performed audits on datasets and documented edge cases impacting AI training projects. Identified and reported inconsistencies to help improve data reliability. • Conducted systematic evaluations of AI outputs for correctness. • Assessed compliance with specified project standards and safety requirements. • Audited training data for gaps or anomalies. • Provided actionable insights to enhance labeling accuracy.

2023 - 2024

Freelance Data Annotator

OtherTextClassification
This role focused on annotating datasets for machine learning intent classification and sentiment tasks. Maintaining high annotation accuracy was a priority. Quality assurance was integral to every assignment to meet project standards. • Processed large-scale text datasets for training purposes. • Performed intent and sentiment annotation tasks. • Ensured dataset quality through self-review and correction. • Successfully achieved above 98% accuracy rates.

This role focused on annotating datasets for machine learning intent classification and sentiment tasks. Maintaining high annotation accuracy was a priority. Quality assurance was integral to every assignment to meet project standards. • Processed large-scale text datasets for training purposes. • Performed intent and sentiment annotation tasks. • Ensured dataset quality through self-review and correction. • Successfully achieved above 98% accuracy rates.

2023 - 2023

AI Data Annotation Intern

OtherTextEntity Ner ClassificationClassification
I was responsible for annotating and evaluating large volumes of Python code and NLP datasets for AI training and model evaluation. My tasks included reviewing AI-generated responses for reasoning accuracy, instruction adherence, and identifying inconsistencies in outputs. Structured feedback was provided regularly to refine and improve model performance. • Annotated and evaluated 1,000+ Python/NLP tasks • Assessed AI responses for logic and accuracy • Identified consistency issues in AI code outputs • Enhanced dataset quality through targeted annotation

I was responsible for annotating and evaluating large volumes of Python code and NLP datasets for AI training and model evaluation. My tasks included reviewing AI-generated responses for reasoning accuracy, instruction adherence, and identifying inconsistencies in outputs. Structured feedback was provided regularly to refine and improve model performance. • Annotated and evaluated 1,000+ Python/NLP tasks • Assessed AI responses for logic and accuracy • Identified consistency issues in AI code outputs • Enhanced dataset quality through targeted annotation

2021 - 2023

AI Prompt Evaluation Intern

OtherText
This role involved reviewing prompts and evaluating responses from conversational AI systems. I applied safety guidelines and documented failures to enhance AI training. Structured feedback was provided to improve prompt effectiveness and compliance. • Analyzed AI-generated outputs for accuracy and logical reasoning. • Identified policy-violating or ambiguous content in system interactions. • Maintained diligent records of prompt issues and workflow gaps. • Contributed to refining guidelines based on observed edge cases.

This role involved reviewing prompts and evaluating responses from conversational AI systems. I applied safety guidelines and documented failures to enhance AI training. Structured feedback was provided to improve prompt effectiveness and compliance. • Analyzed AI-generated outputs for accuracy and logical reasoning. • Identified policy-violating or ambiguous content in system interactions. • Maintained diligent records of prompt issues and workflow gaps. • Contributed to refining guidelines based on observed edge cases.

2021 - 2021

Education

U

University of Nigeria, Nsukka

Bachelor of Science, Electronic Engineering

Bachelor of Science
2014 - 2020

Work History

T

Turing Developers

Python & AI Systems Engineer

Abuja
2025 - Present
N

N/A

AI Tool Workflow Analyst

Abuja
2025 - 2026