For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
P

Peter Oburah

AI Engineer, Google (Gemini Project)

USA flagNairobi, Usa
$50.00/hrExpertScale AISuperannotateData Annotation Tech

Key Skills

Software

Scale AIScale AI
SuperAnnotateSuperAnnotate
Data Annotation TechData Annotation Tech

Top Subject Matter

Multimodal AI (image, video, text alignment and training)
Large Language Models (GPT-4/LLM Alignment and Safety)
Language Model Quality Control and Evaluation

Top Data Types

VideoVideo
TextText
ImageImage
DocumentDocument

Top Task Types

Fine-tuningFine-tuning
RLHFRLHF

Freelancer Overview

AI Engineer, Google (Gemini Project). Brings 1+ years of professional experience across complex professional workflows, research, and quality-focused execution. Core strengths include Internal, Proprietary Tooling, and Whimsical. Education includes Master of Science, Grand Valley State University (2027) and Bachelor of Science, Multimedia University of Kenya (2023). AI-training focus includes data types such as Video and Text and labeling workflows including Fine-tuning, RLHF, and Evaluation.

ExpertEnglish

Labeling Experience

AI Engineer, Google (Gemini Project)

VideoFine Tuning
I led the training and fine-tuning of multimodal AI models on large-scale video, image, and text datasets. My work included dataset curation, validation, and visual reasoning assessment to enhance model performance and ethical alignment. I collaborated cross-functionally to align models and improve outcomes for the Gemini Project at Google.• Curated and validated diverse dataset sources for improved model diversity and compliance. • Conducted structured multimodal evaluations to assess reasoning and quality. • Used internal proprietary tools for dataset management and training workflows. • Achieved measurable improvements in visual understanding and multimodal alignment.

I led the training and fine-tuning of multimodal AI models on large-scale video, image, and text datasets. My work included dataset curation, validation, and visual reasoning assessment to enhance model performance and ethical alignment. I collaborated cross-functionally to align models and improve outcomes for the Gemini Project at Google.• Curated and validated diverse dataset sources for improved model diversity and compliance. • Conducted structured multimodal evaluations to assess reasoning and quality. • Used internal proprietary tools for dataset management and training workflows. • Achieved measurable improvements in visual understanding and multimodal alignment.

2025 - 2026

Quality Control Analyst, OpenAI

Text
As a Quality Control Analyst, I evaluated training and evaluation data for consistency, quality, and ethical standards. I built and applied structured evaluation rubrics using Whimsical to ensure reliable LLM performance. My primary responsibility was reviewing LLM outputs for policy compliance and providing actionable feedback.• Designed and implemented scoring rubrics for text and image prompt outputs. • Reviewed responses for alignment quality and policy adherence in LLMs. • Used Whimsical and internal review tools for data assessment. • Facilitated cross-functional feedback exchange among trainers, users, and developers.

As a Quality Control Analyst, I evaluated training and evaluation data for consistency, quality, and ethical standards. I built and applied structured evaluation rubrics using Whimsical to ensure reliable LLM performance. My primary responsibility was reviewing LLM outputs for policy compliance and providing actionable feedback.• Designed and implemented scoring rubrics for text and image prompt outputs. • Reviewed responses for alignment quality and policy adherence in LLMs. • Used Whimsical and internal review tools for data assessment. • Facilitated cross-functional feedback exchange among trainers, users, and developers.

2024 - 2024

Senior AI Research Engineer, OpenAI

TextRLHF
I supported the alignment and evaluation of the GPT-4 LLM using Reinforcement Learning from Human Feedback (RLHF). I performed adversarial testing, structured model safety evaluations, and accuracy reviews to inform iterative improvement. My work guided the transition and refinement of OpenAI’s language models.• Applied adversarial testing to reveal model vulnerabilities and steerability limitations. • Conducted RLHF annotation tasks and evaluated LLM adherence to factual accuracy. • Worked collaboratively with research teams to operationalize feedback into training pipelines. • Improved safety and alignment metrics for the GPT-4 transition.

I supported the alignment and evaluation of the GPT-4 LLM using Reinforcement Learning from Human Feedback (RLHF). I performed adversarial testing, structured model safety evaluations, and accuracy reviews to inform iterative improvement. My work guided the transition and refinement of OpenAI’s language models.• Applied adversarial testing to reveal model vulnerabilities and steerability limitations. • Conducted RLHF annotation tasks and evaluated LLM adherence to factual accuracy. • Worked collaboratively with research teams to operationalize feedback into training pipelines. • Improved safety and alignment metrics for the GPT-4 transition.

2023 - 2024

Junior AI Research Engineer, OpenAI

TextRLHF
I trained language models using RLHF on diverse text datasets as a Junior AI Research Engineer. My role involved annotating responses, supporting model reasoning, and evaluating early-stage model alignment and usefulness. I worked closely with senior researchers to continuously refine training and evaluation workflows.• Annotated and rated text data for supervised LLM training with RLHF. • Evaluated model responses for reliability, usefulness, and factuality. • Assisted in quality and alignment control for language model outputs. • Used feather and internal tooling for annotation and workflow management.

I trained language models using RLHF on diverse text datasets as a Junior AI Research Engineer. My role involved annotating responses, supporting model reasoning, and evaluating early-stage model alignment and usefulness. I worked closely with senior researchers to continuously refine training and evaluation workflows.• Annotated and rated text data for supervised LLM training with RLHF. • Evaluated model responses for reliability, usefulness, and factuality. • Assisted in quality and alignment control for language model outputs. • Used feather and internal tooling for annotation and workflow management.

2023 - 2023

Education

G

Grand Valley State University

Master of Science, Data Science and Analytics

Master of Science
2025 - 2027
M

Multimedia University of Kenya

Bachelor of Science, Electrical and Telecommunications Engineering

Bachelor of Science
2019 - 2023

Work History

N

Next Technologies Limited

Network Engineer Intern

Nairobi
2022 - 2022