For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
Ana

Ana

AI Evaluation & LLM Quality Specialist – Technology

CANADA flag
North Vancouver, Canada
$25.00/hrIntermediateInternal Proprietary Tooling

Key Skills

Software

Internal/Proprietary Tooling

Top Subject Matter

No subject matter listed

Top Data Types

TextText

Top Label Types

Evaluation Rating

Freelancer Overview

I am a technical professional with over 10 years of experience in software engineering, technical leadership, and systems design, now specializing in AI evaluation and the creation of high-quality training data. My background in computer science and my expertise in structured problem solving allow me to break down complex tasks, identify edge cases, and ensure accuracy and consistency in data annotation projects. I have a strong track record of producing clear technical documentation, analyzing LLM behavior, and evaluating model outputs to improve reliability. I am fluent in both English and Spanish, comfortable working independently and asynchronously, and experienced in using tools and APIs to augment data labeling workflows. My focus on quality, bias detection, and detailed step-by-step reasoning makes me well-suited for roles involving data labeling, annotation, and AI training data for domains such as NLP, enterprise systems, and digital platforms.

IntermediateEnglishSpanish

Labeling Experience

LLM Output Evaluation and Annotation

Internal Proprietary ToolingTextEvaluation Rating
Participated in the evaluation and annotation of Large Language Model (LLM) outputs, focusing on response quality, reasoning accuracy, factual correctness, and alignment with user intent. Tasks included classifying responses, identifying reasoning errors, bias, hallucinations, and policy compliance issues, as well as providing structured feedback to improve model performance. Emphasis was placed on consistency, quality standards, and careful judgment rather than high-volume mechanical labeling. Quality assurance was maintained through guideline adherence and iterative review.

Participated in the evaluation and annotation of Large Language Model (LLM) outputs, focusing on response quality, reasoning accuracy, factual correctness, and alignment with user intent. Tasks included classifying responses, identifying reasoning errors, bias, hallucinations, and policy compliance issues, as well as providing structured feedback to improve model performance. Emphasis was placed on consistency, quality standards, and careful judgment rather than high-volume mechanical labeling. Quality assurance was maintained through guideline adherence and iterative review.

2024 - 2025

Education

C

Complutense University of Madrid

Bachelor of Science, Computer Science

Bachelor of Science
2010 - 2014

Work History

I

Independent

Project Manager

San José
2024 - Present
V

Vena Solutions

Software Development Manager

Vancouver
2022 - Present