For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
L
Lucian-Alin Balint

Lucian-Alin Balint

AI Specialist | Model Evaluation • Prompt Engineering • Red Teaming

United Kingdom flagLondon, United Kingdom
$25.00/hrIntermediateMicro1AppenScale AI

Key Skills

Software

Micro1
AppenAppen
Scale AIScale AI
RemotasksRemotasks
MindriftMindrift
TolokaToloka
LabelboxLabelbox
MercorMercor
Surge AISurge AI
TelusTelus
Other

Top Subject Matter

AI Ethics
LLM Safety Evaluation
Fashion Image Generation & Annotation

Top Data Types

TextText
ImageImage
DocumentDocument
AudioAudio

Top Task Types

Object DetectionObject Detection
RLHFRLHF
Bounding BoxBounding Box
TranscriptionTranscription
Red TeamingRed Teaming
Evaluation/RatingEvaluation/Rating
Data CollectionData Collection
Question AnsweringQuestion Answering
Text GenerationText Generation
Text SummarizationText Summarization
ClassificationClassification

Freelancer Overview

Interactive Prompt Specialist (Red Teaming). Core strengths include Mindrift and Gemini. AI-training focus includes data types such as Text and Image and labeling workflows including Red Teaming and Classification.

IntermediateEnglishRomanian

Labeling Experience

Mindrift

AI Safety & Red Teaming - LLM Evaluation

MindriftTextRed Teaming
Conducted red teaming and adversarial prompt testing on large language models to identify safety weaknesses and failure cases. Designed challenging prompts to evaluate instruction following, reasoning reliability, and policy compliance. Tested model responses for hallucinations, unsafe outputs, logical inconsistencies, and prompt injection vulnerabilities. Documented model weaknesses and contributed feedback to improve model robustness and alignment. Designed adversarial prompts targeting reasoning, safety, and instruction-following behavior Evaluated model outputs using structured rubric-based criteria Performed iterative prompt refinement to surface edge-case failures

Conducted red teaming and adversarial prompt testing on large language models to identify safety weaknesses and failure cases. Designed challenging prompts to evaluate instruction following, reasoning reliability, and policy compliance. Tested model responses for hallucinations, unsafe outputs, logical inconsistencies, and prompt injection vulnerabilities. Documented model weaknesses and contributed feedback to improve model robustness and alignment. Designed adversarial prompts targeting reasoning, safety, and instruction-following behavior Evaluated model outputs using structured rubric-based criteria Performed iterative prompt refinement to surface edge-case failures

2025 - Present

AI Model Evaluation Specialist – Outlier AI

OtherAudioTranscription
Evaluated AI-generated audio transcriptions for accuracy and alignment with spoken recordings. Reviewed model outputs to identify transcription errors, missing words, and contextual inconsistencies. Performed quality validation by comparing AI-generated transcripts against original audio inputs and documenting discrepancies. Applied structured evaluation guidelines to ensure consistency, accuracy, and reliable labeling across transcription validation tasks.

Evaluated AI-generated audio transcriptions for accuracy and alignment with spoken recordings. Reviewed model outputs to identify transcription errors, missing words, and contextual inconsistencies. Performed quality validation by comparing AI-generated transcripts against original audio inputs and documenting discrepancies. Applied structured evaluation guidelines to ensure consistency, accuracy, and reliable labeling across transcription validation tasks.

2025 - Present

AI Model Evaluation Specialist – Outlier AI

OtherImageBounding Box
Visual verification and bounding box annotation were performed on image datasets to assess AI object detection accuracy against ground-truth references. Tasks involved identifying objects and people, reviewing model-generated detections, and validating whether outputs matched the visual content correctly. Applied structured quality standards to ensure consistent object identification, accurate annotation, and reliable verification results across computer vision workflows. Maintained careful attention to detail while performing image review and quality control tasks.

Visual verification and bounding box annotation were performed on image datasets to assess AI object detection accuracy against ground-truth references. Tasks involved identifying objects and people, reviewing model-generated detections, and validating whether outputs matched the visual content correctly. Applied structured quality standards to ensure consistent object identification, accurate annotation, and reliable verification results across computer vision workflows. Maintained careful attention to detail while performing image review and quality control tasks.

2025 - Present

AI Product Prototype – Generative AI Image Fashion Imaging

OtherImageEvaluation RatingClassification
Developed generative AI pipelines for automated fashion image generation using advanced prompt engineering. Designed negative-constraint prompts to minimize visual artifacts and improve dataset quality. Tested scalable workflow integrations for AI-based image processing at production scale. • Created and annotated fashion datasets with high realism standards. • Eliminated anatomical anomalies through iterative prompt optimization. • Automated bounding box generation for object detection enhancement. • Leveraged workflow automation tools for dataset throughput scaling.

Developed generative AI pipelines for automated fashion image generation using advanced prompt engineering. Designed negative-constraint prompts to minimize visual artifacts and improve dataset quality. Tested scalable workflow integrations for AI-based image processing at production scale. • Created and annotated fashion datasets with high realism standards. • Eliminated anatomical anomalies through iterative prompt optimization. • Automated bounding box generation for object detection enhancement. • Leveraged workflow automation tools for dataset throughput scaling.

2025 - Present

AI Model Evaluation Specialist – Outlier AI

Micro1TextRLHF
Structured RLHF evaluation was performed on AI model outputs across text, audio, image, and reasoning tasks. Responses were compared and ranked using rubric-based criteria within human-in-the-loop evaluation workflows. Evaluated outputs for instruction following, factual accuracy, reasoning quality, safety compliance, and tone consistency. Executed structured reviews on math and logical reasoning chains. Conducted red teaming and adversarial prompt generation to surface model weaknesses and safety vulnerabilities.

Structured RLHF evaluation was performed on AI model outputs across text, audio, image, and reasoning tasks. Responses were compared and ranked using rubric-based criteria within human-in-the-loop evaluation workflows. Evaluated outputs for instruction following, factual accuracy, reasoning quality, safety compliance, and tone consistency. Executed structured reviews on math and logical reasoning chains. Conducted red teaming and adversarial prompt generation to surface model weaknesses and safety vulnerabilities.

2025 - Present

Education

U

University of the West of Scotland

Bachelor of Arts with Honours, International Business – Marketing Pathway

Bachelor of Arts with Honours
2022 - 2025

Work History

O

OroVun

Founder

London
2025 - Present
C

Cozyield

Founder & Operator

London
2024 - Present