AI Engineer – Data Annotation & LLM Evaluation
In this role, I designed structured pipelines to transform raw model outputs into annotatable formats for evaluation and verification. I built and maintained validation systems to score and compare large language model (LLM) outputs, ensuring accuracy and consistency. I developed annotation guidelines and quality benchmarks to standardize labeling processes across teams. • Defined annotation rules and structured JSON schemas for labeling workflows. • Implemented validation layers to detect hallucinations and inconsistency in model outputs. • Created iterative feedback loops to improve dataset quality and model accuracy. • Collaborated cross-functionally to refine annotation standards for quality control.