Freelancer Overview
I have hands-on experience in AI training data creation, labeling, and human evaluation across multiple real-world projects involving speech, audio, and large language models. At Alignerr, I worked on Project Vera – VAD Audio Annotation v2, producing millisecond-level (1 ms) word-aligned transcriptions, labeling speech types (standard speech, acknowledgments, interruptions), and annotating background noise events. I also contributed to Project Human Evals v2.3, where I conducted live, multi-turn audio conversations with AI models in noisy environments to evaluate speech recognition, responsiveness, task completion, context awareness, and emotional empathy in both customer support and companionship scenarios. In addition, through Cypher RLHF at Outlier, I supported LLM improvement via human feedback by writing prompts and performing pairwise evaluations of AI-generated responses. I assessed outputs for truthfulness, clarity, grammatical quality, text structure, relevance, and alignment with user intent, providing consistent, high-quality annotations used for model fine-tuning. Together, these projects demonstrate strong skills in attention to detail, critical evaluation, conversational AI understanding, RLHF workflows, and human-in-the-loop training, with a proven ability to generate high-quality data that improves AI accuracy, reliability, and real-world performance.