Remote Freelance AI Data Annotator & Model Evaluation Specialist
As a Remote Freelance AI Data Annotator & Model Evaluation Specialist at Mindrift, I specialize in red teaming, data labeling, and benchmarking of Large Language Models. I design and label datasets to enhance model reasoning and safety in logic-oriented and creative contexts. My work includes detailed adversarial testing and the creation of high-quality instruction datasets for model evaluation and improvement. • Conduct adversarial tests to identify LLM vulnerabilities and bias issues. • Curate and annotate diverse instruction datasets for enhanced reasoning. • Execute side-by-side model comparison for 3H benchmarks (helpfulness, honesty, harmlessness). • Enhance data quality and diversity for model fine-tuning and evaluation.