AI Research Scientist, Data Annotation and Model Evaluation
Developed and tested NLP models as part of the AI research team, focusing on data annotation and evaluation. Participated in benchmarking and quality assurance for large language model (LLM) outputs. Responsibilities included designing annotation workflows, providing detailed feedback, and curating datasets for research advancements. • Designed and annotated text datasets for classification and summarization tasks. • Evaluated and rated model outputs to improve model alignment and accuracy. • Collaborated on quality assurance processes and bias mitigation efforts. • Utilized Python, Hugging Face, and proprietary internal tools for annotation and evaluation.