Multilingual Data Labeling and AI Training with Appen
I have worked on several projects focused on training and evaluating AI models for code generation and developer assistance. My responsibilities included writing high-quality prompts and responses for supervised fine-tuning (SFT), generating reference code solutions, and creating structured function-calling datasets. A major part of my role involved evaluating and ranking model outputs for correctness, efficiency, readability, and security, ensuring that LLMs could produce reliable code across multiple programming languages such as Python, C++, and JavaScript. These projects supported the development of AI coding assistants and copilots by refining their ability to complete functions, debug errors, and follow best practices. I consistently delivered high-quality annotations under strict accuracy and review standards, helping improve the performance and trustworthiness of production-level coding LLMs.