AI Training, Evaluation & Multilingual Localization Projects
Worked on AI training, evaluation and quality assurance projects across multiple task types, supporting the validation and improvement of large language model outputs. My work included evaluating responses for factual accuracy, reasoning quality, completeness, relevance and alignment with user intent, following detailed, guideline-based criteria. I participated in factual verification and freshness evaluation projects, checking whether AI-generated content was correct, up to date, and contextually appropriate, including the identification of outdated, incomplete or misleading information. I also contributed to multilingual content verification and localization, ensuring outputs were properly adapted to the user’s locale and use case. This included language-specific quality checks, cultural and regional adaptation, and validation of country-specific conventions such as units of measurement, currency and terminology, particularly for Spanish-Spain, Catalan and other regional contexts.