LLM Prompt Evaluation and Toxicity Annotation (Meta AI)
Contracted through a global vendor to support Meta’s LLaMA model training pipeline. Evaluated model responses to user prompts and labeled completions for safety, toxicity, factual consistency, and helpfulness. Used structured rubrics to rank outputs and flag problematic completions. Played a key role in refining model alignment with human values and ethical AI guidelines.