Zero Shot Transformer based 167B MoE LLM
Designed and executed data annotation workflows for fine-tuning a 167B MoE language model via RLHF. Created labeling guidelines for preference ranking, question-answering validation, and text generation quality assessment using AWS SageMaker. Annotated 50K+ prompt-response pairs with >0.75 inter-annotator agreement, improving model alignment and reducing hallucinations by 23%.