AI Model Trainer & LLM Evaluator
Served as an AI Model Trainer and LLM Evaluator focused on ranking, assessing, and improving AI-generated text outputs. Evaluated responses for accuracy, logical consistency, bias, and hallucination using RLHF workflows. Annotated, curated, and quality-checked NLP datasets for instruction tuning and benchmarking. • Ranked and compared LLM outputs for correctness.• Annotated and curated datasets for fine-tuning.• Provided structured feedback on LLM response quality.• Collaborated to maintain high annotation standards.