LLM Response Evaluation & Prompt Quality Analysis for Safety-Aligned NLP Models
I worked on a large-scale LLM evaluation project focused on improving model accuracy and safety. My tasks included rating model responses for correctness, reasoning quality, and policy compliance, as well as writing improved prompts and responses for SFT. I also performed red-teaming, content safety classification, and ambiguity resolution. Over the project, I evaluated 10,000+ responses and consistently maintained 95%+ QA accuracy.