Prompt Engineer & LLM Response Rater
I executed few-shot prompting strategies and supervised prompt and response writing tasks to fine-tune LLM classifiers. My responsibilities included curating example datasets, writing prompts, generating and rating LLM outputs, and maintaining prompt/response quality for sub-300ms API endpoints. This experience covered prompt engineering, output review, and structured fast feedback to enhance classifier accuracy. • Designed and tested prompts for LLM classifiers using few-shot learning techniques. • Rated LLM-generated outputs for relevance, accuracy, and coherence. • Maintained quality standards for prompt/response datasets. • Supported rapid inference and continuous LLM evaluation with custom metrics.