AI Tool User, Prompt Engineer, and RLHF Evaluator
I have daily hands-on experience using advanced AI tools such as Claude, Grok, and ChatGPT to generate, evaluate, and refine content for research and workflow automation. My work involves prompt engineering, iterative feedback, and quality assessment to optimize responses and train language models. I focus on clarity, accuracy, and audience-appropriate communication for effective AI output evaluation and tuning. • Conducted detailed prompt engineering and iterative testing with LLMs • Evaluated LLM outputs, providing structured annotation and feedback • Applied analytical judgment in data annotation to improve LLM performance • Developed and documented best practices for AI content generation