Prompt Engineer and LLM Trainer
I was responsible for training and fine-tuning large language models for text, audio, and application-based outputs. Tasks included applying reinforcement learning from human feedback (RLHF) and supervised fine-tuning (SFT) to optimize model alignment with business goals and accuracy. I designed prompts, evaluated model responses, conducted sentiment analysis and classification, and validated model output for safety, creativity, and contextual accuracy. • Applied RLHF and SFT techniques to improve AI model performance and reliability. • Designed and optimized prompts for diverse subject matter, ensuring output safety and relevance. • Evaluated and rated AI responses for accuracy, creativity, and alignment to business needs. • Conducted classification and sentiment analysis as part of the AI model training process.