Prompt Engineer, Soul AI
Applied Reinforcement Learning from Human Feedback (RLHF) to train AI models for improved alignment and accuracy. Fine-tuned an AI model across multiple parameters, enhancing overall response accuracy. Utilized advanced prompt engineering strategies to optimize AI-generated outputs. • Enhanced model alignment using RLHF methodologies. • Fine-tuned models leading to a significant increase in prediction accuracy. • Designed and iterated on prompts to improve AI reliability. • Contributed directly to AI training pipelines for better user outcomes.