Coding evaluation
Create complex computer science prompts and evaluate the LLM response with an elaborate rubric system.
Hire this AI Trainer
Sign in or create an account to invite AI Trainers to your job.
No subject matter listed
Experience in AI model evaluation and data annotation, specializing in stress-testing large language models (LLMs) for safety and technical reasoning. My work has involved red teaming, authoring vulnerability reports, and executing high-precision data labeling across advanced NLP projects. I have a strong background in optimizing model performance using RLHF, designing complex multi-turn prompts, and ensuring rigorous quality assurance for training datasets. I am a native French speaker with advanced English knowledge. I also have advanced knowledge of coding (C#) and Unity (game development).
Create complex computer science prompts and evaluate the LLM response with an elaborate rubric system.
Evaluate LLM responses by crafting prompts in specific categories. Write detailed reports on the responses. Create golden response when both models fail.
Evaluate video edits for social media appeal.
Bachelor of Science, Business Administration
Game Developer