LLM Prompt/Response Evaluation and Rating
Rate and evaluate LLM prompt/response pairs using structured 6-category scoring frameworks on both Handshake AI and Innodata, adhering to strict quality guidelines.
Hire this AI Trainer
Sign in or create an account to invite AI Trainers to your job.
Freelance Data Annotator & AI Generalist. Brings 31+ years of professional experience across financial operations, contract review, compliance, and structured data analysis. Core strengths include Labelbox, Label Studio, and Mercor RL Studio. and UHRS Education includes Bachelor of Arts, Pennsylvania State University. AI-training focus includes data types such as Image, Text, and Video labeling workflows including RLHF and Entity Recognition.
Rate and evaluate LLM prompt/response pairs using structured 6-category scoring frameworks on both Handshake AI and Innodata, adhering to strict quality guidelines.
Engaged in Fine Tuning LLM responses of an Emotional Support Chatbot to rate responses and correct/rewrite undesirable/dangerous responses based on very strict guidelines.
Listening to short Data Rows to discern and transcribe communications between Air Traffic Controllers, Pilots, and Ground Personnel, using specific guidelines.
Researched searched and located social media videos, segmented selected videos based on guideline directions, then created Prompts for the LLM to reproduce the segmented video sample.
Labeled and classified images based on complex guidelines
Bachelor of Arts, Speech Communications and Business
AI Response Rater
T2 Generalist