LLM Response Evaluation/AI Trainer (Text)
I participated in LLM response evaluation by checking AI-generated textual responses for relevance, factual accuracy, and user intent alignment. This involved systematic rating and annotation of model outputs according to project guidelines. My efforts supported dataset improvement for conversational AI systems and generative LLM models. • Rated model responses for factual consistency and relevance • Evaluated user intent alignment in AI-generated text • Ensured adherence to quality standards within text datasets • Supported training and validation of conversational AI models