LLM Output Evaluation & Prompting Practice (Independent Project)
I performed systematic evaluation of large language model (LLM) outputs, rating their accuracy, coherence, and correctness. I created structured prompts covering mathematical explanations, story writing, and logical reasoning tasks to assess model responses. I compared and identified the best outputs and documented my findings to support model improvement. • Wrote and refined prompts for mathematical and creative tasks. • Assessed LLM-generated text for factual accuracy and clarity. • Identified logical sequence errors and provided focused feedback. • Compared outputs and selected the most appropriate responses.