AI text annotations and LLM Evaluation
Worked on AI data labeling and evaluation projects focused on text annotation and large language model output review. Tasks included classifying responses, comparing multiple AI outputs, identifying factual errors or misleading content, and ensuring strict adherence to detailed guidelines. The project required high attention to detail, consistency across large volumes of data, and clear reasoning when evaluating AI-generated content, particularly in finance and stock market–related contexts