AI Prompt Engineer & LLM Tester
Tested and evaluated proprietary and open-source Large Language Models for logical limits, biases, hallucinations, and consistency. Designed complex system prompts and structured tasks for evaluating AI reasoning and output quality. Used advanced LLM techniques and multi-model comparisons to ensure accuracy and reliability in AI-generated analysis of trading and technical data. • Performed daily systematic assessment of AI-generated texts and algorithms. • Annotated errors and categorized LLM outputs for statistical and manual review. • Monitored, flagged, and described indicators of AI hallucination and bias. • Applied prompt engineering and hands-on annotation techniques to raw financial and technical data.