Outlier Ai
1. Rubric design, System Prompt development, and agent/tool-call evaluations & Quality Assurance Conducted rigorous evaluations of AI-generated responses for RLHF tasks, ensuring adherence to strict guidelines. Provided detailed feedback on accuracy, instruction compliance, and reasoning errors, enhancing the reliability and performance of AI models. 2. Prompt Engineering & Adversarial Testing Designed and optimized complex logic-based prompts to assess AI reasoning capabilities, leveraging the STIRR methodology. Performed adversarial testing to identify and rectify reasoning flaws, improving model alignment with factual correctness and user intent. 3. Code Review & Debugging for AI Systems Reviewed and debugged Python and React codebases, identifying logic errors and optimizing workflows in AI-powered applications. Ensured backend API integrity and enhanced code efficiency, contributing to high-quality, production-ready AI solutions.