For employers

Hire this AI Trainer

Sign in or create an account to invite AI Trainers to your job.

Invite to Job
Anish Krishnan

Anish Krishnan

AI Evaluation and Code Assessment (Freelance Software Engineer)

India flagChandigarh, India
$35.00/hrExpertAws SagemakerAxiom AICloudfactory

Key Skills

Software

AWS SageMakerAWS SageMaker
Axiom AI
CloudFactoryCloudFactory
Data Annotation TechData Annotation Tech
LabelboxLabelbox
MercorMercor
Micro1
Scale AIScale AI
SuperAnnotateSuperAnnotate
TagtogTagtog
TolokaToloka
TelusTelus

Top Subject Matter

AI Model Evaluation (Software Engineering)

Top Data Types

TextText
Computer Code ProgrammingComputer Code Programming

Top Task Types

Red Teaming
Computer Programming Coding
Prompt Response Writing SFT
RLHF
Fine Tuning

Freelancer Overview

AI Evaluation and Code Assessment (Freelance Software Engineer). Brings 4+ years of professional experience across complex professional workflows, research, and quality-focused execution. Core strengths include Internal and Proprietary Tooling. Education includes Bachelor of Technology, Chitkara University (2022). AI-training focus includes data types such as Computer Code and Programming and labeling workflows including Evaluation and Rating.

ExpertEnglish

Labeling Experience

claude-hfi

Computer Code ProgrammingComputer Programming Coding
Evaluated 100+ AI-generated backend and full-stack solutions against real-world engineering standards, flagging correctness, safety, and scalability issues in Python and Node.js codebases. • Assessed front-end behavior including state assumptions, data consistency, and integration boundaries affecting user trust and system reliability. • Compare multiple implementations and selected solutions based on trade-offs in performance, scalability, and risk, improving average solution quality scores by 20%. • Designed and validated prompts simulating attack vectors, edge cases, and misuse scenarios, improving robustness of AI-driven systems. • Applied structured reasoning around system failures, ambiguous inputs, and adversarial behavior, aligned with Trust & Risk Engineering principles.

Evaluated 100+ AI-generated backend and full-stack solutions against real-world engineering standards, flagging correctness, safety, and scalability issues in Python and Node.js codebases. • Assessed front-end behavior including state assumptions, data consistency, and integration boundaries affecting user trust and system reliability. • Compare multiple implementations and selected solutions based on trade-offs in performance, scalability, and risk, improving average solution quality scores by 20%. • Designed and validated prompts simulating attack vectors, edge cases, and misuse scenarios, improving robustness of AI-driven systems. • Applied structured reasoning around system failures, ambiguous inputs, and adversarial behavior, aligned with Trust & Risk Engineering principles.

2025 - Present

AI Evaluation and Code Assessment (Freelance Software Engineer)

Evaluated AI-generated backend and full-stack code solutions for correctness, safety, and scalability in various programming languages. Compared and selected AI-suggested implementations based on real-world engineering standards and trade-offs. Designed prompts to simulate adversarial scenarios improving robustness of AI-driven systems.• Assessed the reliability of Python and Node.js code created by AI models.• Conducted structured reasoning on ambiguous and failure-prone system behaviors.• Advanced prompt engineering for evaluating edge cases and misuse.• Improved solution quality scores by 20% through iterative evaluation and rating processes.

Evaluated AI-generated backend and full-stack code solutions for correctness, safety, and scalability in various programming languages. Compared and selected AI-suggested implementations based on real-world engineering standards and trade-offs. Designed prompts to simulate adversarial scenarios improving robustness of AI-driven systems.• Assessed the reliability of Python and Node.js code created by AI models.• Conducted structured reasoning on ambiguous and failure-prone system behaviors.• Advanced prompt engineering for evaluating edge cases and misuse.• Improved solution quality scores by 20% through iterative evaluation and rating processes.

2025 - Present

Mercor

Computer Code ProgrammingRed Teaming
Red Teaming AI via Adversarial Probing, Vulnerability Assessment and Reporting

Red Teaming AI via Adversarial Probing, Vulnerability Assessment and Reporting

2025 - 2026

Education

C

Chitkara University

Bachelor of Technology, Computer Science

Bachelor of Technology
2018 - 2022

Work History

M

Macys

Full Stack Engineer

Chandigarh
2025 - 2025
D

Directv

Full Stack and DevOps Engineer

Chandigarh
2023 - 2025