AI Annotator
My work includes Human-in-the-Loop Model Evaluation, LLM Response Evaluation, and Preference Judgments to compare, rate, and refine large language model outputs for usefulness, safety, and alignment with detailed project guidelines. I support Prompt Iteration and Prompt-Based Tuning by generating and improving prompts and responses (SFT), ensuring models learn from high‑quality, diverse examples. I perform Training Data Quality Assurance and Annotation QA, checking for Consistency, Guideline Adherence, and Reliable Ground-Truth Creation across labeling tasks such as Bounding Box annotation, Text Generation, Evaluation/Rating, Prompt + Response Writing, and Audio Recording. I also handle Edge-Case Handling and Model Output Review, performing Safety/Accuracy Checks and contributing to Continuous Feedback Loops that systematically improve both data pipelines and downstream AI model performance over time.