LLM Alignment & Response Quality Annotation – Alignerr
Contributed to alignment and evaluation tasks aimed at improving large language model safety, reasoning, and instruction-following capabilities. Responsibilities included evaluating model-generated responses for accuracy, coherence, and policy compliance. Performed pairwise ranking of responses, corrected flawed outputs, and generated improved prompt-response pairs used for supervised fine-tuning. Participated in adversarial testing and red-teaming tasks to identify potential model weaknesses and safety issues. Additional Information Worked on complex prompts requiring analytical reasoning, structured explanations, and domain-specific knowledge in technical and analytical subjects.