LLM Evaluation and Text Annotation
Contributed to Project Nightingale, a large-scale language model evaluation and annotation initiative. Tasks included ranking AI-generated responses, identifying factual accuracy, improving fluency and coherence, and labeling text for sentiment and named entities. The project also involved prompt and response writing to support supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF). Ensured consistency and quality by adhering to platform-specific guidelines, achieving accuracy scores above 95% across multiple review cycles.