Federated Learning Model Output Evaluation & Preference Ranking
Evaluated and ranked model outputs across privacy-utility tradeoff scenarios in a Vertical Federated Learning research project. Tasks included assessing response quality, accuracy, and behavioral alignment with expected outcomes — directly mirroring RLHF-style preference ranking workflows. Maintained detailed evaluation logs and applied consistent scoring criteria across experimental runs. Manuscript currently under peer review.