/DPRM

Our paper "Aligning Crowd Feedback via Distributional Preference Reward modelling"

@inproceedings{
li2024aligning,
title={Aligning Crowd Feedback via Distributional Preference Reward Modeling},
author={Dexun Li and Cong Zhang and Kuicai Dong and Derrick Goh Xin Deik and Ruiming Tang and Yong Liu},
booktitle={ICML 2024 Workshop on Models of Human Feedback for AI Alignment},
year={2024},
url={https://openreview.net/forum?id=HHtV1kshHP}
}