preference-alignment
There are 6 repositories under preference-alignment topic.
princeton-nlp/SimPO
[NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward
zjukg/KnowPAT
[Paper][ACL 2024 Findings] Knowledgeable Preference Alignment for LLMs in Domain-specific Question Answering
junkangwu/beta-DPO
[NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$
Shentao-YANG/Dense_Reward_T2I
Source code for "A Dense Reward View on Aligning Text-to-Image Diffusion with Preference" (ICML'24).
junkangwu/Dr_DPO
Towards Robust Alignment of Language Models: Distributionally Robustifying Direct Preference Optimization
BARUDA-AI/Awesome-Preference-Optimization
Survey of preference alignment algorithms