/dpo

Implementation of Direct Preference Optimization

Primary LanguageJupyter Notebook

Stargazers

No one’s star this repository yet.