virajmehta/direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
Jupyter NotebookApache-2.0
Stargazers
No one’s star this repository yet.
Reference implementation for DPO (Direct Preference Optimization)
Jupyter NotebookApache-2.0
No one’s star this repository yet.