human-feedback
There are 15 repositories under human-feedback topic.
lucidrains/PaLM-rlhf-pytorch
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
opendilab/awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
conceptofmind/LaMDA-rlhf-pytorch
Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.
wxjiao/ParroT
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
xrsrke/instructGOOSE
Implementation of Reinforcement Learning from Human Feedback (RLHF)
huggingface/data-is-better-together
Let's build better datasets, together!
trubrics/trubrics-sdk
Product analytics for AI Assistants
yk7333/d3po
[CVPR 2024] Code for the paper "Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model"
PKU-Alignment/beavertails
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
HannahKirk/prism-alignment
The Prism Alignment Project
gao-g/prelude
Aligning LLM Agents by Learning Latent Preference from User Edits
AlaaLab/pathologist-in-the-loop
[ NeurIPS 2023 ] Official Codebase for "Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback"
victor-iyi/rlhf-trl
Reinforcement Learning from Human Feedback with 🤗 TRL
ZiyiZhang27/tdpo
[ICML 2024] Code for the paper "Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases"
01Kevin01/awesome-RLHF-Turkish
A curated list of reinforcement learning with human feedback resources[awesome-RLHF-Turkish] (continually updated)