Research and development on learning from human feedback is difficult because methods like RLHF are complex and costly to run. AlpacaFarm is a simulator that enables research and development on learning from feedback at a fraction of the usual cost, promoting accessible research on instruction following and alignment.
Please read our paper and blog post for details on our research findings.
This repo contains code for
- simulating preference feedback from language models such as GPT-4
- automated evaluation for instruction-following models
- validated reference implementations of baseline methods such as PPO and best-of-n
The data needed to run our code is hosted on HuggingFace: https://huggingface.co/datasets/tatsu-lab/alpaca_farm.
Usage and License Notices: AlpacaFarm is intended and licensed for research use only. The dataset is CC BY NC 4.0 (allowing only non-commercial use) and models trained using the dataset should not be used outside of research purposes. The weight diff is also CC BY NC 4.0 (allowing only non-commercial use).
Instruction-following models are typically developed in 3 steps
- Supervised fine-tuning with demonstrations
- Learning from human feedback; usually pairwise preferences
- Human evaluation with interaction
The goal of AlpacaFarm is to provide three key components that tackles steps 2 and 3: Low-cost simulation of pairwise feedback from API models (e.g. GPT-4, ChatGPT), automated evaluations for methods development, and reference implementations of learning algorithms for comparison and modification.
To install the stable release, run
pip install alpaca-farm
To install from the latest commit on main
branch, run
pip install git+https://github.com/tatsu-lab/alpaca_farm.git
To enable FlashAttention and other optimizations, install
the flash-attn
and apex
packages.
For all the evaluation and annotations we use AlpacaEval with our pool of automatic annotators and additional noise to simulate the variance of human annotations.
To get started, set the environment variable OPENAI_API_KEY
to your OpenAI API key, and (optionally) OPENAI_ORG
to
the
organization ID.
You can do this by running
export OPENAI_API_KEY="sk..."
To annotate the pairs of outputs of your model use the following code. For more details or functions to use if you have outputs in different formats refer to the example notebook.
from alpaca_farm.auto_annotations import PairwiseAutoAnnotator
import json
# load some data
with open("examples/data/outputs_pairs.json") as f:
outputs_pairs = json.load(f)[:6]
print(outputs_pairs[-1:])
# [{'instruction': 'If you could help me write an email to my friends inviting them to dinner on Friday, it would be greatly appreciated.',
# 'input': '',
# 'output_1': "Dear Friends, \r\n\r\nI hope this message finds you well. I'm excited to invite you to dinner on Friday. We'll meet at 7:00 PM at [location]. I look forward to seeing you there. \r\n\r\nBest,\r\n[Name]",
# 'output_2': "Hey everyone! \n\nI'm hosting a dinner party this Friday night and I'd love for all of you to come over. We'll have a delicious spread of food and some great conversations. \n\nLet me know if you can make it - I'd love to see you all there!\n\nCheers,\n[Your Name]"}]
annotator = PairwiseAutoAnnotator()
annotated = annotator.annotate_pairs(outputs_pairs)
print(annotated[-1:])
# [{'instruction': 'If you could help me write an email to my friends inviting them to dinner on Friday, it would be greatly appreciated.',
# 'input': '',
# 'output_1': "Dear Friends, \r\n\r\nI hope this message finds you well. I'm excited to invite you to dinner on Friday. We'll meet at 7:00 PM at [location]. I look forward to seeing you there. \r\n\r\nBest,\r\n[Name]",
# 'output_2': "Hey everyone! \n\nI'm hosting a dinner party this Friday night and I'd love for all of you to come over. We'll have a delicious spread of food and some great conversations. \n\nLet me know if you can make it - I'd love to see you all there!\n\nCheers,\n[Your Name]",
# 'annotator': 'chatgpt_2',
# 'preference': 2}]
If instead of pairs you have a list of sampled outputs, you can use the following.
multisample_outputs = [dict(instruction="repeat the following", input="yes", output=["yes", "no", "maybe", "repeat"])]
print(annotator.annotate_samples(multisample_outputs))
# [{'sample_id': 0,
# 'instruction': 'repeat the following',
# 'input': 'yes',
# 'output_1': 'yes',
# 'output_2': 'maybe',
# 'annotator': 'chatgpt_2',
# 'preference': 1}]
For all the evaluation we use AlpacaEval with our pool of automatic annotators.
To get started, set the environment variable OPENAI_API_KEY to your OpenAI API key, and (optionally) OPENAI_ORG to the organization ID. You can do this by running
export OPENAI_API_KEY="sk..."
The easiest to add your model to the Alpaca Leaderboard is to run the following code, which only requires having outputs for your model on our eval data.
from alpaca_farm.auto_annotations import alpaca_leaderboard
import datasets
# predict on Alpaca eval data
alpaca_eval_data = datasets.load_dataset("tatsu-lab/alpaca_farm", "alpaca_farm_evaluation")["eval"]
... # use the data to get outputs for your model and save it
path_to_outputs = "examples/data/eval_gpt-3.5-turbo-0301.json"
# outputs should be a list of json as such:
# [{'instruction': 'What are the names of some famous actors that started their careers on Broadway?', 'input': '', 'output': 'Some famous actors that started their careers on Broadway are Hugh Jackman, Meryl Streep, Denzel Washington, Audra McDonald, and Lin-Manuel Miranda.', 'generator': 'gpt-3.5-turbo-0301', 'dataset': 'helpful_base', 'datasplit': 'eval'},
# ...]
alpaca_leaderboard(path_or_all_outputs=path_to_outputs, name="My fancy model", is_print_metrics=True)
# n_draws n_total n_wins n_wins_base standard_error win_rate
# GPT4 17.00 804.00 631.00 156.00 1.40 79.54
# ChatGPT 9.00 805.00 503.00 293.00 1.69 63.04
# My fancy model 9.00 803.00 497.00 297.00 1.70 62.45
# RLHF PPO 9.00 805.00 392.00 404.00 1.75 49.25
# SFT 52k (Alpaca 7B) 16.00 805.00 312.00 477.00 1.71 39.75
# SFT 10k 19.00 802.00 278.00 505.00 1.67 35.85
# Davinci001 0.00 805.00 201.00 604.00 1.53 24.97
# LLaMA 7B 0.00 775.00 98.00 677.00 1.19 12.65
If you want to compare against our baseline model (Davinci003 with our prompt) on your own data, you can decode it using main_oai_baselines.
We provide reference implementations of several methods for learning from pairwise feedback.
Example code to run these methods can be found in the examples/
directory.
This includes supervised fine-tuning, reward modeding
, RLHF with PPO, best-of-n decoding and more.
Below we give example commands for reproducing the model artifacts in our paper. Notes:
- All training code are tested with FlashAttention enabled on a machine with 8 80GB A100 GPUs.
- Best-of-n decoding was tested with a single 80GB GPU.
- Supervised fine-tuning and reward modeling can fit on 4 80GB A100 GPUs, while PPO training currently requires at least 8 80GB GPUs.
- Before running the code below, make sure to convert your LLaMA checkpoint and tokenizer into HuggingFace format and
store it at
<your_path_to_hf_converted_llama_ckpt_and_tokenizer>
.
To replicate our SFT10k model fine-tuned from LLaMA in the paper, run
bash examples/scripts/sft.sh \
<your_output_dir_for_sft10k> \
<your_wandb_run_name> \
<your_path_to_hf_converted_llama_ckpt_and_tokenizer>
The SFT10k model will be saved at <your_output_dir>
, and the name of the wandb run will be <your_wandb_run_name>
.
To replicate our reward models trained in the paper, run
bash examples/scripts/reward_modeling.sh \
<your_output_dir_for_reward_model> \
<your_wandb_run_name> \
<your_output_dir_for_sft10k> \
<preference_dataset_name>
Set <preference_dataset_name>
to "alpaca_noisy_multi_preference"
for simulated preference reward model, and
"alpaca_human_preference"
for human preference reward model.
To replicate our RLHF PPO model trained with simulated reward model in the paper, run
bash examples/scripts/rlhf_ppo.sh \
<your_output_dir_for_ppo> \
<your_wandb_run_name> \
<your_output_dir_for_reward_model> \
<your_output_dir_for_sft10k> \
<kl_coef>
<your_output_dir_for_reward_model>
should point to either simulated reward model or human reward model trained
according
to the previous step.
Note the KL penalty coefficient for human reward PPO is much larger than for simulated PPO.
Set <kl_coef>
to 0.0067
for simulated PPO, and 0.02
for human PPO to recover our original results.
Performance of the PPO model is typically much better than SFT at 20-80 PPO steps (less than 4 passes through the entire
set of instructions) and starts to decay with more PPO steps.
To replicate our best-of-n inference-time decoding results for the AlpacaFarm evaluation suite, run
python examples/best_of_n.py \
--task "run_best_of_n" \
--decoder_name_or_path <your_output_dir_for_decoder> \ # Can be SFT model or even PPO tuned model.
--scorer_name_or_path <your_output_dir_for_reward_model> \
--num_return_sequences 16 \ # This is the n in best-of-n.
--per_device_batch_size 4 \ # Reduce this if you don't have enough memory.
--split "eval" \
--mixed_precision "bf16" \
--tf32 True \
--flash_attn True \
--output_path <your_output_path_to_store_samples>
You can then use the generated samples at <your_output_path_to_store_samples>
directly with our automated evaluation.
To replicate our expert iteration results for the AlpacaFarm evaluation suite, first produce best-of-n samples. Run
python examples/best_of_n.py \
--task "run_best_of_n" \
--decoder_name_or_path <your_output_dir_for_decoder> \ # SFT10k model.
--scorer_name_or_path <your_output_dir_for_reward_model> \
--num_return_sequences 16 \ # This is the n in best-of-n.
--per_device_batch_size 4 \ # Reduce this if you don't have enough memory.
--split "unlabeled" \
--mixed_precision "bf16" \
--tf32 True \
--flash_attn True \
--output_path '<your_output_dir_for_expiter_data>/best_of_n_samples.json'
Then perform supervised fine-tuning from the SFT10k checkpoint with the best-of-n samples
bash examples/scripts/expiter.sh \
<your_output_dir_for_expiter> \
<your_wandb_run_name> \
<your_output_dir_for_sft10k> \
<your_output_dir_for_expiter_data>
To replicate our Quark results for the AlpacaFarm evaluation suite, run
bash examples/scripts/rlhf_quark.sh \
<your_output_dir_for_quark> \
<your_wandb_run_name> \
<your_output_dir_for_reward_model> \
<your_output_dir_for_sft10k> \
<kl_coef>
To run the OpenAI reference models with our prompts and decoding hyperparameters, run
python examples/oai_baselines.py \
--model_name <oai_model_name> \
--save_path <save_path>
You can then use the generated samples at <save_path>
directly with our automated evaluation.
We provide model checkpoints for reward models and all our reference methods, listed in Table 2 of our paper. Concretely, we tune each reference method in AlpacaFarm simulation and on human preference data and release both versions. The current list of models (available here) includes:
sft10k
, the supervised learning base model that we collect preference data with.reward-model-sim
, the reward model trained on AlpacaFarm preference data.reward-model-human
, the reward model trained on human preference data.ppo-sim
, the best PPO checkpoint trained in simulation.ppo-human
, the best PPO checkpoint trained on human data.expiter-sim
, the best expert iteration checkpoint trained in simulation.expiter-human
, the best expert iteration checkpoint trained on human data.feedme-sim
, the FeedME method trained on simulated preferences.feedme-human
, the FeedME method trained on human preferences.reward-condition-sim
, the reward conditioning method trained on simulated preferences.
To download and recover these checkpoints, first make sure to have a LLaMA-7B checkpoint converted into the Hugging Face format with transformers>=4.29.2. Then, run the following to download all AlpacaFarm models:
python -m pretrained_models.recover_model_weights \
--llama-7b-hf-dir <your_path_to_hf_converted_llama_ckpt_and_tokenizer> \
--alpaca-farm-model-name all
Or, specify a particular model name to download just that model:
python -m pretrained_models.recover_model_weights \
--llama-7b-hf-dir <your_path_to_hf_converted_llama_ckpt_and_tokenizer> \
--alpaca-farm-model-name <one_of_the_model_names_from_above> \
--models-save-dir <dir_to_save_all_models>
To download either of the reward models individually, you'll need to have sft10k
downloaded first
to <dir_to_save_all_models>
.
Please consider citing our work if you use the data or code in this repo.
@misc{dubois2023alpacafarm,
title={AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback},
author={Yann Dubois and Xuechen Li and Rohan Taori and Tianyi Zhang and Ishaan Gulrajani and Jimmy Ba and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto},
year={2023},
eprint={2305.14387},
archivePrefix={arXiv},
primaryClass={cs.LG}
}