/LLM-Adapters

Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"

Primary LanguagePythonApache License 2.0Apache-2.0

LLM-Adapters

LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models

LLM-Adapters is an easy-to-use framework that integrates various adapters into LLMs and can execute adapter-based PEFT methods of LLMs for different tasks. LLM-Adapter is an extension of HuggingFace's PEFT library, many thanks for their amazing work! Please find our paper at this link: https://arxiv.org/abs/2304.01933.

The framework includes state-of-the-art open-access LLMs: LLaMa, OPT, BLOOM, and GPT-J, as well as widely used adapters such as Bottleneck adapters, Parallel adapters, and LoRA.

Supported Adapters:

  1. LoRA: LORA: LOW-RANK ADAPTATION OF LARGE LANGUAGE MODELS
  2. AdapterH: Parameter-Efficient Transfer Learning for NLP
  3. AdapterP: GMAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer
  4. Parallel: TOWARDS A UNIFIED VIEW OF PARAMETER-EFFICIENT TRANSFER LEARNING
  5. Prefix Tuning: Prefix-Tuning: Optimizing Continuous Prompts for Generation, P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
  6. P-Tuning: GPT Understands, Too
  7. Prompt Tuning: The Power of Scale for Parameter-Efficient Prompt Tuning

Latest News 🔥🔥

  • [2023-08-10] LLM-Adapters has been accepted by EMNLP 2023.
  • [2023-07-16] we released commonsense170k dataset and the The LLaMA-13B-Parallel model outformances ChatGPT on 8 commonsense benchmarks.
  • [2023-04-21] We released math10k dataset and the LLaMA-13B adapter checkpoints. The LLaMA-13B-Parallel model achieves 91% of GPT-3.5 performance!
  • [2023-04-10] We can support GPT-Neo and ChatGLM now!
  • [2023-04-04] Release code and dataset

Special Announcement

The math_10k.json data is collected with the training sets of GSM8K, MAWPS, and AQuA(1000 examples). However, MAWPS consists of AddSub, MultiArith, SingleOp, SingleEq, SimulEq-S, SimulEq-L. Thus, we can't utilize MultiArith, AddSub, and SingleEq as evaluation benchmarks with models trained with math_10k.json. We evaluate the PEFT methods on the MAWPS test set instead, and the result table has been updated (The findings in the paper are consistent). Furthermore, two variations of math_10k.json have been uploaded, math_7K.json where the MAWPS samples have been deleted, and math_14k.json where the MAWPS samples have been deleted as well and we combine ChatGPT and GPT-4 rationales. Sincerely apologize for any inconvenience!

Setup

  1. Install dependencies
pip install -r requirements.txt
  1. Set environment variables, or modify the files referencing BASE_MODEL:
# Files referencing `BASE_MODEL`
# export_hf_checkpoint.py
# export_state_dict_checkpoint.py

export BASE_MODEL=yahma/llama-7b-hf

Both finetune.py and generate.py use --base_model flag as shown further below.

  1. If bitsandbytes doesn't work, install it from source. Windows users can follow these instructions.

Training(finetune.py)

This file contains some code related to prompt construction and tokenization.In this file, specify different adapters and different sets of data, so that different models can be trained.

Example usage for multiple GPUs:

WORLD_SIZE=2 CUDA_VISIBLE_DEVICES=0,1 torchrun --nproc_per_node=2 --master_port=3192 finetune.py \
  --base_model 'yahma/llama-7b-hf' \
  --data_path 'math_10k.json' \
  --output_dir './trained_models/llama-lora' \
  --batch_size 16 \
  --micro_batch_size 4 \
  --num_epochs 3 \
  --learning_rate 3e-4 \
  --cutoff_len 256 \
  --val_set_size 120 \
  --adapter_name lora

The math_10k.json data is collected with the training sets of GSM8K, MAWPS, and AQuA(1000 examples). yahma/llama-7b-hf is a base model, LLaMa-7B. Add lora adapter to this model.

Example usage for Single GPUs:

CUDA_VISIBLE_DEVICES=0 python finetune.py \
  --base_model 'yahma/llama-7b-hf' \
  --data_path 'math_10k.json' \
  --output_dir './trained_models/llama-lora' \
  --batch_size 16 \
  --micro_batch_size 4 \
  --num_epochs 3 \
  --learning_rate 3e-4 \
  --cutoff_len 256 \
  --val_set_size 120 \
  --adapter_name lora

Moreover, you can use --use_gradient_checkpointing to save more GPU memory, but it will increase the training time.

To use the AdapterH, just add the following arguments:

--adapter_name bottleneck # use the bottleneck adapter, refers to AdapterH in the result table

To use the AdapterP, just add the following arguments:

--adapter_name bottleneck 
--use_adapterp  # use the AdapterP, refers to AdapterP in the result table

To use parallel adapter, just add the following arguments:

--adapter_name bottleneck
--use_parallel_adapter

Note that, In order to facilitate INT8 training of large models with parallel adapters, we have adopted a technique whereby the parallel adapter layers are incorporated into multi-head attention layers and MLP layers, in parallel with Linear layers. It is different from Hu et al. (2021).

Inference (generate.py)

This file reads the foundation model from the Hugging Face model hub and the LoRA weights from './trained_models/llama-lora' , and runs a Gradio interface for inference on a specified input. Users should treat this as example code for the use of the model, and modify it as needed. Example usage:

CUDA_VISIBLE_DEVICES=0 torchrun generate.py \
    --base_model 'yahma/llama-7b-hf' \
    --lora_weights './trained_models/llama-lora'

Evaluation (evaluate.py)

To evaluate the performance of the finetuned model on the Arithmetic Reasoning tasks, you can use the following command:

CUDA_VISIBLE_DEVICES=0 python evaluate.py 
    --model LLaMA-7B \ #specify the base model
    --adapter LoRA \   #specify the adapter name ["LoRA", "AdapterH", "AdapterP", "Parallel", "Scaled_Parallel""]
    --dataset SVAMP \  #specify the test dataset
    --base_model 'yahma/llama-7b-hf' \
    --lora_weights './trained_models/llama-lora'

Finetune Result

There are the finetune results in different models with 4 math reasoning datasets, which contains GSM8K, AQuA, SVAMP, and MAWPS. In this table, we use the optimal configuration and placement of Prefix-Tuning, Series Adapter, LoRA, and Parallel Adapter according to the empirical study in our paper.

Model GSM8K AQuA MAWPS SVAMP Average
GPT-3.5 56.4 38.9 87.4 69.9 63.2
BLOOMz-7B-Prefix 13.8 12.5 47.5 24.1 24.5
BLOOMz-7B-Series 14.3 20.5 62.2 38.1 33.8
BLOOMz-7B-Parallel 18.5 18.9 70.6 36.4 36.1
BLOOMz-7B-LoRA 17.4 21.3 70.2 41.0 37.5
GPT-j-6B-Prefix 16.0 14.7 59.2 31.0 30.2
GPT-j-6B-Series 19.5 15.0 80.3 43.6 39.6
GPT-j-6B-Parallel 18.9 17.9 78.2 41.1 39.0
GPT-j-6B-LoRA 23.0 16.1 79.4 46.0 41.1
LLaMA-7B-Prefix 24.4 14.2 63.4 38.1 35.0
LLaMA-7B-Series 33.3 15.0 77.7 52.3 44.6
LLaMA-7B-Parallel 35.3 18.1 82.4 49.6 46.4
LLaMA-7B-LoRA 37.5 18.9 79.0 52.1 46.9
LLaMA-13B-Prefix 31.1 15.7 66.8 41.4 38.8
LLaMA-13B-Series 44.0 22.0 78.6 50.8 48.9
LLaMA-13B-Parallel 43.3 20.5 81.1 55.7 50.2
LLaMA-13B-LoRA 47.5 18.5 83.6 54.6 51.1

There are the finetune results in different models with eight commonsense reasoning datasets.

Model BoolQ PIQA SIQA HellaSwag WinoGrande ARC-e ARC-c OBQA Average
ChatGPT 73.1 85.4 68.5 78.5 66.1 89.8 79.9 74.8 77.0
BLOOMz-7B-Prefix 45.6 53.7 46.3 26.7 49.5 52.1 39.7 44.3 44.7
BLOOMz-7B-Series 65.4 70.4 73.6 53.4 69.3 72.3 55.9 68.0 66.0
BLOOMz-7B-Parallel 64.1 71.5 72.1 52.9 67.0 70.5 54.7 69.6 65.3
BLOOMz-7B-LoRA 65.9 75.3 74.5 57.3 72.5 74.6 57.8 73.4 68.9
GPT-j-6B-Prefix 63.1 66.9 68.7 34.4 64.5 64.4 46.8 59.0 58.5
GPT-j-6B-Series 62.1 63.5 72.3 30.6 68.0 63.9 48.1 63.8 59.0
GPT-j-6B-Parallel 62.2 69.7 70.0 41.7 65.0 60.2 44.6 58.2 59.0
GPT-j-6B-LoRA 62.4 68.6 49.5 43.1 57.3 43.4 31.0 46.6 50.2
LLaMA-7B-Prefix 64.3 76.8 73.9 42.1 72.1 72.9 54.0 60.6 64.6
LLaMA-7B-Series 63.0 79.2 76.3 67.9 75.7 74.5 57.1 72.4 70.8
LLaMA-7B-Parallel 67.9 76.4 78.8 69.8 78.9 73.7 57.3 75.2 72.3
LLaMA-7B-LoRA 68.9 80.7 77.4 78.1 78.8 77.8 61.3 74.8 74.7
LLaMA-13B-Prefix 65.3 75.4 72.1 55.2 68.6 79.5 62.9 68.0 68.4
LLaMA-13B-Series 71.8 83.0 79.2 88.1 82.4 82.5 67.3 81.8 79.5
LLaMA-13B-Parallel 72.5 84.8 79.8 92.1 84.7 84.2 71.2 82.4 81.5
LLaMA-13B-LoRA 72.1 83.5 80.5 90.5 83.7 82.8 68.3 82.4 80.5

Adapter support matrix

This metrix shows whether different models can use LoRA,AdapterH,AdapterP,Parallel and Scaled Parallel adapters.

Adapter LoRA AdapterH AdapterP Parallel Prefix Tuning P-Tuning Prompt Tuning
LLaMA
BLOOM
GPT-J
OPT
GPT-2 🔧Developing 🔧Developing 🔧Developing
GPT-Neo
GPT-NeoX-20B 🔧Developing 🔧Developing 🔧Developing
ChatGLM

TODO List

  • Add AdapterH
  • Add AdapterP
  • Add Parallel Adapter
  • Support More LLMs
  • Support Multiple Adapter
  • Support Adapter Composition
  • Support Adapter Fusion

⭐ Star History

Star History Chart

Citing LLM-Adapter

If you use LLM-Adapters in your publication, please cite it by using the following BibTeX entry.

@article{hu2023llm,
  title={LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models},
  author={Hu, Zhiqiang and Lan, Yihuai and Wang, Lei and Xu, Wanyu and Lim, Ee-Peng and Lee, Roy Ka-Wei and Bing, Lidong and Poria, Soujanya},
  journal={arXiv preprint arXiv:2304.01933},
  year={2023}
}

Acknowledgement

This repo benefits from PEFT, Adapter-Transformer, Alpaca-lora. Thanks for their wonderful works. Additionally, we thank DONG Shan and dream.ai for the exceptional logo design, which has added immense value to our project.