I fine-tuned a 7B LLM on a public instruction dataset using LoRA for 0.5 epochs over 5 hours on a single-GPU g4dn.xlarge
EC2 instance.
Due to the size of the LLM, I had to increase the volume size of the EC2 instance:
(pytorch) du -hs .cache/huggingface/hub/models--Salesforce--xgen-7b-8k-base
26G .cache/huggingface/hub/models--Salesforce--xgen-7b-8k-base
The training code uses the torch, transformers, peft and trl libraries. See requirements.txt
.
The training run is tracked here and the fine-tuned model adaptors are available here.