QLoRA Finetune Example
sr5434 opened this issue · 11 comments
Can you post an example of how to fine-tune a llm with 4-bit QLoRA on a TPU? Thanks for any help you can provide.
Yes soon ill add that but what kinda device you are using? Tpu/gpu/xpu?
I’m using TPU v3
I just change the Params dtype right?
you have to use bits=4 or 8 and pass lora config to Xrapture Module
Ok, thanks
There doesn’t appear to be an argument for bits
Nvm
Nvm
You have to change them inside model config
You can read docs for that or just see examples
Ok, thanks. Also, is there a way to save just the LoRA adapter in the HuggingFace format after training?
Im working on that and im gona build that as soon as possible but i have to finish SFTTrainer and make sure that DPOTrainer doesn't have any more bugs.
Ok, thanks.