erfanzar/EasyDeL

QLoRA Finetune Example

sr5434 opened this issue · 11 comments

Can you post an example of how to fine-tune a llm with 4-bit QLoRA on a TPU? Thanks for any help you can provide.

Yes soon ill add that but what kinda device you are using? Tpu/gpu/xpu?

I’m using TPU v3

I just change the Params dtype right?

you have to use bits=4 or 8 and pass lora config to Xrapture Module

Ok, thanks

There doesn’t appear to be an argument for bits

Nvm

Nvm

You have to change them inside model config

You can read docs for that or just see examples

Ok, thanks. Also, is there a way to save just the LoRA adapter in the HuggingFace format after training?

Im working on that and im gona build that as soon as possible but i have to finish SFTTrainer and make sure that DPOTrainer doesn't have any more bugs.

Ok, thanks.