BAAI-DCAI/Bunny

Training the model throws an error after quantization

Opened this issue · 1 comments

When I use 8-bit quantization in the pre-training process, the code throws an error.

You cannot perform fine-tuning on purely quantized models. Please attach trainable adapters on top of the quantized model to correctly perform fine-tuning. Please see: https://huggingface.co/docs/transformers/peft for more details

I delete model = model.merge_and_unload(), then everything is fine.