OptimalScale/LMFlow

[BUG]Model size change

Mikivishy opened this issue · 5 comments

I was wondering why the gpt2-large model I downloaded from huggingface was 3.1G, but after run_fineturn_with_lora_save_aggregated_weightss.sh it was only 1.5G. This may be a defect in my professional knowledge. Sorry to disturb you. I hope you can get an answer. Here's how I ran it:

./scripts/run_finetune_with_lora_save_aggregated_weights.sh
--model_name_or_path gpt2-large
--dataset_path /data1/LMFlow/data/integrated-data1/dataset6
--output_model_path output_models/gpt2-large-inte6

Thanks for your interest in LMFlow! I am wondering if @hendrydong could help look into this problem? It can be caused by transformers version upgrades. To the best of our knowledge, huggingface is using different model card format after transformers >= 4.30.x, so the merging script may not be functioning well for the latest version of transformers.

Could you please provide your transformers versions so we could help you locate the problem? Thanks 😄

Thanks for your interest in LMFlow! I am wondering if @hendrydong could help look into this problem? It can be caused by transformers version upgrades. To the best of our knowledge, huggingface is using different model card format after transformers >= 4.30.x, so the merging script may not be functioning well for the latest version of transformers.

Could you please provide your transformers versions so we could help you locate the problem? Thanks 😄

thanks for your reply,the version of my transformers library is 4.32.1

Given your model size, I think the precision may play a role. FP32 vs FP16/BF16?