Issues
- 9
python train.py执行训练报错,求解。
#214 opened by pan365wang - 0
请问adalora能用deepspeed训练吗
#266 opened by Yu-Yuqing - 4
LoRA和ptv2微调均发生OOM
#264 opened by shenzhyzzz - 2
INFO:lightning_fabric.utilities.rank_zero:`Trainer.fit` stopped: No training batches.
#261 opened by hasakikiki - 0
有谁用过Mac Studio微调的
#260 opened by xsailor511 - 4
如何使用evaluate.py对测试集进行验证
#236 opened by lawrencelxy - 1
- 4
修改max_seq_length好像并没有生效?
#256 opened by tjulh - 1
怎么控制每训练n轮就保存一次模型呢
#259 opened by tjulh - 5
- 2
推理时怎么指定用哪几张卡
#257 opened by tjulh - 2
您好,非常感谢您的工作。我进行全参数微调后,调用infer_finetuning.py的时候,出现Missing key(s) in state_dict: "_TransformerLightningModule__backbone.transformer.lm_head.weight".请问您有遇到过这个问题吗
#252 opened by Xuan-ZW - 0
显示可训练参数数量问题
#254 opened by xxll88 - 0
缺省Lora训练显存消耗 60G
#253 opened by is - 3
关于数据的instruction,input,output
#230 opened by sanwei111 - 1
模型训练只使用到了单个GPU
#247 opened by GZJAS - 1
ptuning v2 如何启动quantization_bit 4
#245 opened by xxll88 - 1
用单轮数据集。 p-tuning微调chatGLM之后出现的问题。
#243 opened by SMR-S - 7
使用ptv2的方式进行finetune,总是OOM(正常微调、lora方式都没问题)
#218 opened by lxw0109 - 1
请问一下,mac系统装不了deep_training?
#217 opened by WHJTC - 11
- 14
运行lora的训练代码, int8=True ,推理时报错 RuntimeError: expected scalar type Half but found Float,这是什么原因?
#239 opened by MathamPollard - 1
should be load_sft_weight?
#241 opened by HenryYuxuanWang - 0
单轮的QA回答的还不错,但是多轮对话时,就会出现瞎回答的情况。我感觉应该是多轮对话的history记录干扰了回答的效果?目前还不确定是不是这样的情况,有兄弟遇到类似的情况吗?交流一波啊!
#242 opened by SMR-S - 4
Lora int8微调,推理时出错
#202 opened by crellian - 5
整体微调以后,领域内的知识记住了,但是问常规问题,比如你好,你叫什么,他也回答领域内的知识
#206 opened by heiheiwangergou - 1
关于需要多少显卡资源
#234 opened by sanwei111 - 21
- 8
执行infer_lora_finetuning.py报错:‘NoneType’ objectg has no attribute 'learning_rate'
#238 opened by paizhongxing - 0
input_ids格式是否需要<CLS>
#237 opened by Jong-Won - 0
Overriding torch_dtype=None with `torch_dtype=torch.float16` due to requirements of `bitsandbytes` to enable model loading in mixed int8. Either pass torch_dtype=torch.float16 or don't pass this argument at all to remove this warning.
#235 opened by sanwei111 - 2
- 6
- 3
- 5
大佬 ,能讲一下如何合并lora权重到原来的模型中吗?
#203 opened by cywjava - 4
设置 LoRa微调的 'target_modules' 后,运行报错 "AssertionError"
#213 opened by ngbruce - 4
大佬好,请问关于scheduler
#210 opened by IamRoBota - 0
加载lora模型出错~
#223 opened by zlht812 - 1
请问如何试用一般新闻语料对ChatGLM进行继续finetuing呢?
#221 opened by yang9112 - 3
- 2
Lora推理2分30s正常吗?
#216 opened by jikhunb - 0
web/api_lora_demo.py 如何多张卡推理
#220 opened by lxw0109 - 2
deepspeed如何设置可以避免OOM
#209 opened by lianrzh - 2
Deepspeed stage3保存模型权重维度为0
#212 opened by Jong-Won - 2
Lora训练后推理问题
#215 opened by jikhunb - 19
大佬好,请问使用lora和ptv2进行微调分别需要修改哪些配置?
#211 opened by mircop1t - 2
大佬好,请问下数据构造中的特殊token
#208 opened by IamRoBota - 0
数据集
#207 opened by renmengjie7 - 0
训练数据集的q和a有长度限制吗,和max_seq_length是什么关系
#205 opened by lancexiao - 4
infer_lora_finetunning.py 报错
#204 opened by philipyao