hiyouga/ChatGLM-Efficient-Tuning
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
PythonApache-2.0
Issues
- 1
微调后不起效果,是我哪步有问题吗
#406 opened by RemSynch - 1
- 1
epoch的选择&数据集构造
#404 opened by 1028686314 - 1
step和epoch一样
#401 opened by qingjiaozyn - 1
用默认参数微调chatglm2之后对话能力大幅度下降
#402 opened by caoxueqian19901214 - 2
仓库过大,git下载慢
#398 opened by JackKuo666 - 1
请问下在4-bit量化模式Lora微调最小的显存要求
#394 opened by NJU201220184 - 4
4*V100 32g配置,满足zero3全量微调吗?
#392 opened by gongjunjin - 0
- 0
- 0
ptuning后,模型参数不匹配
#383 opened by HLF217 - 4
为什么加载sft权重计算的指标和sft+ppo微调的权重计算的指标是一样的?
#382 opened by JingyuHuAtVoxelCloud - 1
- 4
为什么glm2训练需要较大的学习率才有效
#379 opened by JingBob - 2
lora后模型的泛化能力如何增强呢?
#377 opened by mawenju203 - 0
多卡微调错误:Exception: Could not find the transformer layer class to wrap in the model.
#378 opened by lrx1213 - 5
请问用Lora微调2万条医疗问答,为什么Loss只能降低到2.7,效果不佳
#375 opened by tongjialiang - 2
- 1
LoRA微调已经训练好了,关于Evaluate和Predict的几个疑问
#368 opened by FreeYiran - 2
多机多卡训练是否有相关资料呢?
#367 opened by yifan-chen-2020 - 2
模型训练的问题? 模型出现过拟合
#360 opened by mawenju203 - 4
训练成功 启动测试的时候报错了
#356 opened by bubblebyfly - 3
关于sft之后,奖励模型和rlhf
#372 opened by lzh1998-jansen - 0
数据集oaast_rm_zh问题
#403 opened by PluginsKers - 3
有关ChatGLM-Efficient-Tuning + lora微调出现过拟合的讨论
#365 opened by mawenju203 - 1
请问lora训练是只会训练注意力层,在注意力层加适配器吗?而不训练前馈层。然后freeze是训练前馈层
#387 opened by DRZJ1 - 2
lora微调之后导出的完整模型文件应该要如何调用 我使用transformers的接口加载模型报错
#388 opened by ecent2000 - 4
请问如何制作自己的数据集?
#376 opened by andyzhou1982 - 2
lora_target 的可用值有哪些?
#400 opened by godcrying - 5
when `per_device_eval_batch_size` > 1 and launch by deepspeed, RuntimeError: Tensors must be contiguous
#385 opened by jiahuanluo - 0
- 2
跑sft阶段出现这个问题,环境应该是ok的
#397 opened by wangweihua11 - 1
2400条数据,10个epoch,pre_seq_len=128,lora训练,为什么推理时显示trainable params: 0 || all params: 6243584000 || trainable%: 0.0000,是数据量不够吗,下边贴出训练参数
#396 opened by zeze813 - 2
Learning Scheduler Issue
#393 opened by XingzhiZhou - 2
collator.py的第126行与preprocess.py的preprocess_supervised_dataset是不是存在冲突?input_ids拼接了2次labels
#391 opened by jennfer0808 - 5
如何设置输入长度
#363 opened by OpenHuShen - 2
chatglm2进行lora微调,怎么执行断点续训呢?
#384 opened by Qiang-HU - 3
单机4卡3090训练lora报错
#371 opened by 1006076811 - 1
- 3
- 9
关于PPO中ref_model的问题
#380 opened by JoyYChen - 2
- 4
在服务器上执行 src/train_web.py 报错
#361 opened by FreeYiran - 1
- 0
datasets.builder.DatasetGenerationError: An error occurred while generating the dataset
#374 opened by ArtificialZeng - 3
建议对 data/*.json 文件使用 lfs 管理
#366 opened by Haskely - 2
请问int8版本的ppo怎么运行
#373 opened by couldn - 9
训练不小心中断后,如何设置进行断点续训?
#362 opened by FreeYiran - 1
微信群满啦,麻烦更新下二维码
#359 opened by hongshuo-wang - 0
在web_demo_finetuned.py运行ChatGLM(不是ChatGLM2)时报ValueError: not enough values to unpack (expected 3, got 2)
#358 opened by yug030