Issues
- 0
请问可以训练chatGLM2-6B-32K的模型嘛
#23 opened by GuocaiL - 0
- 8
- 0
ChatGLM数据处理中input_ids的padding问题
#21 opened by nostalgiaer - 1
微调时在runing tokenizer on train dataset这一步卡住
#20 opened by underwoodnoble - 2
我只有200多条多轮对话的数据,去做全参微调能有效果吗?
#19 opened by ymmbb8882ymmbb - 1
- 4
模型训练,出现deepspeed_init() got an unexpected keyword argument 'resume_from_checkpoint'
#12 opened by reyvand14 - 0
max_len 768 这个长度在多轮对话中,是单次输入的长度,还是history+这次输入的长度? 如果是单次输入的长度,那么history的长度在哪里?
#16 opened by ymmbb8882ymmbb - 0
- 1
- 1
有考虑 加入lora训练吗
#9 opened by nsl2014fm - 0
请问这个问题怎么解决OSError: Can't get source for <function apply_rotary_pos_emb at 0x7fef8c15f790>. TorchScript requires source access in order to carry out compilation, make sure original .py files are available.
#14 opened by ymmbb8882ymmbb - 9
请问全参数微调需要什么配置呢?
#10 opened by bigbigwatermalon - 1
请问这个项目的依赖是什么,安装的deepspeed版本好像跑不了。
#11 opened by Hollywood3 - 0
训练数据过多时报错Socket Timeout
#13 opened by onair1314 - 2
模型训练,input_ids出现None type
#7 opened by Fanshell2333 - 0
为什么不用新版本的transformer么?
#8 opened by sunlei198911 - 1
CUDA out of memory. Tried to allocate 11.63 GiB (GPU 0; 23.69 GiB total capacity; 11.63 GiB already allocated; 11.28 GiB free
#5 opened by harbor1981 - 0
多轮数据下载后要如何处理?
#4 opened by wac81 - 1
do_eval时报错
#3 opened by kpg-pgk - 1