Issues
- 1
TypeError: transformers.generation.utils.GenerationMixin.generate() argument after ** must be a mapping, not Tensor
#178 opened by LSK-1 - 1
- 1
index-1.9b-chat 接入langchain报错
#177 opened by jiangzhaokun - 1
05-Qwen2-7B-Instruct Lora 微调报错
#176 opened by hgsw - 3
qwen2 lora微调报错:RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
#172 opened by ArlanCooper - 3
04-GLM-4-9B-Chat vLLM 部署调用显存溢出问题
#162 opened by SongHY-13 - 1
Index-1.9B-Chat Lora 微调合并模型出现ValueError: Can't find 'adapter_config.json' at './output/Index-1.9B-Chat-lora/checkpoint-600'
#175 opened by tangyipeng100 - 3
LlaMA3在英文或者中文上tokenizer是否需要加bos token?
#140 opened by sugarandgugu - 2
GLM-4-9b-chat web版运行时报属性messges不存在
#173 opened by eagleLiu82 - 0
BUG:按照GLM-4示例不能启动API服务
#174 opened by eternal-bug - 4
使用阿里云部署glm-4-9b api报gpu内存不足
#169 opened by eagleLiu82 - 2
如何获取一个大模型的stop_token_ids
#170 opened by jdm4pku - 1
- 1
deepseek-llm-7b-chat微调报错
#171 opened by lzh123415 - 3
Index-1.9B-Chat webDemo部署报错问题
#166 opened by gzhuuser - 1
llama3 langchain
#163 opened by Alexa2077 - 1
有考虑添加多模态大模型的内容吗?
#164 opened by sunhaha123 - 2
深圳市原子同学?
#165 opened by hunchunchao23 - 4
使用 llama3 的 lora 微调报错:NotImplementedError: Cannot copy out of meta tensor; no data!
#115 opened by chenmonster - 0
数据集如何构造?
#161 opened by chopin1998 - 4
微调出来会有不礼貌或攻击性的言语
#108 opened by 1737686924 - 2
DeepSeek-7B-chat-Lora微调训练时报错
#139 opened by 3Justice - 1
可以出一版验证集的process-func吗(Qwen、llama模型的数据集格式)?
#151 opened by demouo - 1
2 warnings: MatMul8bitLt: inputs will be cast from {A.dtype} to float16 during quantization") / kernel version 4.18.0
#152 opened by x6p2n9q8a4 - 2
Qwen1.5-7B Lora微调报错
#107 opened by JeckerWen - 6
peft微调llama3 8b,从第10补开始loss一直都是0
#124 opened by ykallan - 2
04-Qwen-7B-Chat Lora 微调时报错
#119 opened by FisherAndHunter - 2
- 2
华为昇腾服务器的部署方案
#147 opened by dayphosphor - 2
WebDemo部署对话如何逐字逐句显示而不是一次性生成
#149 opened by eternal-bug - 1
Qwen1.5-7B-Chat vLLM 部署调用-速度测试 hf命令错误
#101 opened by ericalduo - 3
【非Issues!讨论帖】通过lora微调的qianwen和直接使用system来预设的区别貌似不是很大
#100 opened by Roger-G - 6
- 1
为什么数据格式化要把输入和输出的合并起来放到input_id?
#129 opened by lmh0921 - 1
01-Qwen1.5-7B-Chat FastApi 部署调用.md 传入数据错误
#128 opened by cqray1990 - 1
考虑出一些LLM时序预测模型的相关教程吗
#127 opened by xaxK - 0
如懿传
#126 opened by xixihahahhhhhhhhh - 1
请问有多模态LLM的部署/微调文档吗,未来有相关更新计划吗
#121 opened by ybshaw - 1
模型微调时报错,报内核版本问题 Detected kernel version 5.4.0
#125 opened by hgsw - 5
微调Qwen1.5-0.5b报错 PermissionError: [Errno 13] Permission denied: './output/Qwen1.5\checkpoint-100'
#123 opened by ykallan - 1
llama3 api报错
#122 opened by samprasgit - 2
- 2
deepseek-v2部署请求
#120 opened by airsxue - 1
在纯 CPU 上可以运行吗?比如苹果电脑没有 cuda?
#117 opened by aa03146 - 1
chatglm3,lora微调报错
#116 opened by Jsonzhang20 - 8
- 2
Qwen1.5-7B Lora微调报错:RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
#112 opened by feifeifei-hue - 1
llama3 API调用的问题
#114 opened by 1148514800 - 4
- 2
InternLM2 缺少包
#98 opened by JinProton