Issues
- 2
deepseek lora
#91 opened by xionghao132 - 3
【非Issues!讨论帖】通过lora微调的qianwen和直接使用system来预设的区别貌似不是很大
#100 opened by Roger-G - 1
请问有多模态LLM的部署/微调文档吗,未来有相关更新计划吗
#121 opened by ybshaw - 2
peft微调llama3 8b,从第10补开始loss一直都是0
#124 opened by ykallan - 5
微调Qwen1.5-0.5b报错 PermissionError: [Errno 13] Permission denied: './output/Qwen1.5\checkpoint-100'
#123 opened by ykallan - 3
使用 llama3 的 lora 微调报错:NotImplementedError: Cannot copy out of meta tensor; no data!
#115 opened by chenmonster - 1
llama3 api报错
#122 opened by samprasgit - 2
- 2
deepseek-v2部署请求
#120 opened by airsxue - 1
04-Qwen-7B-Chat Lora 微调时报错
#119 opened by FisherAndHunter - 1
在纯 CPU 上可以运行吗?比如苹果电脑没有 cuda?
#117 opened by aa03146 - 1
chatglm3,lora微调报错
#116 opened by Jsonzhang20 - 8
- 2
Qwen1.5-7B Lora微调报错:RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
#112 opened by feifeifei-hue - 1
llama3 API调用的问题
#114 opened by 1148514800 - 2
LLaMA3-8B-Instruct+lora使用A800(80GB显存)微调长度8192
#93 opened by 12915494174 - 1
Qwen1.5-7B-Chat vLLM 部署调用-速度测试 hf命令错误
#101 opened by ericalduo - 3
微调出来会有不礼貌或攻击性的言语
#108 opened by 1737686924 - 1
Qwen1.5-7B Lora微调报错
#107 opened by JeckerWen - 3
qwen1.5 量化版本的部署方案有么?
#54 opened by chongqingwei - 0
- 4
- 2
peft训练完成,参考的是04-Qwen-7B-Chat Lora 微调.ipynb,但是重新载入模型时,提示peft版本问题,用的是model = AutoModelForCausalLM.from_pretrained("../output/Qwen/checkpoint-1300/", trust_remote_code=True).eval(),提示的错误时ValueError: The version of PEFT you are using is not compatible, please use a version that is greater than 0.5.0
#95 opened by Roger-G - 2
InternLM2 缺少包
#98 opened by JinProton - 6
- 1
想问下在这个项目下的lora微调和Chatglm3官方微调的demo的数据格式怎么不一样呀
#96 opened by McRays - 6
请问LLAMA3,里面是按1.2.3.4的顺序来分别执行吗?
#94 opened by YadiHe - 0
- 3
请问ChatGLM3微调中的数据集huanhuan的在哪获取?
#81 opened by haory-95 - 3
多卡报错,Qwen1.5-7B-Chat FastApi 部署调用
#79 opened by linzhonghong - 8
请问chatglm模型Lora微调完成之后,如何加载新模型?
#72 opened by waynetest2024 - 4
ChatGLM3-6B微调后成哑巴了(字面意思)
#65 opened by Tangent-90C - 0
与FastChat的区别
#78 opened by deeomnjobs - 1
建议按照顺序撰写README部分教程
#75 opened by EuthL - 1
chatglm3-6b fastapi调用
#74 opened by caixiongjiang - 3
Qwen-1.5-4B LLM推理bug
#73 opened by caixiongjiang - 1
- 7
deepseek官方readme,loss第二轮开始就是
#63 opened by liuyongjie985 - 1
我想我可不可提交给PR支持一下BlueLM我们的蓝心大模型
#66 opened by acwwt - 3
chatglm搭建知识库读取文件出错
#53 opened by AHPUymhd - 3
帮大忙了
#51 opened by sssggghhh007 - 5
想问下部署chatglm出现了问题
#55 opened by DJzing - 0
- 2
这不来个llama的微调教程
#56 opened by Tangent-90C - 1
Qwen1.5-7B推理部分,为什么我在modelscope上的GPU调试,fastapi报错。
#58 opened by tking007 - 4
- 3
- 3
- 8
Failed to import transformers.models.qwen2
#44 opened by lhtpluto - 2
代码笔误pd.read_json('../dataset/huanhuan.jsonl'),应该是pd.read_json('../dataset/huanhuan.json')
#45 opened by Weihong-Liu