datawhalechina/self-llm
《开源大模型食用指南》针对**宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
Jupyter NotebookApache-2.0
Issues
- 1
qwen2-vl微调代码有问题
#296 opened by Zyvpeng - 2
自定义视觉模型和语言模型组合
#295 opened by ander008 - 15
GLM-4-9B-chat FastApi 部署调用报错“Method not allowed”
#293 opened by Ch1r3 - 2
qwen2.5 lora微调报错
#271 opened by qihouji - 1
关于没有GPU的情况咨询(队友呢?救一下)
#292 opened by 912100012 - 1
小白不懂本地部署
#278 opened by belo-belove - 4
AutoDL平台Qwen2.5的环境镜像报错
#290 opened by netAir - 1
LLaMA3_1-8B-Instruct Lora 微调 数据格式化问题
#275 opened by Evilxya - 1
lora微调 没有用peft_model包装base pre train model
#277 opened by MXD6 - 0
使用langchain接入模型,如何适配with_structured_output
#274 opened by Und3r1ine - 7
uvicorn running之后,点击网页,之后显示”目前无法处理此请求“,有没有赛博神医救救
#270 opened by crow-1412 - 5
GLM-4-9B-chat FastApi报AttributeError: 'ChatGLMForConditionalGeneration' object has no attribute 'chat'
#249 opened by hongmin118 - 1
使用fastapi 的GLM4. 出现 AttributeError: 'ChatGLMForConditionalGeneration' object has no attribute 'stream_generate'. Did you mean: 'can_generate'?
#239 opened by triumph - 0
请问可以出一个mistral-instruct-v0.1版本的sft模型吗
#269 opened by drewqueaz - 1
AutoDL的llama3环境镜像transformers版本需要更新
#268 opened by HackerGary - 0
The attention mask is not set and cannot be inferred from input because pad token is same as eos token. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
#267 opened by yangyang1-0 - 0
关于GPU利用率低的问题
#266 opened by Walking-Z - 3
Qwen1.5微调报错
#265 opened by Archerer - 1
使用AutoDL报错
#264 opened by cyk2018 - 1
Llama3 Lora报错:RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
#229 opened by AndyLuo1029 - 1
输出一直在重复
#237 opened by quanquan07 - 1
GLM4私有部署如果要用langchain搭建agent,还需要做哪些改进?
#250 opened by Jimmy-L99 - 1
xtuner 微调报错,求指点
#260 opened by zhanbaohang - 1
要是能出个多模态的就好了
#261 opened by robin-programmer - 1
vllm能够直接调用gguf格式的权重吗
#262 opened by liguoyu666 - 0
[解决方案]Qwen-7B-Chat FastApi 部署调用
#263 opened by YuGuilliman - 0
xtuner 微调internlm2.5出错
#259 opened by zhanbaohang - 1
Qwen2-7B-Instruct Lora微调的问题
#230 opened by yinzih - 5
使用 8 x 4090 微调 llama3-1-8B 后,询问无输出反馈。
#241 opened by Sakkana - 2
glm4 AttributeError: 'ChatGLMForConditionalGeneration' object has no attribute 'chat'
#245 opened by lydd2024 - 3
llama3.1-8B微调后推理没有结果
#231 opened by Telogen - 1
- 1
Yuan2.0的python request调用方法的示例图片不对
#247 opened by Aomferni - 4
LLaMA3_1-8B-Instruct WebDemo 部署 打开Web页面报错
#235 opened by vistar-terry - 0
GLM4教程中“02-GLM-4-9b-Chat 接入 LangChain”代码运行报错
#246 opened by meisongzhu - 0
glm4教程中小问题
#243 opened by warmth27 - 0
DeepSeek-Coder-V2推理警告
#242 opened by Qlalq - 2
通过vllm部署的GLM-4-9B-Chat模型效果很拉垮
#240 opened by Duanexiao - 2
deepseek-chat模型LorA微调完没有adapter_config.json
#238 opened by HolyCrazy - 1
GLM-4-9B-chat FastApi,curl调用时抛错
#236 opened by hongmin118 - 2
- 4
大模型Prompt Template格式
#233 opened by fang-siqi - 0
请问langchain如何实现自定义开源LLM的 stream 响应模式呢?
#232 opened by yuanasdfgh - 0
'NoneType' object has no attribute 'shape'
#228 opened by jjzhu0579 - 6
Qwen2-7B-Instruct WebDemo部署的问题
#227 opened by yinzih - 4
Qwen2-7B-Instruct Langchain 接入的问题
#225 opened by yinzih - 1
Qwen2-7B-Instruct Langchain 问题
#226 opened by yinzih - 1
llama3 FastAPI部署报错怎么解决哇ww
#222 opened by UandUfindU - 0
- 3
采用多张显卡进行GLM-4微调时报错RuntimeError
#217 opened by shutter-cp