Issues
- 4
[Bug]: 对lora merge后的模型量化,量化后模型输出一直出现human:
#1029 opened by shenshaowei - 4
[Bug]: qwen2.5 tools输出格式问题
#1072 opened by kwjlhh - 0
[QwQ-Preview]: LiveCodeBench复现
#1121 opened by mattgithub1919 - 1
[QwQ-Preview]: 请问Qwen2.5-7B-Instruct默认是FP32还是FP16
#1120 opened by lckj2009 - 5
您好,请问QwQ-32B-Preview发布的榜单中LiveCodeBench测试集中2024.08-2024.11数据集是从哪里的?LiveCodeBench最新开源数据集只截至到2024.09
#1112 opened by 13416157913 - 2
[Badcase]: Generate json format error
#1095 opened by jxrjlxc02 - 4
[Bug]: 在 4 卡 16GB V100 机器上采用 lmdeploy 部署 qwen2.5-32b-instruct-gptq-int4 模型,最高输出速度只有 80token/s ,请问这个速度正常吗?
#1023 opened by SolomonLeon - 0
[Badcase]: Qwen2.5-7B-Instruct 中德、中意翻译 不遵循指令、code switch
#1097 opened by pio57019 - 2
[Bug]: AttributeError: Model Qwen2ForCausalLM does not support BitsAndBytes quantization yet.
#1049 opened by yananchen1989 - 1
[Bug]: Qwen 2.5 32B/14B keep repeating input prompt
#1089 opened by youyouisme - 5
[BUG]: autoawq 0.2.7量化得到模型,无法使用vllm进行推理
#1117 opened by dcdmm - 4
"llama_model_load: error loading model: check_tensor_dims: tensor 'token_embd.weight' not found" after using "Quantizing the GGUF with AWQ Scale"
#1101 opened by Autism-al - 2
- 1
[Bug]: Poor support for Hebrew
#1114 opened by shohamyamin - 4
- 2
A simplified version of the inference code ?
#1057 opened by weizhenhuan - 6
[Badcase]: openai.BadRequestError: Error code: 400 - {'error': {'message': 'unexpected EOF', 'type': 'invalid_request_error', 'param': None, 'code': None}}
#1026 opened by XyLove0223 - 8
[Bug]: Qwen2.5-14B-Instruct-GPTQ-Int4存在严重的复读机和幻觉现象
#1024 opened by yang-collect - 1
[Badcase]:
#1027 opened by zhuzcalex - 16
[Bug]: vllm infer Qwen2.5-32B-Instruct-AWQ with 2 * Nvidia-L20, output repeat !!!!
#1090 opened by RoyaltyLJW - 4
[Bug]: Qwen2.5-32b-int4用vllm跑 好像只会生成感叹号
#1103 opened by ciaoyizhen - 0
[Badcase]: Twenty-four game, how to get 24 from 23, 3, 11, 16 Not possible with the given constraints
#1109 opened by novohool - 0
- 3
[Bug]: 模型回答时极大概率出现�
#1039 opened by GitHub-lql - 4
[Bug]: Inference code that works locally (M1 Mac) fails on cloud infrastructure (Linux box)
#1102 opened by nathancarter - 3
- 5
- 4
[Badcase]: loss unstable
#1074 opened by Solo4working - 2
- 3
您好,请问一下为什么Qwen2.5官方发布的榜单中72B和32B的LiveCodeBench分数跟LiveCodeBench作者公开72B和32B的分数差异这么大?
#1079 opened by 13416157913 - 1
[Badcase]: 使用Reasoning数据微调2.0成功,但是2.5失败
#1085 opened by UESTCthb - 3
[Badcase]: 用mlx框架把Qwen2.5-3B-Instruct转换为mlx版本后,又用mlx_lm.convert把mlx版本的模型进行int4量化,最后进行推理的时候报错:ValueError: [dequantize] The matrix should be given as a uint32
#1041 opened by ghoshadow - 1
[Badcase]: 使用llamafactory量化gptq时,使用2个GPU时出错
#1084 opened by czhcc - 0
- 1
[Badcase]: 抢占式实例部署qwen2.5-72B成功,调用失败
#1088 opened by ZX1998-12 - 2
[Badcase]: 相同的微调数据,Qwen1.5 14B准确率比Qwen2.5 14B高20%左右,这是什么原因
#1016 opened by Jayc-Z - 9
[Bug]: The final reason why you will get a model that cannot stop generation when you fine-tune the Qwen2.5-7b-base use Lora and a non-<|endoftext|> token as eos_token.
#1064 opened by hxs91 - 2
[Badcase]: 函数调用结果和目标预期不符
#1075 opened by akai-shuuichi - 1
- 1
[Bug]: 使用Qwen2.5-72B-Instruct-AWQ和Qwen2.5-32B-Instruct-AWQ官方给的演示例子,运行的时候提示:returned non-zero exit status 1
#1059 opened by SuSuStarSmile - 1
- 7
[Bug]: qwen2.5-72b-insruct math 自测分数和榜单分数差异较大
#1020 opened by tianshiyisi - 3
[Bug]: Use of the term "open source" to describe qwen when the training data is not open
#1055 opened by phly95 - 1
[REQUEST]:
#1040 opened by DAAworld - 3
[Bug]: vllm启动大模型,超过一定的上下文长度导致大模型回答答非所问
#1031 opened by Ave-Maria - 4
[Bug]: Model name error in vllm deployment
#1052 opened by JulioZhao97 - 1
[Bug]: can not deploy qwen2.5 with vllm
#1048 opened by joyyyhuang - 1
[Badcase]: qwen2.5一定概率生成\\n
#1021 opened by 520jefferson - 0
- 0
Qwen2.5-1.5b用LLaMA-factory lora微调后,vLLM加载模型报错,求教~
#1022 opened by Jimmy-L99