Issues
- 3
有没有llama 的onnx inference 脚本
#7 opened by hujuntao123 - 1
请问一下对应llama3的transformers版本是多少
#19 opened by sihouzi21c - 3
执行export_llama3.py时为什么会导出一大堆中间文件(参数)?
#16 opened by Lingzzyy - 4
请问Qwen转换出错问题:RuntimeError: Sizes of tensors must match except in dimension 2. Expected size 28 but got size 4 for tensor number 1 in the list.
#17 opened by yanxiao1930 - 1
Please uninstall/disable FlashAttention (and maybe xformers) before model conversion。请问这句话的意思是在模型转换前必须重新训练一个不使用flashattention的模型吗?
#18 opened by PeterXingke - 1
Can you give an example code of how to deduce onnx model after qwen switched to onnx
#15 opened by Pengjie-W - 0
有导出Qwen-VL的7B模型转onnx的程序吗?
#14 opened by chantjhang - 7
使用onnx库读取转换好的onnx模型报错
#13 opened by L1-M1ng - 1
用3090 导出7b 和13b llama2 报oom
#6 opened by hujuntao123 - 1
- 4
convert Qwen question
#1 opened by OneStepAndTwoSteps - 2
转换QWen-7B错误
#12 opened by L1-M1ng - 2
转换qwen模型的时候,提示atten_mask:5 error.
#10 opened by louwangzhiyuY - 1
转换llama也提示错误。 AttributeError: 'tuple' object has no attribute 'get_usable_length'
#11 opened by louwangzhiyuY - 3
使用3090导出 QWen-7b,报OOM问题。
#9 opened by linthy94 - 1
单卡a6000 50g会oom
#8 opened by 77281900000 - 1
- 1