kvcache-ai/ktransformers
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
PythonApache-2.0
Issues
- 0
复现InternLM2.5-7B-Chat-1M报错
#114 opened by Cherishyt - 1
- 0
RuntimeError CUDA error when running Infinite Bench
#113 opened by Flitieter - 1
Detailed specification of the computer hardware to run 236B DeepSeek-Coder-V2
#108 opened by atomlayer - 0
- 0
install error on windows, need help
#109 opened by gaowayne - 1
how to implement new algorithm in this repo?
#105 opened by lumiere-ml - 0
feature request: support internvl2
#107 opened by kolinfluence - 11
- 1
- 1
- 5
- 1
How to infer quantized models on CPU&GPU
#103 opened by shuzhang-pku - 2
Does ktransformers support deepseek V2.5?
#100 opened by huliangbing - 2
Specify MAX_NEW_TOKENS for ktransformers server
#92 opened by arthurv - 5
Busy loop in cpu_backend/task_queue.cpp keeps 1 thread at 100% CPU when queue is empty
#80 opened by sayap - 5
are marline and q4k totally equivalent?
#87 opened by Eutenacity - 1
How can I use opencompass benchmark tools to test ktransformers in long context?
#91 opened by AsVoider - 7
- 3
Deepseekv2推理速度很慢,看样子似乎在cpu上做推理,gpu利用率很低
#93 opened by Chain-Mao - 7
ImportError: DLL load failed while importing KTransformersOps: The specified module was not found.
#94 opened by SCP12rs - 4
Suggestion to add DeepSeek v2.5 support
#95 opened by arisau - 1
Installation Problem
#90 opened by Chain-Mao - 4
Installation requirements
#89 opened by arthurv - 2
可以给出详细的硬件配置清单吗?
#84 opened by qixing-ai - 2
Seg Fault on long replies
#82 opened by matthusby - 1
- 8
8-GPU configuration on L40 OOM
#76 opened by fengyang95 - 9
Is deepseek-ai/DeepSeek-V2.5 supported?
#79 opened by AshD - 1
- 4
- 2
Missing pip packages flash_attn and wheel
#69 opened by bitbottrap - 4
- 2
UnboundLocalError: cannot access local variable 'chunck_mask' where it is not associated with a value
#70 opened by fengyang95 - 1
Would you support glm4-chat-1m
#65 opened by choyakawa - 5
- 4
More Efficient Layer Distribution for DeepSeek Coder v2 on Multiple GPUs and CPUs
#49 opened by BGFGB - 2
Support for Mistral-Large-Instruct-2407-GGUF ?
#53 opened by LIUKAI0815 - 2
Add a instruction for configuring CUDA_HOME and CUDA_PATH to the install section of README.md.
#54 opened by hyx1999 - 5
- 1
- 1
Mixtral-8x7B-v0.1 GGUF file error
#42 opened by RealLittleXian - 2
- 3
Ubuntu 24.04 GLIBCXX version fail
#37 opened by ELigoP - 1
- 1
- 2
ollama chat not realised
#32 opened by xldistance - 1
- 0
Unable to use the web interface
#33 opened by xldistance - 3
using docker got errors
#28 opened by goldenquant