/exllamav2-KTransformers

A fast inference library for running LLMs locally on modern consumer-class GPUs, supporting DeepSeek and Qwen2 MoE

Primary LanguagePythonMIT LicenseMIT

Stargazers

No one’s star this repository yet.