libraryofcelsus/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
PythonMIT
No issues in this repository yet.
A fast inference library for running LLMs locally on modern consumer-class GPUs
PythonMIT
No issues in this repository yet.