Pinned Repositories
.github
Meta-Github repository for all GPUStack repositories.
fastfetch
Like neofetch, but much faster because written mostly in C.
gguf-packer-go
Deliver LLMs of GGUF format via Dockerfile.
gguf-parser-go
Review/Check GGUF files and estimate the memory usage and maximum tokens per second.
gpustack
Manage GPU clusters for running AI models
gpustack-ui
gpustack.github.io
llama-box
LM inference server implementation based on *.cpp.
vox-box
A text-to-speech and speech-to-text server compatible with the OpenAI API, supporting Whisper, FunASR, Bark, and CosyVoice backends.
GPUStack's Repositories
gpustack/gpustack
Manage GPU clusters for running AI models
gpustack/gguf-parser-go
Review/Check GGUF files and estimate the memory usage and maximum tokens per second.
gpustack/llama-box
LM inference server implementation based on *.cpp.
gpustack/vox-box
A text-to-speech and speech-to-text server compatible with the OpenAI API, supporting Whisper, FunASR, Bark, and CosyVoice backends.
gpustack/gguf-packer-go
Deliver LLMs of GGUF format via Dockerfile.
gpustack/gpustack-ui
gpustack/.github
Meta-Github repository for all GPUStack repositories.
gpustack/fastfetch
Like neofetch, but much faster because written mostly in C.
gpustack/gpustack.github.io