mostlygeek/llama-swap
On-demand model switching with llama.cpp (or other OpenAI compatible backends)
GoMIT
Issues
- 3
- 3
- 3
problems unloading the model
#28 opened - 2
Add Homebrew Support for macOS
#26 opened - 2
- 3
- 0
- 5
[Feature] Queue requests
#19 opened - 0
List in llama.cpp readme
#18 opened - 3
- 3
Support more API end points.
#12 opened - 5
Proxy does not set content length.
#11 opened - 5
- 1
- 1
Support routing to multiple backends
#7 opened - 7
Container execution
#5 opened - 0
Support the v1/embedding endpoint
#4 opened - 0
- 1