tensorchord/modelz-llm
OpenAI compatible API for LLMs and embeddings (LLaMA, Vicuna, ChatGLM and many others)
PythonApache-2.0
Issues
- 2
Local custom .gguf modals supported?
#102 opened by sengiv - 1
- 3
Cant use it on Windows
#98 opened by yogeshhk - 1
Missing a LICENSE
#99 opened by loleg - 2
bug: Failed to generate outputs
#97 opened by gaocegege - 1
Function calling feature
#96 opened by willswordh - 1
add llama-2
#90 opened by antonkulaga - 1
do we support vicuna 13b, chatglm2 ?
#88 opened by timiil - 1
feat: provide instructions on how community members can wrap models for this project
#87 opened by PaulConyngham - 1
- 0
bug: Completion request returns wrong response
#83 opened by gaocegege - 0
chore: Fix vicuna 7b
#82 opened by gaocegege - 1
feat: support chatgpt web
#79 opened by dgqyushen - 0
feat: Support falcon 7b
#77 opened by gaocegege - 2
bug: Unexpected OOM in ChatGLM 6B
#69 opened by gaocegege - 3
feat: Support more models
#11 opened by gaocegege - 2
bug: Vicuna performance is not great
#73 opened by gaocegege - 2
bug: RuntimeError: "LayerNormKernelImpl" not implemented for 'Half' in chatglm int4
#62 opened by gaocegege - 4
bug: AttributeError: 'ChatGLMForConditionalGeneration' object has no attribute 'encoder'
#63 opened by arugal - 2
bug: Install cuda again in the image
#55 opened by gaocegege - 0
bug: RuntimeError: Only Tensors of floating point and complex dtype can require gradients
#60 opened by gaocegege - 2
- 0
bug: Object missing required field `model`
#50 opened by gaocegege - 0
test: check if the embedding API is compatible
#45 opened by kemingy - 0
bug: Extra blank space in output
#48 opened by gaocegege - 0
feat: Support flag to use CPU/GPU
#13 opened by gaocegege - 0
- 2
- 0
feat: Add CLI argument int8 and int4
#38 opened by gaocegege - 0
- 2
feat: Support embedding API
#14 opened by gaocegege - 2
bug: 500 with langchain sdk
#16 opened by gaocegege - 0
feat: Remove v1 from URI
#8 opened by gaocegege