OpenCSGs/llm-inference
llm-inference is a platform for publishing and managing llm inference, providing a wide range of out-of-the-box features for model deployment, such as UI, RESTful API, auto-scaling, computing resource management, monitoring, and more.
PythonApache-2.0
Stargazers
- c4fun
- chhao
- crazypangpang
- depenglee1707Bei Jing
- echolvmx
- fanchileChina
- ganisback
- hanxiaofeiEed
- hiveerOpenCSG
- huangfudeSuZhou
- HuFeiHu
- ijaa
- jasonhe258
- jony4Sensetime
- KinglyWayne
- MrFengJian西安
- Pang-GJHarbin Institute of Technology
- qiulint
- rainprob
- RickyFeng1997
- samchen8008
- SeanHH86
- tezhengMicrosoft
- TuanAnhNguyen14111998Hà Nội - Việt Nam
- wanggxa
- wayneliu0019
- WENTAOWU-Marc
- xiagujinqiao
- Xiaolong-RRLPKU
- Xuquansheng
- yakami129shanghai
- yongfengdu
- youngbeom-shin@OpenCSGs
- yxxhero@kubernetes @helmfile @gitlab.cn @helm @dragonflyoss
- zhenrong-wanghpc-now technologies
- zrl4836