wangcx18/llm-vscode-inference-server
An endpoint server for efficiently serving quantized open-source LLMs for code.
PythonApache-2.0
Stargazers
- abutbulIsrael
- adriancboNew York, NY
- AIStorm
- appanpQuadrinous
- arkii
- bellyfat
- bertugmirasyediIstanbul, Turkey
- BluejacketScottBluejacket Software
- bosdhillArista Networks
- chemmyoneUnsure
- Cookiebreed127.0.0.1
- d5423197
- DanFitzgibbon@Gousto
- deba-iitbhIIT Bhilai
- ejalonasGartner
- evolu8
- HennerM@Speechmatics
- Jahysama@genxnetwork
- jcuditCanada
- lyzf
- melMassFR / US
- mikeshi80Shanghai Hyron Software Co Ltd.
- mkozarslanIstanbul
- MzicoDim
- noxaschKuala Lumpur, Malaysia
- Pent
- prawel
- ricardo-lisboasantosPortugal
- saxelsso
- tmur4t4
- umm-maybe
- Unicron187
- wangcx18Tsinghua University
- ZeroCool2u@FRBNY
- zouyzh
- zyxcambridgeshanghai