Pinned Repositories
360zhinao
360zhinao
AutoGPTQ
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
etcd-cpp-apiv3
C++ API for etcd
garycao
programming
go-fs-secretary-prototype
Secretary script prototype for FreeSWITCH
handson-ml2
A series of Jupyter notebooks that walk you through the fundamentals of Machine Learning and Deep Learning in Python using Scikit-Learn, Keras and TensorFlow 2.
hugegraph
HugeGraph Database core component, including graph engine, API, and built-in backends
kubernetes
Production-Grade Container Scheduling and Management
libco
libco is a coroutine library which is widely used in wechat back-end service. It has been running on tens of thousands of machines since 2013.
neo4j-go-driver
Neo4j Bolt Driver for Go
garycaokai's Repositories
garycaokai/neo4j-go-driver
Neo4j Bolt Driver for Go
garycaokai/360zhinao
360zhinao
garycaokai/AutoGPTQ
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
garycaokai/etcd-cpp-apiv3
C++ API for etcd
garycaokai/garycao
programming
garycaokai/go-fs-secretary-prototype
Secretary script prototype for FreeSWITCH
garycaokai/handson-ml2
A series of Jupyter notebooks that walk you through the fundamentals of Machine Learning and Deep Learning in Python using Scikit-Learn, Keras and TensorFlow 2.
garycaokai/hugegraph
HugeGraph Database core component, including graph engine, API, and built-in backends
garycaokai/kubernetes
Production-Grade Container Scheduling and Management
garycaokai/libco
libco is a coroutine library which is widely used in wechat back-end service. It has been running on tens of thousands of machines since 2013.
garycaokai/lightllm
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
garycaokai/redis-plus-plus
Redis client written in C++
garycaokai/serve
Model Serving on PyTorch
garycaokai/text-generation-inference
Large Language Model Text Generation Inference
garycaokai/transformer-deploy
Efficient, scalable and enterprise-grade CPU/GPU inference server for 🤗 Hugging Face transformer models 🚀
garycaokai/trt-samples-for-hackathon-cn
Simple samples for TensorRT programming
garycaokai/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs