Pinned Repositories
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
123
Git-Github-notes-for-study
学习Git和GitHub过程中的学习总结和思考(开源特训营)
llm-compressor
Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
java_project
CharlesRiggins's Repositories
CharlesRiggins/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs