jiqing-feng's Stars
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
huggingface/optimum-intel
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
pytorch/PiPPy
Pipeline Parallelism for PyTorch
PKUFlyingPig/cs-self-learning
计算机自学指南