Pinned Repositories
Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models
cse234-w25
Website for CSE 234, Winter 2025
dsc204a-w24
Website for DSC 204a, Winter 2024
dsc291-PA
FastVideo
FastVideo is a lightweight framework for accelerating large video diffusion models.
hao-ai-lab.github.io
LookaheadDecoding
[ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
MuxServe
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
vllm-ltr
[NeurIPS 2024] Efficient LLM Scheduling by Learning to Rank
Hao AI Lab's Repositories
hao-ai-lab/LookaheadDecoding
[ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
hao-ai-lab/FastVideo
FastVideo is a lightweight framework for accelerating large video diffusion models.
hao-ai-lab/Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models
hao-ai-lab/MuxServe
hao-ai-lab/vllm-ltr
[NeurIPS 2024] Efficient LLM Scheduling by Learning to Rank
hao-ai-lab/dsc204a-w24
Website for DSC 204a, Winter 2024
hao-ai-lab/dsc291-PA
hao-ai-lab/cse234-w25
Website for CSE 234, Winter 2025
hao-ai-lab/dsc291-s24
Website for DSC 291, Spring 2024
hao-ai-lab/hao-ai-lab.github.io
hao-ai-lab/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs