Pinned Repositories
Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models
dsc204a-w24
Website for DSC 204a, Winter 2024
dsc291-PA
dsc291-s24
Website for DSC 291, Spring 2024
hao-ai-lab.github.io
LookaheadDecoding
[ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
MuxServe
vllm-ltr
[NeurIPS 2024] Efficient LLM Scheduling by Learning to Rank
hao-ai-lab's Repositories
hao-ai-lab/LookaheadDecoding
[ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
hao-ai-lab/Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models
hao-ai-lab/MuxServe
hao-ai-lab/vllm-ltr
[NeurIPS 2024] Efficient LLM Scheduling by Learning to Rank
hao-ai-lab/dsc204a-w24
Website for DSC 204a, Winter 2024
hao-ai-lab/dsc291-PA
hao-ai-lab/dsc291-s24
Website for DSC 291, Spring 2024
hao-ai-lab/hao-ai-lab.github.io