llm-compression
There are 7 repositories under llm-compression topic.
horseee/Awesome-Efficient-LLM
A curated list for Efficient Large Language Models
pprp/Pruner-Zero
Evolving Symbolic Pruning Metric from scratch
Picovoice/llm-compression-benchmark
LLM Compression Benchmark
VITA-Group/llm-kick
[ICLR 2024] Jaiswal, A., Gan, Z., Du, X., Zhang, B., Wang, Z., & Yang, Y. Compressing llms: The truth is rarely pure and never simple.
Picovoice/serverless-picollm
LLM Inference on AWS Lambda
GongCheng1919/bias-compensation
[CAAI AIR'24] Minimize Quantization Output Error with Bias Compensation
bupt-ai-club/llm-compression-papers
papers of llm compression