lzd19981105/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
PythonMIT
Watchers
No one’s watching this repository yet.
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
PythonMIT
No one’s watching this repository yet.