mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
PythonMIT
Watchers
- 0xezMelbourne
- AlbertBJBeijing, China
- cetiumBeijing ,China
- ctejada85
- doctorimage
- eemailme
- ghchris2021
- gonnavis@webaverse
- jamesdborin
- jnulzlGuangZhou China
- kdy1999
- liuguoyou
- LuukvEFreelance
- meenchen
- nanggiomientrungVietnam
- nirvanesqueAccenture
- QubitiumModelCloud.ai
- SakitsMIT, EECS
- shayanc1985
- songhanMIT, NVIDIA
- songkq
- tonylinsMIT, EECS
- vince62sSeedfall
- wx-bRIOS