/MoA

The official implementation of the paper <MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression>

Primary LanguagePythonMIT LicenseMIT

Stargazers

No one’s star this repository yet.