/mixture-of-attention

Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts

Primary LanguagePythonMIT LicenseMIT

Stargazers