davidmrau/mixture-of-experts
PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538
PythonGPL-3.0
Stargazers
- abursucvaleo.ai
- AlexandruBurlacuChisinau, Republic of Moldova
- apeterswuChina Beijing
- arturbegReya Labs
- asifrNew York
- bionicles@BitPharma
- caoshijie0501
- chenghuigebeijing
- ChriskueiInstitute of Computing Technology, Chinese Academy of Sciences
- davidmrauUniversity of Amsterdam
- dpoulopoulosHewlett Packard Enterprise
- fly51flyPRIS
- GabrielLin
- gegallego@mt-upc
- gemireZhejiang University
- Hakbin
- hwenjun18
- imirzadehUnited States
- infynite
- jerusalemcShanghai Jiao Tong University
- Jiahuan-PeiVU Amsterdam
- KaleidophonCopenhagen
- kiminh
- Kulbear@AltaML
- lxueaaClear Water Bay, NT, Hong Kong
- mnmjh1215
- prm10
- razhangweiFacebook
- StillerPatrickHelmholtz AI
- Usernamezhx
- vedant
- wanghailiang
- XingLuxiBeijing, China
- yiskw713
- zhaohengyuan1Singapore
- zzachwUIUC