SkyworkAI/Skywork-MoE
Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models
Stargazers
- Ai-trainee
- BBufSkyWork
- BeyonderXXFudan University
- boscotba
- CaoWGGnanjing
- chengtbfSkywork
- chuanmingliuWesteros
- cygwynd
- Davidciw
- EatenBagpipeBeijing, China
- EdTechoolsShanxi Normal University
- GanjinZeroDAMO Academy
- GarrickLinHangZhou
- gm8xx8
- gtromboni98Sapienza University
- gyunggyung@forus-ai
- imisszxq
- Josephrp@Tonic-AI
- jpthu17Peking University
- jwwangchnShenzhen
- kugwzk
- l-dawei
- MarcusOS1
- peterjc123Shanghai, China
- RunxinXuPeking University
- SamaelChen
- Sosky-Hwang
- SparkJiaoNTU-NLP & I2R, A*STAR, Singapore
- Spico197Soochow University
- truskovskiyk@kyryl-opens-ml
- truthless11CoAI Group, Tsinghua University
- Upcreat
- utopic-dev
- xiaomin-DTHU
- xiaox1311
- zhao1iang@lancopku