Pinned Repositories
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
api_toolkits
Collie
fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
fairseq-moe
gpt-moe
An MoE training framework based on gpt-neox and deepspeed
llm_shearing
clone from https://github.com/princeton-nlp/LLM-Shearing
opencompass
patched_deepspeed
ZhiYuanZeng.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
ZhiYuanZeng's Repositories
ZhiYuanZeng/fairseq-moe
ZhiYuanZeng/api_toolkits
ZhiYuanZeng/Collie
ZhiYuanZeng/fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
ZhiYuanZeng/gpt-moe
An MoE training framework based on gpt-neox and deepspeed
ZhiYuanZeng/llm_shearing
clone from https://github.com/princeton-nlp/LLM-Shearing
ZhiYuanZeng/opencompass
ZhiYuanZeng/patched_deepspeed
ZhiYuanZeng/ZhiYuanZeng.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes