Pinned Repositories
LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
MAP-NEO
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
MINT-1T
MINT-1T: A one trillion token multimodal interleaved dataset.
L-hongbin's Repositories
L-hongbin/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
L-hongbin/MAP-NEO