Pinned Repositories
horovod
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
ColossalAI
Making large AI models cheaper, faster and more accessible
accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
training_results_v1.1