distributed-deep-learning
There are 37 repositories under distributed-deep-learning topic.
intel/BigDL
BigDL: Distributed TensorFlow, Keras and PyTorch on Apache Spark/Flink & Ray
dkeras-project/dkeras
Distributed Keras Engine, Make Keras faster with only one line of code.
dyadxmachina/Applied-Deep-Learning-with-TensorFlow
Learn applied deep learning from zero to deployment using TensorFlow 1.8+
zoranzhao/DeepThings
A Portable C Library for Distributed CNN Inference on IoT Edge Clusters
GuanhuaWang/sensAI
sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data
ParCIS/Chimera
Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.
vdutts7/dnn-distributed
Distributed training of DNNs • C++/MPI Proxies (GPT-2, GPT-3, CosmoFlow, DLRM)
rocketmlhq/rmldnn
RocketML Deep Neural Networks
intel/e2eAIOK
Intel® End-to-End AI Optimization Kit
rkhan055/SHADE
SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training
ParCIS/Ok-Topk
Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k communication volume which is asymptotically optimal) with the decentralized parallel Stochastic Gradient Descent (SGD) optimizer, and its convergence is proved theoretically and empirically.
gsyang33/Driple
🚨 Prediction of the Resource Consumption of Distributed Deep Learning Systems
christianramsey/Tensorflow-for-Distributed-Deep-Learning
TensorFlow (1.8+) Datasets, Feature Columns, Estimators and Distributed Training using Google Cloud Machine Learning Engine
ravenprotocol/ravnest
Decentralized Asynchronous Training on Heterogeneous Devices
Shigangli/eager-SGD
Eager-SGD is a decentralized asynchronous SGD. It utilizes novel partial collectives operations to accumulate the gradients across all the processes.
ray-project/anyscale-workshop-nyc-2023
Scalable NLP model fine-tuning and batch inference with Ray and Anyscale
Shigangli/WAGMA-SGD
WAGMA-SGD is a decentralized asynchronous SGD based on wait-avoiding group model averaging. The synchronization is relaxed by making the collectives externally-triggerable, namely, a collective can be initiated without requiring that all the processes enter it. It partially reduces the data within non-overlapping groups of process, improving the parallel scalability.
deepspark/deepspark_java
Java based Convolutional Neural Network package running on Apache Spark framework
amirhosein-mesbah/Deep_Learning
This repository contains the implementation of a wide variety of Deep Learning Projects in different applications of computer vision, NLP, federated, and distributed learning. These projects include university projects and projects implemented due to interest in Deep Learning.
lancelee82/necklace
Distributed deep learning framework based on pytorch/numba/nccl and zeromq.
StefanoFioravanzo/distributed-deeplearning-kubernetes
Collection of resources for automatic deployment of distributed deep learning jobs on a Kubernetes cluster
AmrMKayid/KayDDRL
Distributed Deep Reinforcement Learning for Large Scale Robotic Simulations 👨💻🤖🕸🕹🕷❤️👨🔬
explcre/SHUKUN-Technology-AlgorithmIntern-MultiNodeTraining-for-DLmodels-Horovod-ConfigurationTutorial-Perf
SHUKUN Technology Co.,Ltd Algorithm intern (2020/12-2021/5). Multi-GPU, Multi-node training for deep learning models. Horovod, NVIDIA clara train sdk, configuration tutorial,performance testing.
trilliwon/pytorch-examples
PyTorch Examples for Beginners
veritas9872/Horovod-Pytorch-Tutorial
Horovod Tutorial for Pytorch using NVIDIA-Docker.
hkvision/analytics-zoo
Distributed Tensorflow, Keras and BigDL on Apache Spark
mma735/TFM-DS
Comparison of distributed machine learning techniques applied to openly available datasets
siddhanthiyer-99/Distributed-Training-of-GANs
Implemented training strategies to help improve bottlenecks and to improve the training speed while maintaining the quality of our GANs.
sqaz91819/Blockchain-NAS
A blockchain based neural architecture search project.
thanoskaravangelis/distributed-deep-learning-ntua
Distributed Deep Learning experiments with the BigDL framework over Databricks
bilalsp/yelp-distributed-DL
Yelp review classification using CNN model with horovod on HPC cluster
ch3njust1n/smpl
Simultaneous Multi-Party Learning Framework
hyunnnchoi/google-t5-fsdp-kubeflow
A foundational repository for setting up distributed training jobs using Kubeflow and PyTorch FSDP.
pierric/Mnist-Caffe-MPI
mnist, using caffe and openmpi
smmehrab/distributed-deep-learning
Distributed Deep Learning
sotheanithsok/Image-Recognition-using-Distributed-ResNet-Model
An implementation of a distributed ResNet model for classifying CIFAR-10 and MNIST datasets.