TheSequence

TheSequence

Share this post

TheSequence
TheSequence
🕸 Edge#185: Centralized vs. Decentralized Distributed Training Architectures

🕸 Edge#185: Centralized vs. Decentralized Distributed Training Architectures

Apr 26, 2022
∙ Paid
5

Share this post

TheSequence
TheSequence
🕸 Edge#185: Centralized vs. Decentralized Distributed Training Architectures
Share

In this issue: 

  • we overview Centralized vs. Decentralized Distributed Training Architectures; 

  • we explain GPipe, an Architecture for Training Large Scale Neural Networks;  

  • we explore TorchElastic, a Distributed Training Framework for PyTorch.  

Enjoy the learning!  

💡 ML Concept of the Day: Centralized vs. Decentralized Distributed Training Architectures  

In Edge#183, we discussed data and model parallelism as a fundamental taxonomy to classify distributed training techniques. Both data and model parallelism rely on partitioning tasks across different nodes that need to coordinate updates with each other. This mechanic is the essence of other criteria to classify distributed training architectures: centralized vs. decentralized training.  

The notion of centralization in distributed training has to do with the parameters of a deep neural network. In a centralized training architecture,

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Jesus Rodriguez
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share