3.8 Proceedings Paper

Communication Algorithm-Architecture Co-Design for Distributed Deep Learning

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/ISCA52012.2021.00023

Keywords

distributed deep learning; data-parallel training; all-reduce; interconnection network; algorithm-architecture co-design

Funding

  1. TAMU Dissertation Fellowship

Ask authors/readers for more resources

The study introduces an efficient all-reduce algorithm, MULTITREE, which achieves efficient and scalable communication operations under different interconnect topologies. Through the co-design of algorithm and architecture, it reduces communication time and training time effectively.
Large-scale distributed deep learning training has enabled developments of more complex deep neural network models to learn from larger datasets for sophisticated tasks. In particular, distributed stochastic gradient descent intensively invokes all-reduce operations for gradient update, which dominates communication time during iterative training epochs. In this work, we identify the inefficiency in widely used all-reduce algorithms, and the opportunity of algorithm-architecture co-design. We propose MULTITREE all-reduce algorithm with topology and resource utilization awareness for efficient and scalable all-reduce operations, which is applicable to different interconnect topologies. Moreover, we co-design the network interface to schedule and coordinate the all-reduce messages for contention-free communications, working in synergy with the algorithm. The flow control is also simplified to exploit the bulk data transfer of big gradient exchange. We evaluate the co-design using different all-reduce data sizes for synthetic study, demonstrating its effectiveness on various interconnection network topologies, in addition to state-of-the-art deep neural networks for real workload experiments. The results show that MULTITREE achieves 2.3x and 1.56x communication speedup, as well as up to 81% and 30% training time reduction compared to ring all-reduce and state-of-the-art approaches, respectively.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available