Performance Modeling for Distributed Training of Convolutional Neural Networks

Published: 01 Jan 2021, Last Modified: 15 May 2025PDP 2021EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: We perform a theoretical analysis comparing the scalability of data versus model parallelism, applied to the distributed training of deep convolutional neural networks (CNNs), along five axes: batch size, node (floating-point) arithmetic performance, node memory bandwidth, network link bandwidth, and cluster dimension. Our study relies on analytical performance models that can be configured to reproduce the components and organization of the CNN model as well as the hardware configuration of the target distributed platform. In addition, we provide evidence of the accuracy of the analytical models by performing a validation against a Python library for distributed deep learning training.
Loading