IEEE Access (Jan 2022)

A Hybrid Parallelization Approach for Distributed and Scalable Deep Learning

  • Samson B. Akintoye,
  • Liangxiu Han,
  • Xin Zhang,
  • Haoming Chen,
  • Daoqiang Zhang

DOI
https://doi.org/10.1109/ACCESS.2022.3193690
Journal volume & issue
Vol. 10
pp. 77950 – 77961

Abstract

Read online

Recently, Deep Neural Networks (DNNs) have recorded significant success in handling medical and other complex classification tasks. However, as the sizes of DNN models and the available datasets increase, the training process becomes more complex and computationally intensive, usually taking longer to complete. In this work, we have proposed a generic full end-to-end hybrid parallelization approach combining model and data parallelism for efficiently distributed and scalable training of DNN models. We have also proposed a Genetic Algorithm Based Heuristic Resources Allocation (GABRA) mechanism for optimal distribution of partitions on the available GPUs for computing performance optimization. We have applied our proposed approach to a real use case based on 3D Residual Attention Deep Neural Network (3D-ResAttNet) for efficient Alzheimer Disease (AD) diagnosis on multiple GPUs and compared with the existing state-of-the-art parallel methods. The experimental evaluation shows that our proposed approach is 20% averagely better than existing parallel methods in terms of training time and achieves almost linear speedup with little or no differences in accuracy performance when compared with the existing non-parallel DNN models.

Keywords