Applied Sciences (Dec 2020)

Communication Optimization Schemes for Accelerating Distributed Deep Learning Systems

  • Jaehwan Lee,
  • Hyeonseong Choi,
  • Hyeonwoo Jeong,
  • Baekhyeon Noh,
  • Ji Sun Shin

DOI
https://doi.org/10.3390/app10248846
Journal volume & issue
Vol. 10, no. 24
p. 8846

Abstract

Read online

In a distributed deep learning system, a parameter server and workers must communicate to exchange gradients and parameters, and the communication cost increases as the number of workers increases. This paper presents a communication data optimization scheme to mitigate the decrease in throughput due to communication performance bottlenecks in distributed deep learning. To optimize communication, we propose two methods. The first is a layer dropping scheme to reduce communication data. The layer dropping scheme we propose compares the representative values of each hidden layer with a threshold value. Furthermore, to guarantee the training accuracy, we store the gradients that are not transmitted to the parameter server in the worker’s local cache. When the value of gradients stored in the worker’s local cache is greater than the threshold, the gradients stored in the worker’s local cache are transmitted to the parameter server. The second is an efficient threshold selection method. Our threshold selection method computes the threshold by replacing the gradients with the L1 norm of each hidden layer. Our data optimization scheme reduces the communication time by about 81% and the total training time by about 70% in a 56 Gbit network environment.

Keywords