Journal of Big Data (Feb 2019)

A parallel and distributed stochastic gradient descent implementation using commodity clusters

  • Robert K. L. Kennedy,
  • Taghi M. Khoshgoftaar,
  • Flavio Villanustre,
  • Timothy Humphrey

DOI
https://doi.org/10.1186/s40537-019-0179-2
Journal volume & issue
Vol. 6, no. 1
pp. 1 – 23

Abstract

Read online

Abstract Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerating the training process of Deep Learning using cluster computers faces many challenges ranging from distributed optimizers to the large communication overhead specific to systems with off the shelf networking components. In this paper, we present a novel distributed and parallel implementation of stochastic gradient descent (SGD) on a distributed cluster of commodity computers. We use high-performance computing cluster (HPCC) systems as the underlying cluster environment for the implementation. We overview how the HPCC systems platform provides the environment for distributed and parallel Deep Learning, how it provides a facility to work with third party open source libraries such as TensorFlow, and detail our use of third-party libraries and HPCC functionality for implementation. We provide experimental results that validate our work and show that our implementation can scale with respect to both dataset size and the number of compute nodes in the cluster.

Keywords