A parallel and distributed stochastic gradient descent implementation using commodity clusters

Abstract Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerati...

Full description

Bibliographic Details
Main Authors: Robert K. L. Kennedy, Taghi M. Khoshgoftaar, Flavio Villanustre, Timothy Humphrey
Format: Article
Language:English
Published: SpringerOpen 2019-02-01
Series:Journal of Big Data
Subjects:
Online Access:http://link.springer.com/article/10.1186/s40537-019-0179-2
_version_ 1819265818914979840
author Robert K. L. Kennedy
Taghi M. Khoshgoftaar
Flavio Villanustre
Timothy Humphrey
author_facet Robert K. L. Kennedy
Taghi M. Khoshgoftaar
Flavio Villanustre
Timothy Humphrey
author_sort Robert K. L. Kennedy
collection DOAJ
description Abstract Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerating the training process of Deep Learning using cluster computers faces many challenges ranging from distributed optimizers to the large communication overhead specific to systems with off the shelf networking components. In this paper, we present a novel distributed and parallel implementation of stochastic gradient descent (SGD) on a distributed cluster of commodity computers. We use high-performance computing cluster (HPCC) systems as the underlying cluster environment for the implementation. We overview how the HPCC systems platform provides the environment for distributed and parallel Deep Learning, how it provides a facility to work with third party open source libraries such as TensorFlow, and detail our use of third-party libraries and HPCC functionality for implementation. We provide experimental results that validate our work and show that our implementation can scale with respect to both dataset size and the number of compute nodes in the cluster.
first_indexed 2024-12-23T20:51:26Z
format Article
id doaj.art-e0f68ce3a7914c8b8832fe995e7ab6ac
institution Directory Open Access Journal
issn 2196-1115
language English
last_indexed 2024-12-23T20:51:26Z
publishDate 2019-02-01
publisher SpringerOpen
record_format Article
series Journal of Big Data
spelling doaj.art-e0f68ce3a7914c8b8832fe995e7ab6ac2022-12-21T17:31:40ZengSpringerOpenJournal of Big Data2196-11152019-02-016112310.1186/s40537-019-0179-2A parallel and distributed stochastic gradient descent implementation using commodity clustersRobert K. L. Kennedy0Taghi M. Khoshgoftaar1Flavio Villanustre2Timothy Humphrey3Florida Atlantic UniversityFlorida Atlantic UniversityLexisNexis Business Information SolutionsLexisNexis Business Information SolutionsAbstract Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerating the training process of Deep Learning using cluster computers faces many challenges ranging from distributed optimizers to the large communication overhead specific to systems with off the shelf networking components. In this paper, we present a novel distributed and parallel implementation of stochastic gradient descent (SGD) on a distributed cluster of commodity computers. We use high-performance computing cluster (HPCC) systems as the underlying cluster environment for the implementation. We overview how the HPCC systems platform provides the environment for distributed and parallel Deep Learning, how it provides a facility to work with third party open source libraries such as TensorFlow, and detail our use of third-party libraries and HPCC functionality for implementation. We provide experimental results that validate our work and show that our implementation can scale with respect to both dataset size and the number of compute nodes in the cluster.http://link.springer.com/article/10.1186/s40537-019-0179-2Parallel stochastic gradient descentParallel and distributed processingDeep learningBig dataNeural networkCluster computer
spellingShingle Robert K. L. Kennedy
Taghi M. Khoshgoftaar
Flavio Villanustre
Timothy Humphrey
A parallel and distributed stochastic gradient descent implementation using commodity clusters
Journal of Big Data
Parallel stochastic gradient descent
Parallel and distributed processing
Deep learning
Big data
Neural network
Cluster computer
title A parallel and distributed stochastic gradient descent implementation using commodity clusters
title_full A parallel and distributed stochastic gradient descent implementation using commodity clusters
title_fullStr A parallel and distributed stochastic gradient descent implementation using commodity clusters
title_full_unstemmed A parallel and distributed stochastic gradient descent implementation using commodity clusters
title_short A parallel and distributed stochastic gradient descent implementation using commodity clusters
title_sort parallel and distributed stochastic gradient descent implementation using commodity clusters
topic Parallel stochastic gradient descent
Parallel and distributed processing
Deep learning
Big data
Neural network
Cluster computer
url http://link.springer.com/article/10.1186/s40537-019-0179-2
work_keys_str_mv AT robertklkennedy aparallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT taghimkhoshgoftaar aparallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT flaviovillanustre aparallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT timothyhumphrey aparallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT robertklkennedy parallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT taghimkhoshgoftaar parallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT flaviovillanustre parallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters
AT timothyhumphrey parallelanddistributedstochasticgradientdescentimplementationusingcommodityclusters