ELASTIC GOSSIP: DISTRIBUTING NEURAL NETWORK TRAINING USING GOSSIP-LIKE PROTOCOLS

dc.contributor.advisorOates, Tim
dc.contributor.authorPramod, Siddharth
dc.contributor.departmentComputer Science and Electrical Engineering
dc.contributor.programComputer Science
dc.date.accessioned2021-01-29T18:12:35Z
dc.date.available2021-01-29T18:12:35Z
dc.date.issued2018-01-01
dc.description.abstractDistributing Neural Network training is of particular interest for several reasons including scaling using computing clusters, training at data sources such as IOT devices and edge servers, utilizing underutilized resources across heterogeneous environments, and so on. Most contemporary approaches primarily address scaling using computing clusters and require high network bandwidth and frequent communication. This theses presents an overview of standard approaches to distribute training and proposes a novel technique involving pairwise-communication using Gossip-like protocols, called Elastic Gossip. This approach builds upon an existing technique known as Elastic Averaging SGD (EASGD), and is similar to another technique called Gossiping SGD which also uses Gossip-like protocols. Elastic Gossip is empirically evaluated against Gossiping SGD using the MNIST digit recognition and CIFAR-10 classification tasks, using commonly used Neural Network architectures spanning Multi-Layer Perceptrons (MLPs) and Convolutional Neural Networks (CNNs). It is found that Elastic Gossip, Gossiping SGD, and All-reduce SGD perform quite comparably, even though the latter entails a substantially higher communication cost. While Elastic Gossip performs better than Gossiping SGD in these experiments, it is possible that a more thorough search over hyper-parameter space, specific to a given application, may yield configurations of Gossiping SGD that work better than Elastic Gossip.
dc.formatapplication:pdf
dc.genretheses
dc.identifierdoi:10.13016/m2cl1x-yatq
dc.identifier.other11890
dc.identifier.urihttp://hdl.handle.net/11603/20724
dc.languageen
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.relation.ispartofUMBC Theses and Dissertations Collection
dc.relation.ispartofUMBC Graduate School Collection
dc.relation.ispartofUMBC Student Collection
dc.sourceOriginal File Name: Pramod_umbc_0434M_11890.pdf
dc.subjectDeep Learning
dc.subjectDistributed systems
dc.subjectDistributed training
dc.subjectGossip
dc.subjectNeural Networks
dc.titleELASTIC GOSSIP: DISTRIBUTING NEURAL NETWORK TRAINING USING GOSSIP-LIKE PROTOCOLS
dc.typeText
dcterms.accessRightsDistribution Rights granted to UMBC by the author.
dcterms.accessRightsAccess limited to the UMBC community. Item may possibly be obtained via Interlibrary Loan thorugh a local library, pending author/copyright holder's permission.
dcterms.accessRightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Pramod_umbc_0434M_11890.pdf
Size:
1.2 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
PramodSElastic_Open.pdf
Size:
44.35 KB
Format:
Adobe Portable Document Format
Description: