Accelerating Randomly Projected Gradient with Variance Reduction

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 120
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKim, SeongYoonko
dc.contributor.authorYun, SeYoungko
dc.date.accessioned2021-11-04T06:49:35Z-
dc.date.available2021-11-04T06:49:35Z-
dc.date.created2021-10-19-
dc.date.issued2020-02-
dc.identifier.citationIEEE International Conference on Big Data and Smart Computing (BigComp), pp.531 - 534-
dc.identifier.issn2375-933X-
dc.identifier.urihttp://hdl.handle.net/10203/288836-
dc.description.abstractParallel training methods of stochastic gradient descent (SGD) for deep learning have attracted a huge attention because of excellent scalability properties. In the parallel training methods, workers and a server exchange their gradient vectors (or parameter vectors) each other. However, the communications between workers and server take a lot of time since the dimension of the gradient vectors is extremely high. Transmission needs to cater for network bandwidth, so message compression is necessary. Although message compression algorithms have been proposed for the parallel training methods, there are concerns about the stability and the performance of the compressed training algorithms. hi general, the compressed training algorithms generate random message vectors, and the variance hinders the training. In this paper, we propose a novel compressed training algorithm that uses random projection with a variance reduction trick. Here, we introduce an average gradient vector that reduces the variance. We test the proposed compression method with AlexNet, ResNet20 models with CIFAR100 data. The proposed algorithm shows almost the same performance as the original SGD while uses 16 times compressed message.-
dc.languageEnglish-
dc.publisherIEEE-
dc.titleAccelerating Randomly Projected Gradient with Variance Reduction-
dc.typeConference-
dc.identifier.wosid000569987500097-
dc.identifier.scopusid2-s2.0-85084374585-
dc.type.rimsCONF-
dc.citation.beginningpage531-
dc.citation.endingpage534-
dc.citation.publicationnameIEEE International Conference on Big Data and Smart Computing (BigComp)-
dc.identifier.conferencecountryKO-
dc.identifier.conferencelocationBusan, SOUTH KOREA-
dc.identifier.doi10.1109/BigComp48618.2020.00-11-
dc.contributor.localauthorYun, SeYoung-
Appears in Collection
RIMS Conference Papers
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0