Communication optimization for deep learning in distributed processing environments분산 처리 환경에서 딥 러닝을 위한 통신 기법 최적화

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 88
  • Download : 0
Machine learning has recently been in the spotlight as a solution to problems that were difficult to solve. Distributed processing techniques using graphic processing devices are widely used to deal with the vast amount of data needed to learn neural networks in deep learning, the most widely used type of machine learning, and thus collective communication within distributed systems exists as the main performance bottleneck and impairs the scalability of the system. In this dissertation, we would like to propose techniques and computer architectures that address communication bottlenecks and improve learning performance, considering the characteristics of the intermediate values of collective communication, the multilayer properties of deep learning, and the structure of distributed systems that perform actual computations.
Advisors
김동준researcherKim, Johnresearcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2021
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2021.2,[iii, 27 p. :]

Keywords

Collective Communication▼aDeep Learning▼aDistributed Processing▼aGraphics Processing Unit▼aInterconnect; 그래픽 처리 장치▼a딥 러닝▼a분산 처리▼a상호 연결▼a집단 통신

URI
http://hdl.handle.net/10203/295930
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=948980&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0