Expediting split learning via intermediate activation budgeting strategy중간 계층 활성화 값 선별 전략을 통한 분할 학습 가속화

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 48
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorMoon, Jaekyun-
dc.contributor.advisor문재균-
dc.contributor.authorKousar, Humaira-
dc.date.accessioned2023-06-26T19:34:00Z-
dc.date.available2023-06-26T19:34:00Z-
dc.date.issued2022-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=997258&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/309900-
dc.description학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2022.2,[iv, 28 p. :]-
dc.description.abstractSplit learning (SL) is a recently introduced distributed machine learning technique that enables training of a deep neural network without accessing the massively produced data on the edge devices. SL has a great potential for resource-constraint devices as the network is divided into chunks for the clients and the server complying with their computation powers. While this is lucrative for the clients to work on the smaller model, the server is held accountable for processing the dominant share of the neural network. This computation burden at the server-side becomes onerous especially when a considerable number of clients participate in a training round. Also, the communication requirement from the split client-side network becomes paramount with more data and large network sizes. Additionally, the large data size at the clients may not be equally important for the network training. Data selection is an innate approach to reduce the computation burden and revamp the performance of a neural network. However, existing data selection approaches are limited in the context of SL as the neural network is decentralized, distributed and the client data is also private. This work is the first attempt to consider activation selection in the framework of split neural networks such as SL. The proposed technique works by selecting uncertain activations generated at the client-side network in SL through a small auxiliary network trained on the client's data. The selected subset of these activations is then sent to the server for training the whole neural network. Extensive experimentation and empirical results corroborates that the proposed technique significantly reduces computation burden at the server end and the communication requirement between the server and the clients. Meanwhile, the low-level embeddings learned at the client-side with its local data contributes to better performance (+3\%) as compared to vanilla SL.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.titleExpediting split learning via intermediate activation budgeting strategy-
dc.title.alternative중간 계층 활성화 값 선별 전략을 통한 분할 학습 가속화-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthor쿠사르 후마이라-
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0