Quantized neural network training and inference in practical scenarios실용적인 시나리오에서의 인공신경망 학습과 추론의 양자화

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 204
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorHwang, Sung Ju-
dc.contributor.advisor황성주-
dc.contributor.authorPark, Geon-
dc.date.accessioned2023-06-22T19:31:30Z-
dc.date.available2023-06-22T19:31:30Z-
dc.date.issued2023-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1032321&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/308234-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.2,[iv, 36 p. :]-
dc.description.abstractNeural Network Quantization aims to reduce the size and computational complexity of a neural network for more efficient training and inference of neural networks. However, existing methods often render themselves impractical in real-world scenarios, such as On-device Federated Learning, and with compact models such as MobileNet. In this paper, We show that applying Neural Network quantization in these scenarios are difficult. In On-device Federated Learning scenarios, many diverse devices with different hardware constraints can participate in the same Federated Learning, which leads to degenerate performance in the high-performance devices. With compact models with less redundancies in the weights, it is much more difficult to find quantized weights that do not incur a drop in the model's accuracy. We tackle these challenges by introducing two novel methods for practical neural network quantization: Bitwidth Heterogeneous Federated Learning with Progressive Weight Dequantizer, and Neural Network Binarization with Task-dependent Aggregated Transform. We show that these methods are effective at applying Neural Network quantization in the aforementioned practical scenarios.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectNeural network quantization▼aEfficient neural network training▼aEfficient neural network inference▼aFederated learning▼aBitwidth-heterogeneous federated learning-
dc.subject인공신경망 양자화▼a효율적인 인공신경망 학습▼a효율적인 인공신경망 추론▼a연합학습▼a이종 정밀도 연합학습-
dc.titleQuantized neural network training and inference in practical scenarios-
dc.title.alternative실용적인 시나리오에서의 인공신경망 학습과 추론의 양자화-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthor박건-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0