Quantized neural network training and inference in practical scenarios실용적인 시나리오에서의 인공신경망 학습과 추론의 양자화

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 201
  • Download : 0
Neural Network Quantization aims to reduce the size and computational complexity of a neural network for more efficient training and inference of neural networks. However, existing methods often render themselves impractical in real-world scenarios, such as On-device Federated Learning, and with compact models such as MobileNet. In this paper, We show that applying Neural Network quantization in these scenarios are difficult. In On-device Federated Learning scenarios, many diverse devices with different hardware constraints can participate in the same Federated Learning, which leads to degenerate performance in the high-performance devices. With compact models with less redundancies in the weights, it is much more difficult to find quantized weights that do not incur a drop in the model's accuracy. We tackle these challenges by introducing two novel methods for practical neural network quantization: Bitwidth Heterogeneous Federated Learning with Progressive Weight Dequantizer, and Neural Network Binarization with Task-dependent Aggregated Transform. We show that these methods are effective at applying Neural Network quantization in the aforementioned practical scenarios.
Advisors
Hwang, Sung Juresearcher황성주researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.2,[iv, 36 p. :]

Keywords

Neural network quantization▼aEfficient neural network training▼aEfficient neural network inference▼aFederated learning▼aBitwidth-heterogeneous federated learning; 인공신경망 양자화▼a효율적인 인공신경망 학습▼a효율적인 인공신경망 추론▼a연합학습▼a이종 정밀도 연합학습

URI
http://hdl.handle.net/10203/308234
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1032321&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0