4-bit quantized training of deep neural networks심층 신경망의 4-bit 양자화 학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 124
  • Download : 0
Deep Neural Network(DNN) has been widely used to build A.I. in various fields such as computer vision and natural language process and outperformed other classical techniques. However, its massive computation and memory requirement hinder the implementation of DNNs on edge devices that have low computing power and smaller memory. Quantization of DNN is a network compression technique that reduces the precision of data involved only in the inference, or in both the inference and training. The quantized inference has been studied by lots of researchers, but the quantized training has had little focus and nothing has dealt with the utilization of only 4 or fewer-bit for all data objects in the training. Here I propose a novel quantized training algorithm GSNet (Gradient Stabilizer Network), that utilizes only 4 or fewer-bit data using a 4-bit additional object `stabilizer'. The stabilizer securely accumulates 2-bit gradients which are ternarized using a novel gradient scaling method. GSNet was evaluated in the image classification task and showed competitive performance compared with other quantization techniques that utilize even higher bit-precision.
Advisors
Moon, Jaekyunresearcher문재균researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2021
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2021.2,[III, 23p :]

Keywords

Deep learning▼aOn-device machine learninig▼aNetwork compression▼aQuantized DNN▼aLow-precision training; 심층학습; 온디바이스 기계학습▼a신경망 압축▼a신경망 양자화▼a저정밀도 학습

URI
http://hdl.handle.net/10203/295984
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=948995&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0