Retrain-less quantization for multiplier-less convolutional neural networks곱셈기 없는 CNN 모델 구현을 위한 재훈련 없는 양자화 기법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 144
  • Download : 0
This paper presents an approximate signed digit representation (ASD) which quantizes the weights of convolutional neural networks (CNNs) in order to make multiplier-less CNNs without performing any retraining process. Unlike the existing methods that necessitate retraining for weight quantization, the proposed method directly converts full-precision weights of CNN models into low-precision ones, attaining accuracy comparable to that of full-precision models on the Image classification tasks without going through retraining. Therefore, it is effective in saving the retraining time as well as the related computational cost. As the proposed method simplifies the weights to have up to two non-zero digits, multiplication can be realized with only add and shift operations, resulting in a speed-up of inference time and a reduction of energy consumption and hardware complexity. Experiments conducted for famous CNN architectures, such as AlexNet, VGG-16, ResNet-18 and SqueezeNet, show that the proposed method reduces the model size by 73% at the cost of a little increase of error rate, which ranges from 0.09% to 1.5% on ImageNet dataset. Compared to the previous architecture built with multipliers, the proposed multiplier-less convolution architecture reduces the critical-path delay by 52% and mitigates the hardware complexity and power consumption by more than 50%. In addition, we briefly present new floating-point quantization methods, which quantizes all parameters and feature maps of CNN models.
Advisors
박인철researcherPark, In-Cheolresearcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2020
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.2,[iii, 39 p. :]

Keywords

Convolutional neural networks (CNNs)▼aconvolution architecture▼amultiplier-less structure▼acanonical signed digit▼aweight quantization▼aretraining; 컨볼루션 뉴럴 네트워크▼a컨볼루션 구조▼a곱셈기 없는 구조▼acanonical signed digit▼a파라미터 양자화▼a재훈련

URI
http://hdl.handle.net/10203/284786
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=911416&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0