Retrain-less quantization for multiplier-less convolutional neural networks곱셈기 없는 CNN 모델 구현을 위한 재훈련 없는 양자화 기법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 145
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor박인철-
dc.contributor.advisorPark, In-Cheol-
dc.contributor.authorChoi, Jaewoong-
dc.date.accessioned2021-05-13T19:34:29Z-
dc.date.available2021-05-13T19:34:29Z-
dc.date.issued2020-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=911416&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/284786-
dc.description학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.2,[iii, 39 p. :]-
dc.description.abstractThis paper presents an approximate signed digit representation (ASD) which quantizes the weights of convolutional neural networks (CNNs) in order to make multiplier-less CNNs without performing any retraining process. Unlike the existing methods that necessitate retraining for weight quantization, the proposed method directly converts full-precision weights of CNN models into low-precision ones, attaining accuracy comparable to that of full-precision models on the Image classification tasks without going through retraining. Therefore, it is effective in saving the retraining time as well as the related computational cost. As the proposed method simplifies the weights to have up to two non-zero digits, multiplication can be realized with only add and shift operations, resulting in a speed-up of inference time and a reduction of energy consumption and hardware complexity. Experiments conducted for famous CNN architectures, such as AlexNet, VGG-16, ResNet-18 and SqueezeNet, show that the proposed method reduces the model size by 73% at the cost of a little increase of error rate, which ranges from 0.09% to 1.5% on ImageNet dataset. Compared to the previous architecture built with multipliers, the proposed multiplier-less convolution architecture reduces the critical-path delay by 52% and mitigates the hardware complexity and power consumption by more than 50%. In addition, we briefly present new floating-point quantization methods, which quantizes all parameters and feature maps of CNN models.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectConvolutional neural networks (CNNs)▼aconvolution architecture▼amultiplier-less structure▼acanonical signed digit▼aweight quantization▼aretraining-
dc.subject컨볼루션 뉴럴 네트워크▼a컨볼루션 구조▼a곱셈기 없는 구조▼acanonical signed digit▼a파라미터 양자화▼a재훈련-
dc.titleRetrain-less quantization for multiplier-less convolutional neural networks-
dc.title.alternative곱셈기 없는 CNN 모델 구현을 위한 재훈련 없는 양자화 기법-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthor최재웅-
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0