DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | 박인철 | - |
dc.contributor.advisor | Park, In-Cheol | - |
dc.contributor.author | Choi, Jaewoong | - |
dc.date.accessioned | 2021-05-13T19:34:29Z | - |
dc.date.available | 2021-05-13T19:34:29Z | - |
dc.date.issued | 2020 | - |
dc.identifier.uri | http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=911416&flag=dissertation | en_US |
dc.identifier.uri | http://hdl.handle.net/10203/284786 | - |
dc.description | 학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.2,[iii, 39 p. :] | - |
dc.description.abstract | This paper presents an approximate signed digit representation (ASD) which quantizes the weights of convolutional neural networks (CNNs) in order to make multiplier-less CNNs without performing any retraining process. Unlike the existing methods that necessitate retraining for weight quantization, the proposed method directly converts full-precision weights of CNN models into low-precision ones, attaining accuracy comparable to that of full-precision models on the Image classification tasks without going through retraining. Therefore, it is effective in saving the retraining time as well as the related computational cost. As the proposed method simplifies the weights to have up to two non-zero digits, multiplication can be realized with only add and shift operations, resulting in a speed-up of inference time and a reduction of energy consumption and hardware complexity. Experiments conducted for famous CNN architectures, such as AlexNet, VGG-16, ResNet-18 and SqueezeNet, show that the proposed method reduces the model size by 73% at the cost of a little increase of error rate, which ranges from 0.09% to 1.5% on ImageNet dataset. Compared to the previous architecture built with multipliers, the proposed multiplier-less convolution architecture reduces the critical-path delay by 52% and mitigates the hardware complexity and power consumption by more than 50%. In addition, we briefly present new floating-point quantization methods, which quantizes all parameters and feature maps of CNN models. | - |
dc.language | eng | - |
dc.publisher | 한국과학기술원 | - |
dc.subject | Convolutional neural networks (CNNs)▼aconvolution architecture▼amultiplier-less structure▼acanonical signed digit▼aweight quantization▼aretraining | - |
dc.subject | 컨볼루션 뉴럴 네트워크▼a컨볼루션 구조▼a곱셈기 없는 구조▼acanonical signed digit▼a파라미터 양자화▼a재훈련 | - |
dc.title | Retrain-less quantization for multiplier-less convolutional neural networks | - |
dc.title.alternative | 곱셈기 없는 CNN 모델 구현을 위한 재훈련 없는 양자화 기법 | - |
dc.type | Thesis(Master) | - |
dc.identifier.CNRN | 325007 | - |
dc.description.department | 한국과학기술원 :전기및전자공학부, | - |
dc.contributor.alternativeauthor | 최재웅 | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.