Bin-Specific Quantization in Spectral-Domain Convolutional Neural Network Accelerators

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 45
  • Download : 0
Spectral-domain convolution engines can effectively reduce the computational complexity of convolution operations. In these engines, however, element-wise multiplications of the spectral representations dominate the multiply and accumulate (MAC) operations. In light of this, we propose bin-specific quantization (BSQ), which is to judiciously allocate varying bit width to each spectral bin in overlap-save. This allows efficient computation of the Hadamard product since the magnitude of the high-frequency components in image features is significantly smaller than that of the low-frequency counterparts. Using the statistics from spectral representations of feature maps, we also delineate methods for properly allocating bit precision to those spectral bins. When BSQ is applied, the average bit precisions of the arithmetic operators in spectral-domain convolvers, without the requirement of network re-training, were lowered by 24 % (AlexNet), 20% (VGG-16), and 22% (ResNet-18) while having no significant reduction (< 1%) on classification accuracy on the ImageNet dataset.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2022-06
Language
English
Citation

4th IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022, pp.407 - 410

DOI
10.1109/AICAS54282.2022.9869971
URI
http://hdl.handle.net/10203/312726
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0