Cluster-Promoting Quantization with Bit-Drop for Minimizing Network Quantization Loss

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 119
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorLee, Jung Hyunko
dc.contributor.authorYun, Jihunko
dc.contributor.authorHwang, Sung Juko
dc.contributor.authorYang, Eunhoko
dc.date.accessioned2021-12-14T06:51:35Z-
dc.date.available2021-12-14T06:51:35Z-
dc.date.created2021-12-03-
dc.date.created2021-12-03-
dc.date.created2021-12-03-
dc.date.created2021-12-03-
dc.date.issued2021-10-14-
dc.identifier.citation18th IEEE/CVF International Conference on Computer Vision (ICCV), pp.5370 - 5379-
dc.identifier.urihttp://hdl.handle.net/10203/290618-
dc.description.abstractNetwork quantization, which aims to reduce the bit-lengths of the network weights and activations, has emerged for their deployments to resource-limited devices. Although recent studies have successfully discretized a full-precision network, they still incur large quantization errors after training, thus giving rise to a significant performance gap between a full-precision network and its quantized counterpart. In this work, we propose a novel quantization method for neural networks, Cluster-Promoting Quantization (CPQ) that finds the optimal quantization grids while naturally encouraging the underlying full-precision weights to gather around those quantization grids cohesively during training. This property of CPQ is thanks to our two main ingredients that enable differentiable quantization: i) the use of the categorical distribution designed by a specific probabilistic parametrization in the forward pass and ii) our proposed multi-class straight-through estimator (STE) in the backward pass. Since our second component, multi-class STE, is intrinsically biased, we additionally propose a new bit-drop technique, DropBits, that revises the standard dropout regularization to randomly drop bits instead of neurons. As a natural extension of DropBits, we further introduce the way of learning heterogeneous quantization levels to find proper bit-length for each layer by imposing an additional regularization on DropBits. We experimentally validate our method on various benchmark datasets and network architectures, and also support a new hypothesis for quantization: learning heterogeneous quantization levels outperforms the case using the same but fixed quantization levels from scratch.-
dc.languageEnglish-
dc.publisherComputer Vision Foundation, IEEE Computer Society-
dc.titleCluster-Promoting Quantization with Bit-Drop for Minimizing Network Quantization Loss-
dc.typeConference-
dc.identifier.wosid000797698905057-
dc.identifier.scopusid2-s2.0-85127806375-
dc.type.rimsCONF-
dc.citation.beginningpage5370-
dc.citation.endingpage5379-
dc.citation.publicationname18th IEEE/CVF International Conference on Computer Vision (ICCV)-
dc.identifier.conferencecountryCN-
dc.identifier.conferencelocationVirtual-
dc.identifier.doi10.1109/ICCV48922.2021.00532-
dc.contributor.localauthorHwang, Sung Ju-
dc.contributor.localauthorYang, Eunho-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0