MixedNet: Network Design Strategies for Cost-Effective Quantized CNNs

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 251
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorChang, Dong-Jinko
dc.contributor.authorNam, Byeong-Gyuko
dc.contributor.authorRyu, Seung-Takko
dc.date.accessioned2021-09-08T07:50:20Z-
dc.date.available2021-09-08T07:50:20Z-
dc.date.created2021-09-08-
dc.date.created2021-09-08-
dc.date.created2021-09-08-
dc.date.issued2021-
dc.identifier.citationIEEE ACCESS, v.9, pp.117554 - 117564-
dc.identifier.issn2169-3536-
dc.identifier.urihttp://hdl.handle.net/10203/287676-
dc.description.abstractThis paper proposes design strategies for a low-cost quantized neural network. To prevent the classification accuracy from being degraded by quantization, a structure-design strategy that utilizes a large number of channels rather than deep layers is proposed. In addition, a squeeze-and-excitation (SE) layer is adopted to enhance the performance of the quantized network. Through a quantitative analysis and simulations of the quantized key convolution layers of ResNet and MobileNets, a low-cost layer-design strategy for use when building a neural network is proposed. With this strategy, a low-cost network referred to as a MixedNet is constructed. A 4-bit quantized MixedNet example achieves an on-chip memory size reduction of 60% and fewer memory access by 53% with negligible classification accuracy degradation in comparison with conventional networks while also showing classification accuracy rates of approximately 73% for Cifar-100 and 93% for Cifar-10.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleMixedNet: Network Design Strategies for Cost-Effective Quantized CNNs-
dc.typeArticle-
dc.identifier.wosid000690439700001-
dc.identifier.scopusid2-s2.0-85113892311-
dc.type.rimsART-
dc.citation.volume9-
dc.citation.beginningpage117554-
dc.citation.endingpage117564-
dc.citation.publicationnameIEEE ACCESS-
dc.identifier.doi10.1109/ACCESS.2021.3106658-
dc.contributor.localauthorRyu, Seung-Tak-
dc.contributor.nonIdAuthorNam, Byeong-Gyu-
dc.description.isOpenAccessY-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorQuantization (signal)-
dc.subject.keywordAuthorConvolution-
dc.subject.keywordAuthorNetwork architecture-
dc.subject.keywordAuthorHardware-
dc.subject.keywordAuthorDegradation-
dc.subject.keywordAuthorConvolutional neural networks-
dc.subject.keywordAuthorSystem-on-chip-
dc.subject.keywordAuthorConvolutional neural network-
dc.subject.keywordAuthordeep neural network-
dc.subject.keywordAuthormemory access number-
dc.subject.keywordAuthormemory cost-
dc.subject.keywordAuthoron-chip memory size-
dc.subject.keywordAuthorquantized neural networks-
dc.subject.keywordPlusMEMORY-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0