An Overview of Sparsity Exploitation in CNNs for On-Device Intelligence With Software-Hardware Cross-Layer Optimizations

Cited 10 time in webofscience Cited 0 time in scopus
  • Hit : 240
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKang, Sanghoonko
dc.contributor.authorPark, Gwangtaeko
dc.contributor.authorKim, Sangjinko
dc.contributor.authorKim, Soyeonko
dc.contributor.authorHan, Donghyeonko
dc.contributor.authorYoo, Hoi-Junko
dc.date.accessioned2022-01-04T06:40:34Z-
dc.date.available2022-01-04T06:40:34Z-
dc.date.created2022-01-04-
dc.date.created2022-01-04-
dc.date.issued2021-12-
dc.identifier.citationIEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, v.11, no.4, pp.634 - 648-
dc.identifier.issn2156-3357-
dc.identifier.urihttp://hdl.handle.net/10203/291474-
dc.description.abstractThis paper presents a detailed overview of sparsity exploitation in deep neural network (DNN) accelerators. Despite the algorithmic advancements which drove DNNs to become the standard of artificial intelligence (AI), DNNs' computational and memory overhead limits the deployment of off-the-shelf models on edge devices. Numerous optimizations have been widely studied to efficiently run DNN models on performance and energy-limited mobile devices, from both software's and hardware's perspective. Sparsity exploitation one of the mainstream optimization techniques, whose objective is to achieve higher efficiency and speed through avoiding redundant multiply-and-accumulate (MAC) operations cause by zero operands. This paper overviews previous contributions on sparsity exploitation from both the software and hardware side of views, with the newly proposed taxonomy to categorize and analyze the works. On the software side, different sparsification algorithms are explained, including pruning and output speculation. From the hardware's perspective, advancements in architectures to efficiently handle sparse DNN computation are elaborated. The proposed taxonomy will help classify previous accelerators easily, by which sparsity they exploit and how. In addition, related works on sparse processing-in-memory (PIM) architectures and similarity exploitation are briefly introduced.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleAn Overview of Sparsity Exploitation in CNNs for On-Device Intelligence With Software-Hardware Cross-Layer Optimizations-
dc.typeArticle-
dc.identifier.wosid000730514000013-
dc.identifier.scopusid2-s2.0-85117790606-
dc.type.rimsART-
dc.citation.volume11-
dc.citation.issue4-
dc.citation.beginningpage634-
dc.citation.endingpage648-
dc.citation.publicationnameIEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS-
dc.identifier.doi10.1109/JETCAS.2021.3120417-
dc.contributor.localauthorYoo, Hoi-Jun-
dc.contributor.nonIdAuthorKim, Soyeon-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorCircuits and systems-
dc.subject.keywordAuthorTraining-
dc.subject.keywordAuthorTensors-
dc.subject.keywordAuthorHardware-
dc.subject.keywordAuthorComputer architecture-
dc.subject.keywordAuthorOptimization-
dc.subject.keywordAuthorConvolution-
dc.subject.keywordAuthorOn-device intelligence-
dc.subject.keywordAuthordeep neural network (DNN) processor-
dc.subject.keywordAuthorneural processing unit-
dc.subject.keywordAuthorsparsity exploitation-
dc.subject.keywordAuthorzero-skipping-
dc.subject.keywordAuthorsoftware-hardware co-design-
dc.subject.keywordPlusCONVOLUTIONAL NEURAL-NETWORK-
dc.subject.keywordPlusACCELERATOR-
dc.subject.keywordPlusENERGY-
dc.subject.keywordPlusPROCESSOR-
dc.subject.keywordPlusDESIGN-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 10 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0