A Channel Pruning Optimization With Layer-Wise Sensitivity in a Single-Shot Manner Under Computational Constraints

Cited 1 time in webofscience Cited 0 time in scopus
  • Hit : 100
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorJeon, Minsuko
dc.contributor.authorKim, TaeWooko
dc.contributor.authorLee, Changhako
dc.contributor.authorYoun, Chan-Hyunko
dc.date.accessioned2023-03-06T06:00:38Z-
dc.date.available2023-03-06T06:00:38Z-
dc.date.created2023-03-06-
dc.date.created2023-03-06-
dc.date.issued2023-
dc.identifier.citationIEEE ACCESS, v.11, pp.7043 - 7055-
dc.identifier.issn2169-3536-
dc.identifier.urihttp://hdl.handle.net/10203/305472-
dc.description.abstractIn the constrained computing environments such as mobile device or satellite on-board system, various computational factors of hardware resource can restrict the processing of deep learning (DL) services. Recent DL models such as satellite image analysis mainly require larger resource memory occupation for intermediate feature map footprint than the given memory specification of hardware resource and larger computational overhead (in FLOP) to meet service-level objective in the sense of hardware accelerator. As one of the solutions, we propose a new method of controlling the layer-wise channel pruning in a single-shot manner that can decide how much channels to prune in each layer by observing dataset once without full pretraining. To improve the robustness of the performance degradation, we also propose a layer-wise sensitivity and formulate the optimization problems for deciding layer-wise pruning ratio under target computational constraints. In the paper, the optimal conditions are theoretically derived, and the practical optimum searching schemes are proposed using the optimal conditions. On the empirical evaluation, the proposed methods show robustness on performance degradation, and present feasibility on DL serving under constrained computing environments by reducing memory occupation, providing acceleration effect and throughput improvement while keeping the accuracy performance.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleA Channel Pruning Optimization With Layer-Wise Sensitivity in a Single-Shot Manner Under Computational Constraints-
dc.typeArticle-
dc.identifier.wosid000926019300001-
dc.identifier.scopusid2-s2.0-85146235424-
dc.type.rimsART-
dc.citation.volume11-
dc.citation.beginningpage7043-
dc.citation.endingpage7055-
dc.citation.publicationnameIEEE ACCESS-
dc.identifier.doi10.1109/ACCESS.2022.3232566-
dc.contributor.localauthorYoun, Chan-Hyun-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorComputational modeling-
dc.subject.keywordAuthorSensitivity-
dc.subject.keywordAuthorDegradation-
dc.subject.keywordAuthorMemory management-
dc.subject.keywordAuthorSatellites-
dc.subject.keywordAuthorAnalytical models-
dc.subject.keywordAuthorAdaptation models-
dc.subject.keywordAuthorSingle-shot pruning-
dc.subject.keywordAuthorchannel pruning-
dc.subject.keywordAuthorlottery ticket hypothesis-
dc.subject.keywordAuthorDL model compression-
dc.subject.keywordPlusNETWORKS-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 1 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0