(A) heterogeneous computing-in-memory and neural-processing-unit architecture for an energy efficient floating-point DNN acceleration에너지 효율적인 부동 소수점 연산을 위한 메모리 내 컴퓨팅-심층 신경망 가속기 이기종 아키텍쳐

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 3
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor유회준-
dc.contributor.authorPark, Wonhoon-
dc.contributor.author박원훈-
dc.date.accessioned2024-07-25T19:31:12Z-
dc.date.available2024-07-25T19:31:12Z-
dc.date.issued2023-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1045893&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/320664-
dc.description학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2023.8,[iii, 17 p. :]-
dc.description.abstracthis work presents an energy-efficient digital-based computing-in-memory (CIM) processor to support floatingpoint (FP) deep neural network (DNN) acceleration. Previous FP-CIM processors have two limitations. Processors with post-alignment shows low throughput due to serial operation, and the other processor with prealignment incurs truncation error. To resolve these problems, we focus on the statistics that outlier exists according to shift amount in pre-alignment-based FP operation. As those outlier decreases energy efficiency due to long operation cycles, it needs to be processed separately. The proposed Hetero-FP-CIM integrates both CIM arrays and shared NPU, so they compute both dense inlier and sparse outlier respectively. It also includes efficient weight caching system to avoid entire weight copy in shared NPU. The proposed Hetero-FP-CIM is simulated in 28 nm CMOS technology and occupies 2.7 mm$^2$. As a result, it achieves 5.99 TOPS/W at ImageNet (ResNet50) with bfloat16 representation.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject메모리 내 컴퓨팅▼aSRAM▼a심층 신경망▼a부동 소수점▼a캐시 시스템▼a이기종 연산 가속기-
dc.subjectComputing-in-memory▼aSRAM▼aDeep neural network▼aFloating-point▼aCache system▼aOutlier-handling▼aHeterogeneous processor-
dc.title(A) heterogeneous computing-in-memory and neural-processing-unit architecture for an energy efficient floating-point DNN acceleration-
dc.title.alternative에너지 효율적인 부동 소수점 연산을 위한 메모리 내 컴퓨팅-심층 신경망 가속기 이기종 아키텍쳐-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthorYoo, Hoi-jun-
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0