(A) heterogeneous computing-in-memory and neural-processing-unit architecture for an energy efficient floating-point DNN acceleration에너지 효율적인 부동 소수점 연산을 위한 메모리 내 컴퓨팅-심층 신경망 가속기 이기종 아키텍쳐

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 2
  • Download : 0
his work presents an energy-efficient digital-based computing-in-memory (CIM) processor to support floatingpoint (FP) deep neural network (DNN) acceleration. Previous FP-CIM processors have two limitations. Processors with post-alignment shows low throughput due to serial operation, and the other processor with prealignment incurs truncation error. To resolve these problems, we focus on the statistics that outlier exists according to shift amount in pre-alignment-based FP operation. As those outlier decreases energy efficiency due to long operation cycles, it needs to be processed separately. The proposed Hetero-FP-CIM integrates both CIM arrays and shared NPU, so they compute both dense inlier and sparse outlier respectively. It also includes efficient weight caching system to avoid entire weight copy in shared NPU. The proposed Hetero-FP-CIM is simulated in 28 nm CMOS technology and occupies 2.7 mm$^2$. As a result, it achieves 5.99 TOPS/W at ImageNet (ResNet50) with bfloat16 representation.
Advisors
유회준researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2023.8,[iii, 17 p. :]

Keywords

메모리 내 컴퓨팅▼aSRAM▼a심층 신경망▼a부동 소수점▼a캐시 시스템▼a이기종 연산 가속기; Computing-in-memory▼aSRAM▼aDeep neural network▼aFloating-point▼aCache system▼aOutlier-handling▼aHeterogeneous processor

URI
http://hdl.handle.net/10203/320664
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1045893&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0