(An) energy-efficient deep reinforcement learning processor with dual-mode weight compression and floating-point computing-in-memory이중방식 가중치 압축과 부동소수점 인메모리 연산 구조를 활용한 에너지 효율적 심층 강화학습 가속기

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 98
  • Download : 0
The deep neural network (DNN) has shown impressive performance in various applications such as computer vision, and natural language processing. However, such a high performance of DNN comes at a high computational cost. For calculating DNN, a large size of weight and feature map should be stored in memory, and multiplied by each other after being loaded from memory. Recently, the portion of memory access for the total time and total power consumption has been increased in the DNN accelerators, due to the stagnant improvement of memory, which is known as the "memory wall" problem. Especially in the case of deep reinforcement learning, memory optimization is essential because it has mainly utilized multiple numbers of fully-connected layers simultaneously. In this paper, we propose a deep reinforcement learning accelerator that optimizes memory bandwidth and memory power consumption. The proposed accelerator optimizes memory bandwidth by dual-mode weight compression and optimizes memory power consumption with floating-point in-memory computing architecture.
Advisors
Yoo, Hoi-Junresearcher유회준researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(박사) - 한국과학기술원 : 전기및전자공학부, 2023.2,[ix, 152 p. :]

Keywords

Deep neural network accelerator▼aMobile device▼aDeep reinforcement learning▼aMemory power consumption optimization▼aFloating-point in-memory computing; 심층 신경망 가속기▼a메모리 전력소모 최적화▼a심층 강화학습▼a부동소수점 인메모리 컴퓨팅

URI
http://hdl.handle.net/10203/309157
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1030565&flag=dissertation
Appears in Collection
EE-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0