Exploiting implicit pipeline parallelism in GPU-accelerated workloadsGPU로 가속화된 워크로드에서 암시적 파이프라인 병렬화 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 497
  • Download : 0
While GPGPU programming models such as CUDA and OpenCL are good for exploiting data parallelism, it is difficult to exploit pipeline parallelism with them. Since there are many workloads that spend a large portion of runtime on I/O device access, serial CPU thread execution and/or data transfer through PCIe, performance can be significantly improved if pipeline parallelism between those components is properly leveraged. Unfortunately, current GPGPU programming models can require a significant programmer effort to leverage the parallelism due to complex data dependency. In this work, we propose a framework to exploit implicit pipeline parallelism, without requiring a programmer to explicitly specify data dependency. We propose hardware-based dynamic dependency tracking mechanism to overlap different stages of GPU-accelerated workloads and reduce runtime. Also, our framework does not require any kernel modification or complex dependency tracking by programmer. Our evaluation results show that the proposed framework significantly reduces overall runtime that includes not only kernel execution time but also I/O and data transfer time by up to 40% and by 24% overall.
Advisors
Kim, Dong Junresearcher김동준researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2015
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2015.8,[iv, 30 p. :]

Keywords

CUDA; pipeline; GPU; GPU-accelerated Workload; Overlap stage; 파이프라인; 가속화된 GPU 워크로드; 단계 오버랩

URI
http://hdl.handle.net/10203/243397
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=669192&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0