DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Kim, Junmo | - |
dc.contributor.advisor | 김준모 | - |
dc.contributor.author | Cho, Yooshin | - |
dc.date.accessioned | 2021-05-13T19:34:18Z | - |
dc.date.available | 2021-05-13T19:34:18Z | - |
dc.date.issued | 2020 | - |
dc.identifier.uri | http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=911406&flag=dissertation | en_US |
dc.identifier.uri | http://hdl.handle.net/10203/284776 | - |
dc.description | 학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.2,[iii, 19 p. :] | - |
dc.description.abstract | Spatial attention module is widely used in deep neural networks. The monumental attention module,Transformer [11] was proposed as self-attention and encoder-decoder frameworks in machine translationtask, and improving capability of learning long-range dependencies. After that, spatial attention modulessuch as Non-local block [12] and Criss-cross attention block [5] were also proposed and improving per-formance in vision fields including action recognition, segmentation, and object detection. Despite greatsuccess, spatial attention can only be used in a limited way due to expensive computation and memorycosts. To reduce overhead, segmentation or pooling should be used and several methods were proposed,but stacking enough blocks is still limited. Also, multi-head attention is not used in vision and videotasks due to the memory limitation. In this paper, we analyze which factors play an important role inlearning spatial attention module in the view of geometric definition, and propose ‘Transposed attentionmodule’ that is faster and smaller than Non-local block [12] in the same number of blocks and heads. Thebiggest advantage of our module is that memory and speed are maintained when the number of headsis increasing. Also, we introduce new approach to interpret spatial attention module and get superiorperformance compared to Non-local block [12] on CIFAR-10, CIFAR-100 [9], and Tiny-ImageNet. | - |
dc.language | eng | - |
dc.publisher | 한국과학기술원 | - |
dc.subject | Spatial attention▼ageometric definition▼amulti-head▼afaster and smaller▼aTransposed attention | - |
dc.subject | 공간주의 모듈▼a다중머리 구조▼a기하학적인 정의 | - |
dc.title | (A) study for efficient spatial attention module | - |
dc.title.alternative | 효율적인 공간적 주의 모듈을 위한 연구 | - |
dc.type | Thesis(Master) | - |
dc.identifier.CNRN | 325007 | - |
dc.description.department | 한국과학기술원 :전기및전자공학부, | - |
dc.contributor.alternativeauthor | 조유신 | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.