Regularized stochastic representation for reliable contrastive learning신뢰할 수 있는 대조 학습을 위한 정규화된 확률적 표현 방법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 140
  • Download : 0
Contrastive learning, which has recently received a lot of attention, exploits multi-views of instances (images) to learn view-invariant representation. Conventionally, the multi-views are generated by composition of multiple stochastic augmentations. Since the contrastive learning methods so far implicitly assumed that the generated multi-views are always an appropriate positive pair, their representations were made to be recklessly close in the representation space. However, in the case of complex images, which is common in real-world, inappropriate view pairs are likely to be generated. It likely results in learning problematic representation by encouraging representations to be closer even though the pair is not appropriate. To alleviate this problem, we propose to use a regularized stochastic representation considering the adequacy of the given view pair. With this, we devised a novel method that the model can attenuate the effect from inappropriate pairs. The proposed method consistently outperforms baselines for various downstream tasks (image classification, object detection) on various benchmark datasets (CIFAR-100, ImageNet-100, COCO).
Advisors
Yang, Eunhoresearcher양은호researcher
Description
한국과학기술원 :AI대학원,
Publisher
한국과학기술원
Issue Date
2021
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : AI대학원, 2021.8,[iii, 22 p. :]

Keywords

Contrastive learning▼aRepresentation learning▼aSelf-supervised learning▼aStochastic representation▼aRegularization; 대조 학습▼a표현 학습▼a자기 지도 학습▼a확률적 표현▼a정규화

URI
http://hdl.handle.net/10203/292502
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=964744&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0