Learning visual representations from uncurated data정제되지 않은 데이터로부터 시각적 표현 학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 2
  • Download : 0
The learning of visual representations is a crucial problem in machine learning and computer vision. However, previous studies have focused mainly on improving model performance on benchmark datasets like ImageNet, limiting their applicability in real-world scenarios. In this paper, we propose two solutions to overcome these challenges and address the difficulties in representation learning from uncurated datasets. Firstly, we propose learning object-centric representations by separating objects from backgrounds in multi-object images. This enables us to remove scene biases and enhance the robustness of the model. Secondly, we employ semi-supervised learning when dealing with uncurated, unlabeled data. This allows us to improve the model's performance by leveraging large amounts of unlabeled data. Specifically, for the first problem, we propose object-centric learning techniques in unsupervised and patch-based models. For the second problem, we propose semi-supervised learning techniques in image classification and image-to-text models. Through our proposed techniques, we demonstrate excellent performance by efficiently utilizing uncurated data in various experimental settings.
Advisors
신진우researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(박사) - 한국과학기술원 : 전기및전자공학부, 2023.8,[v, 93 p. :]

Keywords

딥러닝▼a머신러닝▼a인공지능▼a컴퓨터비전; Deep learning▼aMachine learning▼aArtificial intelligence▼aComputer vision

URI
http://hdl.handle.net/10203/320934
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1047224&flag=dissertation
Appears in Collection
EE-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0