Dataset condensation with atypical samples특이한 샘플들을 이용한 데이타셋 압축

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 1
  • Download : 0
The advancement of deep learning relies on huge datasets, which involves extensive training cost. To reduce the size of dataset, dataset distillation emerges, which aims to synthesize a small dataset encoding the information of the entire training set. However, we argue that existing dataset condensation method excessively focus on synthesizing typical samples rather than atypical samples, thereby diminish the network’s generalization capacity. To address the limitation, we propose a new strategy, Select-and-Match (SelMatch), that combines dataset condensation and selection. Our method selects important samples first, and perform condensation with regard to selected samples, which enriches the overall information within the synthetic dataset. We evaluate our method on the CIFAR-100 benchmark and demonstrate its superiority over state-of-the-art selection-only and condensation-only methods.
Advisors
정혜원researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2024.2,[iii, 17 p. :]

Keywords

데이터 효율적 학습▼a데이터셋 축소▼a최적화▼a데이터 선택; Data-efficient learning▼aDataset condensation▼aOptimization▼aSubset selection

URI
http://hdl.handle.net/10203/321566
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1096784&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0