Diversity matters when learning from ensembles다양성을 고려한 앙상블 학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 59
  • Download : 0
Deep ensembles excel in large-scale image classification tasks both in terms of prediction accuracy and calibration. Despite being simple to train, the computation and memory cost of deep ensembles limits their practicability. While some recent works propose to distill an ensemble model into a single model to reduce such costs, there is still a performance gap between the ensemble and distilled models. We propose a simple approach for reducing this gap, i.e., making the distilled performance close to the full ensemble. Our key assumption is that a distilled model should absorb as much function diversity inside the ensemble as possible. We first empirically show that the typical distillation procedure does not effectively transfer such diversity, especially for complex models that achieve near-zero training error. To fix this, we propose a perturbation strategy for distillation that reveals diversity by seeking inputs for which ensemble member outputs disagree. We empirically show that a model distilled with such perturbed samples indeed exhibits enhanced diversity, leading to improved performance.
Advisors
Lee, Juhoresearcher이주호researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2022
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2022.8,[iii, 24 p. :]

Keywords

Deep Ensemble▼aEnsemble Learning▼aKnowledge Distillation; 딥 앙상블▼a앙상블 학습▼a지식 증류

URI
http://hdl.handle.net/10203/308199
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1008202&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0