Maximizing discrimination capability of knowledge distillation with energy function에너지 함수를 활용한 지식 증류의 분별 능력 극대화 방법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 3
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor김대식-
dc.contributor.authorKIM, SeongHak-
dc.contributor.author김성학-
dc.date.accessioned2024-07-30T19:31:37Z-
dc.date.available2024-07-30T19:31:37Z-
dc.date.issued2024-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1097210&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/321638-
dc.description학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2024.2,[vi, 52 p. :]-
dc.description.abstractTo apply the latest computer vision techniques that require a large computational cost in real industrial applications, knowledge distillation methods (KDs) are essential. Existing logit-based KDs apply the constant temperature scaling to all samples in dataset, limiting the utilization of knowledge inherent in each sample individually. In our approach, we classify the dataset into two categories (i.e., low energy and high energy samples) based on their energy score. Through experiments, we have confirmed that low energy samples exhibit high confidence scores, indicating certain predictions, while high energy samples yield low confidence scores, meaning uncertain predictions. To distill optimal knowledge by adjusting non-target class predictions, we apply a higher temperature to low energy samples to create smoother distributions and a lower temperature to high energy samples to achieve sharper distributions. When compared to previous logit-based and feature-based methods, our energy-based KD (Energy KD) achieves better performance on various datasets. Especially, Energy KD shows significant improvements on CIFAR$-$100$-$LT and ImageNet datasets, which contain many challenging samples. Furthermore, we propose high energy-based data augmentation (HE$-$DA) for further improving the performance. We demonstrate that meaningful performance improvement could be achieved by augmenting only $20\%\sim50\%$ of dataset, suggesting that it can be employed on resource-limited devices.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject딥러닝▼a컴퓨터 비전▼a모델 경량화▼a지식 증류 기법▼a데이터 증강 기법▼a에너지 함수-
dc.subjectDeep Learning▼aComputer Vision▼aModel Compression▼aKnowledge Distillation▼aData Augmentation▼aEnergy Function-
dc.titleMaximizing discrimination capability of knowledge distillation with energy function-
dc.title.alternative에너지 함수를 활용한 지식 증류의 분별 능력 극대화 방법-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthorKim, Dae-Shik-
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0