Orthogonal feature regularization : a novel approach for training robust models특징 직교 정규화 : 강건한 모델 훈련을 위한 새로운 접근법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 271
  • Download : 0
Many deep neural networks (DNN) are easily fooled with adversarial examples designed to cause incorrect classification. Many researchers have tried to find a training recipe for defending against adversaries. However, most of these recipes, such as adversarial training, have the drawback that they are only robust for specific adversaries. Recent works have proposed many training algorithms with regularization, such as weight orthogonalization, penalizing the $l_2$ norm of the input gradient, and controlling the Lipschitz constants of each layer, however, they all have some limitations in terms of computational costs and efficacy. To address this problem, we propose a new approach with “resemble regularization”, which can be different from the concept that we generally believe. Our key idea is to encourage each layer’s outputs from different classes to resemble each other. The advantages of this method are that it helps the model to become robust against $l_\infty$ adversarial perturbations while requiring little computational cost, and it can be used with other robust regularization methods simultaneously, resulting in higher robustness. Our method is verified on MNIST and CIFAR-10. On CIFAR-10, we achieve state-of-the- art performance, which substantially improves the accuracy from 38.58% to 65.92% when tested on the adversaries with $l_\infty$ perturbations of $\epsilon$ = 0.1. In addition, robust adversarial errors against most of the adversaries are improved with a large margin of more than 20%. Through analysis of our method, we expect that this approach can reveal the fundamental reasons for the vulnerability of adversarial examples.
Advisors
Yun, Seyoungresearcher윤세영researcher
Description
한국과학기술원 :지식서비스공학대학원,
Publisher
한국과학기술원
Issue Date
2020
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 지식서비스공학대학원, 2020.2,[iii, 33 p. :]

Keywords

deep neural network▼aadversarial example▼aorthogonalization▼arobust▼aregularization; 깊은 신경망▼a적대적 예시▼a직교화▼a강건함▼a정규화

URI
http://hdl.handle.net/10203/283954
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=910739&flag=dissertation
Appears in Collection
KSE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0