Anchor-guided feature refinement for adversarial robustness and out-of-distribution detection앵커가 인도하는 특징 개선을 통한 적대적 견고성 향상 및 분포외 데이터 검출

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 3
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor노용만-
dc.contributor.authorLee, Hakmin-
dc.contributor.author이학민-
dc.date.accessioned2024-08-08T19:31:33Z-
dc.date.available2024-08-08T19:31:33Z-
dc.date.issued2024-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1100040&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/322140-
dc.description학위논문(박사) - 한국과학기술원 : 전기및전자공학부, 2024.2,[vi, 57 p. :]-
dc.description.abstractDeep neural networks (DNNs) are currently performing well in computer vision. Given that applications applying DNNs are utilized by a broad spectrum of users in real-world settings, variety of input data is also extensive. However, recent research has shown that DNNs are highly vulnerable to certain types of input data. Adversarial attack is one type of input data that causes DNNs to malfunction, causing imperceptible changes in the input data that lead to significant differences from human perception. A malicious attacker can easily provoke DNNs to make incorrect outcomes, compromising the reliability of the application. In another case, Out-of-Distribution (OOD) occurs when DNNs receive inputs that do not follow same probability distribution as training data. While DNNs should ideally express uncertainty when faced with unsuitable input data (OOD), they often produce incorrect results with a high level of confidence in practice. Therefore, these types of inputs contribute to the unreliability of DNNs, limiting their application in real-world environments. In this work, we propose frameworks to achieve robustness against adversarial attacks and OOD data through anchor-guided feature refinement. First, for adversarial attacks, the proposed anchors help adversarial training to separate normal and adversarial examples and acquire appropriate statistical characteristics for each. Even if an adversarial user attacks the DNNs, the anchor guides correct statistical processing to refine feature of adversarial example to achieve high performance. Second, Anchor provides a threshold to distinguish between OOD inputs. With the help of visual prompts, the anchor serves as a reference point to refine pre-trained model features for effective OOD detection. At inference time, the anchor is transformed into OOD scores, which are then utilized to establish thresholds for OOD detection. The effectiveness of the proposed method is validated through performance comparisons with state-of-the-art methods, and extensive qualitative and quantitative experiments were also conducted.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject적대적 강인성▼a분포외 검출▼a앵커▼a특징 개선▼a적대적 학습▼a시각적 프롬프트-
dc.subjectAdversarial robustness▼aOut-of-distribution detection▼aAnchor▼aFeature refinement▼aAdversarial training▼aVisual prompt-
dc.titleAnchor-guided feature refinement for adversarial robustness and out-of-distribution detection-
dc.title.alternative앵커가 인도하는 특징 개선을 통한 적대적 견고성 향상 및 분포외 데이터 검출-
dc.typeThesis(Ph.D)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthorRo, Yong Man-
Appears in Collection
EE-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0