CAP-GAN: Towards Adversarial Robustness with Cycle-consistent Attentional Purification

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 90
  • Download : 0
Adversarial attack is aimed at fooling a target classifier with imperceptible perturbation. Adversarial examples, which are carefully crafted with a malicious purpose, can lead to erroneous predictions, resulting in catastrophic accidents. To mitigate the effect of adversarial attacks, we propose a novel purification model called CAP-GAN. CAP-GAN considers the idea of pixel-level and feature-level consistency to achieve reasonable purification under cycle-consistent learning. Specifically, we utilize a guided attention module and knowledge distillation to convey meaningful information to the purification model. Once the model is fully trained, inputs are projected into the purification model and transformed into clean-like images. We vary the capacity of the adversary to argue the robustness against various types of attack strategies. On CIFAR-10 dataset, CAP-GAN outperforms other pre-processing based defenses under both black-box and white-box settings.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2021-07
Language
English
Citation

2021 International Joint Conference on Neural Networks, IJCNN 2021

DOI
10.1109/IJCNN52387.2021.9533322
URI
http://hdl.handle.net/10203/288646
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0