MAT: A Multi-strength Adversarial Training Method to Mitigate Adversarial Attacks
Citations Over TimeTop 11% of 2018 papers
Abstract
Some recent work revealed that deep neural networks (DNNs) are vulnerable to so-called adversarial attacks where input examples are intentionally perturbed to fool DNNs. In this work, we revisit the DNN training process that includes adversarial examples into the training dataset so as to improve DNN's resilience to adversarial attacks, namely, adversarial training. Our experiments show that different adversarial strengths, i.e., perturbation levels of adversarial examples, have different working ranges to resist the attacks. Based on the observation, we propose a multi-strength adversarial training method (MAT) that combines the adversarial training examples with different adversarial strengths to defend adversarial attacks. Two training structures-mixed MAT and parallel MAT-are developed to facilitate the tradeoffs between training time and hardware cost. Our results show that MAT can substantially minimize the accuracy degradation of deep learning systems to adversarial attacks on MNIST, CIFAR-10, CIFAR-100, and SVHN. The tradeoffs between training time, robustness, and hardware cost are also well discussed on a FPGA platform.
Related Papers
- → Adversarial Perturbation Defense on Deep Neural Networks(2021)31 cited
- → Explainable AI for Inspecting Adversarial Attacks on Deep Neural Networks(2020)6 cited
- → Generating adversarial examples for DNN using pooling layers(2019)3 cited
- → Global Adversarial Attacks for Assessing Deep Learning Robustness(2019)3 cited
- → Developing and Defeating Adversarial Examples(2020)