Adversarial Defense via Learning to Generate Diverse Attacks
Title | Adversarial Defense via Learning to Generate Diverse Attacks |
Publication Type | Conference Paper |
Year of Publication | 2019 |
Authors | Jang, Yunseok, Zhao, Tianchen, Hong, Seunghoon, Lee, Honglak |
Conference Name | 2019 IEEE/CVF International Conference on Computer Vision (ICCV) |
Date Published | oct |
Keywords | adversarial defense, Adversarial training, black-box attacks, classification performance, composability, compositionality, Deep Learning, deep neural networks, deterministic generator, Generators, learning (artificial intelligence), machine learning, malicious attacks, Metrics, neural nets, Neural networks, one-shot perturbation, Optimization, pattern classification, Perturbation methods, pubcrawl, recursive generator, resilience, Resiliency, Robustness, security of data, stochastic generator, Training, White Box Security |
Abstract | With the remarkable success of deep learning, Deep Neural Networks (DNNs) have been applied as dominant tools to various machine learning domains. Despite this success, however, it has been found that DNNs are surprisingly vulnerable to malicious attacks; adding a small, perceptually indistinguishable perturbations to the data can easily degrade classification performance. Adversarial training is an effective defense strategy to train a robust classifier. In this work, we propose to utilize the generator to learn how to create adversarial examples. Unlike the existing approaches that create a one-shot perturbation by a deterministic generator, we propose a recursive and stochastic generator that produces much stronger and diverse perturbations that comprehensively reveal the vulnerability of the target classifier. Our experiment results on MNIST and CIFAR-10 datasets show that the classifier adversarially trained with our method yields more robust performance over various white-box and black-box attacks. |
URL | https://ieeexplore.ieee.org/document/9008544/ |
DOI | 10.1109/ICCV.2019.00283 |
Citation Key | jang_adversarial_2019 |
- neural nets
- White Box Security
- Training
- stochastic generator
- security of data
- Robustness
- Resiliency
- resilience
- recursive generator
- pubcrawl
- Perturbation methods
- pattern classification
- optimization
- one-shot perturbation
- Neural networks
- adversarial defense
- Metrics
- malicious attacks
- machine learning
- learning (artificial intelligence)
- Generators
- deterministic generator
- deep neural networks
- deep learning
- Compositionality
- composability
- classification performance
- black-box attacks
- Adversarial training