Visible to the public Representation of White- and Black-Box Adversarial Examples in Deep Neural Networks and Humans: A Functional Magnetic Resonance Imaging Study

TitleRepresentation of White- and Black-Box Adversarial Examples in Deep Neural Networks and Humans: A Functional Magnetic Resonance Imaging Study
Publication TypeConference Paper
Year of Publication2019
AuthorsHan, Chihye, Yoon, Wonjun, Kwon, Gihyun, Kim, Daeshik, Nam, Seungkyu
Conference Name2019 International Joint Conference on Neural Networks (IJCNN)
PublisherIEEE
Keywordsadversarial example, Biological neural networks, biomedical MRI, black-box adversarial examples, brain, brain-inspired deep neural networks, complex level visual tasks, composability, Computer vision, DNNs, Facial animation, fMRI, Functional magnetic resonance imaging, functional magnetic resonance imaging study, high-level visual tasks, human performance, human vision, human visual system, medical image processing, Metrics, neural nets, Neural Network, neurophysiology, Noise, pubcrawl, representation patterns, representational similarity, resilience, Resiliency, Task Analysis, vision, visual perception, visual representation, Visual systems, visualization, white box cryptography
Abstract

The recent success of brain-inspired deep neural networks (DNNs) in solving complex, high-level visual tasks has led to rising expectations for their potential to match the human visual system. However, DNNs exhibit idiosyncrasies that suggest their visual representation and processing might be substantially different from human vision. One limitation of DNNs is that they are vulnerable to adversarial examples, input images on which subtle, carefully designed noises are added to fool a machine classifier. The robustness of the human visual system against adversarial examples is potentially of great importance as it could uncover a key mechanistic feature that machine vision is yet to incorporate. In this study, we compare the visual representations of white- and black-box adversarial examples in DNNs and humans by leveraging functional magnetic resonance imaging (fMRI). We find a small but significant difference in representation patterns for different (i.e. white- versus black-box) types of adversarial examples for both humans and DNNs. However, human performance on categorical judgment is not degraded by noise regardless of the type unlike DNN. These results suggest that adversarial examples may be differentially represented in the human visual system, but unable to affect the perceptual experience.

URLhttps://ieeexplore.ieee.org/document/8851763/
DOI10.1109/IJCNN.2019.8851763
Citation Keyhan_representation_2019