Visible to the public Adversarial Attack on Fake-Faces Detectors Under White and Black Box Scenarios

TitleAdversarial Attack on Fake-Faces Detectors Under White and Black Box Scenarios
Publication TypeConference Paper
Year of Publication2021
AuthorsWang, Xiying, Ni, Rongrong, Li, Wenjie, Zhao, Yao
Conference Name2021 IEEE International Conference on Image Processing (ICIP)
Keywordsadversarial attack, black-box, composability, Conferences, Detectors, Faked faces, feature extraction, forensic models, Forensics, generative adversarial networks, Generators, Metrics, pubcrawl, Resiliency, security, White Box Security, white-box
AbstractGenerative Adversarial Network (GAN) models have been widely used in various fields. More recently, styleGAN and styleGAN2 have been developed to synthesize faces that are indistinguishable to the human eyes, which could pose a threat to public security. But latest work has shown that it is possible to identify fakes using powerful CNN networks as classifiers. However, the reliability of these techniques is unknown. Therefore, in this paper we focus on the generation of content-preserving images from fake faces to spoof classifiers. Two GAN-based frameworks are proposed to achieve the goal in the white-box and black-box. For the white-box, a network without up/down sampling is proposed to generate face images to confuse the classifier. In the black-box scenario (where the classifier is unknown), real data is introduced as a guidance for GAN structure to make it adversarial, and a Real Extractor as an auxiliary network to constrain the feature distance between the generated images and the real data to enhance the adversarial capability. Experimental results show that the proposed method effectively reduces the detection accuracy of forensic models with good transferability.
DOI10.1109/ICIP42928.2021.9506273
Citation Keywang_adversarial_2021