Visible to the public Defense through diverse directionsConflict Detection Enabled

TitleDefense through diverse directions
Publication TypeConference Paper
Year of Publication2020
AuthorsBender, Christopher M., Li, Yang, Shi, Yifeng, Reiter, Michael K., Oliva, Junier B.
Conference NameProceedings of the 37th International Conference on Machine Learning
Date Published07/2020
Conference LocationVirtual
Keywords2020: October, CMU, Securing Safety-Critical Machine Learning Algorithms
Abstract

In this work we develop a novel Bayesian neural network methodology to achieve strong adversarial robustness without the need for online adversarial training. Unlike previous efforts in this direction, we do not rely solely on the stochasticity of network weights by minimizing the divergence between the learned parameter distribution and a prior. Instead, we additionally require that the model maintain some expected uncertainty with respect to all input covariates. We demonstrate that by encouraging the network to distribute evenly across inputs, the network becomes less susceptible to localized, brittle features which imparts a natural robustness to targeted perturbations. We show empirical robustness on several benchmark datasets.

Citation Keynode-74178

Other available formats:

Bender_Defense_thru_Directions_Bauer.pdf
AttachmentTaxonomyKindSize
Bender_Defense_thru_Directions_Bauer.pdfPDF document1.15 MBDownloadPreview
AttachmentSize
bytes