Visible to the public Mitigating Poisoning Attacks on Machine Learning Models: A Data Provenance Based Approach

TitleMitigating Poisoning Attacks on Machine Learning Models: A Data Provenance Based Approach
Publication TypeConference Paper
Year of Publication2017
AuthorsBaracaldo, Nathalie, Chen, Bryant, Ludwig, Heiko, Safavi, Jaehoon Amir
Conference NameProceedings of the 10th ACM Workshop on Artificial Intelligence and Security
PublisherACM
Conference LocationNew York, NY, USA
ISBN Number978-1-4503-5202-4
KeywordsAdversarial Machine Learning, AI Poisoning, causative attacks, Human Behavior, internet of the things, IoT, poisoning attacks, Provenance, pubcrawl, resilience, Resiliency, Scalability, security
AbstractThe use of machine learning models has become ubiquitous. Their predictions are used to make decisions about healthcare, security, investments and many other critical applications. Given this pervasiveness, it is not surprising that adversaries have an incentive to manipulate machine learning models to their advantage. One way of manipulating a model is through a poisoning or causative attack in which the adversary feeds carefully crafted poisonous data points into the training set. Taking advantage of recently developed tamper-free provenance frameworks, we present a methodology that uses contextual information about the origin and transformation of data points in the training set to identify poisonous data, thereby enabling online and regularly re-trained machine learning applications to consume data sources in potentially adversarial environments. To the best of our knowledge, this is the first approach to incorporate provenance information as part of a filtering algorithm to detect causative attacks. We present two variations of the methodology - one tailored to partially trusted data sets and the other to fully untrusted data sets. Finally, we evaluate our methodology against existing methods to detect poison data and show an improvement in the detection rate.
URLhttp://doi.acm.org/10.1145/3128572.3140450
DOI10.1145/3128572.3140450
Citation Keybaracaldo_mitigating_2017