Visible to the public Feature Vulnerability and Robustness Assessment against Adversarial Machine Learning Attacks

TitleFeature Vulnerability and Robustness Assessment against Adversarial Machine Learning Attacks
Publication TypeConference Paper
Year of Publication2021
AuthorsMcCarthy, Andrew, Andriotis, Panagiotis, Ghadafi, Essam, Legg, Phil
Conference Name2021 International Conference on Cyber Situational Awareness, Data Analytics and Assessment (CyberSA)
Keywordsadversarial learning, attack surface, denial-of-service attack, face recognition, feature extraction, Intrusion detection, machine learning, Metrics, network traffic analysis, Perturbation methods, pubcrawl, resilience, Resiliency, Roads, Scalability, telecommunication traffic
AbstractWhilst machine learning has been widely adopted for various domains, it is important to consider how such techniques may be susceptible to malicious users through adversarial attacks. Given a trained classifier, a malicious attack may attempt to craft a data observation whereby the data features purposefully trigger the classifier to yield incorrect responses. This has been observed in various image classification tasks, including falsifying road sign detection and facial recognition, which could have severe consequences in real-world deployment. In this work, we investigate how these attacks could impact on network traffic analysis, and how a system could perform misclassification of common network attacks such as DDoS attacks. Using the CICIDS2017 data, we examine how vulnerable the data features used for intrusion detection are to perturbation attacks using FGSM adversarial examples. As a result, our method provides a defensive approach for assessing feature robustness that seeks to balance between classification accuracy whilst minimising the attack surface of the feature space.
DOI10.1109/CyberSA52016.2021.9478199
Citation Keymccarthy_feature_2021