Protecting Intellectual Property of Deep Neural Networks with Watermarking
Title | Protecting Intellectual Property of Deep Neural Networks with Watermarking |
Publication Type | Conference Paper |
Year of Publication | 2018 |
Authors | Zhang, Jialong, Gu, Zhongshu, Jang, Jiyong, Wu, Hui, Stoecklin, Marc Ph., Huang, Heqing, Molloy, Ian |
Conference Name | Proceedings of the 2018 on Asia Conference on Computer and Communications Security |
Date Published | May 2018 |
Publisher | ACM |
Conference Location | New York, NY, USA |
ISBN Number | 978-1-4503-5576-6 |
Keywords | Artificial neural networks, Collaboration, cyber physical systems, Deep Neural Network, Metrics, ownership verification, policy-based governance, pubcrawl, Resiliency, Watermarking |
Abstract | Deep learning technologies, which are the key components of state-of-the-art Artificial Intelligence (AI) services, have shown great success in providing human-level capabilities for a variety of tasks, such as visual analysis, speech recognition, and natural language processing and etc. Building a production-level deep learning model is a non-trivial task, which requires a large amount of training data, powerful computing resources, and human expertises. Therefore, illegitimate reproducing, distribution, and the derivation of proprietary deep learning models can lead to copyright infringement and economic harm to model creators. Therefore, it is essential to devise a technique to protect the intellectual property of deep learning models and enable external verification of the model ownership. In this paper, we generalize the "digital watermarking'' concept from multimedia ownership verification to deep neural network (DNNs) models. We investigate three DNN-applicable watermark generation algorithms, propose a watermark implanting approach to infuse watermark into deep learning models, and design a remote verification mechanism to determine the model ownership. By extending the intrinsic generalization and memorization capabilities of deep neural networks, we enable the models to learn specially crafted watermarks at training and activate with pre-specified predictions when observing the watermark patterns at inference. We evaluate our approach with two image recognition benchmark datasets. Our framework accurately (100$\backslash$%) and quickly verifies the ownership of all the remotely deployed deep learning models without affecting the model accuracy for normal input data. In addition, the embedded watermarks in DNN models are robust and resilient to different counter-watermark mechanisms, such as fine-tuning, parameter pruning, and model inversion attacks. |
URL | https://dl.acm.org/doi/10.1145/3196494.3196550 |
DOI | 10.1145/3196494.3196550 |
Citation Key | zhang_protecting_2018 |