NNSplitter: An Active Defense Solution to DNN Model via Automated Weight Obfuscation

04/28/2023
by   Tong Zhou, et al.
0

As a type of valuable intellectual property (IP), deep neural network (DNN) models have been protected by techniques like watermarking. However, such passive model protection cannot fully prevent model abuse. In this work, we propose an active model IP protection scheme, namely NNSplitter, which actively protects the model by splitting it into two parts: the obfuscated model that performs poorly due to weight obfuscation, and the model secrets consisting of the indexes and original values of the obfuscated weights, which can only be accessed by authorized users. NNSplitter uses the trusted execution environment to secure the secrets and a reinforcement learning-based controller to reduce the number of obfuscated weights while maximizing accuracy drop. Our experiments show that by only modifying 313 out of over 28 million (i.e., 0.001 can drop to 10 against potential attack surfaces, including norm clipping and fine-tuning attacks.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset