Neural network fragile watermarking with no model performance degradation

08/16/2022
by   Zhaoxia Yin, et al.
0

Deep neural networks are vulnerable to malicious fine-tuning attacks such as data poisoning and backdoor attacks. Therefore, in recent research, it is proposed how to detect malicious fine-tuning of neural network models. However, it usually negatively affects the performance of the protected model. Thus, we propose a novel neural network fragile watermarking with no model performance degradation. In the process of watermarking, we train a generative model with the specific loss function and secret key to generate triggers that are sensitive to the fine-tuning of the target classifier. In the process of verifying, we adopt the watermarked classifier to get labels of each fragile trigger. Then, malicious fine-tuning can be detected by comparing secret keys and labels. Experiments on classic datasets and classifiers show that the proposed method can effectively detect model malicious fine-tuning with no model performance degradation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/25/2022

Memorization in NLP Fine-tuning Methods

Large language models are shown to present privacy risks through memoriz...
research
04/24/2023

Enhancing Fine-Tuning Based Backdoor Defense with Sharpness-Aware Minimization

Backdoor defense, which aims to detect or mitigate the effect of malicio...
research
04/13/2022

Sapinet: A sparse event-based spatiotemporal oscillator for learning in the wild

We introduce Sapinet – a spike timing (event)-based multilayer neural ne...
research
09/10/2019

What do Deep Networks Like to Read?

Recent research towards understanding neural networks probes models in a...
research
10/28/2019

IPGuard: Protecting the Intellectual Property of Deep Neural Networks via Fingerprinting the Classification Boundary

A deep neural network (DNN) classifier represents a model owner's intell...
research
01/12/2021

DeepiSign: Invisible Fragile Watermark to Protect the Integrityand Authenticity of CNN

Convolutional Neural Networks (CNNs) deployed in real-life applications ...
research
06/15/2023

OVLA: Neural Network Ownership Verification using Latent Watermarks

Ownership verification for neural networks is important for protecting t...

Please sign up or login with your details

Forgot password? Click here to reset