Towards Probabilistic Verification of Machine Unlearning

03/09/2020
by   David Marco Sommer, et al.
2

Right to be forgotten, also known as the right to erasure, is the right of individuals to have their data erased from an entity storing it. The General Data Protection Regulation in the European Union legally solidified the status of this long held notion. As a consequence, there is a growing need for the development of mechanisms whereby users can verify if service providers comply with their deletion requests. In this work, we take the first step in proposing a formal framework to study the design of such verification mechanisms for data deletion requests – also known as machine unlearning – in the context of systems that provide machine learning as a service. We propose a backdoor-based verification mechanism and demonstrate its effectiveness in certifying data deletion with high confidence using the above framework. Our mechanism makes a novel use of backdoor attacks in ML as a basis for quantitatively inferring machine unlearning. In our mechanism, each user poisons part of its training data by injecting a user-specific backdoor trigger associated with a user-specific target label. The prediction of target labels on test samples with the backdoor trigger is then used as an indication of the user's data being used to train the ML model. We formalize the verification process as a hypothesis testing problem, and provide theoretical guarantees on the statistical power of the hypothesis test. We experimentally demonstrate that our approach has minimal effect on the machine learning service but provides high confidence verification of unlearning. We show that with a 30% poison ratio and merely 20 test queries, our verification mechanism has both false positive and false negative ratios below 10^-5. Furthermore, we also show the effectiveness of our approach by testing it against an adaptive adversary that uses a state-of-the-art backdoor defense method.

READ FULL TEXT
research
02/07/2020

Machine Unlearning: Linear Filtration for Logit-based Classifiers

Recently enacted legislation grants individuals certain rights to decide...
research
06/29/2022

Approximate Data Deletion in Generative Models

Users have the right to have their data deleted by third-party learned s...
research
02/25/2020

Formalizing Data Deletion in the Context of the Right to be Forgotten

The right of an individual to request the deletion of their personal dat...
research
08/30/2022

On the Trade-Off between Actionable Explanations and the Right to be Forgotten

As machine learning (ML) models are increasingly being deployed in high-...
research
07/11/2019

Making AI Forget You: Data Deletion in Machine Learning

Intense recent discussions have focused on how to provide individuals wi...
research
08/15/2019

Probabilistic Verification in Mechanism Design

We introduce a model of probabilistic verification in a mechanism design...
research
08/05/2023

PromptCARE: Prompt Copyright Protection by Watermark Injection and Verification

Large language models (LLMs) have witnessed a meteoric rise in popularit...

Please sign up or login with your details

Forgot password? Click here to reset