Multi-step Jailbreaking Privacy Attacks on ChatGPT

04/11/2023
by   Haoran Li, et al.
0

With the rapid progress of large language models (LLMs), many downstream NLP tasks can be well solved given good prompts. Though model developers and researchers work hard on dialog safety to avoid generating harmful content from LLMs, it is still challenging to steer AI-generated content (AIGC) for the human good. As powerful LLMs are devouring existing text data from various domains (e.g., GPT-3 is trained on 45TB texts), it is natural to doubt whether the private information is included in the training data and what privacy threats can these LLMs and their downstream applications bring. In this paper, we study the privacy threats from OpenAI's model APIs and New Bing enhanced by ChatGPT and show that application-integrated LLMs may cause more severe privacy threats ever than before. To this end, we conduct extensive experiments to support our claims and discuss LLMs' privacy implications.

READ FULL TEXT

page 3

page 7

page 8

research
02/14/2022

Threats to Pre-trained Language Models: Survey and Taxonomy

Pre-trained language models (PTLMs) have achieved great success and rema...
research
01/14/2021

Privacy Analysis in Language Models via Training Data Leakage Report

Recent advances in neural network based language models lead to successf...
research
04/26/2022

You Don't Know My Favorite Color: Preventing Dialogue Representations from Revealing Speakers' Private Personas

Social chatbots, also known as chit-chat chatbots, evolve rapidly with l...
research
06/06/2020

Privacy threats in intimate relationships

This article provides an overview of intimate threats: a class of privac...
research
05/24/2023

Trade-Offs Between Fairness and Privacy in Language Modeling

Protecting privacy in contemporary NLP models is gaining in importance. ...
research
05/22/2023

Watermarking Text Data on Large Language Models for Dataset Copyright Protection

Large Language Models (LLMs), such as BERT and GPT-based models like Cha...
research
09/11/2023

Privacy Side Channels in Machine Learning Systems

Most current approaches for protecting privacy in machine learning (ML) ...

Please sign up or login with your details

Forgot password? Click here to reset