Protecting personal data in the era of AI

18/03/2024 | Stanford University

The rapid advancement of artificial intelligence (AI), particularly through large language models (LLMs) and chatbots, has introduced significant privacy concerns. Questions arise regarding the use of personal information in AI training data, the sharing of user prompts with law enforcement, and the potential for chatbots to link various aspects of individuals' online lives. 

In response to these challenges, Jennifer King, a privacy and data policy fellow at Stanford University's Institute for Human-Centered Artificial Intelligence, and Caroline Meinhardt, the policy research manager at Stanford HAI, collaborated on a white paper: Rethinking Privacy in the AI Era: Policy Provocations for a Data-Centric World. The paper outlines that AI systems pose many of the same risks we've been dealing with in privacy over the last 20 years of internet use and unrestrained data collection. The difference is the scale in which AI systems consume data.

Read Full Story
Digital privacy, binary image

What is this page?

You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.

The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.

Freevacy has been shortlisted in the Best Educator category.
The PICCASO Privacy Awards recognise the people making an outstanding contribution to this dynamic and fast-growing sector.