A nationally representative UK survey conducted by the Ada Lovelace Institute and the Alan Turing Institute reveals evolving public attitudes towards artificial intelligence (AI). The 2024-25 survey of 3,513 UK residents shows increased public awareness of AI, particularly regarding large language models (LLMs), with 61% having heard of them and 40% having used them.
Despite growing familiarity, public concern about AI's potential harms has also risen across various applications. Common worries include overreliance on technology, potential errors, and a lack of transparency in AI-driven decisions. More than two-thirds (67%) of the UK public are concerned about AI systems making decisions about their lives, 83% are concerned about public sector data sharing with private companies for AI training, and an overwhelming 94% are concerned about the spread of AI-generated harms online.
There's a strong public call for regulation, with 72% expressing that they would feel more comfortable with AI if laws and regulations were in place, up from 62% in the previous survey. Meanwhile, 88% of people believe the government or regulators should have the authority to stop AI products from posing serious harm.

What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.