The UK government is preparing to publish a set of criteria that must be met before any new laws governing the use of artificial intelligence (AI) are passed. The criteria, expected to be published in the coming weeks, would determine when the government would act to place restrictions on powerful AI models created by companies such as OpenAI and Google. Some of the “key tests” that would trigger legislative action include the failure of the UK’s new AI Safety Institute to identify risks around the technology and whether AI companies fail to uphold voluntary commitments to avoid harm.
The news further emphasises the UK’s differing approach from other countries such as the EU, US, and China, which have created more specific legislation and guidance on the technology. Instead, the UK has opted for a light-touch regime, fearing that stricter regulation would stifle industry growth.
£ - This article requires a subscription.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.