The Organisation for Economic Co-operation and Development (OECD) has published an article evaluating the effectiveness of national artificial intelligence (AI) safety institutes. The article highlights the importance of governments understanding the potential of AI in order to manage associated risks and harness its advantages.
In recent months, several countries, including the UK, US, Japan, Canada, and Singapore, have established specialised AI Safety Institutes (AISIs) to keep pace with the rapid advancement of AI technology. These institutes are responsible for assessing the capabilities and risks of AI systems, conducting essential safety research, sharing their findings, and promoting collaboration among stakeholders. The EU's AI Office will undertake similar duties through its AI Safety Unit in addition to its regulatory function. The OECD writes that the establishment of these institutes represents a significant stride towards developing a unified global strategy to ensure the safety and beneficial use of AI.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.