A recent report by the World Privacy Forum has highlighted shortcomings in the AI governance tools currently being used to prevent bias and discrimination. The study found that more than one-third of the 18 AI governance tools used in countries such as Australia, Canada, Dubai, Ghana, Singapore, the UK and the US contained flaws. The tools were not always suitable for the organisation's specific use of AI, were not thoroughly vetted before being introduced, and in many cases, did not document what the tools were and how they worked. The report concludes this could lead to a mismatch in applications and emphasises that the dangers of rushing to govern AI without tailoring the framework to an organisation's specific needs. The report also warns against using such tools without adopting guidelines suited to the country's regulatory landscape.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.