An internal assessment has revealed that an artificial intelligence (AI) system used by the Department for Work and Pensions (DWP) to detect welfare fraud shows bias related to age, disability, marital status, and nationality. Documents obtained from the DWP via a Freedom of Information request indicate that the machine-learning programme disproportionately flagged certain demographic groups for fraud investigations.
The revelation comes after previous assurances from the government that the system posed no discrimination concerns. However, the "statistically significant outcome disparity" from a "fairness analysis" indicates this is a misleading claim. Furthermore, it has been revealed that no evaluations have been made concerning biases related to race, sex, and other protected characteristics. Campaigners criticised the government for a "hurt first, fix later" approach and called for greater transparency about which groups might be unfairly targeted.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.