The National Institute of Standards and Technology (NIST) has published a report concerning the challenges in mitigating attacks on artificial intelligence (AI) and machine learning (ML) systems. The NIST report, developed in association with the UK AI Security Institute (AISI), is intended to assist those responsible for designing, developing, deploying, evaluating, and governing AI systems in devising improved mitigations for adversarial ML (AML).
The report highlights that the data-driven nature of ML systems introduces new attack vectors that traditional software systems do not face, posing threats to security, privacy, and safety. These attacks target various stages of ML operations, including manipulating training data, providing adversarial inputs to disrupt system performance, and maliciously modifying models to extract sensitive training data.
NIST stressed that these attacks have been demonstrated in real-world scenarios, with their complexity and impact steadily rising. As AI systems become more prevalent, their security becomes increasingly critical.

What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.