An article from Privacy International (PI) exploring the Humans in the AI loop approach considers the often overlooked role of data labellers in powering large language models (LLMs). The article focuses on the exploitative labour ecosystem behind data labelling and the use of algorithmic decision-making models in digital labour platforms. In particular, the article highlights the lack of transparency and the challenges faced by workers in terms of job stability and wages, as well as the need for better governance and ethical practices in this sector.
In a related article, the IAPP discusses how adopting a human in the loop strategy for AI oversight should not be considered a catch-all solution for potential system risks. Humans are fallible, inherently biased, and can compound technological errors produced by AI instead of mitigating them. The article outlines the context in which humans in the loop can help to solve AI risks, providing that the loop is clearly defined and the right person is chosen for the role.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.