Meta has published a paper about how it balances privacy and the integrity of its technologies in order to limit bad experiences on its platforms. The document includes examples of complex scenarios such as saving records of confirmed hate speech in order to train automated tools to detect other hate speech. Meta explains that personal data about the people who share the post originally can be minimalised for the training process. Other examples include identity verification, nudity, public safety and security, or where people commit severe Community Standards violations.
Have your say on Meta's approach to these issues. Submit your response by email to: privacy4integrity@fb.com before the consultation closes on 30 September.
What is this page?
You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.
The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.