A warning from the inventor of the first chatbot in 1966, AI can be toxic

25/07/2023 | The Guardian

Joseph Weizenbaum, a computer scientist who created the first chatbot in 1966, was adamant that computers should never be confused with humans. His creation, named Eliza, was a psychotherapist that would reply to messages typed on an electric typewriter connected to a mainframe. Recently, the release of ChatGPT has renewed interest in Eliza, as it illuminated a mechanism of the human mind that affects how we relate to computers. Weizenbaum's experience with Eliza marked the beginning of a remarkable journey that made him a heretic among computer scientists in the 1970s. He published articles and books warning of the dangers posed by their work and believed that artificial intelligence was an "index of the insanity of our world." Today, many prominent researchers share Weizenbaum's view that AI can be toxic. There is much in Weizenbaum's thinking that is relevant now, and he would be less lonely in 2023.

Read Full Story
Gas mask, toxic

What is this page?

You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.

The Privacy Newsfeed monitors over 300 global publications, of which more than 5,750 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.

Freevacy has been shortlisted in the Best Educator category.
The PICCASO Privacy Awards recognise the people making an outstanding contribution to this dynamic and fast-growing sector.