Human-like programs abuse our empathy – even Google engineers aren’t immune | Emily M Bender

It ’s easy to be fooled by the mimicry, but consumers need transparency about how such systems are usedThe Google engineer Blake Lemoine wasn ’t speaking for the company officially when he claimed that Google’s chatbotLaMDA was sentient, but Lemoine ’s misconception shows the risks of designing systems in ways that convince humans they see real, independent intelligence in a program. If we believe that text-generating machines are sentient, what actions might we take based on the text they generate? It led Lemoine to leak secret transcripts f rom the program, resulting in his current suspension from the organisation.Google is decidedly leaning in to that kind of design, as seen in Alphabet CEO Sundar Pichai ’sdemo of that same chatbot at Google I/O in May 2021, where he prompted LaMDA to speak in the voice of Pluto and share some fun facts about the ex-planet. As Google plans to make this a core consumer-facing technology, the fact that one of its own engineers was fooled highlights the need for these systems to be transparent.Emily M Bender is a professor of linguistics at the University of Washington andco-author ofseveral papers on the risks of massive deployment of pattern recognition at scaleContinue reading...
Source: Guardian Unlimited Science - Category: Science Authors: Tags: Artificial intelligence (AI) Computing Consciousness Neuroscience Technology Google Alphabet Source Type: news