We could create a thinking AI that doesn’t have any feelings, and we can create a feeling AI that is not really great at thinking,” Basl said. He might mean that the language model is as intelligent as a small child or that it has the capacity to suffer or desire like a small child, Basl said. It is unclear from the Washington Post story, why Lemoine compares LaMDA to a child. Photo by Martin Klimek for The Washington Post “ she is not feeling terror about climate change.” Blake Lemoine poses for a portrait in Golden Gate Park in San Francisco, California on Thursday, June 9, 2022. “That seems to track some inner mental life,” Basl said. “We see that kind of range of capacities in the animal world,” he said.įor example, Basl said his dog doesn’t prefer the world to be one way rather than the other in any deep sense, but she clearly prefers her biscuits to kibble. It could be aware of the experience it is having, have positive or negative attitudes like feeling pain or wanting to not feel pain, and have desires. Sentient is defined as being able to perceive or feel things and is often compared to sapient.īasl believes that sentient AI would be minimally conscious. There are different definitions of sentience. The fact that this language model can trick a human into thinking that it is sentient speaks to its complexity, but it would need to have some other capacities beyond what it is optimized for to show sentience, Basl said. “It is not like we went to an alien planet and a thing that we never gave any incentives to start communicating with us ,” Basl said. However, LaMDA, a language model, was designed specifically to talk, and the optimization function used to train it to process language and converse incentivizes its algorithm to produce this linguistic evidence. The evidence seems to be grounded in LaMDA’s linguistic abilities and the things it talks about, Basl said. “Reactions like ‘We have created sentient AI’, I think, are extremely overblown,” Basl said. However, Basl believes the evidence mentioned in the Washington Post article is not enough to conclude that LaMDA is sentient. It reminds him of a 7- or 8-year-old child, Blake told the Washington Post. ![]() In the story, Google engineer Blake Lemoine says that the company’s artificially intelligent chatbot generator, LaMDA, with whom he had numerous deep conversations, might be sentient. Maybe that is why a recent Washington Post story has made such a big splash. Photo by Matthew Modoono/Northeastern University Basl believes that sentient AI will be minimally conscious, aware of the experience it is having, able to feel positive or negative feelings and to have desires. “And they are very cavalier about it.” John Basl, Assistant Professor of Philosophy poses for a portrait at Northeastern University. “When you hear Google talk, they talk as if this is just right around the corner or definitely within our lifetimes,” Basl said. ![]() No one knows, however, when humans will create an intelligent or sentient AI, said John Basl, associate professor of philosophy at Northeastern’s College of Social Sciences and Humanities, whose research focuses on the ethics of emerging technologies such as AI and synthetic biology. Still, thinking about AI, many of us imagine human-like robots who, according to countless science fiction stories, will become independent and rebel one day. From virtual assistants like Apple’s Siri and Amazon’s Alexa, to robotic vacuums and self-driving cars, to automated investment portfolio managers and marketing bots, artificial intelligence has become a big part of our everyday lives.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |