A.I. Can Pretend to Love Us, but is that Dangerous for Children? With Sherry Turkle | Big Think
I have very strong feelings about a future in which robots become the kind of conversational agent that pretend to have emotional lives. Shortly after I finished "We Can Make Reclaiming Conversation," I was interviewed for an article in the New York Times about Hello Barbie.
So Hello Barbie comes out of the box and says, now I'm just paraphrasing the jest: "Hi, I'm Hello Barbie. I have a sister. You have a sister. I kind of hate my sister. I'm jealous of your sister. Do you hate your sister? Let's talk about how we feel about our sisters." In other words, it just kind of knows stuff about you and is ready to talk about the arc of a human life and sibling rivalry as though it had a life, a mother, the feelings of jealousy about a sister, and was ready to relate to you on that basis.
And it doesn't. It's teaching pretend empathy. It's asking you to relate to an object that has pretend empathy. And this is really not a—in my view—this is really a not good direction for AI to go. There are so many wonderful things for robots to do, so many wonderful things for robots to do. Having pretend empathy, having pretend conversations about caring and love and things that a robot can feel about their bodies and about their lives and about their mothers and about their sisters gets children and gets elders, which are the other target group for these robots, into a kind of fantasy miasma that is not good for anybody.
Children don't need to learn pretend empathy; they need to learn real empathy, which they get from having real conversations with real people who do have sisters, who do have mothers. And I think this is a very dangerous and indeed very toxic direction. We worry so much about whether we can get people to talk to robots. You know, can you get a child to talk to this Hello Barbie? Can you get an elderly person to talk to a sociable robot? What about who's listening?
There's nobody listening. These robots don't know how to listen and understand what you're saying. They know how to respond. They're programmed to make something of what you say and respond, but they don't know what it means if you say, "My sister makes me feel depressed because she's more beautiful than I am, and I feel that my mother loves her more." That robot really does not do anything useful to you with that information. That's not empathy.
And children need to know that they're being heard by a human being that can do this empathy game with them, this empathy dance with them. So, in the middle of a time when we're having this crisis in empathy, to imagine that now we're going to throw in some robots that will do some pretend empathy, I have to say that in all the optimism of my book, this is the pessimistic part.
I really end the book with a kind of call to arms. I call it "What Do We Forget When We Talk to Machines?" And I mean it to be literally a call to arms that this is not a good direction. We don't need to take this direction; we just need to not buy these products. This doesn't take a social revolution; this just takes consumers saying that they're not going to buy these products. They're not bringing them to their homes...