The internet is emotionally abusing us. And we can’t quit it. | Douglas Rushkoff | Big Think
[Music]
I'm actually the guy that came up with Viola with the tire media virus in a book I wrote in 1994. I look at viral media and weaponized mimetics as kind of my problem child. I originally saw it as the province of the counterculture, that what we could do is identify these sort of unresolved issues in society and then nest in them and provoke an immune response.
So, whether it's, you know, racism or poverty or corporate malfeasance, all the kinds of things that don't really get on the tube or don't really get discussed in an appropriate way; the only people who ended up buying that book really were marketers and like Russian propagandists.
They looked at it and used these ideas not to promote cultural growth or awareness, but to just provoke a response, to sensationalize anything by any means necessary. I mean, you know, propagandists ended up really good at this. They look for whatever issues are creating the most tension in America and then how do we spread it?
So, you know, gun control or abortion. You look for things and what you do is really operationalize conflict. How do we get people to look at those who disagree with them as less than human? Whether you're a left blue state person or a red right state person, if you're watching this, I encourage you to think about how it is that you think about those others. Are you thinking about them as less than you? You probably are.
It's really hard not to these days. Well, they're not; they have similar fears, they're just expressing those fears in a different way. If we can begin to see the other not as the way our weaponized memes are encouraging us to, but to see them as humans, well then it gets interesting.
You know, weaponized memes are the clearest example of human beings becoming the medium. We are no longer the user; we are being played. So, we create algorithms to accomplish some human goal, something we want. Go out online and find this—you're my intelligent agent. You're gonna find information and all that, but we've turned those algorithms against humans.
Now what the algorithms are there to do is to find what computer hackers used to call exploits. Only you're not looking for an exploit in a computer program or an exploit on a server; they're looking for exploits in humans. And where do you find those exploits? You find those exploits in our painstakingly evolved social mechanisms for connection.
So, the algorithms look for what it is that we use to establish rapport. What is it that we use to connect with another person? What are the mechanisms that provoke fear or self-defensive measures? And the algorithms will not knowingly, but just because they're trying everything, they will eventually find those and leverage them.
So, the algorithms on Facebook have found out that people click when they see pictures of their ex-lovers having fun. If you see a picture of your ex having fun, you'll click. If you see your ex not having fun, you don't, apparently, but anything positive about an ex—so your feed is gonna get those. Even if you've unsubscribed or whatever to that person, those things are gonna slip in there because they provoke use.
Is it something you need to see or want to see? Is it good for you when you're trying to leave that thing behind? Of course, it's not, but it pulls you in. And all the algorithms want is to get you to do the behavior that their programmers have asked them to.
So, we've spent, and we're all investing through our SP mutual funds, we are investing trillions of dollars in companies that are developing algorithms specifically designed to make us unhappy, to play us, to abuse us, and to compromise our humanity by leveraging our most important social instincts for really isolating, atomizing purposes.
[Music]
Yo...