Why Facebook Needs to Take Responsibility for Fake News | Wesley Lowery | Big Think
I think Facebook is responsible for what exists and what happens on its platform, and I think that Facebook has been negligent in its responsibility to safeguarding and providing a forum in which sane and reasonable interaction happens and that hysterical and untrue interaction does not happen. What we know is that Facebook has the ability to deal with fake news. These are fake sites that they pop up. They’re being spread by specific pages and specific accounts, and rather than address that, Facebook has allowed millions of people to become deeply ill-informed.
Now there’s a question: how democratic should a platform like Facebook be? If everyone wants to share a fake news article, should that be allowed? Or does Facebook have some editorial control of what it allows to be propagated via its own mechanisms and its own channels? I mean, I think that Facebook needs to take greater steps in these spaces because, again, Facebook itself has become a media publisher, and it is now the platform and the canvas for chaos to be created by all of this fake information spreading so quickly without any check and balance and any responsibility being taken by the platform of sorting through what is true and what is not.
I mean, there has to be some editorial infrastructure. And they’ve had in the past some editorial infrastructure. They at one time had a staff that was figuring out what should be in the trending news and what should not. And immediately after they got rid of that staff, all of a sudden there was a bunch of fake news in the trending, right? And that would be one obvious step. But I also think that there has to be – Facebook has remarkable power through its algorithms and through its media partnerships to make decisions about what becomes prominent on its site and what does not.
There’s certainly an organic democratic role to this, but it’s not as if they are just completely sitting back and playing no role over what we see and what we do not. And as soon as they begin playing that role at all, they now take on, I believe, a responsibility to curate this content. Like I said, I just think it’s extremely damaging. You think about newspapers. Newspapers don’t run every letter they get. There’s a decision. There’s a curation, right? We decide this one is just a rambling that has nothing to do with anything. This is just full of falsehoods, right?
When you choose to publish something on your platform, on your canvas, you are making an editorial decision to allow it to exist in a space. Facebook takes down threatening and abusive things all the time. It takes down personal attacks or nudity or pornography, right? It has the ability, when you have specific publishers that are constantly spreading misinformation, it has some ability to undercut those publishers from further spreading that.