Why Facebook Needs to Take Responsibility for Fake News
Washington Post reporter Wesley Lowery says the social media giant isn't excused from making responsible editorial choices just because it wishes to see itself as a technology company first.
Wesley Lowery is a national reporter for the Washington Post who covers law enforcement, justice, race and politics. He previously covered Congress and national politics. Prior to joining the Washington Post in February 2014, he worked as a breaking news and local politics reporter for the Boston Globe, and has also reported for the Los Angeles Times and the Wall Street Journal. In 2014, he was named the National Association of Black Journalists' "Emerging Journalist of the Year." Follow him on Twitter @WesleyLowery.
Wesley Lowery: I think Facebook is responsible for what exists and what happens on its platform and I think that Facebook has been negligent in its responsibility to safeguarding and providing a forum in which sane and reasonable interaction happens and that hysterical and untrue interaction does not happen. What we know is that Facebook has the ability to deal with fake news. These are fake sites that they pop up. They’re being spread by specific pages and specific accounts and rather than address that Facebook has allowed millions of people to become deeply ill informed. Now there’s a question how democratic should a platform like Facebook be? If everyone wants to share a fake news article should that be allowed? Or does Facebook have some editorial control of what it allows to be propagated via its own mechanisms and its own channels. I mean I think that Facebook needs to take greater steps in these spaces because again Facebook itself has become a media publisher and it is now the platform and the canvas for chaos to be created by all of this fake information spreading so quickly without any check and balance and any responsibility being taken by the platform of sorting through what is true and what is not.
I mean there has to be some editorial infrastructure. And they’ve had in the past some editorial infrastructure. They at one time had a staff that was figuring out what should be in the trending news and what should not. And immediately after they got rid of that staff all of a sudden there was a bunch of fake news in the trending, right. And that would be one obvious step. But I also think that there has to be – Facebook has remarkable power through its algorithms and through its media partnerships to make decisions about what becomes prominent on its site and what does not. There’s certainly an organic democratic role to this but it’s not as if they are just completely sitting back and playing no role over what we see and what we do not. And as soon as they begin playing that role at all they now take on I believe a responsibility to curate this content. Like I said I just think it’s extremely damaging, you think about newspapers. Newspapers don’t run every letter they get. There’s a decision. There’s a curation, right. We decide this one is just a rambling that has nothing to do with anything. This is just full of falsehoods, right. When you choose to publish something on your platform on your canvas you are making an editorial decision to allow it to exist in a space. Facebook takes down threatening and abusive things all the time. It takes down personal attacks or nudity or pornography, right. It has the ability when you have specific publishers that are constantly spreading misinformation it has some ability to undercut those publishers from further spreading that.
A recent Buzzfeed investigation found that fake news stories were more popular on Facebook in the 3-month runup to the 2016 presidential election than real news stories from credible sources. The analysis showed the public's voracious appetite for sensational falsehoods. Stories like the Pope endorsing Donald Trump, or Hillary Clinton selling arms to ISIS, gathered large amounts of attention from Facebook users even though they never happened.
Such stories are frequently featured on bogus websites, masquerading as legitimate news sources, that look simply to convert user clicks into advertising revenue. Enter the question of legal versus moral responsibility. Under US law, online platforms are not legally responsible for content they post, but companies as large and as powerful as Facebook do have a responsibility to the public interest, says Washington Post reporter Wesley Lowery.
Over the last years, Lowery has crisscrossed America, covering stories of police violence that often find a sympathetic audience on social platforms like Facebook. And as an increasing number of individuals get their news from social networks, Lowery compares Facebook to traditional newspapers, which make editorial decisions deciding what to publish. If a letter to the paper is full of misinformation, or is simply a rant, the editors will decline to publish it in the paper. Facebook did have a team of editors who decided which stories should be featured in the trending news category, but those editors were laid off after a controversy over what bias they may have in selecting stories.
Of course one bias they did not have was promoting stories that were flagrantly false. Now, Facebook and Google say they will exclude fraudulent media outlets from their advertising network, attempting to de-incentivize their drive to create fake news in the first place.
The controversy around the Torah codes gets a new life.
- Mathematicians claim to see a predictive pattern in the ancient Torah texts.
- The code is revealed by a method found with special computer software.
- Some events described by reading the code took place after the code was written.
A glass of juice has as much sugar, ounce for ounce, as a full-calorie soda. And those vitamins do almost nothing.
Quick: think back to childhood (if you've reached the scary clown you've gone too far). What did your parents or guardians give you to keep you quiet? If you're anything like most parents, it was juice. But here's the thing: juice is bad for you.
Orangutans join humans and bees in a very exclusive club
- Orangutan mothers wait to sound a danger alarm to avoid tipping off predators to their location
- It took a couple of researchers crawling around the Sumatran jungle to discover the phenomenon
- This ability may come from a common ancestor
SMARTER FASTER trademarks owned by The Big Think, Inc. All rights reserved.