Facebook’s fake news excuses wearing thin

The excuse that Facebook is just a technology firm, an aggregator not up to the task of policing its users, started out thin and is only getting thinner, says the writer.

The excuse that Facebook is just a technology firm, an aggregator not up to the task of policing its users, started out thin and is only getting thinner, says the writer.

Published Nov 17, 2016

Share

Facebook must confront the responsibilities of being a media company, says Hayley Tsukayama.

Washington - Since last week’s election, Facebook’s role in policing fake news on its site has become a very hot topic.

And it should be. Throughout the election, Facebook’s behaviour exposed what seems to be a great contradiction. As the social network has pushed hard to dominate new forms of media, it’s also bent over backward to deny that it is a media company and denying that responsibility that comes with that label.

The truth is that Facebook has already taken on one of the functions of a media company: to act as a gatekeeper. It has labelled satire. It takes down “clickbait” articles which, in its own words, have headlines that “intentionally leave out crucial information, or mislead people”. Its algorithms clearly have some standards for content quality.

But Facebook won’t apply those standards to its fake news problem.

In fact, Facebook chief executive Mark Zuckerberg took to his own profile to explain why and to reject the much-discussed idea that false news articles on the network could have affected the election.

I don’t know if we can lay credit or blame for this election’s outcome at social media’s doorstep.

Finding that out would take a lot of research, an army of sociologists and access to a lot of Facebook data I don’t have.

But what is troubling about Zuckerberg’s post is his explanation for why Facebook isn’t tagging or penalising false news: “Identifying the truth is complicated. While some hoaxes can be completely debunked, a greater amount of content, including from mainstream sources, often gets the basic idea right but some details wrong or omitted.

“An even greater volume of stories express an opinion that many will disagree with and flag as incorrect even when factual.

“I am confident we can find ways for our community to tell us what content is most meaningful, but I believe we must be extremely cautious about becoming arbiters of truth ourselves.”

I agree it’s not easy for Facebook to tackle this problem. Worries of a politicised Facebook have dogged its steps before.

It’s true that Facebook alone shouldn’t define what is the truth.

But its prominence as a source of news gives it the responsibility to flag what is false. In his post, Zuckerberg essentially falls back on an old excuse when the fake news issue comes up: that Facebook is just a technology firm and a platform, an aggregator not up to the task of policing its users.

That excuse started out thin and is only getting thinner.

Sure, social media companies may want to deny that they are in the content business. But they are in it and only getting deeper.

Look at where these companies are investing: more photo-sharing, video, virtual reality. Those are all new media products that they say they need to survive.

Yet Zuckerberg is sticking to his tech company script, denying responsibility for what is posted on Facebook - even when it doesn’t make sense.

For example, to prove his point that fake news has little influence on Facebook, Zuckerberg pointed out that there isn’t that much of it on the network. “Of all the content on Facebook, more than 99% of what people see is authentic,” Zuckerberg said in his post. “Only a very small amount is fake news and hoaxes.”

That may be true, but doesn’t reflect how Facebook works.

A piece of writing crafted to generate clicks, likes and shares, by design, gets outsized attention.

Zuckerberg himself said in his post’s comments that specific users may see more false content on their feeds depending on how they and their friends use the site.

He also refuted the idea that slanted news could have on individual voters by citing Facebook’s own research, which shows the network exposes people to a broader set of ideas. Here again the issue is more nuanced.

Yes, a recent Pew Research Center study showed social media posts have changed some people’s minds.

But the same study showed that people often mute, block or otherwise filter out social media opinions they don’t want to hear.

In my mind, killing fake news online is a shared responsibility. Social media users should, of course, be critical of the things they read. But Facebook and other social media sites aren’t picking up their share of the work. Otherwise such obviously false information wouldn’t score so well with those so-called neutral algorithms.

And if these companies continue to push their media efforts, they can’t then shy away from the responsibilities that come with the business.

If quality and truth are what matter when it comes to content, that has to matter all of the time - not just when it’s uncontroversial.

* Tsukayama covers consumer technology for The Washington Post.

** The views expressed here are not necessarily those of Independent Media.

The Washington Post

Related Topics: