/ 8 February 2019

Fake news: Censorship’s no solution

Platforms also have the responsibility of identifying news sources and improve transparency.
Platforms also have the responsibility of identifying news sources and improve transparency. (Toby Melville/Reuters)

MEDIA

Today, debates about public issues play out on social media, people receive their news via digital platforms and politicians pitch their policies using this same media. The internet is our new public square.

In the public square of old, journalists and editors served as gatekeepers and acted as referees. Human news aggregators set the agenda and provided audiences with credible information and diverse views. We trusted them because of the professionalism and integrity of their editorial processes.

In the new public sphere, this model of journalism, and its role in sustaining democracy, has become obsolete. Traditional media no longer play a dominant gatekeeping and agenda-setting role. Fake news reaches myriad jurisdictions at once.

But so can public and private measures that censor speech. The challenge is to redefine the parameters of civil discourse in the new public sphere without restricting pluralism. Recent examples highlight the risk of throwing the baby out with the bathwater.

Despite the ominous headlines, the influence of fake news on political decision-making appears to be limited. The Reuters Institute for the Study of Journalism at the University of Oxford says the reach of such content is largely restricted to groups of believers seeking to reinforce their own views and prejudices. But that does not make digital deception any less dangerous. Fake news feeds — and is fed ­by —polarisation, and, paradoxically, the more it is discussed, the more disruptive it becomes.

Fake news undermines trust in all forms of media and reinforces the view that it is impossible to discern fact from fiction. When people don’t know what they can believe, journalists’ ability to police the powerful is weakened. This trend will only worsen as “deep-fake news” — bogus images and videos that appear real — becomes more ubiquitous.

Clearly, the vulnerabilities of the digital public sphere must be addressed. Some argue that the solution is to block questionable websites or demote search results. Facebook, for example, censors duplicitous posts and has created an election “war room” to fight disinformation. Google and Twitter have considered similar steps, and all three are under pressure to give authorities access to the private data of users who publish fake news or make defamatory statements. But we believe that these steps are deeply misguided.

At the heart of any strong demo-cracy is a political consensus and arbitration that depends on the public’s ability to debate and disagree. It is not up to private entities or public institutions to censor this process. Rather, we should strive to ensure that citizens have access to a broad array of opinions and ideas.

Freedom of expression and media freedom include the right to receive and impart information without interference. Studies show that most people prefer reliable and pluralistic news sources; the policymakers’ job is to enable them to realise this preference.

A March 2018 report to the European Commission by the high-level group on fake news and online disinformation, which one of us (De Cock Buning) chaired, offered a roadmap, and the commission’s recent action plan provides a good starting point. But more needs to be done.

There is no silver bullet to combat disinformation. Only multi-stakeholder approaches that spread responsibility across the news ecosystem, and take into account the fundamental rights involved, can provide adequate defences against disinformation.

For example, professional media must do more to guarantee the veracity of their coverage. Fact-checking technology can help, as long as it is kept free of political and economic influence.

Big Tech is starting to take responsibility by committing to a code of practice based on the 10 key principles from the high-level report. But Big Tech can contribute in other ways, for example by providing client-based interfaces for curating legitimate news, ensuring diversity in social media timelines, and prioritising the reposting of fact-checked information.

Platforms can also improve transparency in how they use data and code algorithms. Ideally, these algorithms should give consumers more control over editorial preferences and integrate editing and fact-checking applications developed by reliable media organisations.

Platforms must clearly identify news sources, especially paid political or commercial content. We also need more international collaboration and better jurisdictional rules to ensure that laws and regulations protect victims of fake and offensive news without restricting free speech or undermining the rights of whistle-blowers. These conflicts should not be legally settled where only one party has effective access to justice.

Platform companies should co-operate with schools, civil society groups and news organisations to strengthen public media literacy, to distinguish fake news from real.

Only consumers can marginalise fake news. We cannot allow private companies or governments to decide what people should know. The history of democracy is clear on this point: pluralism, not private or public censorship, is the best guarantor of truth. — © Project Syndicate

Madeleine de Cock Buning was chair of the European Commission’s high-level group on fake news and online disinformation. Miguel Poiares Maduro was a member of the European Commission’s group on media freedom and pluralism