The widespread existence of fake news
spread by bots and hyperpartisan trolls
across social media only came to light as a result of the shock election of Donald Trump in the 2016 election. The congressional hearings this month revealed that the problem extended to potentially millions of American voters being exposed to political messaging and ads orchestrated by Russia.
While lawmakers consider what to do to curb the possibility of such interference happening again, social media companies have started to change
their own content policies. This raises an urgent question as to whether our media environment needs more regulation and if so, who should do it, the government or the vast privately owned technology companies?
This week, social media platform Twitter started deleting
the "blue check" verification marks from right wing accounts such as
that of white nationalist Richard Spencer. Biz Stone, one of the founders of the social network, tweeted
: "Verification was invented to authenticate but became perceived as an endorsement. We should have acted earlier but we are now working on a new program."
The removal of a blue check from a Twitter account hardly counts as censorship, although there were plenty of opinions
from members of the far right on the platform to suggest they think it does. The episode marks just one of a number of changes we are likely to see from social platforms as they adjust their internal policies to control more of the content published on their sites.
Facebook recently experimented by removing all news sources from its news feed in six smaller markets: Slovakia, Guatemala, Serbia, Bolivia, Sri Lanka and Cambodia. The result was that traffic to the news sites plummeted, alarming American publishers about the possibility of these changes being extended to their market. Google also changed its algorithm for news to stop sites like 4Chan appearing at the top of news results with false stories as it did after the Las Vegas mass shooting. The change
supposedly purged "fake news" culprits but also raised serious concerns
about how Google and Facebook help determine what news people read and whether they should have that power.
News organizations are not the only people concerned about this type of exercise of covert power. In August, after violent far-right protests in Charlottesville, Virginia, the neo-Nazi website The Daily Stormer was essentially removed from the internet when a series of hosting and platform services, including Google, Amazon and Facebook, refused to support it. Cloudflare, a web security service that protects around 10% of internet traffic from hostile attack, also withdrew its services. Cloudflare chief executive Matthew Prince said
at the time that his decision to remove the Daily Stormer from the internet was potentially "dangerous," adding
, "Without a clear framework as a guide for content regulation, a small number of companies will largely determine what can and cannot be online."
Few people would argue that the Daily Stormer's vile brand of content is missed from the internet, but Prince's point remains the central concern of many legitimate journalism and news organizations. Their worry is that platforms can and will exercise an increasing amount of economic or editorial power over their existence without a proper regulatory framework or transparency.
With Facebook and Google alone forecast to take 63% of US digital advertising revenue in 2017, according to
market research company eMarketer, the ability for other types of advertising-based content businesses to thrive on the web is dwindling.
In the same week that the tech companies appeared on the Hill to testify, billionaire Republican businessman Joe Ricketts abruptly shut down
the unprofitable but respected local website networks DNAInfo and Gothamist. After he described unions in September as "corrosive" to business success, it was widely assumed that part of his reasoning
was the unionization of the workforce at the h