As confrontations between Black Lives Matter protesters and police erupted across the country earlier this month, some Oregonians, mostly older people, saw a Facebook ad pushing a headline about how a Republican politician Wants Martial Law To Control The Obama-Soros Antifa Supersoldiers.
Needless to say, there was no army of left-wing supersoldiers marching across Oregon, nor were former president Barack Obama and billionaire George Soros known to be funding anything antifa-related. And the politician in question didnt actually say there were supersoldiers. The headline, originally from the often-sarcastic, progressive blog Wonkette, was never meant to be taken as straight news.
The whole thing was a mishap born of the modern news age, in which what headlines you see is decided not by a hard-bitten front-page editor but instead by layers of algorithms designed to pick whats news and who should be shown it. This system can work fine, but in this instance it fed into a maelstrom of misinformation that was already inspiring some westerners to grab their guns and guard their towns against the largely non-existent threat of out-of-town antifa troublemakers.
This was just one headline that fed into a sense of paranoia reinforced by rumors from many sources. But deconstructing exactly how it came about provides a window into how easy it is for a fringe conspiracy theory to accidentally slip into the ecosystem of mainstream online news.
The trouble started when SmartNews picked up Wonkettes mocking story. SmartNews is a news aggregation app that brings in users by placing nearly a million dollars worth of ads on Facebook, according to Facebooks published data. According to the startups mission statement, its algorithms evaluate millions of articles, social signals and human interactions to deliver the top 0.01% of stories that matter most, right now.
The company, which says that news should be impartial, trending and trustworthy, usually picks ordinary local news headlines for its Facebook adsmaybe from your local TV news station. Users who install the app get headlines about their home area and topics of interest, curated by SmartNewss algorithms. This time, however, the headline was sourced from Wonkette, in a story mocking Jo Rae Perkins, Oregons Republican U.S. Senate nominee, who has sparked controversy for her promotion of conspiracy theories.
In early June, as protests against police violence cropped up in rural towns across the country, Perkins recorded a Facebook Live video calling for hard martial law to squash the antifa thugs supposedly visiting various towns in Oregon. She also linked protesters, baselessly, to common right-wing targets: Many, many people believe that they are being paid for by George Soros, she said, and this is the army that Obama put together a few years ago.
Perkins never said supersoldierthe term is apparently a Twitterverse joke, in this case added to its headline by Wonkette to mock Perkinss apparent fear of protesters. To someone familiar with its deadpan sardonic style, seeing the hyperbolic headline on Wonkettes website wouldnt raise an eyebrowregular readers would know Wonkette was mocking Perkins. But its 2020, and even insular blog headlines can travel outside their readers RSS feeds and wend their way via social media into precincts where Wonkette isnt broadly known. SmartNews, when it automatically stripped Wonkettes headline of its association with Wonkette and presented it neutrally in the ad, epitomized that phenomenon.
SmartNewss algorithms picked that headline for ads to appear on the Facebook feeds of people in almost every Oregon county, with a banner like Charles County news matching the name of the county where the ad was shown. Its a strategy that the company uses thousands of times a day.
SmartNews vice president Rich Jaroslovsky said that in this case its algorithms did nothing wrong by choosing the tongue-in-cheek headline to show to existing readers. The problem, he says, was that the headline was shown to the wrong people.
SmartNews, he said, focuses a huge amount of time, effort and talent on its algorithms for recommending news stories to users of SmartNewss app. Those algorithms would have aimed the antifa story at people who presumably have a demonstrated interest in the kind of stuff Wonkette specializes in. To those readers and in that context, he said, the story wasnt problematic.
The problems occurred when it was pulled out of its context and placed in a different one for Facebook advertising that isnt aimed by any criterion other than geography. Obviously, this shouldnt have happened, and were taking a number of steps to make sure we address the problems you pointed out, Jaroslovsky said.
Jaroslovsky said Wonkette stories wouldnt be used in ads in the future.
SmartNews targets its ads at people in particular geographic areasin this case, 32 of Oregons 36 counties.
But Facebook had other ideas: Its algorithms chose to show the antifa supersoldiers ad overwhelmingly to people over 55 years old, according to Facebooks published data about ads that it considers political. Undoubtedly, many of those viewers ignored the ad, or werent fooled by it, but the demographic Facebook chose is a demographic that a recent New York University study showed tends to share misinformation on social media more frequently.
This choice by Facebooks algorithms is powerful: An academic paper showed that Facebook evaluates the content of ads and then sometimes steers them disproportionately to users with a particular gender, race, or political view. (The paper didnt study age.)
Facebook also doesnt make it possible to know exactly how many people saw SmartNewss antifa supersoldiers ad. The companys transparency portal says the ad was shown between 197 and 75,000 times, across about 75 variations (based on Android and iPhone and number of counties). Facebook declined to provide more specific data.
Facebook doesnt consider the ads to have violated the companys rules. Ads are reviewed primarily by automated mechanisms, Facebook spokesperson Devon Kearns told The Markup, so its unlikely that a human being at Facebook saw the ads before they ran. However, ads that run with satirical headlines that are taken out of context are eligible to be fact-checked, Kearns said, and ads found to be false are taken down. (Usually, satire and opinion in ads are exempt from being marked as misinformation under Facebooks fact-checking policy, unless theyre presented out of context.)
Wonkette publisher Rebecca Schoenkopf told The Markup she wasnt aware SmartNews was promoting her sites content with Facebook ads but wasnt necessarily against it. In theory, at least, it could have the effect of drawing more readers to her site.
Ironically, she says, Wonkette has a limited Facebook presence. In recent years, the reach of Wonkettes posts on the platform had dwindled to almost nothing.
Following the chain of information from Perkinss Facebook video all the way to the SmartNews ad makes it easy to see how a series of actors took the same original piece of contenta video of Perkins espousing conspiracy theoriesand amplified it to suit their own motives. Each of those links created a potential for misinformation, where the context necessary for understanding could be stripped away.
Lindsay Schubiner, a program director at the Portland, Ore., based Western States Center, which works to counter far-right extremism in the Pacific Northwest, told The Markup that, while social media has had a democratizing effect on information, its also been an ideal format for spreading misinformation.
The SmartNews adinadvertently or notjoined in a chorus of false, misleading and racist posts from white nationalists in response to Black Lives Matter protests, Schubiner wrote in an email. Many of these posts traded in anti-semitism, which has long been a go-to response for white nationalists looking to explain their political losses.
In this case the ad, she said, potentially promulgated the common right-wing trope that Soros, who is Jewish, funds mass protests for left-wing causes.
These bigoted conspiracy theories have helped fuel a surge in far-right activity and organizing, she continued. Its certainly possible that the ads contributed to far-right organizing in Oregon in response to false rumors about anti-fascist gatherings in small towns.
Those conspiracy theories have had real-world consequences. Firearm-toting residents in nearby Washington State harassed a multiracial family who were camping in a converted school bus, trapping them with felled trees, apparently mistakenly believing them to be antifa. The family was able to escape only with the help of some local high school students armed with chainsaws to clear their way to freedom.
This article was originally published on The Markup by Jeremy B. Merrill and Aaron Sankinand and was republished under the Creative Commons Attribution-NonCommercial-NoDerivatives license.
Pssst, hey you!
Do you want to get the sassiest daily tech newsletter every day, in your inbox, for FREE? Of course you do: sign up for Big Spam here.read more
The anatomy of a fake news headline
