The “fake news” phenomenon was on vivid display in social media during the bitter 2016 presidential campaign. A typical example of the nonsense was a story, shared a million times on Facebook, stating falsely that Pope Francis had endorsed the Republican candidate, Donald Trump. BuzzFeed reported that during the last three months of the campaign, more Facebook members shared, “liked” or commented on the top 20 fake news stories than engaged with the 20 most important news stories on real news websites.
Bogus information has always been a threat to a democratic society, dependent as it is on an informed citizenry. As falsehoods spread at the speed of light in the digital age, news consumers more than ever must equip themselves to distinguish fact from fiction. Lacking news literacy skills, they might be tempted to disbelieve even news that is reliably reported.
How and why are these stories developed? For some purveyors of phony news, it’s all about the cash: They collect payments from advertisers whose messages are embedded in their stories on Facebook and Twitter – each of which, according to a Pew Research Center survey in 2015, is a source of news for 63 percent of Americans.
Fabricators typically employ a shrewd understanding of their audience. Those insights give them a good sense of what kinds of lies will be eagerly shared, driving up their profits.
Never miss a local story.
Terrence McCoy of The Washington Post tracked down two of them — a team of twentysomethings who post made-up stories from laptops in their home in Long Beach, California.
As McCoy described it, one of the men, Paris Wade, awakens from a nap and decides do a story on President Obama’s advocacy for the Trans-Pacific Partnership. He composes a headline 3 “CAN'T TRUST OBAMA” — and in “ten minutes and nearly 200 words … he is done with a story that is all opinion, innuendo and rumor.” Then he publishes the story to his website (LibertyWritersNews.com), puts it on Facebook, and watches on a monitor as the post goes viral.
The fabricators’ success underscores evidence that many consumers who retweet or share fake news do so not necessarily because they think the stories are true, but because the stories confirm their own biases.
As an example, Sapna Maheshwari of The New York Times investigated a tweet about paid demonstrators being bused to Austin, Texas, to join a protest against Trump’s election. Maheshwari traced the tweet to Eric Tucker, a 35-year-old co-founder of an Austin marketing company. He wrote that Tucker took pictures of buses he happened to see downtown, then “saw reports of the protests against Mr. Trump in the city and decided the two were connected.”
Tucker’s tweet was shared at least 16,000 times on Twitter and more than 350,000 times on Facebook. The president-elect himself tweeted a complaint about “professional protesters, incited by the media.” But the buses had no connection to the protest; they had been rented to bring people to a software conference. Maheshwari said that when Tucker learned he was wrong, he deleted the original tweet and later posted an image of it stamped “false.” By way of explanation, Tucker told the Times reporter he is “a very busy businessman and I don’t have time to fact-check everything that I put out there.”
Another source of bogus stories is the use of automated bots to generate tweets, which often include links to fake news. “Twitter bots are computer programs based on artificial intelligence that are able to mimic tweets by human beings,” Jeff Nesbit, a former journalist and onetime director of legislative and public affairs at the National Science Foundation, wrote in an article for USNews.com. “They are able to retweet and post based on phrases or key words. They are sophisticated enough that it can be difficult to detect whether something has been generated by a human being or an automated robot.”
And why is this important? A study by two University of Southern California researchers, Alessandro Bessi and Emilio Ferrara, showed that almost 20 percent of the 20.7 million political tweets in the month before the presidential election came from 400,000 bots. Their conclusions: “First, influence can be redistributed across suspicious accounts that may be operated with malicious purposes; second, the political conversation can become further polarized; third, the spreading of misinformation and unverified information can be enhanced.”
Illustrating how the reading public is susceptible to fake news, a survey of 7,800 teenagers by Stanford University, conducted between January 2015 and June 2016, showed that over 80 percent had difficulty judging the credibility of news sources. Professor Sam Wineburg, the study’s lead author, described one phase of the study in an interview on NPR: Students were shown a picture of daisies that appeared to be deformed, accompanied by a claim on the web that they were the result of the nuclear disaster at Fukushima in Japan. The students were asked, “Does this photograph provide proof that the kind of nuclear disaster caused these aberrations in nature?”
Although the photograph had no attribution or other documentation, Wineburg said, nearly all the students “had an extremely difficult time making that determination. They didn’t ask where it came from. They didn’t verify it. They simply accepted the picture as fact.” Wineburg said educational programs are needed to help people determine whether information on the internet is believable or not. Ascertaining credibility is, he said, “a new basic skill in our society.”
So how do you determine what’s fake and what’s real? The News Literacy Project has been working to address misinformation — including viral rumors, hoaxes, propaganda and other forms of fake information – in classrooms for years. NLP’s new checkology virtual classroom includes a lesson focused on helping students “immunize” themselves from viral rumors and a “Check Tool” that helps student build the critical thinking skills needed to assess the credibility of any news and information they encounter online. Other NLP resources include “Ten Questions for Fake News Detection,” “Virology Report: Online Rumor Breakdown” and “Photo Fact-Checking in the Digital Age.”
In addition, Lori Robertson and Eugene Kiely of FactCheck have posted guidelines for “how to spot fake news.” Joyce Valenza of School Library Journal has published a comprehensive analysis of fake news that includes “a literacy toolkit for a ‘post-truth' world.” Richard Hornik, who teaches news literacy at Stony Brook University, offers “7 ways to spot and debunk fake news.”
Gene Foreman is a former managing editor at The Philadelphia Inquirer. He wrote this for the News Literacy Project.