Anthropologist Underground

Anthropologist Underground
October 13
I'm Terrie Torgersen Peterson. I hold a BA in Anthropology from the University of Wyoming. I've done archeological field work at Haluzta in Israel, San Juan River cliff dwellings in the American Southwest, and in the Big Horn Canyon in Wyoming. I'm currently a writer and stay-home mom to two gorgeous, laughing children. I enjoy exploring the intersection of science and culture and my own life as ethnography. I also write for and You can email me: anthropologistunderground [at] gmail [dot] com.


MARCH 2, 2011 4:24PM

When Corrections Fail

Rate: 1 Flag
A version of this article originally appeared at Does This Make Sense.
“There are certain points when debating with someone where you sort-of have to walk away. [...] ‘cause you’ll just basically end up getting a concussion.”
George Hrab

We’ve all been there. Someone we know, perhaps a relative, is nattering on about something abjectly false: something so wrong it’s akin to the claim that the Earth is flat. We try to correct the falsehood by presenting empirical, credible evidence for a spherical Earth. To our immense frustration, the friend (or relative) angrily defends flat Earth. It can seem as if presenting more evidence only strengthens their dedication to flat Earth.

Political scientist Brendan Nyhan studied the persistence of false political beliefs, and the results are fascinating.  He reported these results in a paper he coauthored with Jason Reifler, When Corrections Fail: The Persistence of Political Misperceptions. If you don’t have time to read it, there’s a fantastic interview with Nyhan on the Point of Inquiry podcast.  Here are the major points.

Misinformed vs. Uninformed
Nyhan and Reifler were targeting misinfomation in particular. Misinformation is a falsehood that people think they know to be true. Misinformed test subjects tended to substitute political ideology for facts. Both liberals and conservatives resisted corrective information that conflicted with their political opinions.

Interestingly, there seems to be a negative correlation between the level of confidence people have in their particular political beliefs and the likelihood that the beliefs are accurate. Elites, who sometimes have strong incentives to misinform the public, add false legitimacy and intensity to political misinformation.

In contrast to misinformation, Nyhan provides an example of simple ignorance of facts (I’m paraphrasing from memory) :   

Someone asks you to name the Chief Justice of the Supreme Court, and you don’t know but guess Scalia. When the person says, “No, actually, it’s Roberts,” you are fairly likely to just accept the correction because it’s not tied to a particular political ideology.  

Backfire Effect
An utterly fascinating discovery from the paper is the “backfire effect:”

“[...] individuals who receive unwelcome information may not simply resist challenges to their views. Instead, they may come to support their original opinion even more strongly.”

Although both liberals and conservatives had trouble letting go of misinformation, conservatives were the only ones, in this particular study, who exhibited the backfire effect.

More Detail
I emailed Nyhan with a few follow-up questions. Note than Nyhan speaks only for himself and not for his coauthor.

> 1.  Who are the elites?  Can you give me a couple of examples?  It seems to me that they hold sway both in originally providing misinformation to the public and later play an important role in publicly discrediting misinformation.  Who else in society might be a credible source of corrective information?

I'm not quite sure what you mean by "Who are the elites?" For the purposes of the article, we used "political elites" to refer to prominent political figures who were featured in the media, highly placed in political institutions and groups, etc. Elites often play a key role in getting political myths into circulation. It's certainly true that elites *could* also play a role in discrediting misinformation, but the problem is that it's usually one side's elites that try to do so, and their messages are often not persuasive to supporters of the other side. Ideally, we would like to have institutions and elites who are highly credible to both sides, but very few exist who are willing to try to arbitrate these sorts of disputes and set the record straight. In particular, the media is often reluctant to fact-check and is frequently accused of bias when it does so.

> 2.  Do you think the media's attempt to provide balanced reporting and presenting "both sides" of issues has the effect of implying false equivalency?  I'm thinking here of the role of the media in failing to discredit Andrew Wakefield and the shenanigans of the anti-vaccine movement.

Yes. I've been highly critical of the notion that the media has to be balanced with regard to the facts -- see in particular All the President's Spin, which is a book I co-wrote when I was co-editor of Spinsanity. A good critique of artificial balance is here. For an experimental study of its potentially harmful effects, see [this site].

> 3. How much motivated reasoning can be attributed to a basic human reaction to cognitive dissonance?  None of us wants to be wrong, but it seems to me that the current political climate prohibits the possibility of changing paradigms based on credible evidence.  If so, how can we change the climate to be more open to critical thinking?

It's a very good question. Cognitive dissonance is one mechanism for motivated reasoning, but there are others -- one possibility is that it's driven by emotional incongruence rather than contradictory cognitions (i.e. I like person X, and you say they're wrong). Changing the climate is difficult. We're working on understanding the conditions under which people will be more open to unwelcome information, but there are no easy answers. The main strategy I've advocated is to "name and shame" elites who promote misleading claims -- I discuss this issue more in an article on misinformation about health care reform and in a couple of interviews [here and here].

> 4. Given the potential harm of the backfire effect, do you recommend attempting to correct misperceptions IRL when you know going in that no amount of evidence will sway the proponent of misinformation?  What do you do in these situations?

Yes, I think it's still crucial to correct misperceptions in order to change the incentives for elites as I mentioned above, but we shouldn't mislead ourselves about the effectiveness of that strategy for the mass public. Corrections are most likely to be effective in dissuading elites from promoting misleading claims or stopping the spread of a myth early in its lifespan, not in changing the minds of people once a myth is circulating widely.

I’m hopeful that by discussing these and other barriers to critical thinking, we can all get a little closer to rationality ourselves and carry that out into the larger culture. I’m grateful to people like Brendan Nyhan who help us find the path to reason.

Your tags:


Enter the amount, and click "Tip" to submit!
Recipient's email address:
Personal message (optional):

Your email address:


Type your comment below:
Thanks for this informative post, and goodonya for following up with the authors. I regret to say I see no way out of our present dilemma regarding misinformation.

Once upon a time, most Americans got most of their information from three national networks peopled by thoroughly-trained professional journalists working for news depts which were relatively well-insulated from the accounting dept at networks.
Alas, those days are gone, and gone forever.

They have been largely replaced by newsreaders, bimbos and opinionists on cable, bloggers of every stripe on the Internet, and a network that is the political propaganda machine of the Rabid Right wing of the Republican Party. And alas, it is that latter from which a majority of Americans now get their "news". God help us.
Thanks Tom!

I seems much harder to get credible information now b/c we need to fact-check everything. Even things we want to believe!