The Neuroscience of Why People Won’t Budge on Their Beliefs
When it comes to climate change, gun control, and vaccinations, facts don’t change people’s minds—but there is one technique that might.
Tali Sharot: So most of us think that information is the best way to convince people of our truth, and in fact it doesn’t work that well.We see that all the time. We see it with climate change, where there’s tons of data suggesting that climate change is man-made but about 50 percent of the population doesn’t believe it, or with people arguing about things like how many people were in the presidential inauguration. So we have facts but people decide which facts they want to listen to, which facts they want to take and change their opinions, and which they want to disregard. And one of the reasons for this is when something doesn’t conform to what I already believe, what people tend to do is either disregard it or rationalize it away; because information doesn’t take into account what makes us human, which is our emotions, our desires, our motives and our prior beliefs.
So for example, in one study my colleagues and I tried it to see whether we could use science to change people’s opinions about climate change. The first thing we did was ask people, “Do you believe in man-made climate change? Do you support the Paris Agreement?” And based on their answers we divided them into the strong believers and the weak believers. And then we gave them information.
For some people we said that scientists have reevaluated the data and now conclude that things are actually much worse than they thought before, that the temperature would rise by about seven degrees to ten degrees. For some people we said the scientists have reevaluated the data and they now believe that actually this situation is not as bad as they thought, it’s much better, and the rise in temperature would be quite small.
And what we found is that people who did not believe in climate change, when they heard that the scientists are saying, “Actually it’s not that bad,” they changed their beliefs even more in that direction, so they became more extremist in that direction, but when they heard that the scientists think it’s much worse they didn’t nudge.
And the people who already believe that climate change is man-made, when they heard that scientists are saying things are much worse than they said before, they moved more in that direction, so they became more polarized, but when they heard scientists are saying it’s not that bad they didn’t nudge much. So we gave people information and as a result it caused polarization, it didn’t cause people to come together.
So the question is, what’s happening inside our brain that causes this? And in one study my colleagues and I scanned brain activity of two people who were interacting, and what we found was when those two people agreed on a question that we gave them, the brain was really encoding what the other person was saying, the details that they gave; but when the two people disagreed it looked metaphorically as if the brain was switching off and not encoding what the other person was saying.And as a result when the two agreed they became even more confident, but when they disagreed there wasn’t as much of a change in their confidence in their own view.
What has been shown by Kahan and colleagues from Yale University is that the more intelligent you are the more likely you are to change data at will. So what they did is they first gave participants in their experiment analytical and math questions to solve, and then they gave them data about gun control: is gun control actually reducing violence? And they found that more “intelligent” people actually were more likely to twist data at will to make it conform to what they already believed.So it seems that people are using their intelligence not necessarily to find the truth, but to take in the information and change it to conform to what they already believe. So that suggests that just giving people information without considering first where they’re coming from may backfire at us, but we don’t always need to go against someone’s conviction in order to change their behavior, and let me give you an example.
So this is a study that was conducted at UCLA where what they wanted to do is convince parents to vaccinate their kids. And some of the parents didn’t want to vaccinate their kids because they were afraid of the link with autism.So they had two approaches, first they said, “Well the link with autism is actually not real, here’s all the data suggesting there isn’t a link between vaccines and autism.” And it didn’t really work that well. But instead they used another approach. So instead of going that way they used another approach, which was: let’s not talk about autism, we don’t necessarily need to talk about autism to convince you to vaccinate your kids. Instead they said, “Well look, these vaccines protect kids from deadly diseases, from the measles,” and they showed them pictures of what the measles are. Because in this argument about vaccines people actually forgot what the vaccines are for, what are they protecting us from. And they highlighted that and didn’t necessarily go on to discuss autism. That had a much better outcome. The parents were much more likely to say, “Yes we are going to vaccinate our kids.”
So the lesson here is that we need to find the common motive. The common motive in this case was the health of the children, not necessarily going back to the thing that they were arguing about, that they disagreed about.
If you want someone to see an issue rationally, you just show them the facts, right? No one can refute a fact. Well, brain imaging and psychological studies are showing that, society wide, we may be on the wrong path by holding evidence up as an Ace card. Neuroscientist Tali Sharot and her colleagues have proven that reading the same set of facts polarizes groups of people even further, because of our in-built confirmation biases—something we all fall prey to, equally. In fact, Sharot cites research from Yale University that disproves the idea that the social divisions we are experiencing right now—over climate change, gun control, or vaccines—are somehow the result of an intelligence gap: smart people are just as illogical, and what's more, they are even more skilled at skewing data to align with their beliefs. So if facts aren't the way forward, what is? There is one thing that may help us swap the moral high ground for actual progress: finding common motives. Here, Sharot explains why identifying a shared goal is better than winning a fight. Tali Sharot's newest book is out now: The Influential Mind: What the Brain Reveals about Our Power to Change Others.
Swipe right to make the connections that could change your career.
Swipe right. Match. Meet over coffee or set up a call.
No, we aren't talking about Tinder. Introducing Shapr, a free app that helps people with synergistic professional goals and skill sets easily meet and collaborate.
Even some teachers suffer from anxiety about math.
I teach people how to teach math, and I've been working in this field for 30 years. Across those decades, I've met many people who suffer from varying degrees of math trauma – a form of debilitating mental shutdown when it comes to doing mathematics.
She met mere mortals with and without the Vatican's approval.
- For centuries, the Virgin Mary has appeared to the faithful, requesting devotion and promising comfort.
- These maps show the geography of Marian apparitions – the handful approved by the Vatican, and many others.
- Historically, Europe is where most apparitions have been reported, but the U.S. is pretty fertile ground too.
The legacy of Felix Dzerzhinsky, who led Soviet secret police in the "Red Terror," still confounds Russia.
- Felix Dzerzhinsky led the Cheka, Soviet Union's first secret police.
- The Cheka was infamous for executing thousands during the Red Terror of 1918.
- The Cheka later became the KGB, the spy organization where Russia's President Putin served for years.
SMARTER FASTER trademarks owned by The Big Think, Inc. All rights reserved.