New Survey: Scientists and Experts Have Different Views. Well, D’UH! But WHY!?

A new survey confirms that the lay public trusts science and scientists, but that scientists and the public have different views on specific issues. Unfortunately, the survey tells us how people feel, but not why, which we have to understand if we're going to try and narrow the perception gap between what the public believes and what the bulk of the scientific evidence indicates, a gap that cause all kinds of harm.

New Survey: Scientists and Experts Have Different Views. Well, D’UH! But WHY!?

The news and comment-o-sphere are paying a lot attention to a Pew Research survey, done in association with the American Association for the Advancement of Science (AAAS) that finds that people trust science and scientists, and applaud the benefits of science, but when asked about specific science issues, the public has different views than scientists do. While it is important to know that this perception gap exists, and that faith in science in general remains high, the news coverage and punditry and the survey itself are missing one thing; an explanation for what causes those different views. Why the gap?

That’s too bad, because the explanation for why different people see the same issues in such different ways could make surveys like this truly valuable, instead of telling us what we already know, albeit on new issues. And it’s really too bad because those explanations come from cognitive psychology research that began back in the 1970s, and which found precisely the same gap. Too bad the Pew research didn’t use that knowledge to get at explanations for the perception gap that could help identify ways it might be narrowed.

The research that helps explain this perception gap comes from several fields.

1. As the environmental movement fostered rising concern about industrial chemicals and nuclear radiation in the '70s, regulators and scientists in those fields were frustrated by Chemonoia and RadioPhobia. Yes, those technologies posed risks, but people’s fears far exceeded the likelihood and severity of those risks. Nuclear and chemical regulators and scientists wanted to know why.

So pioneers like Paul Slovic and Baruch Fischhoff and Sarah Lichtenstein and others initiated research into the psychology of risk perception (risks to health, as opposed to the study of people's irrational choices about monetary risks, which was already well underway). And they found what the Pew survey has confirmed. Experts and "regular people" see science issues differently. In fact, Slovic et al. found that different groups of regular people saw things differently too.

In a 1979 paper in the journal Environment titled "Rating the Risks," the researchers asked four groups of people to rank the perceived risk of death from 30 activities and technologies; members of the League of Women Voters, college students (always handy for psych experiments), members of business and professional clubs, and scientific experts in risk assessment. They found the same gap the Pew study found, albeit on different issues.  

Courtesy: Paul Slovic, The Perception of Risk

Their research went on to discover why those gaps exist. They revealed a set of psychological characteristics, "fear factors" that make some risks feel scarier and some risks feel less scary. Those factors help explain some of the Pew findings. For example, Slovic et al. established that we’re more afraid of human-made risks than natural ones, which helps explain the gap regarding the safety of human-made GMO foods...


Survey images courtesy: Pew Research Center

... and the gap about the safety of eating food grown with pesticides.


We’re more afraid of risks imposed on us than those we take voluntary, which helps explain the gap on the Pew question about childhood vaccination (note that it asks whether the program should be required.)

Slovic and colleagues also discovered that while experts’ views are affected by these emotional filters too, the views of people with more information and a scientific bent to think about things critically aligned more closely with the evidence. Which helps explain the gap between public and "scientific expert" views in general.

2. Slovic was a colleague for a time of Daniel Kahneman and the cognitive psychology researchers who were studying why people made monetary choices that seemed to defy rationality. They discovered a broad set of heuristics and biases — mental shortcuts — we use to quickly and subconsciously turn partial information into our judgments and decisions.

Kahneman et al. learned, for example, that once we’ve developed an opinion or view, we tend to stick with it, fitting new information into what we already believe and rejecting information that conflicts. This is known as confirmation bias. You can see how that might have played a role in the answer Pew got on evolution, which a third of Americans in this survey still deny.


They also discovered the bias that "framing" imposes on people’s views. Essentially, how we first hear an issue described shapes our initial views, and then we stick with those views. The public’s picture of GMOs has largely been painted by fear-mongering opponents. Small wonder that 63% of the survey respondents say GMO food is unsafe to eat, while 88% of the experts, whose views have been framed more by the scientific evidence rather than the alarms of advocates (in 15 years, no credible studies have ever found human health harm from GMO food), say it's safe.

3. A modern theory of risk perception that builds on Slovic’s work, Cultural Cognition by Dan Kahan, has found that we tend to see the facts through the lenses of the groups we belong to, based on the values shared by those groups. This is particularly true of issues that have become polarized, totems of tribal values. So when Pew asked about climate change...


... or support for offshore drilling...


... those answers in part reflected people’s feelings about environmental issues and their underlying political values.

This all boils down to one simple thing. Cognition is subjective. We all perceive information, about risk or science or anything, through instinctive and emotional filters that help us subconsciously sense how that information feels. Slovic and Melissa Finucane called this the Affect Heuristic, the fast and mostly subconscious mental process that takes what we know (which is usually not a lot, especially for non-experts) and assesses it based on our experiences and life circumstances and education, and those psychological "fear factors," and cognitive processes like confirmation bias and framing, and the values of the groups with which we most closely identify... and from that complex set of influences we come up with our views and opinions and perceptions, and what we tell pollsters.

These insights seem to have been ignored by the survey, which asked people how they felt, but not why. (Pew says a later survey will dive into demographics, but those don't really reveal the cognitive psychological roots of people's views either.) These explanations also seem to have been ignored by the scientists surveyed, who offered trite, old explanations for the perception gap, explanations that don’t display a lot of... well... expertise.

The perception gap identified in the Pew survey is NOT about people not knowing enough, or expecting solutions too quickly. This is the old-school, post-Enlightenment belief that once people are educated enough, we can all get it "right." Which is wrong. And it's a problem, because until we realize that the perception of risk (or anything) is not just a matter of the facts and objective rationality, but also a matter of emotions and instincts, we won’t be able to narrow these perception gaps. And that’s a problem because as individuals and together as a society, we make a lot of dumb and dangerous choices when our perceptions don’t match the evidence.

(top image courtesy

A brief history of human dignity

What is human dignity? Here's a primer, told through 200 years of great essays, lectures, and novels.

Credit: Benjavisa Ruangvaree / AdobeStock
Sponsored by the Institute for Humane Studies
  • Human dignity means that each of our lives have an unimpeachable value simply because we are human, and therefore we are deserving of a baseline level of respect.
  • That baseline requires more than the absence of violence, discrimination, and authoritarianism. It means giving individuals the freedom to pursue their own happiness and purpose.
  • We look at incredible writings from the last 200 years that illustrate the push for human dignity in regards to slavery, equality, communism, free speech and education.
Keep reading Show less

Astrophysicists: Gamma-ray jets exceed the speed of light

Scientists find that bursts of gamma rays may exceed the speed of light and cause time-reversibility.

An artist's drawing of a particle jet emanating from a black hole at the center of a blazar.

Credit: DESY, Science Communication Lab (used with permission by Astronomy Picture of the Day, which is co-managed by Robert Nemiroff at Michigan Tech).
Surprising Science
  • Astrophysicists propose that gamma-ray bursts may exceed the speed of light.
  • The superluminal jets may also be responsible for time-reversibility.
  • The finding doesn't go against Einstein's theory because this effect happens in the jet medium not a vacuum.
Keep reading Show less

Is free will an illusion?

Philosophers have been asking the question for hundreds of years. Now neuroscientists are joining the quest to find out.

Sponsored by John Templeton Foundation
  • The debate over whether or not humans have free will is centuries old and ongoing. While studies have confirmed that our brains perform many tasks without conscious effort, there remains the question of how much we control and when it matters.
  • According to Dr. Uri Maoz, it comes down to what your definition of free will is and to learning more about how we make decisions versus when it is ok for our brain to subconsciously control our actions and movements.
  • "If we understand the interplay between conscious and unconscious," says Maoz, "it might help us realize what we can control and what we can't."

The Arecibo telescope has collapsed: A look at its 57-year history

Puerto Rico's iconic telescope facilitated important scientific discoveries while inspiring young scientists and the public imagination.

The Arecibo radio telescope

Credit: dennisvdwater via Adobe Stock
Surprising Science
  • The Arecibo Observatory's main telescope collapsed on Tuesday morning.
  • Although officials had been planning to demolish the telescope, the accident marked an unceremonious end to a beloved astronomical tool.
  • The Arecibo radio telescope has facilitated many discoveries in astronomy, including the mapping of near-Earth asteroids and the detection of exoplanets.
Keep reading Show less
Technology & Innovation

DeepMind AI solves 50-year-old biology problem in breakthrough advance

The Google-owned company developed a system that can reliably predict the 3D shapes of proteins.

Scroll down to load more…