Big ideas.
Once a week.
Subscribe to our weekly newsletter.
Liberal and conservative brains react to charged words differently
Partisanship can now be seen in brain scans.

- A new study shows brain activity differs between liberals and conservatives when they watch political videos.
- Brain activity differed between partisans when words tied to emotions, morality, or threats were used.
- The findings could help us understand how partisans process information, perhaps leading to new ways to bridge the divide.
People are somewhat politically polarized these days. While the United States' polarization gets most of the press, increasing polarization is also causing problems in the United Kingdom, Turkey, Poland, Brazil, India, and a variety of other countries around the world. Attempts to understand why are endless, with hypotheses being offered from various schools of sociology, political science, and philosophy.
Recently, scientists investigating partisanship have peered into the neurological basis for our partisan alignments, if any. Their discoveries suggest that partisans' brains may process information differently than those of non-partisans or members of another group.
Adding evidence to this way of thinking is a new study out of UC Berkeley, which finds that liberal and conservative brains react differently to political trigger words.
The study, published in the Proceedings of the National Academy of Sciences, compared the stated political opinions of three dozen test subjects to their brain wave patterns while they watched videos about immigration policy.
The researchers, led by Dr. Yuan Chang Leong, determined the participants' ideologies by asking them how much they agreed or disagreed with proposed legislation. Each response was given a score, with lower values attached to stances considered liberal in the United States.
One such question was: "Would you support legislation that funds a wall along the US-Mexico border to reduce illegal immigration?" Those who strongly agreed were given a high score while those who strongly disagreed got a low score. The scores earned over six questions were used to place the participants on a scale from left to right. The questions had previously been tested on 300 people who identified as liberals, conservatives, or centrists to assure their accuracy.
The test subjects then watched the previously mentioned political videos.
While the parts of the brain dedicated to collecting sensory information reacted similarly for all of the test subjects, the dorsomedial prefrontal cortex, a part of the prefrontal cortex that deals with matters of identity, narratives, and morality, of liberals and conservatives reacted at different times.
Using an fMRI, the researchers saw that neural responses differed between liberals and conservatives as the videos' messages changed. More specifically, the brain's activity was stimulated by its response to messages concerning morality, emotions, or threats. The reactions to these terms were the points of greatest divergence.
A morality based message might be something like, "What are the fundamental ethical principles that are the basis of our society? Do no harm, and be compassionate, and this federal policy violates both of these principles." A threat-based statement might resemble, "I think it's very dangerous, because what we want is cooperation amongst the cities and the federal government to ensure that we have safety in our communities, and to ensure that our citizens are protected."
Participants were asked to rate how much they agreed with each video and how likely they were to change their mind on anything after watching them. Curiously, the closer the subject's brain activity was to that of the "average" liberal or conservative of the study, the more likely they were to report that a video supporting those policies could make them change their mind.
Dr. Leong summarized the findings by saying:
"Our study suggests that there is a neural basis to partisan biases, and some language especially drives polarization. In particular, the greatest differences in neural activity across ideology occurred when people heard messages that highlight threat, morality and emotions."
This study suggests that partisanship impacts how our brains process specific terms and that political messaging relying on threat-based or ethics-based language cause partisans to interpret the message in very different ways. This processing also means that people with similar brains to other partisans are likely to be convinced by similar messages.
The location of the differences in brain function, in the later, higher-level processing department of the brain rather than in the earlier, sensory detecting department, implies that polarization does not affect sensory processing. Additionally, the results do not imply that these effects are hardwired in our brains.
How does it interact with what we already know?
These findings can be added to the list of studies that show that our political alignments might have something to do with how our brains process information. Non-partisans, often suggested to not be a real group of people, have measurably different brain activity than partisans. Brain scans show Democrats and Republicans used different parts of their brains when playing a gambling game.
Dr. Leong hopes to use this information to build better models of how the brain processes political information. Perhaps someday, these models can help us understand how to talk to each other without using these trigger words.
Politics is becoming increasingly polarized in several countries all around the world. The causes for it are still up for debate, and ways to help narrow the gaps between people are still being investigated. An increasing number of studies suggest that some of it comes down to how our brains function.
While the idea of polarization being tied to how our brains work probably won't come as a comfort to most people, the ability to identify precisely what is happening when people have polarized reactions is a step forward, as it offers a chance to understand what the other side is doing when we disagree. Perhaps someday soon, this will translate to better ways to reach across the aisle and more productive conversations informed by neuroscience.
- How to get smarter about the partisan press - Big Think ›
- Non-partisan brains differ from those of partisans - Big Think ›
- How to get smarter about the partisan press - Big Think ›
A Cave in France Changes What We Thought We Knew About Neanderthals
A cave in France contains man’s earliest-known structures that had to be built by Neanderthals who were believed to be incapable of such things.
In a French cave deep underground, scientists have discovered what appear to be 176,000-year-old man-made structures. That's 150,000 years earlier than any that have been discovered anywhere before. And they could only have been built by Neanderthals, people who were never before considered capable of such a thing.
This is going to force a major shift in the way we see these early hominids. Researchers had thought that Neanderthals were profoundly primitive, and just barely human. This cave in France's Aveyron Valley changes all that: It's suddenly obvious that Neanderthals were not quite so unlike us.
According to The Atlantic, Bruniquel Cave was first explored in 1990 by Bruno Kowalsczewski, who was 15 at the time. He'd spent three years digging away at rubble covering a space through which his father felt air moving.
Some members of a local caving club managed to squeeze through the narrow, 30-meter long tunnel Kowalsczewski had dug to arrive in a passageway. They followed it past pools of water and old animal bones for over 330 meters before coming into a large chamber and a scene they had no reason to expect: Stalagmites that someone had broken into hundreds of small pieces, most of which were arranged into two rings—one roughly 6 meters across, and one 2 meters wide—with the remaining pieces stacked into one of four piles or leaning against the rings. There were also indications of fires and burnt bones.
Image source: Etienne FABRE - SSAC
What the?
A professional archeologist, Francois Rouzaud, determined with carbon dating that a burnt bear bone found in the chamber was 47,600 years old, which made the stalagmite structures older than any known cave painting. It also put the cave squarely within the age of the Neanderthals since they were the only humans in France that early. No one had suspected them of being capable of constructing complex forms or doing anything that far underground.
After Rouzard suddenly died in 1999, exploration at the cave stopped until life-long caver Sophie Verheyden, vacationing in the area, heard about it and decided to try and uranium-date the stalagmites inside.
The team she assembled eventually determined that the stalagmites had been broken up by people 176,000 years ago, way farther back even than Rouzard had supposed.
There weren't any signs that Neanderthals lived in the cave, so it's a mystery what they were up to down there. Verheyden thinks it's unlikely that a solitary artist created the tableaux, and so an organized group of skilled workers must've been involved. And “When you see such a structure so far into the cave, you think of something cultural or religious, but that's not proven," Verheyden told The Atlantic.
Whatever they built, the Bruniquel Cave reveals some big surprises about Neanderthals: They had fire, they built things, and likely used tools. Add this to recent discoveries that suggest they buried their dead, made art, and maybe even had language, and these mysterious proto-humans start looking a lot more familiar. A lot more like homo sapiens, and a lot more like distant cousins lost to history.
Paul Hudson/Flickr
Psychopath-ish: How “healthy” brains can look and function like those of psychopaths
A recent study used fMRI to compare the brains of psychopathic criminals with a group of 100 well-functioning individuals, finding striking similarities.
Obscure freaky smiling psycho man
- The study used psychological inventories to assess a group of violent criminals and healthy volunteers for psychopathy, and then examined how their brains responded to watching violent movie scenes.
- The fMRI results showed that the brains of healthy subjects who scored high in psychopathic traits reacted similarly as the psychopathic criminal group. Both of these groups also showed atrophy in brain regions involved in regulating emotion.
- The study adds complexity to common conceptions of what differentiates a psychopath from a "healthy" individual.
When considering what precisely makes someone a psychopath, the lines can be blurry.
Psychological research has shown that many people in society have some degree of malevolent personality traits, such as those described by the "dark triad": narcissism (entitled self-importance), Machiavellianism (strategic exploitation and deceit), and psychopathy (callousness and cynicism). But while people who score high in these traits are more likely to end up in prison, most of them are well functioning and don't engage in extreme antisocial behaviors.
Now, a new study published in Cerebral Cortex found that the brains of psychopathic criminals are structurally and functionally similar to many well-functioning, non-criminal individuals with psychopathic traits. The results suggest that psychopathy isn't a binary classification, but rather a "constellation" of personality traits that "vary in the non-incarcerated population with normal range of social functioning."
Assessing your inner psychopath
The researchers used functional magnetic resonance imaging (fMRI) to compare the brains of violent psychopathic criminals to those of healthy volunteers. All participants were assessed for psychopathy through commonly used inventories: the Hare Psychopathy Checklist-Revised and the Levenson Self-Report Psychopathy Scale.
Experimental design and sample stimuli. The subjects viewed a compilation of 137 movie clips with variable violent and nonviolent content.Nummenmaa et al.
Both groups watched a 26-minute-long medley of movie scenes that were selected to portray a "large variability of social and emotional content." Some scenes depicted intense violence. As participants watched the medley, fMRI recorded how various regions of their brains responded to the content.
The goal was to see whether the brains of psychopathic criminals looked and reacted similarly to the brains of healthy subjects who scored high in psychopathic traits. The results showed similar reactions: When both groups viewed violent scenes, the fMRI revealed strong reactions in the orbitofrontal cortex and anterior insula, brain regions associated with regulating emotion.
These similarities manifested as a positive association: The more psychopathic traits a healthy subject displayed, the more their brains responded like the criminal group. What's more, the fMRI revealed a similar association between psychopathic traits and brain structure, with those scoring high in psychopathy showing lower gray matter density in the orbitofrontal cortex and anterior insula.
There were some key differences between the groups, however. The researchers noted that the structural abnormalities in the healthy sample were mainly associated with primary psychopathic traits, which are: inclination to lie, lack of remorse, and callousness. Meanwhile, the functional responses of the healthy subjects were associated with secondary psychopathic traits: impulsivity, short temper, and low tolerance for frustration.
Overall, the study further illuminates some of the biological drivers of psychopathy, and it adds nuance to common conceptions of the differences between psychopathy and being "healthy."
Why do some psychopaths become criminals?
The million-dollar question remains unanswered: Why do some psychopaths end up in prison, while others (or, people who score high in psychopathic traits) lead well-functioning lives? The researchers couldn't give a definitive answer, but they did note that psychopathic criminals had lower connectivity within "key nodes of the social and emotional brain networks, including amygdala, insula, thalamus, and frontal pole."
"Thus, even though there are parallels in the regional responsiveness of the brain's affective circuit in the convicted psychopaths and well-functioning subjects with psychopathic traits, it is likely that the disrupted functional connectivity of this network is specific to criminal psychopathy."
Fighting online misinformation: We're doing it wrong
Counterintuitively, directly combating misinformation online can spread it further. A different approach is needed.
- Like the coronavirus, engaging with misinformation can inadvertently cause it to spread.
- Social media has a business model based on getting users to spend increasing amounts of time on their platforms, which is why they are hesitant to remove engaging content.
- The best way to fight online misinformation is to drown it out with the truth.
A year ago, the Center for Countering Digital Hate warned of the parallel pandemics — the biological contagion of COVID-19 and the social contagion of misinformation, aiding the spread of the disease. Since the outbreak of COVID-19, anti-vaccine accounts have gained 10 million new social media followers, while we have witnessed arson attacks against 5G masts, hospital staff abused for treating COVID patients, and conspiracists addressing crowds of thousands.
Many have refused to follow guidance issued to control the spread of the virus, motivated by beliefs in falsehoods about its origins and effects. The reluctance we see in some to get the COVID vaccine is greater amongst those who rely on social media rather than traditional media for their information. In a pandemic, lies cost lives, and it has felt like a new conspiracy theory has sprung up online every day.
How we, as social media users, behave in response to misinformation can either enable or prevent it from being seen and believed by more people.
The rules are different online
Credit: Pool via Getty Images
If a colleague mentions in the office that Bill Gates planned the pandemic, or a friend at dinner tells the table that the COVID vaccine could make them infertile, the right thing to do is often to challenge their claims. We don't want anyone to be left believing these falsehoods.
But digital is different. The rules of physics online are not the same as they are in the offline world. We need new solutions for the problems we face online.
Now, imagine that in order to reply to your friend, you must first hand him a megaphone so that everyone within a five-block radius can hear what he has to say. It would do more damage than good, but this is essentially what we do when we engage with misinformation online.
Think about misinformation as being like the coronavirus — when we engage with it, we help to spread it to everyone else with whom we come into contact. If a public figure with a large following responds to a post containing misinformation, they ensure the post is seen by hundreds of thousands or even millions of people with one click. Social media algorithms also push content into more users' newsfeeds if it appears to be engaging, so lots of interactions from users with relatively small followings can still have unintended negative consequences.
The trend of people celebrating and posting photos of themselves or loved ones receiving the vaccine has been far more effective than any attempt to disprove a baseless claim about Bill Gates or 5G mobile technology.
Additionally, whereas we know our friend from the office or dinner, most of the misinformation we see online will come from strangers. They often will be from one of two groups — true believers, whose minds are made up, and professional propagandists, who profit from building large audiences online and selling them products (including false cures). Both of these groups use trolling tactics, that is, seeking to trigger people to respond in anger, thus helping them reach new audiences and thereby gaming the algorithm.
On the day the COVID vaccine was approved in the UK, anti-vaccine activists were able to provoke pro-vaccine voices into posting about thalidomide, exposing new audiences to a reason to distrust the medical establishment. Those who spread misinformation understand the rules of the game online; it's time those of us on the side of enlightenment values of truth and science did too.
How to fight online misinformation
Of course, it is much easier for social media companies to take on this issue than for us citizens. Research from the Center for Countering Digital Hate and Anti-Vax Watch last month found that 65% of anti-vaccine content on social media is linked to just twelve individuals and their organizations. Were the platforms to simply remove the accounts of these superspreaders, it would do a huge amount to reduce harmful misinformation.
The problem is that social media platforms are resistant to do so. These businesses have been built by constantly increasing the amount of time users spend on their platforms. Getting rid of the creators of engaging content that has millions of people hooked is antithetical to the business model. It will require intervention from governments to force tech companies to finally protect their users and society as a whole.
So, what can the rest of us do, while we await state regulation?
Instead of engaging, we should be outweighing the bad with the good. Every time you see a piece of harmful misinformation, share advice or information from a trusted source, like the WHO or BBC, on the same subject. The trend of people celebrating and posting photos of themselves or loved ones receiving the vaccine has been far more effective than any attempt to disprove a baseless claim about Bill Gates or 5G mobile technology. In the attention economy that governs tech platforms, drowning out is a better strategy than rebuttal.
Imran Ahmed is CEO of the Center for Countering Digital Hate.
Self-awareness is what makes us human
Because of our ability to think about thinking, "the gap between ape and man is immeasurably greater than the one between amoeba and ape."
