Skip to content
Sign up for the Smarter Faster newsletter
A weekly newsletter featuring the biggest ideas from the smartest people

LAWRENCE KRAUSS: I like to keep an open mind but not so open that my brains fall out. And that's the key point. We have to skeptically assess the information we receive. We can't be gullible because when we get a lot of information, it's absolutely certain that some of that information is wrong. And so we have to always filter what we get. And we have to ask ourselves the following question: "How open does my brain have to be to accept that information? Does it have to fall out?" And by that, I mean when someone tells you something you have to ask "Is this consistent with my experience? Is it consistent with the experience of other people around me?" And if it isn't, then probably there's a good reason to be skeptical about it; it's probably wrong. If it makes predictions that also appear to be in disagreement with things that you observe around you, you should question it.

And so we should never take anything on faith. That's really the mantra of science, if you want, that faith is the enemy of science. We often talk about a loss of faith in the world today. You don't lose anything by losing faith. What you gain is reality. And so skepticism plays a key role in science simply because we also are hard-wired to want to believe. We're hard-wired to want to find reasons for things. In the savanna in Africa, the trees could be rustling and you could choose to say, "Well, there's no reason for that." Or, "Maybe it's due to a lion." And those individuals who thought there might be no reason, never lived long enough to survive to procreate. And so it's not too surprising, we want to find explanations for everything. And we create them if we need to, to satisfy ourselves, because we need to make sense of the world around us. And what we have to understand is that what makes sense to the universe, is not the same as what makes sense to us. And we can't impose our beliefs on the universe. And the way we get around that inherent bias is by constantly questioning both ourselves and all the information we receive from others. That's what we do in science and it works beautifully in the real world as well.

MICHAEL SHERMER: The problem is this. None of us has the truth. The only way to find out if you're deceiving yourself or not, if you've gone off the rails, if you're wrong in some way, is to listen to other people who disagree with you. I started encountering other people that disagreed with me. You know, we-never-went-to-the-moon people, conspiracy people, whatever. And I thought, "Okay, so how do we know, if I don't know what's coming down the pike say in 10 years from now, if I was gonna teach my students how to think critically, what are the key points, like just basic questions they could ask?" So, it begins with one: How reliable is the source of the claim? Here's the claim, how reliable is it? What's the evidence for it? What's the quality of the evidence? Where does it come from? Who said that? Is this some fake news, alternative site thing, or is it The Wall Street Journal or The New York Times? The source really matters. Has anyone tried to disprove the claim? This is super important because everybody thinks they're right and every website has testimonials about this product or that idea. The question is not "What do your supporters think?" but "What do the people who don't agree with you think?" Because that's what I wanna know. Has anyone run an experiment to try to disprove your theory? And so in science, this is as basic as it gets. Karl Popper called this the Principle of Falsification. That is, we can't ever prove a theory correct, but we can disprove it by having an experiment that shows it's wrong.

So, if you can't falsify it, what are you really doing? And my favorite story on this, by the way, let me just have a little sidetrack here from Karl. He's got this great little section in his book "Demon-Haunted World." There's a dragon in my garage. I have a dragon in my garage, do you want to see it? Come here, let me show you. I pull up the garage door. I go, "Look, can you see the dragon?" And you look in and you go, "Well, no, I don't see anything." "Oh, oh, oh, sorry. This is an invisible dragon." "An invisible dragon?" "Yeah, yeah, yeah, he's invisible." "Well, what if we put some flour on the ground and then we'll get the footprints of the dragon?" "Well, no, see, this is a special dragon that hovers above the ground; it floats. It's an invisible floating dragon." "An invisible floating dragon, okay. Oh, wait, I have some infrared cameras here. We can detect the heat from the dragon." "No, see, this is a cold-blooded dragon. It doesn't give off any heat." "What about the fire? We can detect the fire that the dragon spits out." "Nah, it spits out cold fire." You see the problem? If there's no way for me to falsify that there's a dragon there, what's the difference between an invisible floating heatless dragon and no dragon at all? None. And of course we can apply this to God or any other supernatural paranormal type. If I can't debunk it, if I can't falsify it, if there's no way to test it, then how will we ever know it's true? This is the core of the baloney-detection kit. We have to be able to get to whether it's true or not, in some way so it's not just my opinion versus your opinion and we shout at each other.

Then we want to know, does the claimant's personal beliefs somehow enter in? Because of course, we all have personal opinions and beliefs about things. My politics, my religion, my ideology can influence me. It doesn't make it wrong but it's good to know if somebody has an agenda. So, when you watch Fox News, you know that they have an agenda, for sure. Or there's other sources on the left that have a liberal agenda, maybe NPR, who knows. But you see, it's good to know that, just in case, so when you hear the facts, you go, "Well, maybe, but I know this guy has an agenda." So, that's the kinda thing. Does the new idea being proposed account for the same amount of information that the old idea does and some of the new anomalies that the old idea can't explain? So, people offer theories, so-called alternative theories of physics, for example, and they always send them to me going, "Hey, listen, I'm not good at math but if you help me with the math I'll share the Nobel Prize with you." Right. I don't do math and physics so you might take it to the local high school physics teacher before you announce that you've made the greatest discovery since Newton and Einstein. The problem with that is the current theories do pretty well at explaining almost everything. Not everything, so there's always anomalies we can't explain. So maybe there's some new theory coming down the pike that might account for all the old theories, explanations and the new anomalies, the anomalies that the old theory can't explain. Maybe. But again, we gotta be able to test it first.

So, those are the kind of things. Does the claimant play by the rules of science, the rules of the field that you're in, for example. Again, these alternative physics guys come to me. Don't come to me, I'm not a physicist. Did you at least ask the local high school physics teacher if this makes any sense? Because if you have no training, you don't know all the mistakes that people in the past have already made to get to where they are now. And if you're starting here without the background, you're gonna make lots of mistakes. So, these are the sorts of things that any good baloney-detector should know.

BILL NYE: I've been messing around with this idea of cognitive dissonance. This is to say, you have a worldview, you're presented with evidence that conflicts with the world view, so you either have to change your worldview, which is hard because you've lived your whole life with it, or you just dismiss the evidence so that you don't have this discomfort, this conflict in your mind, this dissonance. And that's what I'm working with right now. And apparently the way to overcome that, is to say, "We're all in this together. Let's learn about this together." Present the conspiracy theorist with the idea that he or she may be rejecting evidence because it's just so uncomfortable, and you're in it together, we're in it together. I'm uncomfortable too. And just understand it's a process. Somebody who has a worldview that's inconsistent with evidence, and I may have some, it takes a while for you to turn around. Like the example of palm reading or astrology, it's not something people reverse their ideas about immediately. It takes, in my experience, it takes about two years for somebody to sort of look at palm reading, look at cold reading, or tarot card reading, for a while and then realize that these tarot card readers, palm readers, are just taking information that you've given them, the client's giving them, and feeding it back to you.

We all tend to go, "Well, look at the facts, change your mind," but it takes people a couple years to change their mind. So, my recommendation, stick with it. You'll get frustrated, the person will get frustrated, but present the idea of cognitive dissonance. This is my latest idea about a way to work together to a scientific understanding.

KRAUSS: Science doesn't prove what's absolutely true. What it does, is prove what's absolutely false. What doesn't satisfy the test of experiment, we throw out. What remains may not be true, but we shrink it down, as Sherlock Holmes would say, "And what remains after all of that is done, is likely to be true." So, many sources, question what you see, and whether it's consistent with what you already know, and be suspicious of your own likes and dislikes when you accept information. That's probably the reason we shouldn't, when we turn to the internet, go to echo chambers and just read the sources that we like. Now, having said that, if you look at many sources, you could also quickly decide which ones are not reliable and throw them out. If they're not reliable in one case then you should be highly suspicious of them in the future. So, we all turn to different sources that we think are more or less reliable based on our past experience. Try that, and I think it's one great way to filter out a lot of the nonsense on the internet.

When I talk about being skeptical, it is important to recognize that you can be surprised. And something that you don't think is sensible, can end up being sensible. That's the way we learn things in physics. So, when someone presents you with an idea that may seem strange, it's reasonable to be skeptical of it but it's worth pursuing long enough to see if it might make sense and to listen to arguments that might be convincing, that might cause you to change your mind. In fact, there's a great school of pedagogy that says, "The only way we actually learn anything is by confronting our own misconceptions." So, once again, while it's reasonable to be skeptical of external information, if you're always skeptical of your response to information, and what your misconceptions are and what your prejudices are, then you will both guard yourself not to accept nonsense but also you will be willing to realize that sometimes what you think is skepticism, is really myopia.

DERREN BROWN: Where I think skepticism, in its broad, modern, popular sense, "I just don't believe in God, don't believe in this don't believe in that," where I think it has its limits—and I speak very much as a skeptic myself and as an atheist, as I said—where it is important to realize the edges of its usefulness, is where those things that may not be objectively true but can be psychologically true, in inverted commas. In other words, psychologically resonant to the path of living, and what we take in life, and what's important to us and what's helpful. So, that's what you don't wanna throw out. You don't wanna throw out that baby with the bath water. So, in religion, for example, those things that are easily knocked down, if you're an atheist, they're easy to kind of make fun of and disprove, those things are also, they kind of often are straw men to knock down, but they can often be pointers back to something that is psychologically useful. They're signifiers of something. If you take what happens with religion, is that you have something that happens, an experience of transcendence or a kind of a thing that happened historically. Nothing magical or supernatural, but just for people at that time, a connection to the sense of the transcendent, whatever that was, a message or something. And then as that moves out of living memory, to re-create it, a bunch of practices, and dogmas and things are formed to try and recreate that feeling. And that becomes now a thing of belief rather than a sort of knowledge that it was at the time. And then to sustain and protect that belief, an institution is sort of created, and developed, and becomes politicized, and powerful, and monetized and all of those things. And then it moves into a world where we are nowadays, where things have to be sort of proved with evidence. So, it starts to try and come up with evidential arguments that somehow never quite really sort of work. So, you do end up with a thing that's easy to knock down but that can miss the fact that there's something at the heart of it, which maybe is useful. Maybe those narratives around religion are useful to us psychologically. Maybe they have an archetypal, or a mythological use, that it would be a shame to dismiss because we feel the absence of those things. It's the very fact we turn to psychics, and fortune tellers, and become terrified and lonely around death. Those things happen because we've lost touch with some of those myths and some of those more resonant narratives. So, I think being a little skeptical about skepticism itself and the easy narratives that it forms, is also, I think, very useful.


Related