3 Tips for Avoiding Fake News in Science

Astrophysicist Michael J. I. Brown offers some guidelines for identifying fake or bad science.

There’s been a lot of talk lately about fake news, and many consider it a genuine threat to a well-functioning democracy. Some of it results from political mischief, or worse. It’s also happening in science reporting as researchers, writers, and publications attempt to gain attention with sensational headlines. Inconclusive evidence may be presented as fact, and sometimes it’s just bad science. Check out RetractionWatch, a site whose purpose is trying to keep up with scientific papers that have been recalled after publication — there were 700 retractions in 2015 alone.

Astrophysicist Michael J. I. Brown is concerned with how difficult it’s become to separate real scientific breakthroughs from clickbait. Writing for Science Alert, he shares his three-step checklist for sorting the truth from the trash. Even with his checklist, it’s not always easy for a real scientist like him to be confident of a study’s validity. But nobody likes to be fooled by fake or bad science.

Neatness Counts

This may seem superficial, but Brown asserts that it’s not. Given that carefully conducted research typically takes a long time, even years, a study that looks slapdash may well be. Typos and cruddy-looking graphics can be a clue that a lack of due diligence is at play.

(LAURIE SULLIVAN)

Brown cites a recent, well-publicized paper by E.F. Borra, E. Trottier that claimed in its Comments section: “Signals probably from Extraterrestrial Intelligence.” Brown was curious, so he had a look. “An immediate red flag for me was some blurry graphs, and figures with captions that weren’t on the same page,” Brown writes. Looking further, he discovered that the author’s conclusion was based on a liberal application of Fourier analysis, which Brown says is known to generate data artifacts that skew results. He also learned that Borra and Trottier chose to work with only a tiny subset of data. Together, Brown feels these two factors render the study questionable at best.

He points out, though, that the neatness filter isn’t always a reliable test, since sometimes great science comes in a less-than-great presentation, as with the announcement of the discovery of the Higgs boson.

Obviousness

As Brown puts it, “’That’s obvious, why didn’t someone think of that before?’

Well, perhaps someone did.” When a study announces something really big and basic, Brown suggests you do a search on that announcement — odds are that you’ll find the topic’s been studied many times before. If you find that to be the case, and if you find no one else came to the same conclusion, what you have is a red flag that should have you carefully considering the new study’s methodology.

(MIKE LICHT)

Brown’s example here is a recent study that asserted the universe isn’t expanding at an accelerating rate, contrary to earlier, well-regarded research. Brown found an edifying discussion by experts on Twitter (one of whom referred to their discussion as “headdesking”) that dismissed the study as having been based on incorrect assumptions about the behavior of the supernovae it examined, and on the ignoring of some key counter-evidence.

Who Published It?

(SIGNE KARIN)

Checking it the publisher of a study can sometimes help you ascertain its value. While Brown notes that a journal’s ranking isn’t a perfect indicator of its trustworthiness, it’s a start. He does note that some highly ranked journals, like Nature, an open-source subsidiary of which published the universe-expansion piece noted above, and Science are sometimes, in his opinion, seduced by the lure of an attention-grabbing headline.

Brown points out that some journals accept papers without peer review or other critiquing process altogether — meaning they’re just taking the authors’ word on a study’s veracity. University of Colorado librarian Jeffrey Beall had been hosting a well-known list of predatory publishers a skeptical reader could consult. It was mentioned in a recent New York Times article, though his site is offline as of this writing; maybe it will return.


Many of us are interested in keeping up with the latest developments in science, and we eagerly consume articles that feed our enthusiasm. Brown’s three tips can help us know how seriously to take what we’re reading, though it must be pointed out that his method is, in this case, an imperfect science of its own.

Car culture and suburbs grow right-wing populism, claims study

New research links urban planning and political polarization.

Pixabay
Politics & Current Affairs
  • Canadian researchers find that excessive reliance on cars changes political views.
  • Decades of car-centric urban planning normalized unsustainable lifestyles.
  • People who prefer personal comfort elect politicians who represent such views.
Keep reading Show less

How to split the USA into two countries: Red and Blue

Progressive America would be half as big, but twice as populated as its conservative twin.

Image: Dicken Schrader
Strange Maps
  • America's two political tribes have consolidated into 'red' and 'blue' nations, with seemingly irreconcilable differences.
  • Perhaps the best way to stop the infighting is to go for a divorce and give the two nations a country each
  • Based on the UN's partition plan for Israel/Palestine, this proposal provides territorial contiguity and sea access to both 'red' and 'blue' America
Keep reading Show less

NASA astronomer Michelle Thaller on ​the multiple dimensions of space and human sexuality

Science and the squishiness of the human mind. The joys of wearing whatever the hell you want, and so much more.

Flickr / 13winds
Think Again Podcasts
  • Why can't we have a human-sized cat tree?
  • What would happen if you got a spoonful of a neutron star?
  • Why do we insist on dividing our wonderfully complex selves into boring little boxes
Keep reading Show less