Defense Department project develops first tools to detect ‘deepfake videos’

A Defense Department project has developed some of the first tools able to detect when videos have been digitally manipulated—content often called deepfake videos.


A Defense Department project has developed some of the first tools able to automatically detect a particularly deceptive type of digital content called 'deepfake' videos.

Deepfake videos often feature one person’s face convincingly merged with another. Other videos show one person’s face making movements and speaking words they may have never made or said in real life, so that, combined with audio manipulation, the result can be the likeness of former President Barack Obama saying things actually uttered by someone else in a studio.

The technology uses machine learning processes to learn the details of a person’s face. The A.I. analyzes video footage of the target person to learn as much as it can; the more footage it has to study, the more it learns. That’s why presidents and celebrities are frequently used in deepfake experiments.

It’s a technological evolution that’s alarmed many in media and government, unsurprisingly. The fear is that it could usher in a new era of fake news, one in which it would be virtually impossible to tell whether what you see on a screen is real or fake.

“This is an effort to try to get ahead of something,” said Florida senator Marco Rubio in remarks at the Heritage Foundation. “The capability to do all of this is real. It exists now. The willingness exists now. All that is missing is the execution. And we are not ready for it, not as a people, not as a political branch, not as a media, not as a country.”

In 2014, deepfake technology started to get much better thanks to an innovative approach called generative adversarial networks (GAN). As Martin Giles writes for the MIT Technology Review, the approach is similar to an art forger and an art detective who repeatedly try to outwit one another, resulting in increasingly convincing fakes.

“Both networks are trained on the same data set. The first one, known as the generator, is charged with producing artificial outputs, such as photos or handwriting, that are as realistic as possible. The second, known as the discriminator, compares these with genuine images from the original data set and tries to determine which are real and which are fake. On the basis of those results, the generator adjusts its parameters for creating new images. And so it goes, until the discriminator can no longer tell what’s genuine and what’s bogus.”

Recently, a contest run by the U.S. Defense Advanced Research Projects Agency (DARPA) asked researchers to automate existing forensic tools in an effort to keep up with deepfake technology. The goal was to find areas where deepfake technology is falling short.

One forensics approach exploited a rather simple observation: faces created by deepfake technology hardly ever blink. The reason? Neural networks that create the deepfakes typically study still images of a target face, not blinking ones. It’s a good exploit for now, but deepfake generators could get around it by feeding the networks more images of blinking faces.

Other techniques, as Will Knight wrote for MIT Technology Review, take advantage of the peculiar signatures deepfake technology leaves behind, like unnatural head movements and odd eye color.

“We are working on exploiting these types of physiological signals that, for now at least, are difficult for deepfakes to mimic,” says Hany Farid, a leading digital forensics expert at Dartmouth University.

Still, it’s possible these kinds of forensic approaches will be forever one step behind the evolution of deepfake technology.

“Theoretically, if you gave a GAN all the techniques we know to detect it, it could pass all of those techniques,” David Gunning, the DARPA program manager in charge of the project, told MIT Technology Review. “We don’t know if there's a limit. It’s unclear.”

LinkedIn meets Tinder in this mindful networking app

Swipe right to make the connections that could change your career.

Getty Images
Sponsored
Swipe right. Match. Meet over coffee or set up a call.

No, we aren't talking about Tinder. Introducing Shapr, a free app that helps people with synergistic professional goals and skill sets easily meet and collaborate.

Keep reading Show less

Why 'upgrading' humanity is a transhumanist myth

Upload your mind? Here's a reality check on the Singularity.

Videos
  • Though computer engineers claim to know what human consciousness is, many neuroscientists say that we're nowhere close to understanding what it is, or its source.
  • Scientists are currently trying to upload human minds to silicon chips, or re-create consciousness with algorithms, but this may be hubristic because we still know so little about what it means to be human.
  • Is transhumanism a journey forward or an escape from reality?
Keep reading Show less

Steven Pinker's 13 rules for writing better

The Harvard psychologist loves reading authors' rules for writing. Here are his own.

NEW YORK, NY - JULY 21: Steven Pinker speaks onstage during OZY Fest 2018 at Rumsey Playfield, Central Park on July 21, 2018 in New York City. (Photo by Brad Barket/Getty Images for Ozy Media)
Personal Growth
  • Steven Pinker is many things: linguist, psychologist, optimist, Harvard professor, and author.
  • When it comes to writing, he's a student and a teacher.
  • Here's are his 13 rules for writing better, more simply, and more clearly.
Keep reading Show less

Dead – yes, dead – tardigrade found beneath Antarctica

A completely unexpected discovery beneath the ice.

(Goldstein Lab/Wkikpedia/Tigerspaws/Big Think)
Surprising Science
  • Scientists find remains of a tardigrade and crustaceans in a deep, frozen Antarctic lake.
  • The creatures' origin is unknown, and further study is ongoing.
  • Biology speaks up about Antarctica's history.
Keep reading Show less