There’s no way we could stop a rogue AI

Max Planck Institute scientists crash into a computing wall there seems to be no way around.

There’s no way we could stop a rogue AI
Credit: josefkubes/Adobe Stock
  • Artificial intelligence that's smarter than us could potentially solve problems beyond our grasp.
  • AI that are self-learning can absorb whatever information they need from the internet, a Pandora's Box if ever there was one.
  • The nature of computing itself prevents us from limiting the actions of a super-intelligent AI if it gets out of control.

There have been a fair number of voices—Stephen Hawking among them—raised in warning that a super-intelligent artificial intelligence could one day turn on us and that we shouldn't be in such a hot, unquestioning hurry to develop true AI. Others say, naw, don't worry. Now a new white paper from scientists at the Center for Humans and Machines at the Max Planck Institute for Human Development presents a series of theoretical tests that confirm the threat: Due to the basic concepts underlying computing, we would be utterly unable to control a super-intelligent AI.

"We argue that total containment is, in principle, impossible, due to fundamental limits inherent to computing itself," write the paper's authors.

The white paper is published in the Journal of Artificial Intelligence Research.

Why worry?

Credit: @nt/Adobe Stock

"A super-intelligent machine that controls the world sounds like science fiction," says paper co-author Manuel Cebrian in a press release. "But there are already machines that perform certain important tasks independently without programmers fully understanding how they learned it. The question therefore arises whether this could at some point become uncontrollable and dangerous for humanity."

The lure of AI is clear. Its ability to "see" the patterns in data make it a promising agent for solving problems too complex for us to wrap our minds around. Could it cure cancer? Solve the climate crisis? The possibilities are nearly endless.

Connected to the internet, AI can grab whatever information it needs to achieve its task, and therein lies a big part of the danger. With access to every bit of human data—and responsible for its own education—who knows what lessons it would learn regardless of any ethical constraints built into its programming? Who knows what goals it would embrace and what it might do to achieve them?

Even assuming benevolence, there's danger. Suppose that an AI is confronted by an either/or choice akin to the Trolley Dilemma, maybe even on a grand scale: Might an AI decide to annihilate millions of people if it decided the remaining billions would stand a better chance of survival?

A pair of flawed options

Credit: Maxim_Kazmin/Adobe Stock

The most obvious way to keep a super intelligent AI from getting ahead of us is to limit its access to information by preventing it from connecting to the internet. The problem with limiting access to information, though, is that it would make any problem we assign the AI more difficult for it to solve. We would be weakening its problem-solving promise possibly to a point of uselessness.

The second approach that might be taken is to limit what a super-intelligent AI is capable of doing by programming into it certain boundaries. This might be akin to writer Isaac Asimov's Laws of Robotics, the first of which goes: "A robot may not injure a human being or, through inaction, allow a human being to come to harm."

Unfortunately, says the study, a series of logical tests reveal that it's impossible to create such limits. Any such a containment algorithm, it turns out, would be self-defeating.

Containment is impossible

Credit: UncleFredDesign/Adobe Stock

"If you break the problem down to basic rules from theoretical computer science, it turns out that an algorithm that would command an AI not to destroy the world could inadvertently halt its own operations. If this happened, you would not know whether the containment algorithm is still analyzing the threat, or whether it has stopped to contain the harmful AI. In effect, this makes the containment algorithm unusable."

The team investigated stacking containment algorithms, with each monitoring the behavior of the previous one, but eventually the same problem arises: The final check halts itself, rendering it unreliable.

Too smart?

The Planck researchers also concluded that a similar bit of logic makes it impossible for us to know when a self-learning computer's intelligence has come to exceed our own. Essentially, we're not smart enough to be able to develop tests for intelligence superior to ours.

"Machines take me by surprise with great frequency. This is largely because I do not do sufficient calculation to decide what to expect them to do." — Alan Turing

This means that it's entirely conceivable that an AI capable of self-learning may well quietly ascend to super-intelligence without our even knowing it — a scary reason all by itself to slow down our hurly-burley race to artificial intelligence.

In the end, we're left with a dangerous bargain to make or not make: Do we risk our safety in exchange for the possibility that AI will solve problems we can't?

Every 27.5 million years, the Earth’s heart beats catastrophically

Geologists discover a rhythm to major geologic events.

Credit: desertsolitaire/Adobe Stock
Surprising Science
  • It appears that Earth has a geologic "pulse," with clusters of major events occurring every 27.5 million years.
  • Working with the most accurate dating methods available, the authors of the study constructed a new history of the last 260 million years.
  • Exactly why these cycles occur remains unknown, but there are some interesting theories.
Keep reading Show less

Which countries are voyaging to Mars this decade?

We are likely to see the first humans walk on Mars this decade.

Mars's Twin Peaks

Technology & Innovation
  • Space agencies have successfully sent three spacecraft to Mars this year.
  • The independent missions occurred at around the same time because Earth and Mars were particularly close to each other last summer, providing an opportune time to launch.
  • SpaceX says it hopes to send a crewed mission to Mars by 2026, while the U.S. and China aim to land humans on the planet in the 2030s.
Keep reading Show less

How Apple and Nike have branded your brain

A new episode of "Your Brain on Money" illuminates the strange world of consumer behavior and explores how brands can wreak havoc on our ability to make rational decisions.

Apple logo via Adobe Stock
  • Effective branding can not only change how you feel about a company, it can actually change how your brain is wired.
  • Our new series "Your Brain on Money," created in partnership with Million Stories, recently explored the surprising ways brands can affect our behavior.
  • Brands aren't going away. But you can make smarter decisions by slowing down and asking yourself why you're making a particular purchase.
Keep reading Show less
Sponsored by Singleton

How Apple and Nike have branded your brain

Powerful branding can not only change how you feel about a company, it can actually change how your brain is wired.