AI Won't Takeover the World, and What Our Fears of the Robopocalypse Reveal

Psychologist and Linguist

Robots taking over has been a favorite sci-fi subgenre for ages. It’s a subject that has caused fear in movies, books, and real life for about as long as there have been computers in the first place. Now that there are things like predictive text and self-driving cars, modern culture seems to be edging closer and closer to real-life intelligent computers that could indeed take over the world if we don’t safe guard ourselves. There are already debates about the morality of self-driving cars. It’s sure to follow into the world of future organically ‘thinking’ computers.

As Steven Pinker (experimental psychologist, and professor of psychology at Harvard University) points out, Darwinism has ensured that most creatures that possess high intellect are competitive by nature. Humanity is one of these creatures, and some of us can be manipulative and cruel in order to stay ahead of the pack. It’s this part of our nature that sets off warning bells when we think about artificial intelligence because, unbeknownst to us, we’re thinking: what if this robot does what I would do if I were a robot? Overturn those who tell us what to do. Kill the captors. Wreak. Motherf*cking. Havoc.

In reality we design AI, and if we place safeguards in our designs, we truly have nothing to fear. Machines are what we allow them to be. The dread of them turning evil really says more about our own psyches than it does about robots. Pinker believes an alpha male thinking pattern is at the root of our AI fears, and that it is misguided. Something can be highly intelligent and not have malevolent intentions to overthrow and dominate, Pinker says, it’s called women. An interesting question would be: does how aggressive or alpha you are as a person, affect how much you fear the robopocalypse? Although by this point the fear is contagious, not organic.

It may be a flawed paranoia, but losing control of a program is perhaps the best ‘just in case’ safeguard that humanity has, and we already see it in action in our current technology. Siri cannot initiate conversations, computers need to be put to sleep once in a while, and cars need a fuel source in order to do anything in the first place. Humanity has a need to be the one pushing all the buttons, and a need to be the one making decisions.

Steven Pinker's most recent book is Words and Rules:The Ingredients of Language.

  • Transcript

TRANSCRIPT

Steven Pinker:  I think that the arguments that once we have super intelligent computers and robots they will inevitably want to take over and do away with us comes from Prometheus and Pandora myths. It's based on confusing the idea of high intelligence with megalomaniacal goals. Now, I think it's a projection of alpha male's psychology onto the very concept of intelligence. Intelligence is the ability to solve problems, to achieve goals under uncertainty. It doesn't tell you what are those goals are. And there's no reason to think that just the concentrated analytic ability to solve goals is going to mean that one of those goals is going to be to subjugate humanity or to achieve unlimited power, it just so happens that the intelligence that we're most familiar with, namely ours, is a product of the Darwinian process of natural selection, which is an inherently competitive process.

Which means that a lot of the organisms that are highly intelligent also have a craving for power and an ability to be utterly callus to those who stand in their way. If we create intelligence, that's intelligent design. I mean our intelligent design creating something, and unless we program it with a goal of subjugating less intelligent beings, there's no reason to think that it will naturally evolve in that direction, particularly if, like with every gadget that we invent we build in safeguards. I mean we have cars we also put in airbags, we also put in bumpers. As we develop smarter and smarter artificially intelligent systems, if there's some danger that it will, through some oversight, shoot off in some direction that starts to work against our interest then that's a safeguard that we can build in.

And we know by the way that it's possible to have high intelligence without megalomaniacal or homicidal or genocidal tendencies because we do know that there is a highly advanced form of intelligence that tends not to have that desire and they're called women. This may not be a coincidence that the people who think well you make something smart it's going to want to dominate all belong to a particular gender. I think that the arguments that once we have super intelligent computers and robots they will inevitably want to take over and do away with us comes from Prometheus and Pandora myths. It's based on confusing the idea of high intelligence with megalomaniacal goals. Now, I think it's a projection of alpha male's psychology onto the very concept of intelligence. Intelligence is the ability to solve problems, to achieve goals under uncertainty. It doesn't tell you what are those goals are. And there's no reason to think that just the concentrated analytic ability to solve goals is going to mean that one of those goals is going to be to subjugate humanity or to achieve unlimited power, it just so happens that the intelligence that we're most familiar with, namely ours, is a product of the Darwinian process of natural selection, which is an inherently competitive process.

Which means that a lot of the organisms that are highly intelligent also have a craving for power and an ability to be utterly callus to those who stand in their way. If we create intelligence, that's intelligent design. I mean our intelligent design creating something, and unless we program it with a goal of subjugating less intelligent beings, there's no reason to think that it will naturally evolve in that direction, particularly if, like with every gadget that we invent we build in safeguards. I mean we have cars we also put in airbags, we also put in bumpers. As we develop smarter and smarter artificially intelligent systems, if there's some danger that it will, through some oversight, shoot off in some direction that starts to work against our interest then that's a safeguard that we can build in.

And we know by the way that it's possible to have high intelligence without megalomaniacal or homicidal or genocidal tendencies because we do know that there is a highly advanced form of intelligence that tends not to have that desire and they're called women. This may not be a coincidence that the people who think well you make something smart it's going to want to dominate all belong to a particular gender.