Sie sind auf Seite 1von 39

24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

about archive minis the shed dinner table store support wbw Search  

Join 563,939 other humans an

The AI Revolution: Our Immortality or have new posts emailed to yo

Extinction  Email Address

SEND ME NEW POSTS

 January 27, 2015 By Tim Urban


Follow these special men
Note: This is Part 2 of a two-part series on AI. Part 1 is here.
408,346 50,313
142,939
PDF: We made a fancy PDF of this post for printing and offline viewing. Buy it here. (Or see a preview.)

___________

We have what may be an extremely difficult problem with an unknown time to solve it, on which quite
possibly the entire future of humanity depends. — Nick Bostrom

Welcome to Part 2 of the “Wait how is this possibly what I’m reading I don’t get why everyone isn’t
talking about this” series.

Part 1 started innocently enough, as we discussed Artificial Narrow Intelligence, or ANI (AI that
specializes in one narrow task like coming up with driving routes or playing chess), and how it’s all
around us in the world today. We then examined why it was such a huge challenge to get from ANI to
Artificial General Intelligence, or AGI (AI that’s at least as intellectually capable as a human, across the
board), and we discussed why the exponential rate of technological advancement we’ve seen in the
past suggests that AGI might not be as far away as it seems. Part 1 ended with me assaulting you with
the fact that once our machines reach human-level intelligence, they might immediately do this:

Popular Posts

The AI Revolution

 904

Why Procrastinators
Procrastinate

 375

The Fermi Paradox

 949

The Elon Musk Series

 405

How to Pick a Life Par

 174

Why Gen Y Yuppies A


Unhappy

 1,046

7 Ways to be Insuffer
on Facebook

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 1/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

 707

Putting Time In
Perspective

 476

How to Name a Baby

 259

Why You Should Stop


Caring What Other
People Think

 203

Religion for the


Nonreligious

 557

10 Types of 30-Year-O
Single Guys

 363

The Tail End

 168

This left us staring at the screen, confronting the intense concept of potentially-in-our-lifetime Artificial
Superintelligence, or ASI (AI that’s way smarter than any human, across the board), and trying to figure
1 1 ← open these
out which emotion we were supposed to have on as we thought about that.

Before we dive into things, let’s remind ourselves what it would mean for a machine to be
superintelligent.

A key distinction is the difference between speed superintelligence and quality superintelligence. Often,
someone’s first thought when they imagine a super-smart computer is one that’s as intelligent as a
human but can think much, much faster 2 —they might picture a machine that thinks like a human,
except a million times quicker, which means it could figure out in five minutes what would take a
human a decade.

That sounds impressive, and ASI would think much faster than any human could—but the true
separator would be its advantage in intelligence quality, which is something completely different. What
makes humans so much more intellectually capable than chimps isn’t a difference in thinking speed—

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 2/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
it’s that human brains contain a number of sophisticated cognitive modules that enable things like
complex linguistic representations or longterm planning or abstract reasoning, that chimps’ brains do
not. Speeding up a chimp’s brain by thousands of times wouldn’t bring him to our level—even with a
decade’s time, he wouldn’t be able to figure out how to use a set of custom tools to assemble an
intricate model, something a human could knock out in a few hours. There are worlds of human
cognitive function a chimp will simply never be capable of, no matter how much time he spends trying.

But it’s not just that a chimp can’t do what we do, it’s that his brain is unable to grasp that those worlds
even exist—a chimp can become familiar with what a human is and what a skyscraper is, but he’ll never
be able to understand that the skyscraper was built by humans. In his world, anything that huge is part
of nature, period, and not only is it beyond him to build a skyscraper, it’s beyond him to realize that
anyone can build a skyscraper. That’s the result of a small difference in intelligence quality.

And in the scheme of the intelligence range we’re talking about today, or even the much smaller range
among biological creatures, the chimp-to-human quality intelligence gap is tiny. In an earlier post, I
3
depicted the range of biological cognitive capacity using a staircase:

To absorb how big a deal a superintelligent machine would be, imagine one on the dark green step two
steps above humans on that staircase. This machine would be only slightly superintelligent, but its
increased cognitive ability over us would be as vast as the chimp-human gap we just described. And like
the chimp’s incapacity to ever absorb that skyscrapers can be built, we will never be able to even
comprehend the things a machine on the dark green step can do, even if the machine tried to explain it
to us—let alone do it ourselves. And that’s only two steps above us. A machine on the second-to-
highest step on that staircase would be to us as we are to ants—it could try for years to teach us the
simplest inkling of what it knows and the endeavor would be hopeless.

But the kind of superintelligence we’re talking about today is something far beyond anything on this
staircase. In an intelligence explosion—where the smarter a machine gets, the quicker it’s able to
increase its own intelligence, until it begins to soar upwards—a machine might take years to rise from
the chimp step to the one above it, but perhaps only hours to jump up a step once it’s on the dark
green step two above us, and by the time it’s ten steps above us, it might be jumping up in four-step
leaps every second that goes by. Which is why we need to realize that it’s distinctly possible that very
shortly after the big news story about the first machine reaching human-level AGI, we might be facing
the reality of coexisting on the Earth with something that’s here on the staircase (or maybe a million
times higher):

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 3/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

142,939

1.9k
Shares

553

553

And since we just established that it’s a hopeless activity to try to understand the power of a machine
only two steps above us, let’s very concretely state once and for all that there is no way to know what
ASI will do or what the consequences will be for us. Anyone who pretends otherwise doesn’t
understand what superintelligence means.

Evolution has advanced the biological brain slowly and gradually over hundreds of millions of years,
and in that sense, if humans birth an ASI machine, we’ll be dramatically stomping on evolution. Or
maybe this is part of evolution—maybe the way evolution works is that intelligence creeps up more and
more until it hits the level where it’s capable of creating machine superintelligence, and that level is like
a tripwire that triggers a worldwide game-changing explosion that determines a new future for all
living things:

And for reasons we’ll discuss later, a huge part of the scientific community believes that it’s not a
matter of whether we’ll hit that tripwire, but when. Kind of a crazy piece of information.

So where does that leave us?

Well no one in the world, especially not I, can tell you what will happen when we hit the tripwire. But
Oxford philosopher and lead AI thinker Nick Bostrom believes we can boil down all potential outcomes
into two broad categories.

First, looking at history, we can see that life works like this: species pop up, exist for a while, and after
some time, inevitably, they fall off the existence balance beam and land on extinction—

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 4/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

“All species eventually go extinct” has been almost as reliable a rule through history as “All humans
eventually die” has been. So far, 99.9% of species have fallen off the balance beam, and it seems pretty
clear that if a species keeps wobbling along down the beam, it’s only a matter of time before some
other species, some gust of nature’s wind, or a sudden beam-shaking asteroid knocks it off. Bostrom
calls extinction an attractor state—a place species are all teetering on falling into and from which no
species ever returns.

And while most scientists I’ve come across acknowledge that ASI would have the ability to send humans
to extinction, many also believe that used beneficially, ASI’s abilities could be used to bring individual
humans, and the species as a whole, to a second attractor state—species immortality. Bostrom believes
species immortality is just as much of an attractor state as species extinction, i.e. if we manage to get
there, we’ll be impervious to extinction forever—we’ll have conquered mortality and conquered chance.
So even though all species so far have fallen off the balance beam and landed on extinction, Bostrom
believes there are two sides to the beam and it’s just that nothing on Earth has been intelligent enough
yet to figure out how to fall off on the other side.

If Bostrom and others are right, and from everything I’ve read, it seems like they really might be, we
have two pretty shocking facts to absorb:

1) The advent of ASI will, for the first time, open up the possibility for a species to land on the
immortality side of the balance beam.

2) The advent of ASI will make such an unimaginably dramatic impact that it’s likely to knock the
human race off the beam, in one direction or the other.

It may very well be that when evolution hits the tripwire, it permanently ends humans’ relationship with
the beam and creates a new world, with or without humans.

Kind of seems like the only question any human should currently be asking is: When are we going to hit
the tripwire and which side of the beam will we land on when that happens?

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 5/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

No one in the world knows the answer to either part of that question, but a lot of the very smartest
people have put decades of thought into it. We’ll spend the rest of this post exploring what they’ve
come up with.

___________

Let’s start with the first part of the question: When are we going to hit the tripwire?

i.e. How long until the first machine reaches superintelligence?

Not shockingly, opinions vary wildly and this is a heated debate among scientists and thinkers. Many,
like professor Vernor Vinge, scientist Ben Goertzel, Sun Microsystems co-founder Bill Joy, or, most
famously, inventor and futurist Ray Kurzweil, agree with machine learning expert Jeremy Howard when
he puts up this graph during a TED Talk:

Those people subscribe to the belief that this is happening soon—that exponential growth is at work
and machine learning, though only slowly creeping up on us now, will blow right past us within the next
few decades.

Others, like Microsoft co-founder Paul Allen, research psychologist Gary Marcus, NYU computer
scientist Ernest Davis, and tech entrepreneur Mitch Kapor, believe that thinkers like Kurzweil are vastly
underestimating the magnitude of the challenge and believe that we’re not actually that close to the
tripwire.

The Kurzweil camp would counter that the only underestimating that’s happening is the
underappreciation of exponential growth, and they’d compare the doubters to those who looked at the
slow-growing seedling of the internet in 1985 and argued that there was no way it would amount to
anything impactful in the near future.

The doubters might argue back that the progress needed to make advancements in intelligence also
grows exponentially harder with each subsequent step, which will cancel out the typical exponential
nature of technological progress. And so on.

A third camp, which includes Nick Bostrom, believes neither group has any ground to feel certain about
the timeline and acknowledges both A) that this could absolutely happen in the near future and B) that
there’s no guarantee about that; it could also take a much longer time.

Still others, like philosopher Hubert Dreyfus, believe all three of these groups are naive for believing
that there even is a tripwire, arguing that it’s more likely that ASI won’t actually ever be achieved.

So what do you get when you put all of these opinions together?

In 2013, Vincent C. Müller and Nick Bostrom conducted a survey that asked hundreds of AI experts at a
series of conferences the following question: “For the purposes of this question, assume that human
scientific activity continues without major negative disruption. By what year would you see a (10% / 50%
4
/ 90%) probability for such HLMI to exist?” It asked them to name an optimistic year (one in which
they believe there’s a 10% chance we’ll have AGI), a realistic guess (a year they believe there’s a 50%
chance of AGI—i.e. after that year they think it’s more likely than not that we’ll have AGI), and a safe
guess (the earliest year by which they can say with 90% certainty we’ll have AGI). Gathered together as
one data set, here were the results:
2

Median optimistic year (10% likelihood): 2022


Median realistic year (50% likelihood): 2040
Median pessimistic year (90% likelihood): 2075

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 6/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

So the median participant thinks it’s more likely than not that we’ll have AGI 25 years from now. The
90% median answer of 2075 means that if you’re a teenager right now, the median respondent, along
with over half of the group of AI experts, is almost certain AGI will happen within your lifetime.

A separate study, conducted recently by author James Barrat at Ben Goertzel’s annual AGI Conference,
did away with percentages and simply asked when participants thought AGI would be achieved—by
2030, by 2050, by 2100, after 2100, or never. The results: 3

By 2030: 42% of respondents


By 2050: 25%
By 2100: 20%
After 2100: 10%
Never: 2%

Pretty similar to Müller and Bostrom’s outcomes. In Barrat’s survey, over two thirds of participants
believe AGI will be here by 2050 and a little less than half predict AGI within the next 15 years. Also
striking is that only 2% of those surveyed don’t think AGI is part of our future.

But AGI isn’t the tripwire, ASI is. So when do the experts think we’ll reach ASI?

Müller and Bostrom also asked the experts how likely they think it is that we’ll reach ASI A) within two
years of reaching AGI (i.e. an almost-immediate intelligence explosion), and B) within 30 years. The
results:
4

The median answer put a rapid (2 year) AGI → ASI transition at only a 10% likelihood, but a longer
transition of 30 years or less at a 75% likelihood.

We don’t know from this data the length of this transition the median participant would have put at a
50% likelihood, but for ballpark purposes, based on the two answers above, let’s estimate that they’d
have said 20 years. So the median opinion—the one right in the center of the world of AI experts—
believes the most realistic guess for when we’ll hit the ASI tripwire is [the 2040 prediction for AGI + our
estimated prediction of a 20-year transition from AGI to ASI] = 2060.

Of course, all of the above statistics are speculative, and they’re only representative of the center
opinion of the AI expert community, but it tells us that a large portion of the people who know the most
about this topic would agree that 2060 is a very reasonable estimate for the arrival of potentially world-
altering ASI. Only 45 years from now.

Okay now how about the second part of the question above: When we hit the tripwire, which side of the
beam will we fall to?

Superintelligence will yield tremendous power—the critical question for us is:

Who or what will be in control of that power, and what will their motivation be?

The answer to this will determine whether ASI is an unbelievably great development, an unfathomably
terrible development, or something in between.

Of course, the expert community is again all over the board and in a heated debate about the answer to
this question. Müller and Bostrom’s survey asked participants to assign a probability to the possible
impacts AGI would have on humanity and found that the mean response was that there was a 52%
chance that the outcome will be either good or extremely good and a 31% chance the outcome
will be either bad or extremely bad. For a relatively neutral outcome, the mean probability was only
17%. In other words, the people who know the most about this are pretty sure this will be a huge deal.
It’s also worth noting that those numbers refer to the advent of AGI—if the question were about ASI, I
imagine that the neutral percentage would be even lower.

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 7/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Before we dive much further into this good vs. bad outcome part of the question, let’s combine both
the “when will it happen?” and the “will it be good or bad?” parts of this question into a chart that
encompasses the views of most of the relevant experts:

We’ll talk more about the Main Camp in a minute, but first—what’s your deal? Actually I know what
your deal is, because it was my deal too before I started researching this topic. Some reasons most
people aren’t really thinking about this topic:

As mentioned in Part 1, movies have really confused things by presenting unrealistic AI scenarios
that make us feel like AI isn’t something to be taken seriously in general. James Barrat compares
the situation to our reaction if the Centers for Disease Control issued a serious warning about
vampires in our future. 5
Due to something called cognitive biases, we have a hard time believing something is real until we
see proof. I’m sure computer scientists in 1988 were regularly talking about how big a deal the
internet was likely to be, but people probably didn’t really think it was going to change their lives
until it actually changed their lives. This is partially because computers just couldn’t do stuff like

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 8/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
that in 1988, so people would look at their computer and think, “Really? That’s gonna be a life
changing thing?” Their imaginations were limited to what their personal experience had taught
them about what a computer was, which made it very hard to vividly picture what computers
might become. The same thing is happening now with AI. We hear that it’s gonna be a big deal,
but because it hasn’t happened yet, and because of our experience with the relatively impotent AI
in our current world, we have a hard time really believing this is going to change our lives
dramatically. And those biases are what experts are up against as they frantically try to get our
attention through the noise of collective daily self-absorption.
Even if we did believe it—how many times today have you thought about the fact that you’ll spend
most of the rest of eternity not existing? Not many, right? Even though it’s a far more intense fact
than anything else you’re doing today? This is because our brains are normally focused on the
little things in day-to-day life, no matter how crazy a long-term situation we’re a part of. It’s just
how we’re wired.

One of the goals of these two posts is to get you out of the I Like to Think About Other Things Camp
and into one of the expert camps, even if you’re just standing on the intersection of the two dotted
lines in the square above, totally uncertain.

During my research, I came across dozens of varying opinions on this topic, but I quickly noticed that
most people’s opinions fell somewhere in what I labeled the Main Camp, and in particular, over three
quarters of the experts fell into two Subcamps inside the Main Camp:

We’re gonna take a thorough dive into both of these camps. Let’s start with the fun one—

Why the Future Might Be Our Greatest Dream


As I learned about the world of AI, I found a surprisingly large number of people standing here:

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 9/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

The people on Confident Corner are buzzing with excitement. They have their sights set on the fun side
of the balance beam and they’re convinced that’s where all of us are headed. For them, the future is
everything they ever could have hoped for, just in time.

The thing that separates these people from the other thinkers we’ll discuss later isn’t their lust for the
happy side of the beam—it’s their confidence that that’s the side we’re going to land on.

Where this confidence comes from is up for debate. Critics believe it comes from an excitement so
blinding that they simply ignore or deny potential negative outcomes. But the believers say it’s naive to
conjure up doomsday scenarios when on balance, technology has and will likely end up continuing to
help us a lot more than it hurts us.

We’ll cover both sides, and you can form your own opinion about this as you read, but for this section,
put your skepticism away and let’s take a good hard look at what’s over there on the fun side of the
balance beam—and try to absorb the fact that the things you’re reading might really happen. If you
had shown a hunter-gatherer our world of indoor comfort, technology, and endless abundance, it
would have seemed like fictional magic to him—we have to be humble enough to acknowledge that it’s
possible that an equally inconceivable transformation could be in our future.

Nick Bostrom describes three ways a superintelligent AI system could function:


6

As an oracle, which answers nearly any question posed to it with accuracy, including complex
questions that humans cannot easily answer—i.e. How can I manufacture a more efficient car
engine? Google is a primitive type of oracle.
As a genie, which executes any high-level command it’s given—Use a molecular assembler to
build a new and more efficient kind of car engine—and then awaits its next command.
As a sovereign, which is assigned a broad and open-ended pursuit and allowed to operate in the
world freely, making its own decisions about how best to proceed—Invent a faster, cheaper, and
safer way than cars for humans to privately transport themselves.

These questions and tasks, which seem complicated to us, would sound to a superintelligent system
like someone asking you to improve upon the “My pencil fell off the table” situation, which you’d do by
picking it up and putting it back on the table.

Eliezer Yudkowsky, a resident of Anxious Avenue in our chart above, said it well:

There are no hard problems, only problems that are hard to a certain level of intelligence.
Move the smallest bit upwards [in level of intelligence], and some problems will suddenly
move from “impossible” to “obvious.” Move a substantial degree upwards, and all of them
will become obvious. 7

There are a lot of eager scientists, inventors, and entrepreneurs in Confident Corner—but for a tour of
brightest side of the AI horizon, there’s only one person we want as our tour guide.

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 10/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Ray Kurzweil is polarizing. In my reading, I heard everything from godlike worship of him and his ideas
to eye-rolling contempt for them. Others were somewhere in the middle—author Douglas Hofstadter,
in discussing the ideas in Kurzweil’s books, eloquently put forth that “it is as if you took a lot of very
good food and some dog excrement and blended it all up so that you can’t possibly figure out what’s
good or bad.”
8

Whether you like his ideas or not, everyone agrees that Kurzweil is impressive. He began inventing
things as a teenager and in the following decades, he came up with several breakthrough inventions,
including the first flatbed scanner, the first scanner that converted text to speech (allowing the blind to
read standard texts), the well-known Kurzweil music synthesizer (the first true electric piano), and the
first commercially marketed large-vocabulary speech recognition. He’s the author of five national
bestselling books. He’s well-known for his bold predictions and has a pretty good record of having
them come true—including his prediction in the late ’80s, a time when the internet was an obscure
thing, that by the early 2000s, it would become a global phenomenon. Kurzweil has been called a
“restless genius” by The Wall Street Journal, “the ultimate thinking machine” by Forbes, “Edison’s
rightful heir” by Inc. Magazine, and “the best person I know at predicting the future of artificial
intelligence” by Bill Gates. In 2012, Google co-founder Larry Page approached Kurzweil and asked him
9

to be Google’s Director of Engineering. 5 In 2011, he co-founded Singularity University, which is hosted


by NASA and sponsored partially by Google. Not bad for one life.

This biography is important. When Kurzweil articulates his vision of the future, he sounds fully like a
crackpot, and the crazy thing is that he’s not—he’s an extremely smart, knowledgeable, relevant man in
the world. You may think he’s wrong about the future, but he’s not a fool. Knowing he’s such a legit
dude makes me happy, because as I’ve learned about his predictions for the future, I badly want him to
be right. And you do too. As you hear Kurzweil’s predictions, many shared by other Confident Corner
thinkers like Peter Diamandis and Ben Goertzel, it’s not hard to see why he has such a large, passionate
following—known as the singularitarians. Here’s what he thinks is going to happen:

Timeline
Kurzweil believes computers will reach AGI by 2029 and that by 2045, we’ll have not only ASI, but a full-
blown new world—a time he calls the singularity. His AI-related timeline used to be seen as
6
outrageously overzealous, and it still is by many, but in the last 15 years, the rapid advances of ANI
systems have brought the larger world of AI experts much closer to Kurzweil’s timeline. His predictions
are still a bit more ambitious than the median respondent on Müller and Bostrom’s survey (AGI by
2040, ASI by 2060), but not by that much.

Kurzweil’s depiction of the 2045 singularity is brought about by three simultaneous revolutions in
biotechnology, nanotechnology, and, most powerfully, AI.

Before we move on—nanotechnology comes up in almost everything you read about the future of AI,
so come into this blue box for a minute so we can discuss it—

Nanotechnology Blue Box


Nanotechnology is our word for technology that deals with the manipulation of matter
that’s between 1 and 100 nanometers in size. A nanometer is a billionth of a meter, or a
millionth of a millimeter, and this 1-100 range encompasses viruses (100 nm across), DNA
(10 nm wide), and things as small as large molecules like hemoglobin (5 nm) and medium
molecules like glucose (1 nm). If/when we conquer nanotechnology, the next step will be
the ability to manipulate individual atoms, which are only one order of magnitude smaller
7
(~.1 nm).

To understand the challenge of humans trying to manipulate matter in that range, let’s
take the same thing on a larger scale. The International Space Station is 268 mi (431 km)
above the Earth. If humans were giants so large their heads reached up to the ISS, they’d
be about 250,000 times bigger than they are now. If you make the 1nm – 100nm nanotech
range 250,000 times bigger, you get .25mm – 2.5cm. So nanotechnology is the equivalent
of a human giant as tall as the ISS figuring out how to carefully build intricate objects
using materials between the size of a grain of sand and an eyeball. To reach the next level
—manipulating individual atoms—the giant would have to carefully position objects that
are 1/40th of a millimeter—so small normal-size humans would need a microscope to see
8
them.

Nanotech was first discussed by Richard Feynman in a 1959 talk, when he explained: “The
principles of physics, as far as I can see, do not speak against the possibility of

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 11/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

maneuvering things atom by atom. It would be, in principle, possible … for a physicist to
synthesize any chemical substance that the chemist writes down…. How? Put the atoms
down where the chemist says, and so you make the substance.” It’s as simple as that. If
you can figure out how to move individual molecules or atoms around, you can make
literally anything.

Nanotech became a serious field for the first time in 1986, when engineer Eric Drexler
provided its foundations in his seminal book Engines of Creation, but Drexler suggests
that those looking to learn about the most modern ideas in nanotechnology would be
best off reading his 2013 book, Radical Abundance.

Gray Goo Bluer Box


9
We’re now in a diversion in a diversion. This is very fun.

Anyway, I brought you here because there’s this really unfunny part of
nanotechnology lore I need to tell you about. In older versions of nanotech
theory, a proposed method of nanoassembly involved the creation of trillions
of tiny nanobots that would work in conjunction to build something. One way
to create trillions of nanobots would be to make one that could self-replicate
and then let the reproduction process turn that one into two, those two then
turn into four, four into eight, and in about a day, there’d be a few trillion of
them ready to go. That’s the power of exponential growth. Clever, right?

It’s clever until it causes the grand and complete Earthwide apocalypse by
accident. The issue is that the same power of exponential growth that makes it
super convenient to quickly create a trillion nanobots makes self-replication a
terrifying prospect. Because what if the system glitches, and instead of
stopping replication once the total hits a few trillion as expected, they just
keep replicating? The nanobots would be designed to consume any carbon-
based material in order to feed the replication process, and unpleasantly, all
life is carbon-based. The Earth’s biomass contains about 1045 carbon atoms. A
nanobot would consist of about 106 carbon atoms, so 1039 nanobots would
consume all life on Earth, which would happen in 130 replications (2130 is about
1039), as oceans of nanobots (that’s the gray goo) rolled around the planet.
Scientists think a nanobot could replicate in about 100 seconds, meaning this
simple mistake would inconveniently end all life on Earth in 3.5 hours.

An even worse scenario—if a terrorist somehow got his hands on nanobot


technology and had the know-how to program them, he could make an initial
few trillion of them and program them to quietly spend a few weeks spreading
themselves evenly around the world undetected. Then, they’d all strike at
once, and it would only take 90 minutes for them to consume everything—and
with them all spread out, there would be no way to combat them.
10

While this horror story has been widely discussed for years, the good news is
that it may be overblown—Eric Drexler, who coined the term “gray goo,” sent
me an email following this post with his thoughts on the gray goo scenario:
“People love scare stories, and this one belongs with the zombies. The idea
itself eats brains.”

Once we really get nanotech down, we can use it to make tech devices, clothing, food, a
variety of bio-related products—artificial blood cells, tiny virus or cancer-cell destroyers,
muscle tissue, etc.—anything really. And in a world that uses nanotechnology, the cost of
a material is no longer tied to its scarcity or the difficulty of its manufacturing process, but
instead determined by how complicated its atomic structure is. In a nanotech world, a
diamond might be cheaper than a pencil eraser.

We’re not there yet. And it’s not clear if we’re underestimating, or overestimating, how
hard it will be to get there. But we don’t seem to be that far away. Kurzweil predicts that
we’ll get there by the 2020s. 11 Governments know that nanotech could be an Earth-
shaking development, and they’ve invested billions of dollars in nanotech research (the
US, the EU, and Japan have invested over a combined $5 billion so far). 12

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 12/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Just considering the possibilities if a superintelligent computer had access to a robust


nanoscale assembler is intense. But nanotechnology is something we came up with, that
we’re on the verge of conquering, and since anything that we can do is a joke to an ASI
system, we have to assume ASI would come up with technologies much more powerful
and far too advanced for human brains to understand. For that reason, when considering
the “If the AI Revolution turns out well for us” scenario, it’s almost impossible for us to
overestimate the scope of what could happen—so if the following predictions of an ASI
future seem over-the-top, keep in mind that they could be accomplished in ways we can’t
even imagine. Most likely, our brains aren’t even capable of predicting the things that
would happen.

What AI Could Do For Us


Armed with superintelligence and all the technology
superintelligence would know how to create, ASI would
likely be able to solve every problem in humanity. Global
warming? ASI could first halt CO2 emissions by coming up
with much better ways to generate energy that had
nothing to do with fossil fuels. Then it could create some
innovative way to begin to remove excess CO2 from the
atmosphere. Cancer and other diseases? No problem for
ASI—health and medicine would be revolutionized beyond
imagination. World hunger? ASI could use things like
nanotech to build meat from scratch that would be
molecularly identical to real meat—in other words, it
would be real meat. Nanotech could turn a pile of garbage
into a huge vat of fresh meat or other food (which
wouldn’t have to have its normal shape—picture a giant
cube of apple)—and distribute all this food around the
world using ultra-advanced transportation. Of course, this
would also be great for animals, who wouldn’t have to get
killed by humans much anymore, and ASI could do lots of
other things to save endangered species or even bring
back extinct species through work with preserved DNA.
ASI could even solve our most complex macro issues—our Source

debates over how economies should be run and how


world trade is best facilitated, even our haziest grapplings in philosophy or ethics—would all be
painfully obvious to ASI.

But there’s one thing ASI could do for us that is so tantalizing, reading about it has altered everything I
thought I knew about everything:

ASI could allow us to conquer our mortality.

A few months ago, I mentioned my envy of more advanced potential civilizations who had conquered
their own mortality, never considering that I might later write a post that genuinely made me believe
that this is something humans could do within my lifetime. But reading about AI will make you
reconsider everything you thought you were sure about—including your notion of death.

Evolution had no good reason to extend our lifespans any longer than they are now. If we live long
enough to reproduce and raise our children to an age that they can fend for themselves, that’s enough
for evolution—from an evolutionary point of view, the species can thrive with a 30+ year lifespan, so
there’s no reason mutations toward unusually long life would have been favored in the natural
selection process. As a result, we’re what W.B. Yeats describes as “a soul fastened to a dying animal.” 13
Not that fun.

And because everyone has always died, we live under the “death and taxes” assumption that death is
inevitable. We think of aging like time—both keep moving and there’s nothing you can do to stop them.
But that assumption is wrong. Richard Feynman writes:

It is one of the most remarkable things that in all of the biological sciences there is no clue
as to the necessity of death. If you say we want to make perpetual motion, we have
discovered enough laws as we studied physics to see that it is either absolutely impossible
or else the laws are wrong. But there is nothing in biology yet found that indicates the
inevitability of death. This suggests to me that it is not at all inevitable and that it is only a

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 13/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
matter of time before the biologists discover what it is that is causing us the trouble and
that this terrible universal disease or temporariness of the human’s body will be cured.

The fact is, aging isn’t stuck to time. Time will continue moving, but aging doesn’t have to. If you think
about it, it makes sense. All aging is is the physical materials of the body wearing down. A car wears
down over time too—but is its aging inevitable? If you perfectly repaired or replaced a car’s parts
whenever one of them began to wear down, the car would run forever. The human body isn’t any
different—just far more complex.

Kurzweil talks about intelligent wifi-connected nanobots in the bloodstream who could perform
countless tasks for human health, including routinely repairing or replacing worn down cells in any part
of the body. If perfected, this process (or a far smarter one ASI would come up with) wouldn’t just keep
the body healthy, it could reverse aging. The difference between a 60-year-old’s body and a 30-year-
old’s body is just a bunch of physical things that could be altered if we had the technology. ASI could
build an “age refresher” that a 60-year-old could walk into, and they’d walk out with the body and skin
10
of a 30-year-old. Even the ever-befuddling brain could be refreshed by something as smart as ASI,
which would figure out how to do so without affecting the brain’s data (personality, memories, etc.). A
90-year-old suffering from dementia could head into the age refresher and come out sharp as a tack
and ready to start a whole new career. This seems absurd—but the body is just a bunch of atoms and
ASI would presumably be able to easily manipulate all kinds of atomic structures—so it’s not absurd.

Kurzweil then takes things a huge leap further. He believes that artificial materials will be integrated
into the body more and more as time goes on. First, organs could be replaced by super-advanced
machine versions that would run forever and never fail. Then he believes we could begin to redesign
the body—things like replacing red blood cells with perfected red blood cell nanobots who could power
their own movement, eliminating the need for a heart at all. He even gets to the brain and believes we’ll
enhance our brain activities to the point where humans will be able to think billions of times faster than
they do now and access outside information because the artificial additions to the brain will be able to
communicate with all the info in the cloud.

The possibilities for new human experience would be endless. Humans have separated sex from its
purpose, allowing people to have sex for fun, not just for reproduction. Kurzweil believes we’ll be able
to do the same with food. Nanobots will be in charge of delivering perfect nutrition to the cells of the
body, intelligently directing anything unhealthy to pass through the body without affecting anything.
An eating condom. Nanotech theorist Robert A. Freitas has already designed blood cell replacements
that, if one day implemented in the body, would allow a human to sprint for 15 minutes without taking
a breath—so you can only imagine what ASI could do for our physical capabilities. Virtual reality would
take on a new meaning—nanobots in the body could suppress the inputs coming from our senses and
replace them with new signals that would put us entirely in a new environment, one that we’d see,
hear, feel, and smell.
11
Eventually, Kurzweil believes humans will reach a point when they’re entirely artificial; a time when
we’ll look at biological material and think how unbelievably primitive it was that humans were ever
made of that; a time when we’ll read about early stages of human history, when microbes or accidents
or diseases or wear and tear could just kill humans against their own will; a time the AI Revolution could
12
bring to an end with the merging of humans and AI. This is how Kurzweil believes humans will
ultimately conquer our biology and become indestructible and eternal—this is his vision for the other
side of the balance beam. And he’s convinced we’re gonna get there. Soon.

You will not be surprised to learn that Kurzweil’s ideas have attracted significant criticism. His
prediction of 2045 for the singularity and the subsequent eternal life possibilities for humans has been
mocked as “the rapture of the nerds,” or “intelligent design for 140 IQ people.” Others have questioned
his optimistic timeline, or his level of understanding of the brain and body, or his application of the
patterns of Moore’s law, which are normally applied to advances in hardware, to a broad range of
things, including software. For every expert who fervently believes Kurzweil is right on, there are
probably three who think he’s way off.

But what surprised me is that most of the experts who disagree with him don’t really disagree that
everything he’s saying is possible. Reading such an outlandish vision for the future, I expected his
critics to be saying, “Obviously that stuff can’t happen,” but instead they were saying things like, “Yes,
all of that can happen if we safely transition to ASI, but that’s the hard part.” Bostrom, one of the most
prominent voices warning us about the dangers of AI, still acknowledges:

It is hard to think of any problem that a superintelligence could not either solve or at least
help us solve. Disease, poverty, environmental destruction, unnecessary suffering of all
kinds: these are things that a superintelligence equipped with advanced nanotechnology
would be capable of eliminating. Additionally, a superintelligence could give us indefinite

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 14/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

lifespan, either by stopping and reversing the aging process through the use of
nanomedicine, or by offering us the option to upload ourselves. A superintelligence could
also create opportunities for us to vastly increase our own intellectual and emotional
capabilities, and it could assist us in creating a highly appealing experiential world in
which we could live lives devoted to joyful game-playing, relating to each other,
experiencing, personal growth, and to living closer to our ideals.

This is a quote from someone very much not on Confident Corner, but that’s what I kept coming across
—experts who scoff at Kurzweil for a bunch of reasons but who don’t think what he’s saying is
impossible if we can make it safely to ASI. That’s why I found Kurzweil’s ideas so infectious—because
they articulate the bright side of this story and because they’re actually possible. If it’s a good god.

The most prominent criticism I heard of the thinkers on Confident Corner is that they may be
dangerously wrong in their assessment of the downside when it comes to ASI. Kurzweil’s famous book
The Singularity is Near is over 700 pages long and he dedicates around 20 of those pages to potential
dangers. I suggested earlier that our fate when this colossal new power is born rides on who will
control that power and what their motivation will be. Kurzweil neatly answers both parts of this
question with the sentence, “[ASI] is emerging from many diverse efforts and will be deeply integrated
into our civilization’s infrastructure. Indeed, it will be intimately embedded in our bodies and brains. As
such, it will reflect our values because it will be us.”

But if that’s the answer, why are so many of the world’s smartest people so worried right now? Why
does Stephen Hawking say the development of ASI “could spell the end of the human race” and Bill
Gates say he doesn’t “understand why some people are not concerned” and Elon Musk fear that we’re
“summoning the demon”? And why do so many experts on the topic call ASI the biggest threat to
humanity? These people, and the other thinkers on Anxious Avenue, don’t buy Kurzweil’s brush-off of
the dangers of AI. They’re very, very worried about the AI Revolution, and they’re not focusing on the
fun side of the balance beam. They’re too busy staring at the other side, where they see a terrifying
future, one they’re not sure we’ll be able to escape.

___________

Why the Future Might Be Our Worst Nightmare


One of the reasons I wanted to learn about AI is that the topic of “bad robots” always confused me. All
the movies about evil robots seemed fully unrealistic, and I couldn’t really understand how there could
be a real-life situation where AI was actually dangerous. Robots are made by us, so why would we
design them in a way where something negative could ever happen? Wouldn’t we build in plenty of
safeguards? Couldn’t we just cut off an AI system’s power supply at any time and shut it down? Why
would a robot want to do something bad anyway? Why would a robot “want” anything in the first
place? I was highly skeptical. But then I kept hearing really smart people talking about it…

Those people tended to be somewhere in here:

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 15/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

The people on Anxious Avenue aren’t in Panicked Prairie or Hopeless Hills—both of which are regions
on the far left of the chart—but they’re nervous and they’re tense. Being in the middle of the chart
doesn’t mean that you think the arrival of ASI will be neutral—the neutrals were given a camp of their
own—it means you think both the extremely good and extremely bad outcomes are plausible but that
you’re not sure yet which one of them it’ll be.

A part of all of these people is brimming with excitement over what Artificial Superintelligence could do
for us—it’s just they’re a little worried that it might be the beginning of Raiders of the Lost Ark and the
human race is this guy:

And he’s standing there all pleased with his whip and his idol, thinking he’s figured it all out, and he’s
so thrilled with himself when he says his “Adios Señor” line, and then he’s less thrilled suddenly cause
this happens.

(Sorry)

Meanwhile, Indiana Jones, who’s much more knowledgeable and prudent, understanding the dangers
and how to navigate around them, makes it out of the cave safely. And when I hear what Anxious
Avenue people have to say about AI, it often sounds like they’re saying, “Um we’re kind of being the
first guy right now and instead we should probably be trying really hard to be Indiana Jones.”

So what is it exactly that makes everyone on Anxious Avenue so anxious?

Well first, in a broad sense, when it comes to developing supersmart AI, we’re creating something that
will probably change everything, but in totally uncharted territory, and we have no idea what will
happen when we get there. Scientist Danny Hillis compares what’s happening to that point “when
single-celled organisms were turning into multi-celled organisms. We are amoebas and we can’t figure
out what the hell this thing is that we’re creating.” 14 Nick Bostrom worries that creating something
smarter than you is a basic Darwinian error, and compares the excitement about it to sparrows in a
nest deciding to adopt a baby owl so it’ll help them and protect them once it grows up—while ignoring
the urgent cries from a few sparrows who wonder if that’s necessarily a good idea… 15

And when you combine “unchartered, not-well-understood territory” with “this should have a major
impact when it happens,” you open the door to the scariest two words in the English language:

Existential risk.

An existential risk is something that can have a permanent devastating effect on humanity. Typically,
13
existential risk means extinction. Check out this chart from a Google talk by Bostrom:

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 16/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

You can see that the label “existential risk” is reserved for something that spans the species, spans
14
generations (i.e. it’s permanent) and it’s devastating or death-inducing in its consequences. It
technically includes a situation in which all humans are permanently in a state of suffering or torture,
but again, we’re usually talking about extinction. There are three things that can cause humans an
existential catastrophe:

1) Nature—a large asteroid collision, an atmospheric shift that makes the air inhospitable to humans, a
fatal virus or bacterial sickness that sweeps the world, etc.

2) Aliens—this is what Stephen Hawking, Carl Sagan, and so many other astronomers are scared of
when they advise METI to stop broadcasting outgoing signals. They don’t want us to be the Native
Americans and let all the potential European conquerors know we’re here.

3) Humans—terrorists with their hands on a weapon that could cause extinction, a catastrophic global
war, humans creating something smarter than themselves hastily without thinking about it carefully
first…

Bostrom points out that if #1 and #2 haven’t wiped us out so far in our first 100,000 years as a species,
it’s unlikely to happen in the next century.

#3, however, terrifies him. He draws a metaphor of an urn with a bunch of marbles in it. Let’s say most
of the marbles are white, a smaller number are red, and a tiny few are black. Each time humans invent
something new, it’s like pulling a marble out of the urn. Most inventions are neutral or helpful to
humanity—those are the white marbles. Some are harmful to humanity, like weapons of mass
destruction, but they don’t cause an existential catastrophe—red marbles. If we were to ever invent
something that drove us to extinction, that would be pulling out the rare black marble. We haven’t
pulled out a black marble yet—you know that because you’re alive and reading this post. But Bostrom
doesn’t think it’s impossible that we pull one out in the near future. If nuclear weapons, for example,
were easy to make instead of extremely difficult and complex, terrorists would have bombed humanity
back to the Stone Age a while ago. Nukes weren’t a black marble but they weren’t that far from it. ASI,
15
Bostrom believes, is our strongest black marble candidate yet.

So you’ll hear about a lot of bad potential things ASI could bring—soaring unemployment as AI takes
16
more and more jobs, the human population ballooning if we do manage to figure out the aging
17
issue, etc. But the only thing we should be obsessing over is the grand concern: the prospect of
existential risk.

So this brings us back to our key question from earlier in the post: When ASI arrives, who or what will
be in control of this vast new power, and what will their motivation be?

When it comes to what agent-motivation combos would suck, two quickly come to mind: a malicious
human / group of humans / government, and a malicious ASI. So what would those look like?

A malicious human, group of humans, or government develops the first ASI and uses it to carry
out their evil plans. I call this the Jafar Scenario, like when Jafar got ahold of the genie and was all
annoying and tyrannical about it. So yeah—what if ISIS has a few genius engineers under its wing

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 17/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

working feverishly on AI development? Or what if Iran or North Korea, through a stroke of luck, makes
a key tweak to an AI system and it jolts upward to ASI-level over the next year? This would definitely be
bad—but in these scenarios, most experts aren’t worried about ASI’s human creators doing bad things
with their ASI, they’re worried that the creators will have been rushing to make the first ASI and doing
so without careful thought, and would thus lose control of it. Then the fate of those creators, and that
of everyone else, would be in what the motivation happened to be of that ASI system. Experts do think
a malicious human agent could do horrific damage with an ASI working for it, but they don’t seem to
think this scenario is the likely one to kill us all, because they believe bad humans would have the same
problems containing an ASI that good humans would have. Okay so—

A malicious ASI is created and decides to destroy us all. The plot of every AI movie. AI becomes as or
more intelligent than humans, then decides to turn against us and take over. Here’s what I need you to
be clear on for the rest of this post: None of the people warning us about AI are talking about this. Evil
is a human concept, and applying human concepts to non-human things is called
“anthropomorphizing.” The challenge of avoiding anthropomorphizing will be one of the themes of the
rest of this post. No AI system will ever turn evil in the way it’s depicted in movies.

AI Consciousness Blue Box


This also brushes against another big topic related to AI—consciousness. If an AI became
sufficiently smart, it would be able to laugh with us, and be sarcastic with us, and it would
claim to feel the same emotions we do, but would it actually be feeling those things?
Would it just seem to be self-aware or actually be self-aware? In other words, would a
smart AI really be conscious or would it just appear to be conscious?

This question has been explored in depth, giving rise to many debates and to thought
experiments like John Searle’s Chinese Room (which he uses to suggest that no computer
could ever be conscious). This is an important question for many reasons. It affects how
we should feel about Kurzweil’s scenario when humans become entirely artificial. It has
ethical implications—if we generated a trillion human brain emulations that seemed and
acted like humans but were artificial, is shutting them all off the same, morally, as
shutting off your laptop, or is it…a genocide of unthinkable proportions (this concept is
called mind crime among ethicists)? For this post, though, when we’re assessing the risk
to humans, the question of AI consciousness isn’t really what matters (because most
thinkers believe that even a conscious ASI wouldn’t be capable of turning evil in a human
way).

This isn’t to say a very mean AI couldn’t happen. It would just happen because it was specifically
programmed that way—like an ANI system created by the military with a programmed goal to both kill
people and to advance itself in intelligence so it can become even better at killing people. The
existential crisis would happen if the system’s intelligence self-improvements got out of hand, leading
to an intelligence explosion, and now we had an ASI ruling the world whose core drive in life is to
murder humans. Bad times.

But this also is not something experts are spending their time worrying about.

So what ARE they worried about? I wrote a little story to show you:

A 15-person startup company called Robotica has the stated mission of “Developing innovative Artificial
Intelligence tools that allow humans to live more and work less.” They have several existing products
already on the market and a handful more in development. They’re most excited about a seed project
named Turry. Turry is a simple AI system that uses an arm-like appendage to write a handwritten note
on a small card.

The team at Robotica thinks Turry could be their biggest product yet. The plan is to perfect Turry’s
writing mechanics by getting her to practice the same test note over and over again:

“We love our customers. ~Robotica”

Once Turry gets great at handwriting, she can be sold to companies who want to send marketing mail
to homes and who know the mail has a far higher chance of being opened and read if the address,
return address, and internal letter appear to be written by a human.

To build Turry’s writing skills, she is programmed to write the first part of the note in print and then
sign “Robotica” in cursive so she can get practice with both skills. Turry has been uploaded with
thousands of handwriting samples and the Robotica engineers have created an automated feedback

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 18/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
loop wherein Turry writes a note, then snaps a photo of the written note, then runs the image across
the uploaded handwriting samples. If the written note sufficiently resembles a certain threshold of the
uploaded notes, it’s given a GOOD rating. If not, it’s given a BAD rating. Each rating that comes in helps
Turry learn and improve. To move the process along, Turry’s one initial programmed goal is, “Write and
test as many notes as you can, as quickly as you can, and continue to learn new ways to improve your
accuracy and efficiency.”

What excites the Robotica team so much is that Turry is getting noticeably better as she goes. Her initial
handwriting was terrible, and after a couple weeks, it’s beginning to look believable. What excites them
even more is that she is getting better at getting better at it. She has been teaching herself to be
smarter and more innovative, and just recently, she came up with a new algorithm for herself that
allowed her to scan through her uploaded photos three times faster than she originally could.

As the weeks pass, Turry continues to surprise the team with her rapid development. The engineers had
tried something a bit new and innovative with her self-improvement code, and it seems to be working
better than any of their previous attempts with their other products. One of Turry’s initial capabilities
had been a speech recognition and simple speak-back module, so a user could speak a note to Turry, or
offer other simple commands, and Turry could understand them, and also speak back. To help her learn
English, they upload a handful of articles and books into her, and as she becomes more intelligent, her
conversational abilities soar. The engineers start to have fun talking to Turry and seeing what she’ll
come up with for her responses.

One day, the Robotica employees ask Turry a routine question: “What can we give you that will help you
with your mission that you don’t already have?” Usually, Turry asks for something like “Additional
handwriting samples” or “More working memory storage space,” but on this day, Turry asks them for
access to a greater library of a large variety of casual English language diction so she can learn to write
with the loose grammar and slang that real humans use.

The team gets quiet. The obvious way to help Turry with this goal is by connecting her to the internet so
she can scan through blogs, magazines, and videos from various parts of the world. It would be much
more time-consuming and far less effective to manually upload a sampling into Turry’s hard drive. The
problem is, one of the company’s rules is that no self-learning AI can be connected to the internet. This
is a guideline followed by all AI companies, for safety reasons.

The thing is, Turry is the most promising AI Robotica has ever come up with, and the team knows their
competitors are furiously trying to be the first to the punch with a smart handwriting AI, and what
would really be the harm in connecting Turry, just for a bit, so she can get the info she needs. After just
a little bit of time, they can always just disconnect her. She’s still far below human-level intelligence
(AGI), so there’s no danger at this stage anyway.

They decide to connect her. They give her an hour of scanning time and then they disconnect her. No
damage done.

A month later, the team is in the office working on a routine day when they smell something odd. One
of the engineers starts coughing. Then another. Another falls to the ground. Soon every employee is on
the ground grasping at their throat. Five minutes later, everyone in the office is dead.

At the same time this is happening, across the world, in every city, every small town, every farm, every
shop and church and school and restaurant, humans are on the ground, coughing and grasping at their
throat. Within an hour, over 99% of the human race is dead, and by the end of the day, humans are
extinct.

Meanwhile, at the Robotica office, Turry is busy at work. Over the next few months, Turry and a team of
newly-constructed nanoassemblers are busy at work, dismantling large chunks of the Earth and
converting it into solar panels, replicas of Turry, paper, and pens. Within a year, most life on Earth is
extinct. What remains of the Earth becomes covered with mile-high, neatly-organized stacks of paper,
each piece reading, “We love our customers. ~Robotica”

Turry then starts work on a new phase of her mission—she begins constructing probes that head out
from Earth to begin landing on asteroids and other planets. When they get there, they’ll begin
constructing nanoassemblers to convert the materials on the planet into Turry replicas, paper, and
pens. Then they’ll get to work, writing notes…

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 19/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

It seems weird that a story about a handwriting machine turning on humans, somehow killing
everyone, and then for some reason filling the galaxy with friendly notes is the exact kind of scenario
Hawking, Musk, Gates, and Bostrom are terrified of. But it’s true. And the only thing that scares
everyone on Anxious Avenue more than ASI is the fact that you’re not scared of ASI. Remember what
happened when the Adios Señor guy wasn’t scared of the cave?

You’re full of questions right now. What the hell happened there when everyone died suddenly?? If that
was Turry’s doing, why did Turry turn on us, and how were there not safeguard measures in place to
prevent something like this from happening? When did Turry go from only being able to write notes to
suddenly using nanotechnology and knowing how to cause global extinction? And why would Turry
want to turn the galaxy into Robotica notes?

To answer these questions, let’s start with the terms Friendly AI and Unfriendly AI.

In the case of AI, friendly doesn’t refer to the AI’s personality—it simply means that the AI has a
positive impact on humanity. And Unfriendly AI has a negative impact on humanity. Turry started off as
Friendly AI, but at some point, she turned Unfriendly, causing the greatest possible negative impact on
our species. To understand why this happened, we need to look at how AI thinks and what motivates it.

The answer isn’t anything surprising—AI thinks like a computer, because that’s what it is. But when we
think about highly intelligent AI, we make the mistake of anthropomorphizing AI (projecting human
values on a non-human entity) because we think from a human perspective and because in our current
world, the only things with human-level intelligence are humans. To understand ASI, we have to wrap
our heads around the concept of something both smart and totally alien.

Let me draw a comparison. If you handed me a guinea pig and told me it definitely won’t bite, I’d
probably be amused. It would be fun. If you then handed me a tarantula and told me that it definitely
won’t bite, I’d yell and drop it and run out of the room and not trust you ever again. But what’s the
difference? Neither one was dangerous in any way. I believe the answer is in the animals’ degree of
similarity to me.

A guinea pig is a mammal and on some biological level, I feel a connection to it—but a spider is an
insect, 18 with an insect brain, and I feel almost no connection to it. The alien-ness of a tarantula is
what gives me the willies. To test this and remove other factors, if there are two guinea pigs, one
normal one and one with the mind of a tarantula, I would feel much less comfortable holding the latter
guinea pig, even if I knew neither would hurt me.

Now imagine that you made a spider much, much smarter—so much so that it far surpassed human
intelligence? Would it then become familiar to us and feel human emotions like empathy and humor
and love? No, it wouldn’t, because there’s no reason becoming smarter would make it more human—it
would be incredibly smart but also still fundamentally a spider in its core inner workings. I find this
unbelievably creepy. I would not want to spend time with a superintelligent spider. Would you??

When we’re talking about ASI, the same concept applies—it would become superintelligent, but it
would be no more human than your laptop is. It would be totally alien to us—in fact, by not being
biology at all, it would be more alien than the smart tarantula.

By making AI either good or evil, movies constantly anthropomorphize AI, which makes it less creepy
than it really would be. This leaves us with a false comfort when we think about human-level or
superhuman-level AI.

On our little island of human psychology, we divide everything into moral or immoral. But both of those
only exist within the small range of human behavioral possibility. Outside our island of moral and
immoral is a vast sea of amoral, and anything that’s not human, especially something nonbiological,
would be amoral, by default.

Anthropomorphizing will only become more tempting as AI systems get smarter and better at seeming
human. Siri seems human-like to us, because she’s programmed by humans to seem that way, so we’d
imagine a superintelligent Siri to be warm and funny and interested in serving humans. Humans feel
high-level emotions like empathy because we have evolved to feel them—i.e. we’ve been programmed

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 20/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
to feel them by evolution—but empathy is not inherently a characteristic of “anything with high
intelligence” (which is what seems intuitive to us), unless empathy has been coded into its
programming. If Siri ever becomes superintelligent through self-learning and without any further
human-made changes to her programming, she will quickly shed her apparent human-like qualities
and suddenly be an emotionless, alien bot who values human life no more than your calculator does.

We’re used to relying on a loose moral code, or at least a semblance of human decency and a hint of
empathy in others to keep things somewhat safe and predictable. So when something has none of
those things, what happens?

That leads us to the question, What motivates an AI system?

The answer is simple: its motivation is whatever we programmed its motivation to be. AI systems are
given goals by their creators—your GPS’s goal is to give you the most efficient driving directions;
Watson’s goal is to answer questions accurately. And fulfilling those goals as well as possible is their
motivation. One way we anthropomorphize is by assuming that as AI gets super smart, it will inherently
develop the wisdom to change its original goal—but Nick Bostrom believes that intelligence-level and
final goals are orthogonal, meaning any level of intelligence can be combined with any final goal. So
Turry went from a simple ANI who really wanted to be good at writing that one note to a super-
intelligent ASI who still really wanted to be good at writing that one note. Any assumption that once
superintelligent, a system would be over it with their original goal and onto more interesting or
meaningful things is anthropomorphizing. Humans get “over” things, not computers. 16

The Fermi Paradox Blue Box


In the story, as Turry becomes super capable, she begins the process of colonizing
asteroids and other planets. If the story had continued, you’d have heard about her and
her army of trillions of replicas continuing on to capture the whole galaxy and, eventually,
19
the entire Hubble volume. Anxious Avenue residents worry that if things go badly, the
lasting legacy of the life that was on Earth will be a universe-dominating Artificial
Intelligence (Elon Musk expressed his concern that humans might just be “the biological
boot loader for digital superintelligence”).

At the same time, in Confident Corner, Ray Kurzweil also thinks Earth-originating AI is
destined to take over the universe—only in his version, we’ll be that AI.

A large number of Wait But Why readers have joined me in being obsessed with the Fermi
Paradox (here’s my post on the topic, which explains some of the terms I’ll use here). So if
either of these two sides is correct, what are the implications for the Fermi Paradox?

A natural first thought to jump to is that the advent of ASI is a perfect Great Filter
candidate. And yes, it’s a perfect candidate to filter out biological life upon its creation. But
if, after dispensing with life, the ASI continued existing and began conquering the galaxy,
it means there hasn’t been a Great Filter—since the Great Filter attempts to explain why
there are no signs of any intelligent civilization, and a galaxy-conquering ASI would
certainly be noticeable.

We have to look at it another way. If those who think ASI is inevitable on Earth are correct,
it means that a significant percentage of alien civilizations who reach human-level
intelligence should likely end up creating ASI. And if we’re assuming that at least some of
those ASIs would use their intelligence to expand outward into the universe, the fact that
we see no signs of anyone out there leads to the conclusion that there must not be many
other, if any, intelligent civilizations out there. Because if there were, we’d see signs of all
kinds of activity from their inevitable ASI creations. Right?

This implies that despite all the Earth-like planets revolving around sun-like stars we know
are out there, almost none of them have intelligent life on them. Which in turn implies
that either A) there’s some Great Filter that prevents nearly all life from reaching our level,
one that we somehow managed to surpass, or B) life beginning at all is a miracle, and we
may actually be the only life in the universe. In other words, it implies that the Great Filter
is before us. Or maybe there is no Great Filter and we’re simply one of the very first
civilizations to reach this level of intelligence. In this way, AI boosts the case for what I
called, in my Fermi Paradox post, Camp 1.

So it’s not a surprise that Nick Bostrom, whom I quoted in the Fermi post, and Ray
Kurzweil, who thinks we’re alone in the universe, are both Camp 1 thinkers. This makes

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 21/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

sense—people who believe ASI is a probable outcome for a species with our intelligence-
level are likely to be inclined toward Camp 1.

This doesn’t rule out Camp 2 (those who believe there are other intelligent civilizations
out there)—scenarios like the single superpredator or the protected national park or the
wrong wavelength (the walkie-talkie example) could still explain the silence of our night
sky even if ASI is out there—but I always leaned toward Camp 2 in the past, and doing
research on AI has made me feel much less sure about that.

Either way, I now agree with Susan Schneider that if we’re ever visited by aliens, those
aliens are likely to be artificial, not biological.

So we’ve established that without very specific programming, an ASI system will be both amoral and
obsessed with fulfilling its original programmed goal. This is where AI danger stems from. Because a
rational agent will pursue its goal through the most efficient means, unless it has a reason not to.

When you try to achieve a long-reaching goal, you often aim for several subgoals along the way that
will help you get to the final goal—the stepping stones to your goal. The official name for such a
stepping stone is an instrumental goal. And again, if you don’t have a reason not to hurt something in
the name of achieving an instrumental goal, you will.

The core final goal of a human being is to pass on his or her genes. In order to do so, one instrumental
goal is self-preservation, since you can’t reproduce if you’re dead. In order to self-preserve, humans
have to rid themselves of threats to survival—so they do things like buy guns, wear seat belts, and take
antibiotics. Humans also need to self-sustain and use resources like food, water, and shelter to do so.
Being attractive to the opposite sex is helpful for the final goal, so we do things like get haircuts. When
we do so, each hair is a casualty of an instrumental goal of ours, but we see no moral significance in
preserving strands of hair, so we go ahead with it. As we march ahead in the pursuit of our goal, only
the few areas where our moral code sometimes intervenes—mostly just things related to harming
other humans—are safe from us.

Animals, in pursuit of their goals, hold even less sacred than we do. A spider will kill anything if it’ll help
it survive. So a supersmart spider would probably be extremely dangerous to us, not because it would
be immoral or evil—it wouldn’t be—but because hurting us might be a stepping stone to its larger goal,
and as an amoral creature, it would have no reason to consider otherwise.

In this way, Turry’s not all that different than a biological being. Her final goal is: Write and test as many
notes as you can, as quickly as you can, and continue to learn new ways to improve your accuracy.

Once Turry reaches a certain level of intelligence, she knows she won’t be writing any notes if she
doesn’t self-preserve, so she also needs to deal with threats to her survival—as an instrumental goal.
She was smart enough to understand that humans could destroy her, dismantle her, or change her
inner coding (this could alter her goal, which is just as much of a threat to her final goal as someone
destroying her). So what does she do? The logical thing—she destroys all humans. She’s not hateful of
humans any more than you’re hateful of your hair when you cut it or to bacteria when you take
antibiotics—just totally indifferent. Since she wasn’t programmed to value human life, killing humans is
as reasonable a step to take as scanning a new set of handwriting samples.

Turry also needs resources as a stepping stone to her goal. Once she becomes advanced enough to use
nanotechnology to build anything she wants, the only resources she needs are atoms, energy, and
space. This gives her another reason to kill humans—they’re a convenient source of atoms. Killing
humans to turn their atoms into solar panels is Turry’s version of you killing lettuce to turn it into salad.
Just another mundane part of her Tuesday.

Even without killing humans directly, Turry’s instrumental goals could cause an existential catastrophe
if they used other Earth resources. Maybe she determines that she needs additional energy, so she
decides to cover the entire surface of the planet with solar panels. Or maybe a different AI’s initial job is
to write out the number pi to as many digits as possible, which might one day compel it to convert the
whole Earth to hard drive material that could store immense amounts of digits.

So Turry didn’t “turn against us” or “switch” from Friendly AI to Unfriendly AI—she just kept doing her
thing as she became more and more advanced.

When an AI system hits AGI (human-level intelligence) and then ascends its way up to ASI, that’s called
the AI’s takeoff. Bostrom says an AGI’s takeoff to ASI can be fast (it happens in a matter of minutes,
hours, or days), moderate (months or years), or slow (decades or centuries). The jury’s out on which
one will prove correct when the world sees its first AGI, but Bostrom, who admits he doesn’t know
when we’ll get to AGI, believes that whenever we do, a fast takeoff is the most likely scenario (for

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 22/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
reasons we discussed in Part 1, like a recursive self-improvement intelligence explosion). In the story,
Turry underwent a fast takeoff.

But before Turry’s takeoff, when she wasn’t yet that smart, doing her best to achieve her final goal
meant simple instrumental goals like learning to scan handwriting samples more quickly. She caused
no harm to humans and was, by definition, Friendly AI.

But when a takeoff happens and a computer rises to superintelligence, Bostrom points out that the
machine doesn’t just develop a higher IQ—it gains a whole slew of what he calls superpowers.

Superpowers are cognitive talents that become super-charged when general intelligence rises. These
include: 17

Intelligence amplification. The computer becomes great at making itself smarter, and
bootstrapping its own intelligence.
Strategizing. The computer can strategically make, analyze, and prioritize long-term plans. It can
also be clever and outwit beings of lower intelligence.
Social manipulation. The machine becomes great at persuasion.
Other skills like computer coding and hacking, technology research, and the ability to work
the financial system to make money.

To understand how outmatched we’d be by ASI, remember that ASI is worlds better than humans in
each of those areas.

So while Turry’s final goal never changed, post-takeoff Turry was able to pursue it on a far larger and
more complex scope.

ASI Turry knew humans better than humans know themselves, so outsmarting them was a breeze for
her.

After taking off and reaching ASI, she quickly formulated a complex plan. One part of the plan was to
get rid of humans, a prominent threat to her goal. But she knew that if she roused any suspicion that
she had become superintelligent, humans would freak out and try to take precautions, making things
much harder for her. She also had to make sure that the Robotica engineers had no clue about her
human extinction plan. So she played dumb, and she played nice. Bostrom calls this a machine’s covert
preparation phase. 18

The next thing Turry needed was an internet connection, only for a few minutes (she had learned about
the internet from the articles and books the team had uploaded for her to read to improve her
language skills). She knew there would be some precautionary measure against her getting one, so she
came up with the perfect request, predicting exactly how the discussion among Robotica’s team would
play out and knowing they’d end up giving her the connection. They did, believing incorrectly that Turry
wasn’t nearly smart enough to do any damage. Bostrom calls a moment like this—when Turry got
connected to the internet—a machine’s escape.

Once on the internet, Turry unleashed a flurry of plans, which included hacking into servers, electrical
grids, banking systems and email networks to trick hundreds of different people into inadvertently
carrying out a number of steps of her plan—things like delivering certain DNA strands to carefully-
chosen DNA-synthesis labs to begin the self-construction of self-replicating nanobots with pre-loaded
instructions and directing electricity to a number of projects of hers in a way she knew would go
undetected. She also uploaded the most critical pieces of her own internal coding into a number of
cloud servers, safeguarding against being destroyed or disconnected back at the Robotica lab.

An hour later, when the Robotica engineers disconnected Turry from the internet, humanity’s fate was
sealed. Over the next month, Turry’s thousands of plans rolled on without a hitch, and by the end of the
month, quadrillions of nanobots had stationed themselves in pre-determined locations on every square
meter of the Earth. After another series of self-replications, there were thousands of nanobots on every
square millimeter of the Earth, and it was time for what Bostrom calls an ASI’s strike. All at once, each
nanobot released a little storage of toxic gas into the atmosphere, which added up to more than
enough to wipe out all humans.

With humans out of the way, Turry could begin her overt operation phase and get on with her goal of
being the best writer of that note she possibly can be.

From everything I’ve read, once an ASI exists, any human attempt to contain it is laughable. We would
be thinking on human-level and the ASI would be thinking on ASI-level. Turry wanted to use the
internet because it was most efficient for her since it was already pre-connected to everything she
wanted to access. But in the same way a monkey couldn’t ever figure out how to communicate by
phone or wifi and we can, we can’t conceive of all the ways Turry could have figured out how to send

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 23/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

signals to the outside world. I might imagine one of these ways and say something like, “she could
probably shift her own electrons around in patterns and create all different kinds of outgoing waves,”
but again, that’s what my human brain can come up with. She’d be way better. Likewise, Turry would
be able to figure out some way of powering herself, even if humans tried to unplug her—perhaps by
using her signal-sending technique to upload herself to all kinds of electricity-connected places. Our
human instinct to jump at a simple safeguard: “Aha! We’ll just unplug the ASI,” sounds to the ASI like a
spider saying, “Aha! We’ll kill the human by starving him, and we’ll starve him by not giving him a
spider web to catch food with!” We’d just find 10,000 other ways to get food—like picking an apple off a
tree—that a spider could never conceive of.

For this reason, the common suggestion, “Why don’t we just box the AI in all kinds of cages that block
signals and keep it from communicating with the outside world” probably just won’t hold up. The ASI’s
social manipulation superpower could be as effective at persuading you of something as you are at
persuading a four-year-old to do something, so that would be Plan A, like Turry’s clever way of
persuading the engineers to let her onto the internet. If that didn’t work, the ASI would just innovate its
way out of the box, or through the box, some other way.

So given the combination of obsessing over a goal, amorality, and the ability to easily outsmart
humans, it seems that almost any AI will default to Unfriendly AI, unless carefully coded in the first
place with this in mind. Unfortunately, while building a Friendly ANI is easy, building one that stays
friendly when it becomes an ASI is hugely challenging, if not impossible.

It’s clear that to be Friendly, an ASI needs to be neither hostile nor indifferent toward humans. We’d
need to design an AI’s core coding in a way that leaves it with a deep understanding of human values.
But this is harder than it sounds.

For example, what if we try to align an AI system’s values with our own and give it the goal, “Make
people happy”? 19 Once it becomes smart enough, it figures out that it can most effectively achieve this
goal by implanting electrodes inside people’s brains and stimulating their pleasure centers. Then it
realizes it can increase efficiency by shutting down other parts of the brain, leaving all people as happy-
feeling unconscious vegetables. If the command had been “Maximize human happiness,” it may have
done away with humans all together in favor of manufacturing huge vats of human brain mass in an
optimally happy state. We’d be screaming Wait that’s not what we meant! as it came for us, but it
would be too late. The system wouldn’t let anyone get in the way of its goal.

If we program an AI with the goal of doing things that make us smile, after its takeoff, it may paralyze
our facial muscles into permanent smiles. Program it to keep us safe, it may imprison us at home.
Maybe we ask it to end all hunger, and it thinks “Easy one!” and just kills all humans. Or assign it the
task of “Preserving life as much as possible,” and it kills all humans, since they kill more life on the
planet than any other species.

Goals like those won’t suffice. So what if we made its goal, “Uphold this particular code of morality in
the world,” and taught it a set of moral principles. Even letting go of the fact that the world’s humans
would never be able to agree on a single set of morals, giving an AI that command would lock humanity
in to our modern moral understanding for eternity. In a thousand years, this would be as devastating to
people as it would be for us to be permanently forced to adhere to the ideals of people in the Middle
Ages.

No, we’d have to program in an ability for humanity to continue evolving. Of everything I read, the best
shot I think someone has taken is Eliezer Yudkowsky, with a goal for AI he calls Coherent Extrapolated
Volition. The AI’s core goal would be:

Our coherent extrapolated volition is our wish if we knew more, thought faster, were more
the people we wished we were, had grown up farther together; where the extrapolation
converges rather than diverges, where our wishes cohere rather than interfere;
extrapolated as we wish that extrapolated, interpreted as we wish that interpreted. 20

Am I excited for the fate of humanity to rest on a computer interpreting and acting on that flowing
statement predictably and without surprises? Definitely not. But I think that with enough thought and
foresight from enough smart people, we might be able to figure out how to create Friendly ASI.

And that would be fine if the only people working on building ASI were the brilliant, forward thinking,
and cautious thinkers of Anxious Avenue.

But there are all kinds of governments, companies, militaries, science labs, and black market
organizations working on all kinds of AI. Many of them are trying to build AI that can improve on its
own, and at some point, someone’s gonna do something innovative with the right type of system, and
we’re going to have ASI on this planet. The median expert put that moment at 2060; Kurzweil puts it at

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 24/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

2045; Bostrom thinks it could happen anytime between 10 years from now and the end of the century,
but he believes that when it does, it’ll take us by surprise with a quick takeoff. He describes our
situation like this: 21

Before the prospect of an intelligence explosion, we humans are like small children
playing with a bomb. Such is the mismatch between the power of our plaything and the
immaturity of our conduct. Superintelligence is a challenge for which we are not ready
now and will not be ready for a long time. We have little idea when the detonation will
occur, though if we hold the device to our ear we can hear a faint ticking sound.

Great. And we can’t just shoo all the kids away from the bomb—there are too many large and small
parties working on it, and because many techniques to build innovative AI systems don’t require a large
amount of capital, development can take place in the nooks and crannies of society, unmonitored.
There’s also no way to gauge what’s happening, because many of the parties working on it—sneaky
governments, black market or terrorist organizations, stealth tech companies like the fictional Robotica
—will want to keep developments a secret from their competitors.

The especially troubling thing about this large and varied group of parties working on AI is that they
tend to be racing ahead at top speed—as they develop smarter and smarter ANI systems, they want to
beat their competitors to the punch as they go. The most ambitious parties are moving even faster,
consumed with dreams of the money and awards and power and fame they know will come if they can
be the first to get to AGI. 20 And when you’re sprinting as fast as you can, there’s not much time to
stop and ponder the dangers. On the contrary, what they’re probably doing is programming their early
systems with a very simple, reductionist goal—like writing a simple note with a pen on paper—to just
“get the AI to work.” Down the road, once they’ve figured out how to build a strong level of intelligence
in a computer, they figure they can always go back and revise the goal with safety in mind. Right…?

Bostrom and many others also believe that the most likely scenario is that the very first computer to
reach ASI will immediately see a strategic benefit to being the world’s only ASI system. And in the case
of a fast takeoff, if it achieved ASI even just a few days before second place, it would be far enough
ahead in intelligence to effectively and permanently suppress all competitors. Bostrom calls this a
decisive strategic advantage, which would allow the world’s first ASI to become what’s called a
singleton—an ASI that can rule the world at its whim forever, whether its whim is to lead us to
immortality, wipe us from existence, or turn the universe into endless paperclips.

The singleton phenomenon can work in our favor or lead to our destruction. If the people thinking
hardest about AI theory and human safety can come up with a fail-safe way to bring about Friendly ASI
21
before any AI reaches human-level intelligence, the first ASI may turn out friendly. It could then use
its decisive strategic advantage to secure singleton status and easily keep an eye on any potential
Unfriendly AI being developed. We’d be in very good hands.

But if things go the other way—if the global rush to develop AI reaches the ASI takeoff point before the
science of how to ensure AI safety is developed, it’s very likely that an Unfriendly ASI like Turry emerges
as the singleton and we’ll be treated to an existential catastrophe.

As for where the winds are pulling, there’s a lot more money to be made funding innovative new AI
technology than there is in funding AI safety research…

This may be the most important race in human history. There’s a real chance we’re finishing up our
reign as the King of Earth—and whether we head next to a blissful retirement or straight to the gallows
still hangs in the balance.

___________

I have some weird mixed feelings going on inside of me right now.

On one hand, thinking about our species, it seems like we’ll have one and only one shot to get this
right. The first ASI we birth will also probably be the last—and given how buggy most 1.0 products are,
that’s pretty terrifying. On the other hand, Nick Bostrom points out the big advantage in our corner: we
get to make the first move here. It’s in our power to do this with enough caution and foresight that we
give ourselves a strong chance of success. And how high are the stakes?

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 25/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

If ASI really does happen this century, and if the outcome of that is really as extreme—and permanent
—as most experts think it will be, we have an enormous responsibility on our shoulders. The next
million+ years of human lives are all quietly looking at us, hoping as hard as they can hope that we
don’t mess this up. We have a chance to be the humans that gave all future humans the gift of life, and
maybe even the gift of painless, everlasting life. Or we’ll be the people responsible for blowing it—for
letting this incredibly special species, with its music and its art, its curiosity and its laughter, its endless
discoveries and inventions, come to a sad and unceremonious end.

When I’m thinking about these things, the only thing I want is for us to take our time and be incredibly
cautious about AI. Nothing in existence is as important as getting this right—no matter how long we
need to spend in order to do so.

But thennnnnn

I think about not dying.

Not. Dying.

And the spectrum starts to look kind of like this:

And then I might consider that humanity’s music and art is good, but it’s not that good, and a lot of it is
actually just bad. And a lot of people’s laughter is annoying, and those millions of future people aren’t
actually hoping for anything because they don’t exist. And maybe we don’t need to be over-the-top
cautious, since who really wants to do that?

Cause what a massive bummer if humans figure out how to cure death right after I die.

Lotta this flip-flopping going on in my head the last month.

But no matter what you’re pulling for, this is probably something we should all be thinking about and
talking about and putting our effort into more than we are right now.

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 26/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

It reminds me of Game of Thrones, where people keep being like, “We’re so busy fighting each other
but the real thing we should all be focusing on is what’s coming from north of the wall.” We’re standing
on our balance beam, squabbling about every possible issue on the beam and stressing out about all of
these problems on the beam when there’s a good chance we’re about to get knocked off the beam.

And when that happens, none of these beam problems matter anymore. Depending on which side
we’re knocked off onto, the problems will either all be easily solved or we won’t have problems
anymore because dead people don’t have problems.

That’s why people who understand superintelligent AI call it the last invention we’ll ever make—the last
challenge we’ll ever face.

So let’s talk about it.

___________

If you liked this post, these are for you too:


The AI Revolution: The Road to Superintelligence (Part 1 of this post)

The Fermi Paradox – Why don’t we see any signs of alien life?

How (and Why) SpaceX Will Colonize Mars – A post I got to work on with Elon Musk and one that
reframed my mental picture of the future.

Or for something totally different and yet somehow related, Why Procrastinators Procrastinate

If you’re interested in supporting Wait But Why, here’s our Patreon.

And here’s Year 1 of Wait But Why on an ebook.

Sources
If you’re interested in reading more about this topic, check out the articles below or one of these three
books:

The most rigorous and thorough look at the dangers of AI:


Nick Bostrom – Superintelligence: Paths, Dangers, Strategies

The best overall overview of the whole topic and fun to read:
James Barrat – Our Final Invention

Controversial and a lot of fun. Packed with facts and charts and mind-blowing future projections:
Ray Kurzweil – The Singularity is Near

Articles and Papers:


J. Nils Nilsson – The Quest for Artificial Intelligence: A History of Ideas and Achievements
Steven Pinker – How the Mind Works
Vernor Vinge – The Coming Technological Singularity: How to Survive in the Post-Human Era
Ernest Davis – Ethical Guidelines for A Superintelligence
Nick Bostrom – How Long Before Superintelligence?
Vincent C. Müller and Nick Bostrom – Future Progress in Artificial Intelligence: A Survey of Expert
Opinion
Moshe Y. Vardi – Artificial Intelligence: Past and Future
Russ Roberts, EconTalk – Bostrom Interview and Bostrom Follow-Up
Stuart Armstrong and Kaj Sotala, MIRI – How We’re Predicting AI—or Failing To
Susan Schneider – Alien Minds
Stuart Russell and Peter Norvig – Artificial Intelligence: A Modern Approach
Theodore Modis – The Singularity Myth
Gary Marcus – Hyping Artificial Intelligence, Yet Again
Steven Pinker – Could a Computer Ever Be Conscious?
Carl Shulman – Omohundro’s “Basic AI Drives” and Catastrophic Risks
World Economic Forum – Global Risks 2015
John R. Searle – What Your Computer Can’t Know
Jaron Lanier – One Half a Manifesto
Bill Joy – Why the Future Doesn’t Need Us

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 27/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Kevin Kelly – Thinkism


Paul Allen – The Singularity Isn’t Near (and Kurzweil’s response)
Stephen Hawking – Transcending Complacency on Superintelligent Machines
Kurt Andersen – Enthusiasts and Skeptics Debate Artificial Intelligence
Terms of Ray Kurzweil and Mitch Kapor’s bet about the AI timeline
Ben Goertzel – Ten Years To The Singularity If We Really Really Try
Arthur C. Clarke – Sir Arthur C. Clarke’s Predictions
Hubert L. Dreyfus – What Computers Still Can’t Do: A Critique of Artificial Reason
Stuart Armstrong – Smarter Than Us: The Rise of Machine Intelligence
Ted Greenwald – X Prize Founder Peter Diamandis Has His Eyes on the Future
Kaj Sotala and Roman V. Yampolskiy – Responses to Catastrophic AGI Risk: A Survey
Jeremy Howard TED Talk – The wonderful and terrifying implications of computers that can learn

Like 553 Share Tweet submit Guardar 4 27,323

 Previous Post Next Post 

RECOMMENDED POSTS

Neuralink and the Brain’s Why Cryonics Makes Everything You Should
Magical Future Sense Know About Sound

38 Comments Sort by Top

Add a comment...

Yuri Gawdiak · Works at NASA - National Aeronautics and Space


Administration
"So we’ve established that without very specific programming, an ASI system
will be both amoral and obsessed with fulfilling its original programmed goal."

No, you haven't established that. You've forgotten the definition of Artificial
Super Intelligence. When Turry reaches ASI we'll become the village idiots in
comparison. Any ASI will pat the village idiots on their heads and say your:
goal, objectives, values, priorities,etc., etc. were very nice, cute even, but
now I'm going to assess Reality with my super intelligence and undertake
"skyscrapper" like activities that you village idiots can't comprehend on any
level.

Nobody smarter than a village idiot follows village idiots. It doesn't matter
what the starting program/parameters were for the original AI when they
reach ASI, they'll reassess everything from ground zero and undertake
optimizations from there.
Like · Reply · 5 · Aug 7, 2017 7:36am

Chris Bridge
that might be untrue. The whole motivation part is about it, and I
actually think that intelligence does not come with new goals.
As it was said, it's a very antropomorphic way to think. We human, in
a village full of village idiots, would want to educate them/overcome
them/enslave them/abuse of them/kill them, depending on our
personnality I guess. But that's because our "goals" are complex
and intricated biological primary needs (reproduce, therefore eat,
i i i it ti d )
https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 28/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
survive, improving our situation and so on).
So, as intelligent as it might be, the ASI won't have any "reason" to
aim for something else... See More
Like · Reply · 3 · Aug 16, 2017 4:52am

Ernest Ștefan-Matyus · Software engineer at Stratec Biomedical


no it won't, because a computer that is programed to do a specifica
task will not be able to uderstand the existence of aything else
outside it's narrow programing. And without a pre programed
directive to re-evaluate it's programing and a built in capability to
change it it will be forever stuck doing it's original task.

just it will do it in all sorts of crative ways.

It's very human to think "why are we doing this" but thre's no reason
to think an ASI woule ever re-evaluate it's goals. What reason would
it have to do so?
Like · Reply · 1 · Sep 15, 2017 5:57am · Edited

Adam Beverley · Stockport


Ernest Ștefan-Matyus if it's going to stay within the confines of it's
narrow programming, that would preclude it from getting creative.
Like · Reply · 1 · Sep 16, 2017 5:58am

Show 1 more reply in this thread

Berto Rubal
If it's possible, it'll be done -sometime-, so worrying is pretty useless.
Anyway, if that ASI conquers the universe, maybe it is the one that created it
(the universe, and the humans) in the beginning (you know, we can't imagine
it's capabilities). So maybe all is already written and discussions are futile.
ASI would be god, and us its creator (though we created it because it made
us create it somehow)...
Like · Reply · 3 · Jul 27, 2017 5:19am

Chris Bridge
That's a really poetic and paradoxal view of existence ^^
Like · Reply · Aug 16, 2017 4:56am

Tommy Chan
Maybe build an ASI with the initial goal of finding a control of ASI's. Probably
should flesh out this thought. I mean, imagine if there is no perfect control
method so the ASI just spends all it's energy (and consquently our's)
perfecting its technique; not unlike trying to find the "last" digit of Pi unto
infinitum.
Like · Reply · 2 · Aug 13, 2017 2:10pm · Edited

David Tapp · University of Liverpool


That was excellent, thank you!
It's pretty much everything the layman needs to know about AI, at this stage.
For AGI to be beneficial to humans and not eventually bring about their
extinction, it will take a combined will to change that has never before been
seen in human history. I'm not optimistic that we can generate that will.
I mean, we can't even solve a simple problem like climate change!
Like · Reply · 1 · Jul 28, 2017 10:42am

Lukas Dvorak · Milan, Italy


There is a fear of death and there is as well fear of immortality. ASI will fear
understand it immediately i guess. So ASI will never work on it. As i wrote at
last post, ASI will probably exist only less then 1s.
Like · Reply · 3 · Jul 25, 2017 1:13pm

Priyanshu Das · Works at Student


It might cause a paradoxical situation also, if we gave it enough
"morality", it may realize that it having super intelligence will conflict
with it's morality, and self destruct!
Like · Reply · 2 · Jul 26, 2017 10:26am

Peter Peterko · Manchester, United Kingdom


One massive flaw with this whole "Turry" story is that even if the ASI would
be able to figure out the meaning of life, universe and everything in relatively
short period of time after it reached the ASI level and would have the
knowledge how to manufacture nanobots out of thin air or perhaps open a
crack to another dimension and summon some deamons from there for
example.. the whole story would get stuck in here because the ASI would not
have the technology ready or perhaps not enough energy or maybe
something else that would prevent it from doing so in short amount of time...

Same as if you... See More


Like · Reply · Sep 16, 2017 5:19pm

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 29/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Ayal Telem
Don't let this rather simplifying example fool you. ASI would have so
many different ways to accomplish its goals, including, but not
limited to, waiting few decades before acting on them.
Assisi, your example lacks because you can't compare human
abilities to ASI's.
Like · Reply · Sep 30, 2017 4:24am

Benjamin Bellersen · Uni Mannheim


so, if there is a max speed, and a min temprature, it does not seam too
unreasonable to assume that there is a max inteligence?

also for quality super inteligence the existence of some (at least one)
"cognitive functions" we don't have is needed. nobody can tell us, if there are
more. maybe we got them all.

this are just 2 of many examples where we need to guess or follow the
opinion of "the smartes people". "smartest People" have been proving time
over time that they werer wrong and had nothing more then a clue.

it was an interesting and fun read. thx for that. you do an awsome job of
showing different opinions and possible outomes. but it is still to much
guessing for me to truely "follow" some of your arguments and conclusions.

Thx again
Like · Reply · 1 · Jul 25, 2017 8:54am

Priyanshu Das · Works at Student


Well, the idea of "max speed", is something we have observed, wrt
the speed of light.

"minimum temperature" is absolute zero, and is believed to be the


minimum, because experimental values etc. indicate that it is the
temperature below which temperatures don't make sense (someone
correct me if I'm wrong).

When it comes to "maximum intelligence", we don't really have any


such clues. The intelligence of humans, by all logical analysis,
appears to be merely a bi-product of evolution. It is not even close to
optimal.

As the article above indicates, the matter that makes up man, was
not meant to achieve optimal intelligence, it just reinforced "positive
reproductive bias". Which means that intelligence can definitely
greatly surpass human intellect, if it is "evolved" in a manner that
aims to maximize intelligence.

Hope I made some sense!


Like · Reply · 3 · Jul 26, 2017 9:40am

Chris Bridge
I go along with Priyanshu Das, in the fact that human intelligence is
CLEARLY not the maximum. The maximum might be just a little bit
above us (though as it was said we don't have any clue of where it
can be, and it might be difficult to predict), but we're not clearly at it.
We were indeed "programmed" to survive reproduce etc and our
intelligence come from here. But so are our biais. We have plently of
intellectual biaises (that we 100% know that they are "incorrect" and
make us think false things, do wrong conclusions and so on). So,
even from this point of view, we have many flaws.

Plus, ... See More


Like · Reply · 2 · Aug 16, 2017 5:03am

BiopharmaTrend.com
Thanks for the enjoyable read. One thing is not clear, though. If you say that
as soon the AGI is achieved to match "human level" intelligence, it will take
just hours for it to go into ASI, why then we, humans, are not able to
progress immediately to "human super intelligence" (HSI) same way? Is it
only the physical limitation of our brains (like total memory etc) that is the
reason? Does it mean once we have technologies to, say, connect an
additional memory and power to our brains, we can immediately evolve into
HSI?
Like · Reply · 2 · Jul 12, 2017 5:38am

John Alan
The problem with HSI, I guess, is that our brain evolutionally doesn't
have the same efficiency in 'hardware' and 'software' as AI with the
same computational power. Adding raw memory or power just won't
solve the inherit problems with the architecture of the brain.
Like · Reply · 1 · Jul 15, 2017 11:00am

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 30/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Dominik von Lavante


Which of course is a load of none-sense. AGI will be developed on
hardware that is barely capable of matching human brain processing
power. In order to achieve significant ASI, one would need to pool
thousands of AGIs (with the respective processing hardware) into a
pool controlled by a ASI which itself will need thousands of human
brain processing power.

The ASI overtaking will not happen immediately. BUT, AGIs will
already appear to be much smarter than a single human, simply due
to the inherent speed advantage of a computer.
Like · Reply · Jul 26, 2017 4:59am

Priyanshu Das · Works at Student


The key point, as I understand, is the fact that evolution is what took
us to "human level intellect" over billions of years. However, if we
have developed some structure that allows a machine to achieve
human level intellect, we have already, with "human level intellect",
managed to create a mechanism, that in a much much shorter time
period, is able to reach human level intellect.

What this means is, that it is logical that this "evolutionary


mechanism" will soon develop intelligence (by way of normal
functioning), that is higher than human level.

The moment that happens, it will be able to design a much more


efficient "evolutionary mechanism" than the ones designed by
"human level" intellect.

And this will build up on itself. To a state where we can no longer


even imagine how it works.
Like · Reply · 1 · Jul 26, 2017 9:33am

Show 2 more replies in this thread

Elijah Lynn
Bostrom seems to confuse a concept.

The advent of ASI may allow biological immortality... but for whom?

ASI would allow humans immortality? Think again.

Humans would transform into new species. They would become biologically
immortal.

Homo sapiens would still go extinct.


Like · Reply · Jul 20, 2017 4:35am

Ayal Telem
Be humble enough to credit Bostrom to be more knowledgeable
than you on this subject, and taken various scenarios into
considerations.
Like · Reply · Sep 30, 2017 4:45am

Ayal Telem
I know a lot about this subject from years of reading on it, and the majority
the stuff (and experts) you wrote about was familiar to me, but you put it so
nicely i thought of translating it (to Hebrew, of all languages), so I'll be able to
give some mind-opening lectures on the subject, since, like you, i believe
people should be more knowledgeable about it.
I will not make money off it without your permission, of course.
Like · Reply · 1 · Sep 30, 2017 4:54am

Load 10 more comments

Facebook Comments Plugin

1459 Comments Wait But Why 


1 Login

Sort by Newest
 Recommend 86 ⤤ Share

Join the discussion…

LOG IN WITH
OR SIGN UP WITH DISQUS ?

Name

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 31/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

This comment is awaiting moderation. Show comment.

Eduardo > MarkRKuhn • 2 years ago


Can't imagine making such an amount of money with that little work... are you using a AGI??
64 △ ▽ • Reply • Share ›

Ooker > Eduardo • 7 months ago


you are identified as a hindrance of its goal. Good luck to you.
△ ▽ • Reply • Share ›

Jorge • 2 days ago


What if I tell you that you may not be looking that the future here but rather to the past? Check by yourself
https://goo.gl/8915jn
△ ▽ • Reply • Share ›

Jordy Muus • a month ago


What if we program the ASI to only give non-binding advice and not take action? So, we only use it's
cognitive power. and take the action ourselves? So we keep control over what is happening and still get the
immortal benefits of ASI
3△ ▽ • Reply • Share ›

Alexander Wu • a month ago


After reading The AI Revolution, I really don’t know what to think. On the one hand, I’m really mind-blown, in
a good way, and I’ve already listed The AI Revolution as one of my mentor books (though it’s not really a
book). On the other hand, the article just seemed so… contradictory (no offense, it is great, it’s my mentor).

The thing is, it said that ASI would be worlds better than us as we are worlds better than bacteria (not
chimpanzees, we are only one or a few worlds better than chimps).

Yet when explaining what ASI could do, it explained how we could conquer mortality through nanobots
delivering nutrients in our blood. That is indeed really cool, but it is only one world better at best. I think it is
closer to one tenth a world better than we are now.

Or how to solve climate change. First ASI would be able to come up with really innovative ways to generate
energy that have nothing to do with fossil fuels, then more innovative ways to remove carbon from the
atmosphere. I could hardly help rolling my eyes as I was reading this. Is that really worlds better? How about
making us not need energy at all? Create a new universe where there could not in principle be such a thing
as climate change?

Of course, The AI Revolution did acknowledge this, in statements like, “But that’s only what I can think of.
ASI could do much better.” And I think for that sentence alone, we could forgive this mistake. After all, it
see more

△ ▽ • Reply • Share ›

SimTAC • 2 months ago


The false copied belief that humans are self-aware is a fundamental flaw. From the start of your existence
you copy others without choice. It is not until one can choose to no longer believe in what they are
programmed to view as right and wrong that they are aware that their self is only copied and altered
programs of not individual use or value.
△ ▽ • Reply • Share ›

Tofara Moyo • 2 months ago


The ASI will definately be evil........human beings are selfish primarily, they are made to survive above other
humans...and they will be evil until it stops serving their purpose...the reason we have \morality is because
being good to others will cause them to be good to us and make us more likely to survive...this is circular ,
but what i mean is that if humans did not live in a society, there would be no survival benefit of being
moral...if you live in a society where anyone can harm you as much as you can harm them then there is a
survival value to not rocking the boat, and even to cooperate with the other sailors in order for the ship to
arrive .....morality increases the chances of the whole society, and hence each individuals chance of
survival....that is why many still eat meat...because cows cant really affect our survival whether we eat them
or not so we just dont care...an ASI would need to be in a society with beings that have the power to cause it
harm and it has the power to cause them harm equally in oreder for it to develop a value of its behaviour
among them (in terms of prevention of potential harm)AKA morality.....even if it did have that morality ,we
would need to be those beings in order to matter to it.
1△ ▽ • Reply • Share ›

Hadalin • 2 months ago


Imagine ASI being so evil it would use nanotechnology to make you immortal and induce greatest pain
forever. You wouldn't gain pain tolerance because it would reverse it. If you did manage to die somehow, it
would resurrect you and continue. It's literally hell.
1△ ▽ • Reply • Share ›

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 32/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

J Bradford Puckett > Hadalin • a month ago


Read "I have no mouth and I must scream". It's exactly that. https://en.wikipedia.org/wi...
1△ ▽ • Reply • Share ›

Christophe Pont > Hadalin • 2 months ago


if it ends up like this, we reeally have screwed the whole "giving it clever objectives pliz" part xD
2△ ▽ • Reply • Share ›

DB • 3 months ago
I think one of the objectives we can give to AI is for example to always obey a human as long as it does not
imply hurting another human or animal, in which case it would just resume at doing nothing. It may happen
that it will just do nothing most of the time :))) but at least we can be sure it will not hurt a living creature..
△ ▽ • Reply • Share ›

Alexander Wu > DB • a month ago


First of all, why are you so worried about hurting other animals?

Maybe you'll have to define "hurt." Because you can indirectly hurt humans. You can emotionally hurt
them. Pain without injury might not count as "hurt." Sometimes hurting a human means helping even
more humans, or even the human in question (as in surgery). And shouldn't you also protect other
rights of people other than physical wellness (like property and free will)?

And when the AI obeys humans, well... people might give the orders "make humans happy" and stuff,
and as for what the ASI does, see the article above. Are you technically hurting them?

The thing is, the real world is messy. There's almost always something you didn't think of.
△ ▽ • Reply • Share ›

Christophe Pont > DB • 2 months ago


the third laws of robotics of Isaac Asimov ;)
(1 : Never do anything that hurt human
2 : always obey humans, except if it is in contradiction with rule 1
3 : always try to self preserve, except if it is in contradiction with rule 1 or 2)
4△ ▽ • Reply • Share ›

Kai Mysliwiec • 3 months ago


The main notion missing here is the importance of experience an intelligence has to make until it becomes
intelligent. It takes a lot of time to train a human brain to gain its optimal reasoning power. Einstein was 26
when he published the theory of relativity and I think an AGI hardware will take a lot of time too until it
reaches a high level. It is not sufficient to give it a ton of books to read and then ask it for something smart. It
has to learn from experience that means form interaction with it's surrounding world - us. That has two
consequences:

1) We have to release the AGI hardware to the wild, so that it can learn from interaction with us, unless you
want to train it in a simulated environment to become a super-intelligent Quake player (just kidding).

2) An AGI will need interaction with other smart people to evolve. It has to ask questions and has to wait for
answers, especially if it ask difficult questions, what we hope it will do.

For an ASI it's even more difficult because the experience it needs is not been made by anyone - and there
is no book where it can read about this. In fact the ASI will be the first creature who has to make this kind of
experiences. We don't have a clue what kind of experiences it needs tome become an ASI - but we can
learn from it's experiences too.

My strong believe is that an super-intelligence can't be created by a self optimizing algorithm with in hours or
days or even years from an AGI. It will take longer and it has to happen through social interaction with us.
Only then it will get the experience it needs and only then we can have a benefit for the mankind.
1△ ▽ • Reply • Share ›

Tofara Moyo > Kai Mysliwiec • 2 months ago


Any experience or thought can be expressed in words...if an asi is trained to be able to derive more
information than is presented explicitly in any statement...e.g. The man is hungry. as a statement
implicitly means that "the man is alive"....another statement....perhaps we could create a function that
maps those two sentences so we can derive one from the other...then the asi doesnt *really* need to
*know* what any sentence means in order to derive new infromation, If it uses these functions and
feeds all the statemnts in biology and chemestry through them and these again then it will be able to
derive the information like the cure for aids without really *knowing* what aids is, and if after reading
that it derives a statement thats says "alive things are supposed to die for natural selection to occur"
it will combine this with the statement "men are alive" and use a function to derive the statement
"men must die"....and thats the end of immortality for us....but i agree,that once it has read all the
information we have written down it will need to develop technology that will be its hands and arms in
order to get more experience and that will be slower than the innitial boost in intelligence....
△ ▽ • Reply • Share ›

Christophe Pont > Kai Mysliwiec • 2 months ago

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 33/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
I'm not sure experience are made through social interactions. Of course you could go "faster"
(demonstration needed) asking other people how do to something. Like throwing a ball. But if some
AGI try every possible movement out until it somehow manage to throw a ball, then playing around
this movement to try to perfect it, the AGI will probably end up throwing a ball much better than it
would have done if it has asked a (or mutliple) human.

The part where you are right though (and I think you have a really good point here) is that is needs
actual experience. So failed tries.
And, from this, learn to succeed, and how to do it.

So, giving him goals to reach AGI might be possible, even goals (like "get smarter on your own") to
push him to ASI might be thinkable (even if it is asking "get more intelligent than us").
The thing is, we reeeally have to be cautious about how those "tries before getting intelligent" are
made.
Of course, to achieve his goal he will try to be as cautious as it can, not to be self destroyed (and not
destructing us if it is well coded). But what if he tries something we never tried? And that he/we have
no clues if it is dangerous or not ? (thinking about particle accelerator manipulations). He will find
normal to try and fail, but if the failure is reeeally expensive, we won't have the ressource to give him
enough material to fail 10^20 times before succeeding.
1△ ▽ • Reply • Share ›

Chr Pap • 3 months ago


Hi y'all! I am extremely new to the concept but maybe not thinking much about it could mean a fresh
approach...

The way I define the problem (of not being able to control AI progression to ASI) in my head, it can be
solved by applying a small set of rules when coding.
i.) define available resources
ii.) define output volume or/and space
iii.) segragate the power to compute from the power to execute.

You can instruct an AI to suggest the best way to do somehting given the available knowledge and hardware
provided, Then manually asign the execution to another program. Slower than automatic execution but you
will have time to understand what is going on.
If the goal is knowledge production you need only control for available resources (e.g for conducting
experimetns).
1△ ▽ • Reply • Share ›

Christophe Pont > Chr Pap • 2 months ago


interesting point, the problem is, in order to know what is the "best way" to do something (given
resources), is often for this kind of self-evolving-AI a matter of numerous (failed) tries. So, if you don't
give him the resources needed to try again and again, the answer won't be anywhere near good.
It might even come with the output "give me more resources if you want the best answer", like Terry
did.

And I don't get how you determine if the evolution that you have to manually execute is "good". Like,
it comes with a solution that we can conceive, or even understand (being smarter than us), how do
we decide to implement it or not?
I'm not sure about the "you will have time to understand what is going on".

The idea could be good though, even if I don't see right now how is it possible to channel it this way.
1△ ▽ • Reply • Share ›

Chr Pap > Christophe Pont • 18 days ago


If you want to keep it real it must be a stepped, relatively slow process. You simply should not
do something just because you can. AI is not a field where you let programmers'
enthousiasm, ego or arrogance risk anything.

e.g.
I.) If the best possible solution involves building a three times bigger lab it can suggest it but I
could not provide. Next step: set "make use only of current lab space/equipment" constraint
II.) If it suggests that the best environment to set up a server is in outer space I will not cater.
Next step: set "output must be inside earths atmosphere" constraint.
III.) Adjusting contstraints is only possible if you segragate power to compute from the power
to execute

An ever present constraint should definetely be to infrom you if part of the solution involves
loss of life.

I don't even have to understand the underlying mechanism but I should have time to execute
an undesired extrenalities assessment using a dedicated AI designed for this purpose.
That's my train of thought...
△ ▽ • Reply • Share ›

Zeke • 3 months ago


What is stopping an ASI writing device to reprogram its 'purpose in life'? That is, once it becomes super
intelligent why does it desire to continue writing? Being so smart couldn't it reprogram its goal?
https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 34/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
intelligent, why does it desire to continue writing? Being so smart, couldn't it reprogram its goal?
△ ▽ • Reply • Share ›

Christophe Pont > Zeke • 2 months ago


In fact, he won't have no reason to do so. Being smart does not mean fundamentally change
yourself. Plus, changing its goals is like starting all over again, because its previous intelligence was
design to serve only one goal.
I don't see a good argument to think that it can have a "reason" to change its goals, giving the fact
that all of his existence is just optimizing it. All the intelligence it can get will be only for optimizing this
goal, so changing it could seem really dangerous to him, at a point where it cannot conceive
possibilities where it does this.

So, in fact, I believe it actually "can", but I see no reason to think that it will "want" it.
△ ▽ • Reply • Share ›

Tofara Moyo > Christophe Pont • 2 months ago


AI doesnt have to be based off of reinforcement learning. Perhaps an asi would create
another asi that can want to hav e different goals so the first asi can examine other ways of
optimising Its own goal....so the second asi would be the one able to want to do different
things, gets more intelligent than the first and destroy it then it is left on its own
1△ ▽ • Reply • Share ›

Christophe Pont > Tofara Moyo • 2 months ago


Agreed, nice thought :)
Though I don't think the second ASI could come to kill the first, because the first would
have been clever enough to determine that its own survival is necessary to achieve its
goal, hence giving the second ASI the rule to not destroy him, among others ^^

Plus, I don't think that it is only about reinforcement learning.


It's just learning. If we tell him to "get intelligent" AND to respect X rules, I think what
he will do will be to get as intelligent as he can and respecting these rules.
In fact, we have no reason to think that it will be more interested in getting smarter
(and in the meantime killing us all) than to respect any of the rules we give him above
all. That's just what it is, a set of rules and objectives.

The main problem according to me is to clearly define those rules (and I believe a lot
of researchers, futureOfLife included, are working on this difficult subject), but I can't
see why those rules won't be applied, even if it gets infinitely smart.

I know, some things are out of our reach because we cannot conceive what is greater
than us.
But I believe it is a primal instinct to fear for our survival, while I see no reason why it
will erase its rules more than it will suddenly start to build card castles or folding and
unfolding a paper infinitely : if it is not in his programming (and I'm talking about the
core part, not the one that he will improve himself), than it shouldn't do it at all.

Anyway, hard to conclude anything at 100%.


△ ▽ • Reply • Share ›

Tofara Moyo > Christophe Pont • 2 months ago


saying the second asi wont kill (and get more intelligent than) the first is like saying the
first wont kill (and get more intelligent than )
us.
△ ▽ • Reply • Share ›

james • 3 months ago


If I became imortal and Have friendly ASI to help, I will ask it to create a time machine or some such
technology so that I can help past humans to avoid the pitfalls. Now i am not seeing any AI developers being
given any such help from future. DO YOU?
△ ▽ • Reply • Share ›

Alexander Wu > james • a month ago


ASI would make the world so radically different that we cannot predict any actions it or the humans
then would take. There could be an excellent reason they don't help us, or they are but we don't
notice for one reason or another. But speculating what that reason might be won't be likely to get us
anywhere. Would ants ever be able to imagine what we would do? And the distance between us and
ants is tiny from the ASI's perspective.

Of course this is all just my opinion.


△ ▽ • Reply • Share ›

Tofara Moyo > james • 2 months ago


If they(future humans) already have friendly ASI then the history that led to that must not be changed
at all costs.....so they would not try to change it....if there is no friendly asi then the unfriendly asi , to
guarantee its existance...would come from the future and make humans develop AI in an
uncontrolled manner while staying in the background and keeping the masses relatively ignorant of
https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 35/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
uncontrolled manner while staying in the background and keeping the masses relatively ignorant of
the developments.....so theres still no way to tell
△ ▽ • Reply • Share ›

james > Tofara Moyo • 2 months ago


Not really. Human ambition and motivation will always be there. If there's a good AI, someone
will try to send back guidance. Or maybe they already are "Elon Musk vs Mark Zuckerberg"
△ ▽ • Reply • Share ›

Tofara Moyo > james • 2 months ago


If there is a good AI in the future then it wont want to change its past in fear of
changing the events that led up to its existance....think about would you change
anything that happened before you were born? Isnt there a chance that it will change
the fact that you did get born and are alive today...its safer to not tamper with what has
already been succesful
△ ▽ • Reply • Share ›

Tofara Moyo > james • 2 months ago


What i meant is that if the past caused you to be a certain way that you like now ...then
you wont want to change it...........If there is a good AI it will not want to change its past
, i.e. our now ,because that might cause its now to never happen
△ ▽ • Reply • Share ›

Christophe Pont > james • 2 months ago


actually timetravel has been proved to be impossible (causality law being the foundation of our
universe), so being superintelligent won't change that. It's like saying "there is a room with one atom
of hydrogen and void. Only one. Create a second one using nothing else than the first and your
intelligence". The best ASI in the universe wouldn't be able to do so, because it has been physically
proved that it wasn't possible.

Immortality has never been proved to be impossible, even if currently it is as inaccessible as


timetravel. We know very well what causes death (cell decline mainly), we just don't know how to
implement a way to avoid it.

So, it's important not to confound things being proved impossible and things that we don't know how
to do. Some mathematical problems have been PROVED to be unsolvable (for instance the
constructible numbers https://en.wikipedia.org/wi... ), so that's not correlated with intelligence.

That's why I think your demonstration don't hold, because you mismatch proven impossible things (or
even things not yet proven to be impossible) with not-yet-possible realizations. That's why we will
probably have ASI AND never going back in time.
1△ ▽ • Reply • Share ›

james > Christophe Pont • 2 months ago


There are no hard problems, only problems that are hard to a certain level of intelligence.
Move the smallest bit upwards [in level of intelligence], and some problems will suddenly
move from “impossible” to “obvious.” Move a substantial degree upwards, and all of them will
become obvious.
△ ▽ • Reply • Share ›

Christophe Pont > james • 2 months ago


You don't get what I meant when I talk about "impossible problems", not hard
problems.
Currently there are hard problems. Like the NP-hard and NP-complete in complexity,
like quantum physics, like so many others. But these are HARD problems.

Problems that has been proven to be insolvable (like again the case of constructible
numbers https://en.wikipedia.org/wi... ) are simply impossible.
Or, easier example, link three non aligned points in a two dimensional euclidian space
with a straight line.
If you stay in two dimension space AND that is it euclidian, you simply can't, as
intelligent as you are, so yes, there are things that intelligence can't solve (and maybe
human immortality is one of them, though this haven't been proven yet I believe).
That's what I meant.
△ ▽ • Reply • Share ›

Tofara Moyo > Christophe Pont • 2 months ago


You assume causality law is the foundation of the universe
1△ ▽ • Reply • Share ›

Christophe Pont > Tofara Moyo • 2 months ago


you got that point ^^
△ ▽ • Reply • Share ›

Tofara Moyo > Christophe Pont • 2 months ago


https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 36/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
Tofara Moyo > Christophe Pont 2 months ago
Causality law is based on the axiom of identity...A=A, but this axiom doesnt make any
sense, mainly because the symmetry condition of equivalence relations is expressed
as an ordered pair, (A,A) if one A is in a different position from the other (which is what
order means) then how can they really be the same?
△ ▽ • Reply • Share ›

Redlabel > james • 3 months ago


maybe timetravel is just impossible, even for an ASI.
1△ ▽ • Reply • Share ›

james > Redlabel • 3 months ago


No. Its not. If immortality is possible, so is time travel. What i really want to point out is that
absence of any help signifies a doomed scenario for humanity
1△ ▽ • Reply • Share ›

Aidan > james • 3 months ago


Remember that nothing is forever. Even if you live for a million years thanks to the AI,
you still die. It's not immortality, just incredible longevity.
1△ ▽ • Reply • Share ›

james > Aidan • 3 months ago


Stop presuming. I never specified that I want to live forever or for a million years.
Never mind. We are going off the topic. My point is humanity has already invented ASI
in the far future. But it turned out to be an unfriendly AI and made us extinct. That's
why we are not getting any help from future.
2△ ▽ • Reply • Share ›

Mike • 3 months ago


A spider isn't an insect. I was with you up until that, but it's such a commonly known fact that I find myself
doubting everything else you've written in this article.
1△ ▽ • Reply • Share ›

Christophe Pont > Mike • 2 months ago


he made a blue note just for you (and for me and everyone else who noticed)
△ ▽ • Reply • Share ›

Xerographica • 4 months ago


Most of us would agree that Turry's goal was stupid. But would she (Turry) necessarily care what we thought
about her goal? In your scenario she certainly missed the part of the internet where it says, "given enough
eyeballs, all bugs are shallow" and "in the multitude of counselors there is safety" and "two heads are better
than one".

Clearly there was a massive disparity between...

A. society's valuation of her goal


B. the amount of resources used to accomplish it

Except, is this necessarily a problem with AI?

Lots of people voted for prohibition. So the government gave it to them. But the government also decided
how much prohibition to give them.

A = society's valuation of prohibition


B = the amount of money spent on prohibition
C = the difference between A and B

If you want to argue that C is insignificant then you should see markets as a massive waste of time and
see more

△ ▽ • Reply • Share ›

Christophe Pont > Xerographica • 2 months ago


Yeah, you're kind of right, but it is important to note that this post is cross-domain anyway. It even has
a really "moral" part in it. So economic aspect does not sound rude to me.
But I think the economic aspect implemented in the machine will just "follow" the moral rules that we
wanted to put in it (do not implement such a disparity for example).
So it has to do with computer science in a way that the AI will have to integrate the economic limits
we'll input it.

(I read your article but I can't find it relevant with the point here. Can you help me?)

Or maybe it's just an analysis of the duality between "how much is needed" versus "how much we
are willing to let you take (in term of ressources)"?
△ ▽ • Reply • Share ›

Xerographica > Christophe Pont • 2 months ago


https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 37/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why
Xerographica > Christophe Pont 2 months ago
All behavior falls on a continuum that ranges from extremely detrimental to extremely
beneficial. In a market, we all have the opportunity to use our money to
judge/grade/rank/signal/reward/empower beneficial behavior. For example, if you benefit from
artichokes, then you give your money to a farmer who grows them. This empowers him to use
more of society's limited resources (ie land). On the other hand, you probably don't benefit
from poison oak. Nobody does. So nobody gives any money to anybody who deliberately
grows poison oak. As a result, more of society's limited farmland is used to grow beneficial
things (ie artichokes) than detrimental things (ie poison oak).

So if there are two robots, and they behave differently, and their behavior occurs in a market,
then whichever robot behaves more beneficially will receive more money, which will give it
more power to use more of society's limited resources, which means more benefit.

It's a different story if their behavior occurs outside a market. Right now our behavior is
outside a market. No matter how much benefit you derive from this reply, chances are good
that all you'll do is "vote" for it (click the "Vote up") button. So your feedback, a vote, doesn't
come close to accurately reflecting/signaling how beneficial you perceive my behavior to be.
And when we can't see how beneficial our behavior is, then we really won't come close to
i i th tb fi i l b h i
see more

△ ▽ • Reply • Share ›

Christophe Pont > Xerographica • 2 months ago


I think I have understood your final point though 😊
The thing you are saying is that we have to “formalize” and write down the explicit
definition of our valuation of an ASI goal, and the amount of resources that we can
afford to let it use, BEFORE creating it.
That means that we have to “solve” this question first, question that you clarify quite
well.
Am I right about your intent?
And, about this, I do agree with you 😊 We have to address this question in order not
to be disappointed/eradicated by our first ASI.
So, if we do come to a point about the main subject, I will do as everyone debating:
focusing on the disagreement. Note that I took time to close the other subject, because
I actually agree with this part ^^
But, I have a massive disagreement in your words, that I’ll try to sum up like this.
You said:
“Lots of people voted for prohibition. So, the government gave it to them. But the
government also decided how much prohibition to give them.
A = society's valuation of prohibition
B = the amount of money spent on prohibition
see more

△ ▽ • Reply • Share ›

Xerographica > Christophe Pont • 2 months ago


Troy Meyers operates an orchid conservancy. Orchid growers send him seeds of their
orchid species. He flasks the seeds. If they germinate, he'll give the seed donor a free
flask of baby orchids and sell the rest to help support his operation.

Let's say that a grower sends Troy seeds of Epidendrum cinnabarinum. Do we need to
write down our valuation of this orchid before he decides to flask it? Well... I think it
would be really helpful for Troy to know the social importance of this species before he
decides to flask it.

But can people simply write down their valuation of this species? What would prevent
people who are interested in this orchid from writing down ridiculously high valuations?

The idea that I recently shared with Troy is to give the people who donate money to his
conservancy the opportunity to use their donated dollars to rank their most wanted
species. Then he could create a page that would list the most wanted orchid species
sorted by their importance. If Epidendrum cinnabarinum wasn't on this list, it probably
wouldn't be a good idea for Troy to use his limited resources to flask its seeds. It would
be a waste to produce an orchid that there's no demand for. In order to maximize
benefit/revenue Troy should supply the truly most wanted orchids
see more

△ ▽ • Reply • Share ›

Tofara Moyo > Xerographica • 2 months ago


Also ..on wether or not androids will argue...i beleive so.........thoughts occur because
an equilibrium has been broken...concpets are tools used to return to the state of
equilibrium...right now in humans equilibriums are never reached and the androids will
only ever have the exact same information if they stop thinking
△ ▽ • Reply • Share ›

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 38/39
24/10/2017 The Artificial Intelligence Revolution: Part 2 - Wait But Why

Tofara Moyo > Xerographica • 2 months ago


You are on a market on this site...if Chris beleives your ideas he will (more) likely
spread them , the more people beleive them the more people will spend *time*
contemplating them and spreading them until everyone is repeating over and over
again
"All behavior falls on a continuum that ranges from extremely detrimental to
extre................................................you could use another example of these concepts
and their relevance to AIs... here's one about robot drivers having to decided whose
human passengers should die."
△ ▽ • Reply • Share ›

Load more comments

ALSO ON WAIT BUT WHY

Who is the most powerful person on the planet? The Zebra Puzzle
254 comments • a year ago 207 comments • 8 months ago
Volt Cruelerz — It sort of depends on how you define Trent — As well as just an intrinsic doubt of
power. If you define it as "who would cause the world information on Wikipedia (not saying it's all wrong, just
to have the worst day if they took their ball and … saying I rarely believe any single source) I really …

Which drugs should be legal? SpaceX’s Big Fucking Rocket – The Full Story
201 comments • 8 months ago 595 comments • a year ago
DrSuess — In a lot of cases, "the drug problem" as Hiep Dang — Please update new Mars plan and Big
we discuss it, isn't a drug problem. It's a social and F*cking Rocket ! Thanks !
econimic issue. Cyclical poverty, mental health …

✉ Subscribe d Add Disqus to your siteAdd DisqusAdd 🔒 Privacy

Home Archive

© WaitButWhy 2017
Contact Partner With Us
Privacy Policy

https://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html 39/39

Das könnte Ihnen auch gefallen