Professional Documents
Culture Documents
Max Tegmark
Dept. of Physics & MIT Kavli Institute, Massachusetts Institute of Technology, Cambridge, MA 02139
(Dated: September 4, 2014)
Relentless progress in artificial intelligence (AI) is increasingly raising concerns that machines
will replace humans on the job market, and perhaps altogether. Eliezer Yudkowski and others
have explored the possibility that a promising future for humankind could be guaranteed by a
superintelligent Friendly AI [1], designed to safeguard humanity and its values. I will argue that,
from a physics perspective where everything is simply an arrangement of elementary particles, this
might be even harder than it appears. Indeed, it may require thinking rigorously about the meaning
of life: What is meaning in a particle arrangement? What is life? What is the ultimate ethical
imperative, i.e., how should we strive to rearrange the particles of our Universe and shape its future?
If we fail to answer the last question rigorously, this future is unlikely to contain humans.
arXiv:1409.0813v2 [cs.CY] 3 Sep 2014
AI research and evolutionary psychology shed further arranged in that way at an earlier time, that arrangement
light on how this subversion occurs. When optimizing will typically not last. And what particle arrangement is
a rational agent to attain a goal, limited hardware re- preferable, anyway?
sources may preclude implementing a perfect algorithm, It is important to remember that, according to evo-
so that the best choice involves what AI-researchers term lutionary psychology, the only reason that we humans
limited rationality: an approximate algorithm that have any preferences at all is because we are the solu-
works reasonably well in the restricted context where tion to an evolutionary optimization problem. Thus all
the agent expects to find itself [6]. Darwinian evolu- normative words in our human language, such as deli-
tion has implemented our human inclusive-reproductive- cious, fragrant, beautiful, comfortable, interest-
fitness optimization in precisely this way: rather than ask ing, sexy, good, meaningful and happy, trace
in every situation which action will maximize our number their origin to this evolutionary optimization: there is
of successful offspring, our brains instead implements a therefore no guarantee that a superintelligent AI would
hodgepodge of heuristic hacks (which we call emotional find them rigorously definable. For example, suppose we
preferences) that worked fairly well in most situations in attempt to define a goodness function which the AI
the habitat where we evolved and often fail badly in can try to maximize, in the spirit of the utility functions
other situations that they were not designed to handle, that pervade economics, Bayesian decision theory and
such as todays society. The sub-goal to procreate was AI design. This might pose a computational nightmare,
implemented as a desire for sex rather than as a (highly since it would need to associate a goodness value with ev-
efficient) desire to become a sperm/egg donor and, as ery one of more than a googolplex possible arrangement
mentioned, is subverted by contraceptives. The sub-goal of the elementary particles in our Universe. We would
of not starving to death is implemented in part as a de- also like it to associate higher values with particle ar-
sire to consume foods that taste sweet, triggering todays rangements that some representative human prefers. Yet
diabesity epidemic and subversions such as diet sodas. the vast majority of possible particle arrangements corre-
Why do we choose to trick our genes and subvert spond to strange cosmic scenarios with no stars, planets
their goal? Because we feel loyal only to our hodge- or people whatsoever, with which humans have no expe-
podge of emotional preferences, not to the genetic goal rience, so who is to say how good they are?
that motivated them which we now understand and There are of course some functions of the cosmic par-
find rather banal. We therefore choose to hack our ticle arrangement that can be rigorously defined, and we
reward mechanism by exploiting its loopholes. Analo- even know of physical systems that evolve to maximize
gously, the human-value-protecting goal we program into some of them. For example, a closed thermodynamic
our friendly AI becomes the machines genes. Once this system evolves to maximize (course-grained) entropy. In
friendly AI understands itself well enough, it may find the absence of gravity, this eventually leads to heat death
this goal as banal or misguided as we find compulsive re- where everything is boringly uniform and un-changing.
production, and it is not obvious that it will not find a So entropy is hardly something we would want our AI
way to subvert it by exploiting loopholes in our program- to call utility and strive to maximize. Here are other
ming. quantities that one could strive to maximize and which
appear likely to be rigorously definable in terms of par-
ticle arrangements:
III. THE FINAL GOAL CONUNDRUM
The fraction of all the matter in our Universe that is
in the form of a particular organism, say humans or
Many such challenges have been explored in the E-Coli (inspired by evolutionary inclusive-fitness-
friendly-AI literature (see [5] for a superb review), and so maximization)
far, no generally accepted solution has been found. From
my physics perspective, a key reason for this is that much What Alex Wissner-Gross & Cameron Freer term
of the literature (including Bostroms book [5]) uses the causal entropy [7] (a proxy for future opportuni-
concept of a final goal for the friendly AI, even though ties), which they argue is the hallmark of intelli-
such a notion is problematic. In AI research, intelligent gence.
agents typically have a clear-cut and well-defined final
goal, e.g., win the chess game or drive the car to the des- The ability of the AI to predict the future in the
tination legally. The same holds for most tasks that we spirit of Marcus Hutters AIXI paradigm [8].
assign to humans, because the time horizon and context The computational capacity of our Universe.
is known and limited. But now we are talking about the
entire future of life in our Universe, limited by nothing The amount of consciousness in our Universe,
but the (still not fully known) laws of physics. Quan- which Giulio Tononi has argued corresponds to in-
tum effects aside, a truly well-defined goal would specify tegrated information [9].
how all particles in our Universe should be arranged at
the end of time. But it is not clear that there exists a When one starts with this physics perspective, it is hard
well-defined end of time in physics. If the particles are to see how one rather than another interpretation of
3
utility or meaning would naturally stand out as spe- with a rigorously defined goal will be able to improve its
cial. One possible exception is that for most reasonable goal attainment by eliminating us.
definitions of meaning, our Universe has no meaning This means that to wisely decide what to do about
if it has no consciousness. Yet maximizing consciousness AI-development, we humans need to confront not only
also appears overly simplistic: is it really better to have traditional computational challenges, but also some of
10 billion people experiencing unbearable suffering than the most obdurate questions in philosophy. To program
to have 9 billion people feeling happy? a self-driving car, we need to solve the trolley problem of
In summary, we have yet to identify any final goal for whom to hit during an accident. To program a friendly
our Universe that appears both definable and desirable. AI, we need to capture the meaning of life. What is
The only currently programmable goals that are guar- meaning? What is life? What is the ultimate ethical
anteed to remain truly well-defined as the AI gets pro- imperative, i.e., how should we strive to shape the future
gressively more intelligent are goals expressed in terms of of our Universe? If we cede control to a superintelligence
physical quantities alone: particle arrangements, energy, before answering these questions rigorously, the answer
entropy, causal entropy, etc. However, we currently have it comes up with is unlikely to involve us.
no reason to believe that any such definable goals will be
desirable by guaranteeing the survival of humanity. Con-
trariwise, it appears that we humans are a historical acci-
dent, and arent the optimal solution to any well-defined Acknowledgments: The author wishes to thank
physics problem. This suggests that a superintelligent AI Meia Chita-Tegmark for helpful discussions.
[1] Yudkowski, Eliezer 2001, Creating Friendly AI 1.0: B and Stan Franklin, 483-492, Frontiers in Artificial Intel-
The Analysis and Design of Benevolent Goal Ar- ligence and Applications 171, Amsterdam:IOS
chitectures. Machine Intelligence Research Institute, [5] Bostrom, Nick 2014, Superintelligence: Paths, Dangers,
http://intelligence.org/files/CFAI.pdf Strategies, Oxford: OUP
[2] Good, Irving John 1965, Speculations Concerning the [6] Russell, Stuart & Norvig, Peter 2010: Artificial Intelli-
First Ultraintelligent Machine, in Advances in Comput- gence: A Modern Approach, New York: Prentice Hall
ers, edited by Franz L. Alr and Morris Rubinoff, 6:31-88, [7] Wissner-Gross, Alexander D. & Freer, C. E. 2013, Causal
New York: Academic Press Entropic Forces, Phys. Rev. Lett. 110, 168702
[3] Vinge, Vernor 1993, The Coming Technological Singular- [8] Hutter, Marcus 2000, A Theory of Universal Arti-
ity: How to Survive in the Post-Homan Era, in Vision-21: ficial Intelligence based on Algorithmic Complexity,
Interdisciplinary Science and Engineering in the Era of arXiv:cs/0004001
Cyberspace, 11-22, NASA Conference Publication 10129, [9] Tononi, Giulio 2008, Consciousness as Integrated Infor-
NASA Lewis Research Center mation: a Provisional Manifesto, Biol. Bull. 215, 216,
[4] Omohundro, Stephen M. 2008, The Basic AI Drives, http://www.biolbull.org/content/215/3/216.full
in Artificial General Intelligence 2008: proceedings of the
First AGI Conference, edited by Pei Fang, Ben Goerzel