2,041 473 663MB
English Pages 1152 [1134] Year 2020
Artificialfintelligence Bl
ol
ANVIOAEM APPreacH
Helligigh=eljtjelf)
Artificial Intelligence A Modern Approach Fourth Edition
PEARSON SERIES IN ARTIFICIAL INTELLIGENCE Stuart Russell and Peter Norvig, Editors
FORSYTH & PONCE
Computer Vision: A Modern Approach, 2nd ed.
JURAFSKY & MARTIN
Speech and Language Processing, 2nd ed.
RUSSELL & NORVIG
Artificial Intelligence: A Modern Approach, 4th ed.
GRAHAM
NEAPOLITAN
ANSI Common Lisp
Learning Bayesian Networks
Artificial Intelligence A Modern Approach Fourth Edition Stuart J. Russell and Peter Norvig
Contributing writers: Ming-Wei Chang Jacob Devlin Anca Dragan David Forsyth Tan Goodfellow Jitendra M. Malik Vikash Mansinghka Judea Pearl Michael Wooldridge
@
Pearson
Copyright © 2021, 2010, 2003 by Pearson Education, Inc. or its affiliates, 221 River Street, Hoboken, NJ 07030. All Rights Reserved. Manufactured in the United States of America. This publication is protected by copyright, and permission should be obtained from the publisher prior to any prohibited reproduction, storage in a retrieval system, or transmission in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise. For information regarding permissions, request forms, and the appropriate contacts within the Pearson Education Global Rights and Permissions department, please visit www.pearsoned.com/permissions/. Acknowledgments of third-party content appear on the appropriate page within the text. Cover Images: Alan Turing - Science History Images/Alamy Stock Photo Statue of Aristotle — Panos Karas/Shutterstock Ada Lovelace - Pictorial Press Ltd/Alamy Stock Photo Autonomous cars — Andrey Suslov/Shutterstock Atlas Robot ~ Boston Dynamics, Inc. Berkeley Campanile and Golden Gate Bridge — Ben Chu/Shutterstock Background ghosted nodes — Eugene Sergeev/Alamy Stock Photo Chess board with chess figure — Titania/Shutterstock Mars Rover - Stocktrek Images, Inc./Alamy Stock Photo Kasparov - KATHY WILLENS/AP Images PEARSON, ALWAYS LEARNING is an exclusive trademark owned by Pearson Education, Inc. or its affiliates in the U.S. and/or other countries.
Unless otherwise indicated herein, any third-party trademarks, logos, or icons that may appear in this work are the property of their respective owners, and any references to third-party trademarks, logos, icons, or other trade dress are for demonstrative or descriptive purposes only. Such references are not intended to imply any sponsorship, endorsement, authorization, or promotion of Pearson’s products
by the owners of such marks, or any relationship between the owner and Pearson Education, Inc., or its affiliates, authors, licensees, or distributors.
Library of Congress Cataloging-in-Publication Data Russell, Stuart J. (Stuart Jonathan), author. | Norvig, Peter, author. rtificial intelligence : @ modern approach/ Stuart J. Russell and Peter Norvig. Description: Fourth edition. | Hoboken : Pearson, [2021] | Series: Pearson series in artificial intelligence | Includes bibliographical references and index. | Summary: “Updated edition of popular textbook on Artificial Intelligence.”— Provided by publisher. Identifiers: LCCN 2019047498 | ISBN 9780134610993 (hardcover) Subjects: LCSH: Artificial intelligence. Classification: LCC Q335 .R86 2021 | DDC 006.3-dc23 LC record available at https://lcen.loc.gov/2019047498 ScoutAutomatedPrintCode
@
Pearson ISBN-10: ISBN-13:
0-13-461099-7 978-0-13-461099-3
For Loy, Gordon, Lucy, George, and Isaac — S.J.R. For Kris, Isabella, and Juliet — P.N.
This page intentionally left blank
Preface Artificial Intelligence (Al) is a big field, and this is a big book.
We have tried to explore
the full breadth of the field, which encompasses logic, probability, and continuous mathematics; perception, reasoning, learning, and action; fairness, trust, social good, and safety; and applications that range from microelectronic devices to robotic planetary explorers to online services with billions of users. The subtitle of this book is “A Modern Approach.”
That means we have chosen to tell
the story from a current perspective. We synthesize what is now known into a common
framework, recasting early work using the ideas and terminology that are prevalent today.
We apologize to those whose subfields are, as a result, less recognizable. New to this edition
This edition reflects the changes in Al since the last edition in 2010: « We focus more on machine learning rather than hand-crafted knowledge engineering, due to the increased availability of data, computing resources, and new algorithms.
« Deep learning, probabilistic programming, and multiagent systems receive expanded coverage, each with their own chapter.
« The coverage of natural language understanding, robotics, and computer vision has been revised to reflect the impact of deep learning. + The robotics chapter now includes robots that interact with humans and the application of reinforcement learning to robotics.
« Previously we defined the goal of Al as creating systems that try to maximize expected utility, where the specific utility information—the objective—is supplied by the human
designers of the system. Now we no longer assume that the objective is fixed and known by the Al system;
instead, the system may be uncertain about the true objectives of the
humans on whose behalf it operates. It must learn what to maximize and must function
appropriately even while uncertain about the objective. « We increase coverage of the impact of Al on society, including the vital issues of ethics, fairness, trust, and safety.
+ We have moved the exercises from the end of each chapter to an online site.
This
allows us to continuously add to, update, and improve the exercises, to meet the needs
of instructors and to reflect advances in the field and in Al-related software tools.
* Overall, about 25% of the material in the book is brand new. The remaining 75% has
been largely rewritten to present a more unified picture of the field. 22% of the citations in this edition are to works published after 2010.
Overview of the book The main unifying theme is the idea of an intelligent agent. We define Al as the study of agents that receive percepts from the environment and perform actions. Each such agent implements a function that maps percept sequences to actions, and we cover different ways
to represent these functions, such as reactive agents, real-time planners, decision-theoretic vii
viii
Preface
systems, and deep learning systems. We emphasize learning both as a construction method for competent systems and as a way of extending the reach of the designer into unknown
environments. We treat robotics and vision not as independently defined problems, but as occurring in the service of achieving goals. We stress the importance of the task environment in determining the appropriate agent design.
Our primary aim is to convey the ideas that have emerged over the past seventy years
of Al research and the past two millennia of related work.
We have tried to avoid exces-
sive formality in the presentation of these ideas, while retaining precision. We have included mathematical formulas and pseudocode algorithms to make the key ideas concrete; mathe-
matical concepts and notation are described in Appendix A and our pseudocode is described in Appendix B. This book is primarily intended for use in an undergraduate course or course sequence. The book has 28 chapters, each requiring about a week’s worth of lectures, so working through the whole book requires a two-semester sequence. A one-semester course can use. selected chapters to suit the interests of the instructor and students.
The book can also be
used in a graduate-level course (perhaps with the addition of some of the primary sources suggested in the bibliographical notes), or for self-study or as a reference.
Term
Throughout the book, important points are marked with a triangle icon in the margin.
‘Wherever a new term is defined, it is also noted in the margin. Subsequent significant uses
of the term are in bold, but not in the margin. We have included a comprehensive index and
an extensive bibliography. The only prerequisite is familiarity with basic concepts of computer science (algorithms, data structures, complexity) at a sophomore level. Freshman calculus and linear algebra are useful for some of the topics. Online resources
Online resources are available through pearsonhighered. com/cs-resources or at the book’s Web site, aima. cs . berkeley. edu. There you will find: « Exercises, programming projects, and research projects. These are no longer at the end of each chapter; they are online only. Within the book, we refer to an online exercise with a name like “Exercise 6.NARY.”
Instructions on the Web site allow you to find
exercises by name or by topi Implementations of the algorithms in the book in Python, Java, and other programming languages (currently hosted at github.com/aimacode).
A list of over 1400 schools that have used the book, many with links to online course
materials and syllabi. Supplementary material and links for students and instructors.
Instructions on how to report errors in the book, in the likely event that some exist.
Book cover
The cover depicts the final position from the decisive game 6 of the 1997 chess match in which the program Deep Blue defeated Garry Kasparov (playing Black), making this the first time a computer had beaten a world champion in a chess match. Kasparov is shown at the
Preface
top. To his right is a pivotal position from the second game of the historic Go match between former world champion Lee Sedol and DeepMind’s ALPHAGO program. Move 37 by ALPHAGO violated centuries of Go orthodoxy and was immediately seen by human experts as an embarrassing mistake, but it turned out to be a winning move. At top left is an Atlas humanoid robot built by Boston Dynamics. A depiction of a self-driving car sensing its environment appears between Ada Lovelace, the world’s first computer programmer, and Alan Turing, whose fundamental work defined artificial intelligence.
At the bottom of the chess
board are a Mars Exploration Rover robot and a statue of Aristotle, who pioneered the study of logic; his planning algorithm from De Motu Animalium appears behind the authors’ names. Behind the chess board is a probabilistic programming model used by the UN Comprehensive Nuclear-Test-Ban Treaty Organization for detecting nuclear explosions from seismic signals. Acknowledgments It takes a global village to make a book. Over 600 people read parts of the book and made
suggestions for improvement. The complete list is at aima.cs.berkeley.edu/ack.html;
we are grateful to all of them. We have space here to mention only a few especially important contributors. First the contributing writers:
Judea Pearl (Section 13.5, Causal Network Vikash Mansinghka (Section 15.3, Programs as Probability Models): Michael Wooldridge (Chapter 18, Multiagent Decision Making); Tan Goodfellow (Chapter 21, Deep Learning); Jacob Devlin and Mei-Wing Chang (Chapter 24, Deep Learning for Natural Language); « Jitendra Malik and David Forsyth (Chapter 25, Computer Vision); « Anca Dragan (Chapter 26, Robotics). Then some key roles: « Cynthia Yeung and Malika Cantor (project management); « Julie Sussman and Tom Galloway (copyediting and writing suggestions); « Omari Stephens (illustrations); « Tracy Johnson (editor); « Erin Ault and Rose Kernan (cover and color conversion); « Nalin Chhibber, Sam Goto, Raymond de Lacaze, Ravi Mohan, Ciaran O’Reilly, Amit Patel, Dragomir Radiv, and Samagra Sharma (online code development and mentoring); « Google Summer of Code students (online code development). Stuart would like to thank his wife, Loy Sheflott, for her endless patience and boundless wisdom. He hopes that Gordon, Lucy, George, and Isaac will soon be reading this book after they have forgiven him for working so long on it. RUGS (Russell’s Unusual Group of Students) have been unusually helpful, as always. Peter would like to thank his parents (Torsten and Gerda) for getting him started, and his wife (Kris), children (Bella and Juliet), colleagues, boss, and friends for encouraging and tolerating him through the long hours of writing and rewriting.
ix
About the Authors Stuart Russell was born in 1962 in Portsmouth, England.
He received his B.A. with first-
class honours in physics from Oxford University in 1982, and his Ph.D. in computer science from Stanford in 1986.
He then joined the faculty of the University of California at Berke-
ley, where he is a professor and former chair of computer science, director of the Center for Human-Compatible Al, and holder of the Smith-Zadeh Chair in Engineering. In 1990, he received the Presidential Young Investigator Award of the National Science Foundation, and in 1995 he was cowinner of the Computers and Thought Award. He is a Fellow of the Amer-
ican Association for Artificial Intelligence, the Association for Computing Machinery, and
the American Association for the Advancement of Science, an Honorary Fellow of Wadham College, Oxford, and an Andrew Carnegie Fellow. He held the Chaire Blaise Pascal in Paris
from 2012 to 2014. He has published over 300 papers on a wide range of topics in artificial intelligence. His other books include The Use of Knowledge in Analogy and Induction, Do the Right Thing: Studies in Limited Rationality (with Eric Wefald), and Human Compatible: Ariificial Intelligence and the Problem of Control.
Peter Norvig is currently a Director of Research at Google, Inc., and was previously the director responsible for the core Web search algorithms. He co-taught an online Al class that signed up 160,000 students, helping to kick off the current round of massive open online
classes. He was head of the Computational Sciences Division at NASA Ames Research Center, overseeing research and development in artificial intelligence and robotics. He received
aB.S. in applied mathematics from Brown University and a Ph.D. in computer science from Berkeley. He has been a professor at the University of Southern California and a faculty
member at Berkeley and Stanford. He is a Fellow of the American Association for Artificial
Intelligence, the Association for Computing Machinery, the American Academy of Arts and Sciences, and the California Academy of Science. His other books are Paradigms of Al Programming: Case Studies in Common Lisp, Verbmobil: A Translation System for Face-to-Face Dialog, and Intelligent Help Systems for UNIX. The two authors shared the inaugural AAAI/EAAI Outstanding Educator award in 2016.
Contents Atrtificial Intelligence Introduction L1 WhatTs AI? ... .o 1.2 The Foundations of Artificial Intelligence . 1.3 The History of Artificial Intelligence 1.4 The State of the Art . . . 1.5 Risks and Benefits of AI .
Summary
Bibliographical and Historical Notes
Intelligent Agents 2.1
22
2.3
24
Agentsand Environments
1 1 5 17 27 31
34
. . . . . . .
. . ..
..
... ...
Good Behavior: The Concept of Rationality
35 ...
The Nature of Environments . . . .. .
TheStructure of Agents
36 36
...................
. . . .. .. ... ...
SUMMATY . ..o o e Bibliographical and Historical Notes . . . . .. ..................
it
39
42
47
60 60
Problem-solving
Solving Problems by Searching 3.1 Problem-Solving Agents .
63 63
3.5
84
3.2 3.3 3.4
Example Problems . . . . Search Algorithms . . . Uninformed Search Strategies . . . . . . .
3.6
Heuristic Functions
Informed (Heuristic) Search Strategies . .
Summary
Bibliographical and Historical Notes
. . . . . . .
66 71 76 97
104
106
Search in Complex Environments
110
Summary
141
4.1 4.2 4.3 4.4 4.5
Local Search and Optimization Problems . Local Search in Continuous Spaces . . . . Search with Nondeterministic Actions Search in Partially Observable Environments Online Search Agents and Unknown Environments
Bibliographical and Historical Notes
. . . . .. ..................
Adversarial Search and Games 5.1 Game Theory 5.2 Optimal Decisions in Games
110 119 122 126 134 142
146 146 148 xi
Contents 5.3 54
Heuristic Alpha-Beta Tree Search Monte Carlo Tree Search
156 161
5.6 5.7
Partially Observable Games . Limitations of Game Search Algorithms
168 173
5.5
Stochastic Games
Summary
Bibliographical 6
174
and Historical Notes
175
Constraint Satisfaction Problems
180
6.1
Defining Constraint Satisfaction Problems
180
6.3 64 6.5
Backtracking Search for CSPs Local Search for CSPs . . . The Structure of Problems .
191 197 199
6.2
Constraint Propagation: Inference in CSPs .
Summary
Bibliographical and Historical Notes
III
7
164
. . . . .. ..................
204
Logical Agents 7.1 Knowledge-Based Agents
208 209
7.4 7.5
217 222
7.2 73
The Wumpus World . . Logic...................
7.6
Effective Propositional Model Checking
Propositional Logic: A Very Simple Logic . Propositional Theorem Proving
Bibliographical and Historical Notes
. . . . . ...................
210 214
232
237 246 247
First-Order Logic
251
8.1 8.2 8.3
251 256 265
Representation Revisited . Syntax and Semantics of First-Order Logic Using First-Order Logic . .
.
84 Knowledge Engineering in First-Order Logic Summary Bibliographical and Historical Notes
9
203
Knowledge, reasoning, and planning
7.7 Agents Based on Propositional Logic . . SUMMary ... 8
185
. . . . . .
271 277
278
Inference in First-Order Logic
280
Summary
309
9.1 9.2 9.3 9.4 9.5
Propositional vs. First-Order Inference . . . ... ............. Unification and First-Order Inference . . Forward Chaining . . . Backward Chaining . . Resolution . ......
Lo
Bibliographical and Historical Notes
. . . . . .
280 282 286 293 298
310
Contents
10 Knowledge Representation 10.1 Ontological Engineering . 102 Categories and Object 10.3 10.4
Events Mental Objects and Modal Logic
10.5
Reasoning Systems for Categories
10.6
Reasoning with Default Information
Summary
Bibliographical
and Historical NOes . . . . .. ..o oot oot
11 Automated Planning 11.1
112
11.3 11.4 11.5 11.6
Definition of Classical Planning
Algorithms for Classical Planning
Heuristics for Planning Hierarchical Planning . . Planning and Acting in Nondeterministic Domains Time, Schedules, and Resources
117 Analysis of Planning Approaches Summary ... Bibliographical and Historical Notes . . . . . . ... IV
. . . .
ooor
oot
Uncertain knowledge and reasoning
12 Quantifying Uncertainty
12.1
Acting under Uncertainty
12.2 12.3
Basic Probability Notation Inference Using Full Joint Distributions . .
12.5 12.6 12.7
Bayes’ Rule and Its Use . Naive Bayes Models . . . The Wumpus World Revisited
124
Independence
Summary
Bibliographical and Historical Notes
. . . . .. ..................
13 Probabilistic Reasoning
13.1
13.2 13.3
134
13.5
Representing Knowledge in an Uncertain Domain . . . .
The Semantics of Bayesian Networks . . . .. ... ... Exact Inference in Bayesian Networks
Approximate Inference for Bayesian Networks . . . . . . Causal Networks . . . . .
Summary
... .........
Bibliographical and Historical Notes 14 Probabilistic Reasoning over Time 14.1 Time and Uncertainty . . 14.2 Inference in Temporal Models
. . . . . . .
xiii
xiv
Contents
14.3 Hidden Markov Models 144 Kalman Filters . . . . . 14.5 Dynamic Bayesian Networks Summary Bibliographic and Historical Notes 15 Probabilistic Programming 15.1 Relational Probability Models 152 Open-Universe Probability Models
153
. . .
Keeping Track ofa Complex World . . .
Summary Bibliographic 16 Making Simple Decisions 16.1 16.2 16.3 16.4 16.5 16.6 16.7
Combining Beliefs and Desires under Uncertainty The Basis of Utility Theory .
. . . .. ........
Multiattribute Utility Functions Decision Networks . . . . .. ...... The Value of Information Unknown Preferences . . .
Summary Bibliographical and Historical Notes . . . . . . 17
18
Making Complex Decisions 17.1 Sequential Decision Problems 17.2 Algorithms for MDPs . 17.3 Bandit Problems . . . . 17.4 Partially Observable MDPs 17.5 Algorithms for Solving POMDPs . . . . Summary Bibliographical and Historical Notes . . . . . . Multiagent Decision Making 18.1
182 18.3
18.4
Properties of Multiagent Environments
Non-Cooperative Game Theory Cooperative Game Theory
. . . . .. ... ..........
Making Collective Decisions
Summary Bibliographical and Historical Notes . . . . . . . ... ..............
562 562 572 581 588 590 595 596 599 599 605 626 632 645 646
Machine Learning
19 Learning from Examples 19.1
FormsofLearning
. . .. ..........................
651
Contents
21
22
192 Supervised Learning. . . 193 Learning Decision Trees . 19.4 Model Selection and Optimization 19.5 The Theory of Learning . . . 19.6 Lincar Regression and Classification 19.7 Nonparametric Models 19.8 Ensemble Learning 19.9 Developing Machine Learning Systems . . Summary Bibliographical and Historical Notes Learning Probabilistic Models 20.1 Statistical Leaming . . .« ..o i i it 202 Leaming with Complete Data 203 Leaming with Hidden Variables: The EM Algorithm . . . Summary Bibliographical and Historical Notes Deep Learning 211 Simple Feedforward Networks . . ..« oo oo ooounittt o 212 Computation Graphs for Decp Learning 213 Convolutional Networks . . . . . . . . . . 214 Leaming Algorithms. . . 215 GeneraliZation . . . ... i i i 216 Recurrent Neural Networks - - .« oo« oo otot ot 217 Unsupervised Learning and Transfer Learning . . . . . . . ... ... .. 208 APPHCALONS « .« o« e e e e SUMMATY .« o o oo oo e e e e et Bibliographical and Historical Notes . . . . .« .. ... ..o ooooooo. Reinforcement Learning 22.1 Leaming from Rewards . . . . ..o oo oott it 222 Passive Reinforcement Learning 223 Active Reinforcement Learning 224 Generalization in Reinforcement Learning 225 Policy Search 22.6 Apprenticeship and Inverse Reinforcement Learning . . . 227 Applications of Reinforcement Learning SUMMALY - o o v vveeeerneneenenns Bibliographical and Historical Notes
VI
653 657 665 672 676 686 696 704 714 715 721 721 724 737 746 747 750 751 756 760 765 768 772 775 782 784 785 789 789 791 797 803 810 812 815 818 819
Communicating, perceiving, and acting
23 Natural Language Processing 23.1 Language Models 23.2
Grammar
823 823
833
xv
xvi
Contents 233 23.4
Parsing . ........ Augmented Grammars .
235 Complications of Real Natural Language 23.6 Natural Language Tasks . . . . . . . .. Summary Bibliographical and Historical Notes
24 Deep Learning for Natural Language Processing 24.1 24.2
Word Embeddings . . . .. .............. .. ... ..., Recurrent Neural Networks for NLP
24.4
The Transformer Architecture
243 245
24.6
Sequence-to-Sequence Models
Pretraining and Transfer Learning . . . . State of the art
Summary ... ..... ...
Bibliographical and Historical Notes 25
. . . . .. ..................
835 841
845 849 850 851
856
856 860
864 868
871
875
878
878
Computer Vision 25.1 Introduction 25.2 Image Formation . . . .
881 881 882
256
901
253 254 255
Simple Image Features Classifying Images . . . Detecting Objects . . .
The3D World
888 895 899
. . ...
25.7 Using Computer Vision Summary ... ... ...
906 919
Bibliographical and Historical Notes
920
26 Robotics 26.1 Robots 26.2 Robot Hardware . . . . 26.3 What kind of problem is robotics solving? 26.4 Robotic Perception . . . 26.5 Planning and Control 26.6 Planning Uncertain Movements . . . . . 26.7 Reinforcement Learning in Robotics 26.8 Humans and Robots 26.9 Alternative Robotic Frameworks 26.10 Application Domains
Summary ... ...
...
Bibliographical and Historical Notes
VII
. . . . .. ..................
925 925 926 930 931 938 956 958 961 968 971
974 975
Conclusions
27 Philosophy, Ethics, and Safety of AI 27.1
TheLimitsof AL . . ... ..........................
981
981
Contents 27.2 27.3
Can Machines Really Think? The Ethics of AT . .
.
Bibliographical and Historical Notes
.
Summary
. ... .....
L.
984 986
1005
1006
28 The Future of AT 28.1 28.2
A
B
. .
1012 1018
Mathematical Background
1023
A2 Vectors, Matrices, and Linear Algebra A3 Probability Distributions . . . . ... ... e L .. Bibliographical and Historical Notes . . . . .. ..................
1025 1027 1029
Al
Al Components Al Architectures
1012
Complexity Analysis and O() Notation . .
Notes on Languages and Algorithms B.1 B.2 B.3
Defining Languages with Backus—Naur Form (BNF) Describing Algorithms with Pseudocode . Online Supplemental Material . . . . . . .
Bibliography
Index
.
1023
1030
1030 1031 1032
1033
1069
xvii
This page intentionally left blank
G
1
INTRODUCTION In which we try to explain why we consider artificial intelligence to be a subject most
worthy of study, and in which we try to decide what exactly it is, this being a good thing to
decide before embarking.
We call ourselves Homo sapiens—man the wise—because our intelligence is so important
to us. For thousands of years, we have tried to understand how we think and act—that is,
how our brain, a mere handful of matter, can perceive, understand, predict, and manipulate a world far larger and more complicated than itself. The field of artificial intelligence, or Al
is concerned with not just understanding but also building intelligent entities—machines that
Intelligence Artificial intelligence
can compute how to act effectively and safely in a wide variety of novel situations.
Surveys regularly rank Al as one of the most interesting and fastest-growing fields, and it is already generating over a trillion dollars a year in revenue. Al expert Kai-Fu Lee predicts
that its impact will be “more than anything in the history of mankind.” Moreover, the intel-
lectual frontiers of Al are wide open. Whereas a student of an older science such as physics
might feel that the best ideas have already been discovered by Galileo, Newton, Curie, Ein-
stein, and the rest, Al still has many openings for full-time masterminds.
Al currently encompasses a huge variety of subfields, ranging from the general (learning,
reasoning, perception, and so on) to the specific, such as playing chess, proving mathematical theorems, writing poetry, driving a car, or diagnosing diseases. Al is relevant to any intellectual task; it is truly a universal field.
1.1
What
Is AI?
‘We have claimed that Al is interesting, but we have not said what it is. Historically, researchers have pursued several different versions of Al Some have defined intelligence in
terms of fidelity to human performance, while others prefer an abstract, formal definition of
intelligence called rationality—loosely speaking, doing the “right thing.” The subject matter Rationality itself also varies: some consider intelligence to be a property of internal thought processes
and reasoning, while others focus on intelligent behavior, an external characterization.! From these two dimensions—human
vs. rational® and thought vs. behavior—there are
four possible combinations, and there have been adherents and research programs for all 1" Inthe public eye, there i sometimes confusion between the terms “artificial intelligence” and “machine learning” Machine learning is a subfield of Al that studies the ability to improve performance based on experience. Some Al systems use machine learning methods to achieve competence, but some do not. 2 We are not suggesting that humans are “irrational” in the dictionary sense of “deprived of normal mental clarity” We are merely conceding that human decisions are not always mathematically perfect.
Chapter 1 Introduction
four. The methods used are necessarily different: the pursuit of human-like intelligence must be in part an empirical science related to psychology, involving observations and hypotheses about actual human behavior and thought processes; a rationalist approach, on the other hand, involves a combination of mathematics and engineering, and connecs to statistics, control theory, and economics. The various groups have both disparaged and helped cach other. Let us look at the four approaches in more detail. 1.1.1
Turing test
Acting humanly:
The Turing test approach
The Turing test, proposed by Alan Turing (1950), was designed as a thought experiment that would sidestep the philosophical vagueness of the question “Can a machine think?” A com-
puter passes the test if a human interrogator, after posing some written questions, cannot tell
whether the written responses come from a person or from a computer. Chapter 27 discusses the details of the test and whether a computer would really be intelligent if it passed.
For
now, we note that programming a computer to pass a rigorously applied test provides plenty
Natural language processing Knowledge representation Automated reasoning Machine learning Total Turing test Computer vision Robotics
to work on. The computer would need the following capabilities:
o natural language processing to communicate successfully in a human language;
o knowledge representation to store what it knows or hears; « automated reasoning to answer questions and to draw new conclusions;
« machine learning to adapt to new circumstances and to detect and extrapolate patterns. Turing viewed the physical simulation of a person as unnecessary to demonstrate intelligence.
However, other researchers have proposed a total Turing test, which requires interaction with
objects and people in the real world. To pass the total Turing test, a robot will need * computer vision and speech recognition to perceive the world;
o robotics to manipulate objects and move about.
These six disciplines compose most of AL Yet Al researchers have devoted little effort to
passing the Turing test, believing that it is more important to study the underlying principles of intelligence. The quest for “artificial flight” succeeded when engineers and inventors stopped imitating birds and started using wind tunnels and learning about aerodynamics.
Aeronautical engineering texts do not define the goal of their field as making “machines that
fly so exactly like pigeons that they can fool even other pigeons.” 1.1.2
Thinking humanly:
The cognitive modeling approach
To say that a program thinks like a human, we must know how humans think. We can learn about human thought in three ways:
Introspection
+ introspection—trying to catch our own thoughts as they go by;
Bra
« brain imaging—observing the brain in action.
Psychological experiment
imaging
« psychological experiments—observing a person in action;
Once we have a sufficiently precise theory of the mind, it becomes possible to express the
theory as a computer program. If the program’s input—output behavior matches corresponding human behavior, that is evidence that some of the program’s mechanisms could also be
operating in humans.
For example, Allen Newell and Herbert Simon, who developed GPS, the “General Problem Solver” (Newell and Simon, 1961), were not content merely to have their program solve
Section 1.1
What Is AI?
problems correctly. They were more concerned with comparing the sequence and timing of its reasoning steps to those of human subjects solving the same problems.
The interdisci-
plinary field of cognitive science brings together computer models from Al and experimental techniques from psychology to construct precise and testable theories of the human mind.
Cognitive science
Cognitive science is a fascinating field in itself, worthy of several textbooks and at least
one encyclopedia (Wilson and Keil, 1999). We will occasionally comment on similarities or
differences between Al techniques and human cognition. Real cognitive science, however, is necessarily based on experimental investigation of actual humans or animals. We will leave
that for other books, as we assume the reader has only a computer for experimentation.
In the early days of Al there was often confusion between the approaches. An author would argue that an algorithm performs well on a task and that it is therefore a good model
of human performance, or vice versa. Modern authors separate the two kinds of claims; this
distinction has allowed both Al and cognitive science to develop more rapidly. The two fields fertilize each other, most notably in computer vision, which incorporates neurophysiological evidence into computational models.
Recently, the combination of neuroimaging methods
combined with machine learning techniques for analyzing such data has led to the beginnings
of a capability to “read minds™—that is, to ascertain the semantic content of a person’s inner
thoughts. This capability could, in turn, shed further light on how human cognition works. 1.1.3
Thinking rationally:
The
“laws of thought”
approach
The Greek philosopher Aristotle was one of the first to attempt to codify “right thinking”™— that is, irrefutable reasoning processes. His syllogisms provided patterns for argument struc-
tures that always yielded correct conclusions when given correct premises. The canonical
Syllogism
example starts with Socrates is a man and all men are mortal and concludes that Socrates is
mortal. (This example is probably due to Sextus Empiricus rather than Aristotle.) These laws of thought were supposed to govern the operation of the mind; their study initiated the field called logic.
Logicians in the 19th century developed a precise notation for statements about objects
in the world and the relations among them. (Contrast this with ordinary arithmetic notation,
which provides only for statements about numbers.) By 1965, programs could, in principle, solve any solvable problem described in logical notation.
The so-called logicist tradition
within artificial intelligence hopes to build on such programs to create intelligent systems.
Logicist
Logic as conventionally understood requires knowledge of the world that is certain—
a condition that, in reality, is seldom achieved. We simply don’t know the rules of, say, politics or warfare in the same way that we know the rules of chess or arithmetic. The theory
of probability fills this gap, allowing rigorous reasoning with uncertain information.
In Probability
principle, it allows the construction of a comprehensive model of rational thought, leading
from raw perceptual information to an understanding of how the world works to predictions about the future. What it does not do, is generate intelligent behavior.
theory of rational action. Rational thought, by itself, is not enough. 1.1.4
Acting ra
nally:
The
For that, we need a
rational agent approach
An agent is just something that acts (agent comes from the Latin agere, to do). Of course,
all computer programs do something, but computer agents are expected to do more: operate autonomously, perceive their environment, persist over a prolonged time period, adapt to
Agent
Chapter 1 Introduction Rational agent
change, and create and pursue goals. A rational agent is one that acts so as to achieve the best outcome or, when there is uncertainty, the best expected outcome.
In the “laws of thought” approach to Al the emphasis was on correct inferences. Making correct inferences is sometimes part of being a rational agent, because one way to act rationally is to deduce that a given action is best and then to act on that conclusion.
On the
other hand, there are ways of acting rationally that cannot be said to involve inference. For
example, recoiling from a hot stove is a reflex action that is usually more successful than a slower action taken after careful deliberation.
All the skills needed for the Turing test also allow an agent to act rationally. Knowledge
representation and reasoning enable agents to reach good decisions. We need to be able to
generate comprehensible sentences in natural language to get by in a complex society.
We
need learning not only for erudition, but also because it improves our ability to generate
effective behavior, especially in circumstances that are new.
The rational-agent approach to Al has two advantages over the other approaches. First, it
is more general than the “laws of thought™ approach because correct inference is just one of several possible mechanisms for achieving rationality. Second, it is more amenable to scien-
tific development. The standard of rationality is mathematically well defined and completely
general. We can often work back from this specification to derive agent designs that provably
achieve it—something that is largely impossible if the goal is to imitate human behavior or
thought processes. For these reasons, the rational-agent approach to Al has prevailed throughout most of the field’s history. In the early decades, rational agents were built on logical foundations and formed definite plans to achieve specific goals. Later, methods based on probability Do the
> right thing
Standard model
theory and machine learning allowed the creation of agents that could make decisions under
uncertainty to attain the best expected outcome.
In a nutshell, A/ has focused on the study
and construction of agents that do the right thing. What counts as the right thing is defined
by the objective that we provide to the agent. This general paradigm is so pervasive that we
might call it the standard model. It prevails not only in Al but also in control theory, where a
controller minimizes a cost function; in operations research, where a policy maximizes a sum of rewards; in statistics, where a decision rule minimizes a loss function; and in economics, where a decision maker maximizes utility or some measure of social welfare.
We need to make one important refinement to the standard model to account for the fact
that perfect rationality—always taking the exactly optimal action—is not feasible in complex
Limited rationality
environments. The computational demands are just too high. Chapters 5 and 17 deal with the
issue of limited rationality—acting appropriately when there is not enough time to do all the computations one might like. However, perfect rationality often remains a good starting point for theoretical analysis.
1.1.5
Benef
| machines
The standard model
has been a useful guide for Al research since its inception, but it is
probably not the right model in the long run. The reason is that the standard model assumes that we will supply a fully specified objective to the machine.
For an artificially defined task such as chess or shortest-path computation, the task comes
with an objective built in—so the standard model is applicable.
As we move into the real
world, however, it becomes more and more difficult to specify the objective completely and
Section 1.2
The Foundations of Artificial Intelligence
correctly. For example, in designing a self-driving car, one might think that the objective is to reach the destination safely. But driving along any road incurs a risk of injury due to other
errant drivers, equipment failure, and so on; thus, a strict goal of safety requires staying in the garage. There is a tradeoff between making progress towards the destination and incurring a risk of injury. How should this tradeoff be made? Furthermore, to what extent can we allow the car to take actions that would annoy other drivers? How much should the car moderate
its acceleration, steering, and braking to avoid shaking up the passenger? These kinds of questions are difficult to answer a priori. They are particularly problematic in the general area of human-robot interaction, of which the self-driving car is one example. The problem of achieving agreement between our true preferences and the objective we put into the machine is called the value alignment problem: the values or objectives put into
the machine must be aligned with those of the human. If we are developing an Al system in
Value alignment problem
the lab or in a simulator—as has been the case for most of the field’s history—there is an easy
fix for an incorrectly specified objective: reset the system, fix the objective, and try again. As the field progresses towards increasingly capable intelligent systems that are deployed
in the real world, this approach is no longer viable. A system deployed with an incorrect
objective will have negative consequences. more negative the consequences.
Moreover, the more intelligent the system, the
Returning to the apparently unproblematic example of chess, consider what happens if
the machine is intelligent enough to reason and act beyond the confines of the chessboard. In that case, it might attempt to increase its chances of winning by such ruses as hypnotiz-
ing or blackmailing its opponent or bribing the audience to make rustling noises during its
opponent’s thinking time.>
It might also attempt to hijack additional computing power for
itself. These behaviors are not “unintelligent” or “insane”; they are a logical consequence of defining winning as the sole objective for the machine. It is impossible to anticipate all the ways in which a machine pursuing a fixed objective might misbehave. There is good reason, then, to think that the standard model is inadequate.
‘We don’t want machines that are intelligent in the sense of pursuing their objectives; we want
them to pursue our objectives. If we cannot transfer those objectives perfectly to the machine, then we need a new formulation—one in which the machine is pursuing our objectives, but is necessarily uncertain as to what they are. When a machine knows that it doesn’t know the
complete objective, it has an incentive to act cautiously, to ask permission, to learn more about
our preferences through observation, and to defer to human control. Ultimately, we want agents that are provably beneficial to humans. We will return to this topic in Section 1.5.
The Foundations of Arti
| Intelligence
In this section, we provide a brief history of the disciplines that contributed ideas, viewpoints,
and techniques to Al Like any history, this one concentrates on a small number of people, events, and ideas and ignores others that also were important. We organize the history around
a series of questions. We certainly would not wish to give the impression that these questions are the only ones the disciplines address or that the disciplines have all been working toward Al as their ultimate fruition.
3 In one of the fi opponent’s eyes.”
books on chess, Ruy Lopez (1561) wrote, “Always place the board so the sun
Provably beneficial
Chapter 1 Introduction 1.2.1
Philosophy
« Can formal rules be used to draw valid conclusions? « How does the mind arise from a physical brain? « Where does knowledge come from? + How does knowledge lead to action?
Aristotle (384-322 BCE) was the first to formulate a precise set of laws governing the rational
part of the mind. He developed an informal system of syllogisms for proper reasoning, which in principle allowed one to generate conclusions mechanically, given initial premises. Ramon Llull (c. 1232-1315) devised a system of reasoning published as Ars Magna or The Great Art (1305). Llull tried to implement his system using an actual mechanical device:
a set of paper wheels that could be rotated into different permutations.
Around 1500, Leonardo da Vinci (1452-1519) designed but did not build a mechanical calculator; recent reconstructions have shown the design to be functional. The first known calculating machine was constructed around 1623 by the German scientist Wilhelm Schickard (1592-1635). Blaise Pascal (1623-1662) built the Pascaline in 1642 and wrote that it “produces effects which appear nearer to thought than all the actions of animals.” Gottfried Wilhelm Leibniz (1646-1716) built a mechanical device intended to carry out operations on concepts rather than numbers, but its scope was rather limited. In his 1651 book Leviathan, Thomas Hobbes (1588-1679) suggested the idea of a thinking machine, an “artificial animal”
in his words, arguing “For what is the heart but a spring; and the nerves, but so many strings; and the joints, but so many wheels.” He also suggested that reasoning was like numerical computation: “For ‘reason’ ... is nothing but ‘reckoning,” that is adding and subtracting.”
It’s one thing to say that the mind operates, at least in part, according to logical or nu-
merical rules, and to build physical systems that emulate some of those rules. It’s another to
say that the mind itself is such a physical system. René Descartes (1596-1650) gave the first clear discussion of the distinction between mind and matter. He noted that a purely physical conception of the mind seems to leave little room for free will. If the mind is governed en-
Dualism
tirely by physical laws, then it has no more free will than a rock “deciding” to fall downward. Descartes was a proponent of dualism.
He held that there is a part of the human mind (or
soul or spirit) that is outside of nature, exempt from physical laws. Animals, on the other hand, did not possess this dual quality; they could be treated as machines. An alternative to dualism is materialism, which holds that the brain’s operation accord-
ing to the laws of physics constitutes the mind. Free will is simply the way that the perception
of available choices appears to the choosing entity. The terms physicalism and naturalism are also used to describe this view that stands in contrast to the supernatural.
Empiricism
Given a physical mind that manipulates knowledge, the next problem is to establish the source of knowledge. The empiricism movement, starting with Francis Bacon’s (1561-1626) Novum Organum,* is characterized by a dictum of John Locke (1632-1704): “Nothing is in the understanding, which was not first in the senses.”
Induction
David Hume’s (1711-1776) A Treatise of Human Nature (Hume,
1739) proposed what
is now known as the principle of induction: that general rules are acquired by exposure to
repeated associations between their elements.
4 The Novum Organum is an update of Aristotle’s Organon, or instrument of thought.
Section 1.2
The Foundations of Artificial Intelligence
Building on the work of Ludwig Wittgenstein (1889-1951) and Bertrand Russell (1872—
1970), the famous Vienna Circle (Sigmund, 2017), a group of philosophers and mathemati-
cians meeting in Vienna in the 1920s and 1930s, developed the doctrine of logical positivism. This doctrine holds that all knowledge can be characterized by logical theories connected, ul-
timately, to observation sentences that correspond to sensory inputs; thus logical positivism combines rationalism and empiricism.
The confirmation theory of Rudolf Carnap (1891-1970) and Carl Hempel (1905-1997)
attempted to analyze the acquisition of knowledge from experience by quantifying the degree
Logical positivism Observation sentence
Confirmation theory
of belief that should be assigned to logical sentences based on their connection to observations that confirm or disconfirm them.
Carnap’s book The Logical Structure of the World (1928)
was perhaps the first theory of mind as a computational process. The final element in the philosophical
picture of the mind is the connection
between
knowledge and action. This question is vital to AT because intelligence requires action as well as reasoning. Moreover, only by understanding how actions are justified can we understand how to build an agent whose actions are justifiable (or rational).
Avistotle argued (in De Motu Animalium) that actions are justified by a logical connection
between goals and knowledge of the action’s outcome:
But how does it happen that thinking is sometimes accompanied by action and sometimes not, sometimes by motion, and sometimes not? It looks as if almost the same thing happens as in the case of reasoning and making inferences about unchanging objects. But in that case the end is a speculative proposition ... whereas here the conclusion which results from the two premises is an action. . . I need covering; a cloak is a covering. T need a cloak. What I need, I have to make; I need a cloak. I have to make a cloak. And the conclusion, the “T have to make a cloak.” is an action.
In the Nicomachean Ethics (Book IIL. 3, 1112b), Aristotle further elaborates on this topic, suggesting an algorithm:
‘We deliberate not about ends, but about means. For a doctor does not deliberate whether he shall heal, nor an orator whether he shall persuade, . . They assume the end and consider how and by what means it is attained, and if it scems casily and best produced thereby: while if it is achieved by one means only they consider how it will be achieved by this and by what means this will be achieved, till they come to the first cause, ... and what is last in the order of analysis seems to be first in the order of becoming. And if we come on an impossibility, we give up the search, e.g., if we need money and this cannot be got; but if a thing appears possible we try to do it. Aristotle’s algorithm was implemented 2300 years later by Newell and Simon in their General Problem Solver program. We would now call it a greedy regression planning system (see Chapter 11). Methods based on logical planning to achieve definite goals dominated the first few decades of theoretical research in AL
Thinking purely in terms of actions achieving goals is often useful but sometimes inapplicable. For example, if there are several different ways to achieve a goal, there needs to be some way to choose among them. More importantly, it may not be possible to achieve a goal with certainty, but some action must still be taken. How then should one decide? Antoine Ar-
nauld (1662), analyzing the notion of rational decisions in gambling, proposed a quantitative formula for maximizing the expected monetary value of the outcome. Later, Daniel Bernoulli
(1738) introduced the more general notion of utility to capture the internal, subjective value Utility
Chapter 1 Introduction of an outcome.
The modern notion of rational decision making under uncertainty involves
maximizing expected utility, as explained in Chapter 16.
Utilitarianism
In matters of ethics and public policy, a decision maker must consider the interests of multiple individuals. Jeremy Bentham (1823) and John Stuart Mill (1863) promoted the idea
of utilitarianism: that rational decision making based on maximizing utility should apply to all spheres of human activity, including public policy decisions made on behalf of many
individuals. Utilitarianism is a specific kind of consequentialism: the idea that what is right
and wrong is determined by the expected outcomes of an action. Deontological ethics
In contrast, Immanuel Kant, in 1875 proposed a theory of rule-based or deontological
ethics, in which “doing the right thing” is determined not by outcomes but by universal social
laws that govern allowable actions, such as “don’t lie” or “don’t kill.” Thus, a utilitarian could tell a white lie if the expected good outweighs the bad, but a Kantian would be bound not to,
because lying is inherently wrong. Mill acknowledged the value of rules, but understood them as efficient decision procedures compiled from first-principles reasoning about consequences. Many modern Al systems adopt exactly this approach.
1.2.2
Mathematics
+ What are the formal rules to draw valid conclusions?
« What can be computed?
+ How do we reason with uncertain information? Philosophers staked out some of the fundamental ideas of Al but the leap to a formal science
required the mathematization of logic and probability and the introduction of a new branch
Formal logic
of mathematics: computation. The idea of formal logic can be traced back to the philosophers of ancient Greece, India, and China, but its mathematical development really began with the work of George Boole (1815-1864), who worked out the details of propositional, or Boolean, logic (Boole, 1847). In 1879, Gottlob Frege (1848-1925) extended Boole’s logic to include objects and relations,
creating the first-order logic that is used today.> In addition to its central role in the early period of Al research, first-order logic motivated the work of Godel and Turing that underpinned
Probability
computation itself, as we explain below.
The theory of probability can be seen as generalizing logic to situations with uncertain
information—a consideration of great importance for Al Gerolamo Cardano (1501-1576)
first framed the idea of probability, describing it in terms of the possible outcomes of gam-
bling events.
In 1654, Blaise Pascal (1623-1662), in a letter to Pierre Fermat (1601-1665),
showed how to predict the future of an unfinished gambling game and assign average pay-
offs to the gamblers.
Probability quickly became an invaluable part of the quantitative sci-
ences, helping to deal with uncertain measurements and incomplete theories. Jacob Bernoulli (1654-1705, uncle of Daniel), Pierre Laplace (1749-1827), and others advanced the theory and introduced new statistical methods. Thomas Bayes (1702-1761) proposed a rule for updating probabilities in the light of new evidence; Bayes’ rule is a crucial tool for Al systems. Statistics
The formalization of probability, combined with the availability of data, led to the emergence of statistics as a field. One of the first uses was John Graunt’s analysis of Lon-
5 Frege’s proposed notation for first-order logic—an arcane combination of textual and geometric features— never became popular.
Section 1.2
don census data in 1662.
The Foundations of Artificial Intelligence
Ronald Fisher is considered the first modern statistician (Fisher,
1922). He brought together the ideas of probability, experiment design, analysis of data, and
computing—in 1919, he insisted that he couldn’t do his work without a mechanical calculator called the MILLIONAIRE
(the first calculator that could do multiplication), even though the
cost of the calculator was more than his annual ry (Ross, 2012). The history of computation is as old as the history of numbers, but the first nontrivial
algorithm is thought to be Euclid’s algorithm for computing greatest common divisors. The word algorithm comes from Muhammad
Algorithm
ibn Musa al-Khwarizmi, a 9th century mathemati-
cian, whose writings also introduced Arabic numerals and algebra to Europe. Boole and others discussed algorithms for logical deduction, and, by the late 19th century, efforts were under way to formalize general mathematical reasoning as logical deduction.
Kurt Godel (1906-1978) showed that there exists an effective procedure to prove any true
statement in the first-order logic of Frege and Russell, but that first-order logic could not cap-
ture the principle of mathematical induction needed to characterize the natural numbers. In 1931, Godel showed that limits on deduction do exist. His incompleteness theorem showed
that in any formal theory as strong as Peano arithmetic (the elementary theory of natural
Incompleteness theorem
numbers), there are necessarily true statements that have no proof within the theory.
This fundamental result can also be interpreted as showing that some functions on the
integers cannot be represented by an algorithm—that is, they cannot be computed.
This
motivated Alan Turing (1912-1954) to try to characterize exactly which functions are com-
putable—capable of being computed by an effective procedure. The Church-Turing thesis Computability proposes to identify the general notion of computability with functions computed by a Turing machine (Turing, 1936). Turing also showed that there were some functions that no Turing machine can compute. For example, no machine can tell in general whether a given program will return an answer on a given input or run forever. Although computability is important to an understanding of computation, the notion of
tractability has had an even greater impact on Al Roughly speaking, a problem is called Tractability intractable if the time required to solve instances of the problem grows exponentially with
the size of the instances. The distinction between polynomial and exponential growth in
complexity was first emphasized in the mid-1960s (Cobham, 1964; Edmonds, 1965). It is important because exponential growth means that even moderately large instances cannot be
solved in any reasonable time. The theory of NP-completeness, pioneered by Cook (1971) and Karp (1972), provides a NP-completeness basis for analyzing the tractability of problems: any problem class to which the class of NPcomplete problems can be reduced is likely to be intractable. (Although it has not been proved that NP-complete problems are necessarily intractable, most theoreticians believe it.) These results contrast with the optimism with which the popular press greeted the first computers—
“Electronic Super-Brains” that were “Faster than Einstein!” Despite the increasing speed of computers, careful use of resources and necessary imperfection will characterize intelligent systems. Put crudely, the world is an extremely large problem instance!
1.2.3
Economics
« How should we make decisions in accordance with our preferences? « How should we do this when others may not go along? « How should we do this when the payoff may be far in the future?
10
Chapter 1 Introduction The science of economics originated in 1776, when Adam Smith (1723-1790) published An
Inquiry into the Nature and Causes of the Wealth of Nations. Smith proposed to analyze economies as consisting of many individual agents attending to their own interests. Smith
was not, however, advocating financial greed as a moral position: his earlier (1759) book The
Theory of Moral Sentiments begins by pointing out that concern for the well-being of others is an essential component of the interests of every individual.
Most people think of economics as being about money, and indeed the first mathemati-
cal analysis of decisions under uncertainty, the maximum-expected-value formula of Arnauld (1662), dealt with the monetary value of bets. Daniel Bernoulli (1738) noticed that this formula didn’t seem to work well for larger amounts of money, such as investments in maritime
trading expeditions. He proposed instead a principle based utility, and explained human investment choices by proposing additional quantity of money diminished as one acquired more Léon Walras (pronounced “Valrasse™) (1834-1910) gave
on maximization of expected that the marginal utility of an money. utility theory a more general
foundation in terms of preferences between gambles on any outcomes (not just monetary
outcomes). The theory was improved by Ramsey (1931) and later by John von Neumann
and Oskar Morgenstern in their book The Theory of Games and Economic Behavior (1944).
Decision theory
Economics is no longer the study of money; rather it is the study of desires and preferences.
Decision theory, which combines probability theory with utility theory, provides a for-
mal and complete framework for individual decisions (economic or otherwise) made under
uncertainty—that is, in cases where probabilistic descriptions appropriately capture the de-
cision maker’s environment.
This is suitable for “large” economies where each agent need
pay no attention to the actions of other agents as individuals.
For “small” economies, the
situation is much more like a game: the actions of one player can significantly affect the utility of another (either positively or negatively). Von Neumann and Morgenstern’s develop-
ment of game theory (see also Luce and Raiffa, 1957) included the surprising result that, for some games, a rational agent should adopt policies that are (or least appear to be) randomized. Unlike decision theory, game theory does not offer an unambiguous prescription for selecting actions. In AL decisions involving multiple agents are studied under the heading of multiagent systems (Chapter 18).
Economists, with some exceptions, did not address the third question listed above: how to
Operations research
make rational decisions when payoffs from actions are not immediate but instead result from
several actions taken in sequence. This topic was pursued in the field of operations research,
which emerged in World War II from efforts in Britain to optimize radar installations, and later
found innumerable civilian applications. The work of Richard Bellman (1957) formalized a
class of sequential decision problems called Markov decision processes, which we study in Chapter 17 and, under the heading of reinforcement learning, in Chapter 22. Work in economics and operations research has contributed much to our notion of rational
agents, yet for many years Al research developed along entirely separate paths. One reason was the apparent complexity of making rational decisions. The pioneering Al researcher
Satisfcing
Herbert Simon (1916-2001) won the Nobel Prize in economics in 1978 for his early work
showing that models based on satisficing—making decisions that are “good enough,” rather than laboriously calculating an optimal decision—gave a better description of actual human behavior (Simon, 1947). Since the 1990s, there has been a resurgence of interest in decisiontheoretic techniques for AL
Section 1.2
1.2.4
The Foundations of Artificial Intelligence
11
Neuroscience
+ How do brains process information?
Neuroscience is the study of the nervous system, particularly the brain. Although the exact Neuroscience way in which the brain enables thought is one of the great mysteries of science, the fact that it
does enable thought has been appreciated for thousands of years because of the evidence that strong blows to the head can lead to mental incapacitation. It has also long been known that human brains are somehow different; in about 335 BCE Aristotle wrote, “Of all the animals,
man has the largest brain in proportion to his size.”® Still, it was not until the middle of the
18th century that the brain was widely recognized as the seat of consciousness. Before then,
candidate locations included the heart and the spleen.
Paul Broca’s (1824-1880) investigation of aphasia (speech deficit) in brain-damaged patients in 1861 initiated the study of the brain’s functional organization by identifying a localized area in the left hemisphere—now called Broca’s area—that is responsible for speech production.” By that time, it was known that the brain consisted largely of nerve cells, or neurons, but it was not until 1873 that Camillo Golgi (1843-1926) developed a staining technique Neuron allowing the observation of individual neurons (see Figure 1.1). This technique was used by Santiago Ramon y Cajal (1852-1934) in his pioneering studies of neuronal organization.® It is now widely accepted that cognitive functions result from the electrochemical operation of these structures.
That is, a collection of simple cells can lead to thought, action, and
consciousness. In the pithy words of John Searle (1992), brains cause minds.
Actuators
Agent
Figure 2.14 A model-based. utility-based agent. It uses a model of the world, along with a utility function that measures its preferences among states of the world. Then it chooses the action that leads to the best expected utility, where expected utility is computed by averaging over all possible outcome states, weighted by the probability of the outcome. aim for, none of which can be achieved with certainty, utility provides a way in which the likelihood of success can be weighed against the importance of the goals. Partial observability and nondeterminism are ubiquitous in the real world, and so, there-
fore, is decision making under uncertainty. Technically speaking, a rational utility-based agent chooses the action that maximizes the expected utility of the action outcomes—that
is, the utility the agent expects to derive, on average, given the probabilities and utilities of each outcome. (Appendix A defines expectation more precisely.) In Chapter 16, we show that any rational agent must behave as if it possesses a utility function whose expected value
it tries to maximize. An agent that possesses an explicit utility function can make rational de-
cisions with a general-purpose algorithm that does not depend on the specific utility function
being maximized. In this way, the “global” definition of rationality—designating as rational
those agent functions that have the highest performance—is turned into a “local” constraint
on rational-agent designs that can be expressed in a simple program. The utility-based agent structure appears in Figure 2.14.
Utility-based agent programs
appear in Chapters 16 and 17, where we design decision-making agents that must handle the uncertainty inherent in nondeterministic or partially observable environments. Decision mak-
ing in multiagent environments is also studied in the framework of utility theory, as explained in Chapter 18. At this point, the reader may be wondering, “Is it that simple? We just build agents that maximize expected utility, and we're done?”
It’s true that such agents would be intelligent,
but it’s not simple. A utility-based agent has to model and keep track of its environment,
tasks that have involved a great deal of research on perception, representation, reasoning,
and learning. The results of this research fill many of the chapters of this book. Choosing
the utility-maximizing course of action is also a difficult task, requiring ingenious algorithms
that fill several more chapters.
Even with these algorithms, perfect rationality is usually
Expected utility
56
Chapter 2 Intelligent Agents Performance standard
Cride
|-
Sensors =
Learning element learning goals Problem generator Agent
changes knowledge
Performance element
JUSWIUOIAUF
feedback ‘
Actuators
Figure 2.15 A general leaning agent. The “performance element” box represents what we have previously considered to be the whole agent program. Now, the “learning element” box gets to modify that program to improve its performance. unachievable in practice because of computational complexity, as we noted in Chapter 1. We. Model-free agent
also note that not all utility-based agents are model-based; we will see in Chapters 22 and 26 that a model-free agent can learn what action is best in a particular situation without ever
learning exactly how that action changes the environment.
Finally, all of this assumes that the designer can specify the utility function correctly; Chapters 17, 18, and 22 consider the issue of unknown utility functions in more depth. 2.4.6
Learning agents
We have described agent programs with various methods for selecting actions. We have not, so far, explained how the agent programs come into being. Tn his famous early paper, Turing (1950) considers the idea of actually programming his intelligent machines by hand. He estimates how much work this might take and concludes, “Some more expeditious method seems desirable.”
them.
The method he proposes is to build learning machines and then to teach
In many areas of Al this is now the preferred method for creating state-of-the-art
systems. Any type of agent (model-based, goal-based, utility-based, etc.) can be built as a learning agent (or not). Learning has another advantage, as we noted earlier: it allows the agent to operate in initially unknown environments and to become more competent than its initial knowledge alone
might allow. In this section, we briefly introduce the main ideas of learning agents. Through-
out the book, we comment on opportunities and methods for learning in particular kinds of
Learning element
Performance element
agents. Chapters 19-22 go into much more depth on the learning algorithms themselves. A learning agent can be divided into four conceptual components, as shown in Figure 2.15.
The most important distinction is between the learning element,
which is re-
sponsible for making improvements, and the performance element, which is responsible for selecting external actions. The performance element is what we have previously considered
Section 2.4 The Structure of Agents 1o be the entire agent: it takes in percepts and decides on actions. The learning element uses feedback from the critic on how the agent is doing and determines how the performance element should be modified to do better in the future.
57
Critic
The design of the learning element depends very much on the design of the performance
element. When trying to design an agent that learns a certain capability, the first question is
not “How am I going to get it to learn this?” but “What kind of performance element will my
agent use to do this once it has learned how?” Given a design for the performance element,
learning mechanisms can be constructed to improve every part of the agent.
The critic tells the learning element how well the agent is doing with respect to a fixed
performance standard. The critic is necessary because the percepts themselves provide no indication of the agent’s success. For example, a chess program could receive a percept indicating that it has checkmated its opponent, but it needs a performance standard to know that this is a good thing; the percept itself does not say so. It is important that the performance
standard be fixed. Conceptually, one should think of it as being outside the agent altogether because the agent must not modify it to fit its own behavior.
The last component of the learning agent is the problem generator. It is responsible
for suggesting actions that will lead to new and informative experiences. If the performance
Problem generator
element had its way, it would keep doing the actions that are best, given what it knows, but if the agent is willing to explore a little and do some perhaps suboptimal actions in the short
run, it might discover much better actions for the long run. The problem generator’s job is to suggest these exploratory actions. This is what scientists do when they carry out experiments.
Galileo did not think that dropping rocks from the top of a tower in Pisa was valuable in itself.
He was not trying to break the rocks or to modify the brains of unfortunate pedestrians. His
aim was to modify his own brain by identifying a better theory of the motion of object The learning element can make changes to any of the “knowledge” components shown in the agent diagrams (Figures 2.9, 2.11, 2.13, and 2.14). The simplest cases involve learning directly from the percept sequence. Observation of pairs of successive states of the environ-
ment can allow the agent to learn “What my actions do” and “How the world evolves™ in
response (o its actions. For example, if the automated taxi exerts a certain braking pressure
when driving on a wet road, then it will soon find out how much deceleration is actually
achieved, and whether it skids off the road. The problem generator might identify certain parts of the model that are in need of improvement and suggest experiments, such as trying out the brakes on different road surfaces under different conditions.
Improving the model components of a model-based agent so that they conform better
with reality is almost always a good idea, regardless of the external performance standard.
(In some cases, it is better from a computational point of view to have a simple but slightly inaccurate model rather than a perfect but fiendishly complex model.)
Information from the
external standard is needed when trying to learn a reflex component or a utility function.
For example, suppose the taxi-driving agent receives no tips from passengers who have
been thoroughly shaken up during the trip. The external performance standard must inform
the agent that the loss of tips is a negative contribution to its overall performance; then the
agent might a sense, the (or penalty) performance
be able to learn that violent maneuvers do not contribute to its own utility. In performance standard distinguishes part of the incoming percept as a reward Reward that provides direct feedback on the quality of the agent’s behavior. Hard-wired Penalty standards such as pain and hunger in animals can be understood in this way.
Chapter 2 Intelligent Agents More example, settles on know it’s
generally, human choices can provide information about human preferences. For suppose the taxi does not know that people generally don’t like loud noises, and the idea of blowing its horn continuously as a way of ensuring that pedestrians coming. The consequent human behavior—covering ears, using bad language, and
possibly cutting the wires to the horn—would provide evidence to the agent with which to update its utility function. This
issue is discussed further in Chapter 22.
In summary, agents have a variety of components, and those components
can be repre-
sented in many ways within the agent program, so there appears to be great variety among
learning methods. There is, however, a single unifying theme. Learning in intelligent agents can be summarized as a process of modification of each component of the agent to bring the components into closer agreement with the available feedback information, thereby improv-
ing the overall performance of the agent. 2.4.7
How the components of agent programs work
We have described agent programs (in very high-level terms) as consisting of various compo-
nents, whose function it is to answer questions such as: “What is the world like now?” “What action should I do now?” “What do my actions do?” The next question for a student of Al
is, “How on Earth do these components work?” It takes about a thousand pages to begin to answer that question properly, but here we want to draw the reader’s attention to some basic
distinctions among the various ways that the components can represent the environment that the agent inhabits. Roughly speaking, we can place the representations along an axis of increasing complexity and expressive power—atomic, factored, and structured. To illustrate these ideas, it helps
to consider a particular agent component, such as the one that deals with “What my actions
do.” This component describes the changes that might occur in the environment as the result
(a) Atomic
oIIQ.
of taking an action, and Figure 2.16 provides schematic depictions of how those transitions might be represented.
mII.l
58
(b) Factored
() Structured
Figure 2.16 Three ways to represent states and the transitions between them. (a) Atomic
representation: a state (such as B or C) is a black box with no internal structure; (b) Factored representation: a state consists ofa vector of attribute values; values can be Boolean, real-
valued, or one of a fixed set of symbols. (c) Structured representation: a state includes objects, each of which may have attributes of its own as well s relationships to other objects.
Section 2.4 The Structure of Agents In an atomic representation each state of the world is indivisible—it has no internal
structure. Consider the task of finding a driving route from one end of a country to the other
59
Atomic representation
via some sequence of cities (we address this problem in Figure 3.1 on page 64). For the purposes of solving this problem, it may suffice to reduce the state of the world to just the name of the city we are in—a single atom of knowledge, a “black box whose only discernible property is that of being identical to or different from another black box. The standard algorithms
underlying search and game-playing (Chapters 3-5), hidden Markov models (Chapter 14), and Markov decision processes (Chapter 17) all work with atomic representations.
Factored representation each of which can have a value. Consider a higher-fidelity description for the same driving Variable problem, where we need to be concerned with more than just atomic location in one city or Attribute another; we might need to pay attention to how much gas is in the tank, our current GPS Value A factored representation splits up each state into a fixed set of variables or attributes,
coordinates, whether or not the oil warning light is working, how much money we have for
tolls, what station is on the radio, and so on. While two different atomic states have nothing in
common—they are just different black boxes—two different factored states can share some
attributes (such as being at some particular GPS location) and not others (such as having lots
of gas or having no gas); this makes it much easier to work out how to turn one state into an-
other. Many important areas of Al are based on factored representations, including constraint satisfaction algorithms (Chapter 6), propositional logic (Chapter 7), planning (Chapter 11), Bayesian networks (Chapters 12-16), and various machine learning algorithms. For many purposes, we need to understand the world as having things in it that are related to each other, not just variables with values. For example, we might notice that a large truck ahead of us is reversing into the driveway of a dairy farm, but a loose cow is blocking the truck’s path. A factored representation is unlikely to be pre-equipped with the at-
tribute TruckAheadBackingIntoDairyFarmDrivewayBlockedByLooseCow with value true or
Jfalse. Tnstead, we would need a structured representation, in which objects such as cows Structured representation and trucks and their various and varying relationships can be described explicitly (see Figure 2.16(c)).
Structured representations underlie relational databases and first-order logic
(Chapters 8, 9, and 10), first-order probability models (Chapter 15), and much of natural lan-
guage understanding (Chapters 23 and 24). In fact, much of what humans express in natural language concerns objects and their relationships.
As we mentioned earlier, the axis along which atomic, factored, and structured repre-
sentations lie is the axis of increasing expressiveness. Roughly speaking, a more expressive
representation can capture, at least as concisely, everything a less expressive one can capture,
Expressiveness
plus some more. Ofien, the more expressive language is much more concise; for example, the rules of chess can be written in a page or two of a structured-representation language such
as first-order logic but require thousands of pages when written in a factored-representation
language such as propositional logic and around 10*® pages when written in an atomic language such as that of finite-state automata. On the other hand, reasoning and learning become
more complex as the expressive power of the representation increases. To gain the benefits
of expressive representations while avoiding their drawbacks, intelligent systems for the real world may need to operate at all points along the axis simultaneously.
Another axis for representation involves the mapping of concepts to locations in physical
memory,
whether in a computer or in a brain.
If there is a one-to-one mapping between
concepts and memory locations, we call that a localist representation. On the other hand, Localist representation
Chapter 2 Intelligent Agents if the representation of a concept is spread over many memory locations, and each memory Distributed representation
location is employed as part of the representation of multiple different concepts, we call
that a distributed representation. Distributed representations are more robust against noise and information loss. With a localist representation, the mapping from concept to memory location is arbitrary, and if a transmission error garbles a few bits, we might confuse Truck
with the unrelated concept Truce. But with a
distributed representation, you can think of each
concept representing a point in multidimensional space, and if you garble a few bits you move
to a nearby point in that space, which will have similar meaning.
Summary
This chapter has been something of a whirlwind tour of AL which we have conceived of as the science of agent design. The major points to recall are as follows: « An agent is something that perceives and acts in an environment. The agent function for an agent specifies the action taken by the agent in response to any percept sequence.
+ The performance measure evaluates the behavior of the agent in an environment.
A
rational agent acts so as to maximize the expected value of the performance measure, given the percept sequence it has seen so far.
+ A task environment specification includes the performance measure, the external en-
vironment, the actuators, and the sensors. In designing an agent, the first step must always be to specify the task environment as fully as possible.
« Task environments vary along several significant dimensions. They can be fully or partially observable, single-agent or multiagent, deterministic or nondeterministic, episodic or sequential, static or dynamic, discrete or continuous, and known or unknown.
« In cases where the performance measure is unknown or hard to specify correctly, there
is a significant risk of the agent optimizing the wrong objective. In such cases the agent design should reflect uncertainty about the true objective. + The agent program implements the agent function. There exists a variety of basic
agent program designs reflecting the kind of information made explicit and used in the decision process. The designs vary in efficiency, compactness, and flexibility. The appropriate design of the agent program depends on the nature of the environment.
+ Simple reflex agents respond directly to percepts, whereas model-based reflex agents maintain internal state to track aspects of the world that are not evident in the current
percept. Goal-based agents act to achieve their goals, and utility-based agents try to
maximize their own expected “happiness.”
« All agents can improve their performance through learning.
Bibliographical and Historical Notes The central role of action in intelligence—the notion of practical reasoning—goes back at least as far as Aristotle’s Nicomachean Ethics.
Practical reasoning was also the subject of
McCarthy’s influential paper “Programs with Common Sense” (1958). The fields of robotics and control theory are, by their very nature, concerned principally with physical agents. The
Bibliographical and Historical Notes
61
concept of a controller in control theory is identical to that of an agent in AL Perhaps sur- Controller prisingly, Al has concentrated for most of its history on isolated components of agents—
question-answering systems, theorem-provers, vision systems, and so on—rather than on
whole agents. The discussion of agents in the text by Genesereth and Nilsson (1987) was an
influential exception. The whole-agent view is now widely accepted and is a central theme in
recent texts (Padgham and Winikoff, 2004; Jones, 2007; Poole and Mackworth, 2017).
Chapter 1 traced the roots of the concept of rationality in philosophy and economics. In
AL the concept was of peripheral interest until the mid-1980s, when it began to suffuse many discussions about the proper technical foundations of the field. A paper by Jon Doyle (1983)
predicted that rational agent design would come to be seen as the core mission of Al, while
other popular topics would spin off to form new disciplines. Careful attention to the properties of the environment and their consequences for rational agent design is most apparent in the control theory tradition—for example, cla:
control systems (Dorf and Bishop, 2004; Kirk, 2004) handle fully observable, deterministic environments; stochastic optimal control (Kumar and Varaiya, 1986; Bertsekas and Shreve,
2007) handles partially observable, stochastic environments; and hybrid control (Henzinger and Sastry, 1998; Cassandras and Lygeros, 2006) deals with environments containing both
discrete and continuous elements. The distinction between fully and partially observable en-
vironments is also central in the dynamic programming literature developed in the field of operations research (Puterman, 1994), which we discuss in Chapter 17.
Although simple reflex agents were central to behaviorist psychology (see Chapter 1),
most Al researchers view them as too simple to provide much leverage. (Rosenschein (1985)
and Brooks (1986) questioned this assumption; see Chapter 26.) A great deal of work has gone into finding efficient algorithms for keeping track of complex environments (Bar-
Shalom et al., 2001; Choset et al., 2005; Simon, 2006), most of it in the probabilistic setting. Goal-based agents are presupposed in everything from Aristotle’s view of practical rea-
soning to McCarthy’s early papers on logical Al Shakey the Robot (Fikes and Nilsson, 1971; Nilsson,
1984) was the first robotic embodiment of a logical, goal-based agent.
A
full logical analysis of goal-based agents appeared in Genesereth and Nilsson (1987), and a goal-based programming methodology called agent-oriented programming was developed by Shoham (1993). The agent-based approach is now extremely popular in software engineering (Ciancarini and Wooldridge, 2001). It has also infiltrated the area of operating systems, ‘where autonomic computing refers to computer systems and networks that monitor and con-
trol themselves with a perceive—act loop and machine learning methods (Kephart and Chess, 2003). Noting that a collection of agent programs designed to work well together in a true
multiagent environment necessarily exhibits modularity—the programs share no internal state
and communicate with each other only through the environment—it
is common
within the
field of multiagent systems to design the agent program of a single agent as a collection of autonomous sub-agents.
In some cases, one can even prove that the resulting system gives
the same optimal solutions as a monolithic design.
The goal-based view of agents also dominates the cognitive psychology tradition in the area of problem solving, beginning with the enormously influential Human Problem Solving (Newell and Simon, 1972) and running through all of Newell’s later work (Newell, 1990).
Goals, further analyzed as desires (general) and intentions (currently pursued), are central to the influential theory of agents developed by Michael Bratman (1987).
Autonomic computing
62
Chapter 2 Intelligent Agents As noted in Chapter 1, the development of utility theory as a basis for rational behavior goes back hundreds of years. In Al early research eschewed utilities in favor of goals, with some exceptions (Feldman and Sproull, 1977). The resurgence of interest in probabilistic
methods in the 1980s led to the acceptance of maximization of expected utility as the most general framework for decision making (Horvitz et al., 1988). The text by Pearl (1988) was
the first in Al to cover probability and utility theory in depth; its exposition of practical methods for reasoning and decision making under uncertainty was probably the single biggest factor in the rapid shift towards utility-based agents
in the 1990s (see Chapter 16). The for-
malization of reinforcement learning within a decision-theoretic framework also contributed
to this
shift (Sutton, 1988). Somewhat remarkably, almost all Al research until very recently
has assumed that the performance measure can be exactly and correctly specified in the form
of a utility function or reward function (Hadfield-Menell e al., 2017a; Russell, 2019).
The general design for learning agents portrayed in Figure 2.15 is classic in the machine
learning literature (Buchanan et al., 1978; Mitchell, 1997). Examples of the design, as em-
bodied in programs, go back at least as far as Arthur Samuel’s (1959, 1967) learning program for playing checkers. Learning agents are discussed in depth in Chapters 19-22. Some early papers on agent-based approaches are collected by Huhns and Singh (1998) and Wooldridge and Rao (1999). Texts on multiagent systems provide a good introduction to many aspects of agent design (Weiss, 2000a; Wooldridge, 2009). Several conference series devoted to agents began in the 1990s, including the International Workshop on Agent Theories, Architectures, and Languages (ATAL), the International Conference on Autonomous Agents (AGENTS), and the International Conference on Multi-Agent Systems (ICMAS). In
2002, these three merged to form the International Joint Conference on Autonomous Agents and Multi-Agent Systems (AAMAS).
From 2000 to 2012 there were annual workshops on
Agent-Oriented Software Engineering (AOSE). The journal Autonomous Agents and MultiAgent Systems was founded in 1998. Finally, Dung Beetle Ecology (Hanski and Cambefort, 1991) provides a wealth of interesting information on the behavior of dung beetles. YouTube
has inspiring video recordings of their activities.
TR 3
SOLVING PROBLEMS BY SEARCHING In which we see how an agent can look ahead to find a sequence of actions that will eventually achieve its goal.
‘When the correct action to take is not immediately obvious, an agent may need to to plan
ahead: 10 consider a sequence of actions that form a path to a goal state. Such an agent is called a problem-solving agent, and the computational process it undertakes is called search.
Problem-solving agents use atomic representations, as described in Section 2.4.7—that
is, states of the world are considered as wholes, with no internal structure visible to the
~5/ebiemsolvine
Search
problem-solving algorithms. Agents that use factored or structured representations of states are called planning agents and are discussed in Chapters 7 and 11.
We will cover several search algorithms. In this chapter, we consider only the simplest
environments: known.
episodic, single agent, fully observable, deterministic,
static, discrete, and
We distinguish between informed algorithms, in which the agent can estimate how
far it is from the goal, and uninformed algorithms, where no such estimate is available. Chapter 4 relaxes the constraints on environments, and Chapter 5 considers multiple agents. This chapter uses the concepts of asymptotic complexity (that is, O(n) notation). Readers
unfamiliar with these concepts should consult Appendix A. 3.1
Problem-Solving Agents
Imagine an agent enjoying a touring vacation in Romania. The agent wants to take in the sights, improve its Romanian, enjoy the nightlife, avoid hangovers, and so on. The decision problem is a complex one. Now, suppose the agent is currently in the city of Arad and has a nonrefundable ticket to fly out of Bucharest the following day.
The agent observes
street signs and sees that there are three roads leading out of Arad: one toward Sibiu, one to
Timisoara, and one to Zerind. None of these are the goal, so unless the agent is familiar with
the geography of Romania, it will not know which road to follow."
If the agent has no additional information—that is, if the environment is unknown—then
the agent can do no better than to execute one of the actions at random.
This
sad situation
is discussed in Chapter 4. In this chapter, we will assume our agents always have access to information about the world, such as the map in Figure 3.1. With that information, the agent
can follow this four-phase problem-solving process:
+ Goal formulation: The agent adopts the goal of reaching Bucharest.
Goals organize
behavior by limiting the objectives and hence the actions to be considered.
‘We are assuming that most readers are in the same position and can easily imagine themselves to be as clueles as our agent. We apologize to Romanian readers who are unable to take advantage of this pedagogic: device.
Goal formulation
Chapter 3 Solving Problems by Searching
75 Arad
Sibiu
0
Fagaras
18
Vashui Rimnicu Vileea
Timisoara
Urziceni
75 Drobeta
Bucharest 90
riu Figure 3.1 A simplified road map of part of Romania, with road distances in miles. Craiova
Problem formulation
Hirsova
Eforie
+ Problem formulation: The agent devises a description of the states and actions necessary to reach the goal—an abstract model of the relevant part of the world. For our
agent, one good model is to consider the actions of traveling from one city to an adja-
cent city, and therefore the only fact about the state of the world that will change due to
an action is the current city.
Search
+ Search:
Before taking any action in the real world, the agent simulates sequences of
actions in its model, searching until it finds a sequence of actions that reaches the goal. Such a sequence is called a solution. The agent might have to simulate multiple sequences that do not reach the goal, but eventually it will find a solution (such as going
Solution
from Arad to Sibiu to Fagaras to Bucharest), or it will find that no solution is possible.
Execution
|
+ Execution: The agent can now execute the actions in the solution, one at a time.
It is an important property that in a fully observable, deterministic, known environment, the
solution to any problem is a fixed sequence of actions: drive to Sibiu, then Fagaras, then
Bucharest. If the model is correct, then once the agent has found a solution, it can ignore its
Open-loop
Closed-loop
percepts while it is executing the actions—closing its eyes, so to speak—because the solution
is guaranteed to lead to the goal. Control theorists call this an open-loop system: ignoring the
percepts breaks the loop between agent and environment. If there is a chance that the model
is incorrect, or the environment is nondeterministic, then the agent would be safer using a
closed-loop approach that monitors the percepts (see Section 4.4).
In partially observable or nondeterministic environments, a solution would be a branching
strategy that recommends different future actions depending on what percepts arrive. For
example, the agent might plan to drive from Arad to Sibiu but might need a contingency plan
in case it arrives in Zerind by accident or finds a sign saying “Drum inchis” (Road Closed).
Section 3.1 3.1.1
Problem-Solving Agents
Search problems and solutions
Problem
A search problem can be defined formally as follows: * A set of possible states that the environment can be in. We call this the state space. + The initial state that the agent starts in. For example: Arad.
+ A set of one or more goal states. Sometimes there is one goal state (e.g., Bucharest), sometimes there is a small set of alternative goal states, and sometimes the goal is
defined by a property that applies to many states (potentially an infinite number). For example, in a vacuum-cleaner world, the goal might be to have no dirt in any location, regardless of any other facts about the state.
States State space Initial state Goal states,
We can account for all three of these
possibilities by specifying an Is-GOAL method for a problem. In this chapter we will
sometimes say “the goal” for simplicity, but what we say also applies to “any one of the
possible goal states.”
« The actions available to the agent. Given a state s, ACTIONS(s) returns a finite? set of actions that can be executed in 5. We say that each of these actions is applicable in s.
An example:
Action Applicable
ACTIONS (Arad) = {ToSibiu, ToTimisoara, ToZerind } + A transition model, which describes what each action does.
RESULT(s, a) returns the
state that results from doing action a in state s. For example,
Transition model
RESULT(Arad, ToZerind) = Zerind.
« Anaction cost function, denoted by ACTION-COST(s,a,s") when we are programming or ¢(s,a,s') when we are doing math, that gives the numeric cost of applying action a in state s to reach state s'.
Action cost function
A problem-solving agent should use a cost function that
reflects its own performance measure; for example, for route-finding agents, the cost of
an action might be the length in miles (as seen in Figure 3.1), or it might be the time it
takes to complete the action. A sequence of actions forms a path, and a solution is a path from the initial state to a goal
state. We assume that action costs are additive; that is, the total cost of a path is the sum of the
Path
individual action costs. An optimal solution has the lowest path cost among all solutions. In Optimal solution
this chapter, we assume that all action costs will be positive, to avoid certain complications.®
The state space can be represented as a graph in which the vertices are states and the Graph directed edges between them are actions. The map of Romania shown in Figure 3.1 is such a graph, where each road indicates two actions, one in each direction. 2 For problems with an infinite number of actions we would need techniques that go beyond this chapter. 3 Inany problem with a cycle of net negative cost, the cost-optimal solution is to go around that cycle an infinite number of times. The Bellman—Ford and Floyd-Warshall algorithms (not covered here) handle negative-cost actions, as long as there are no negative cycles. It is easy to accommodate zero-cost actions, as long as the number of consecutive zero-cost actions is bounded. For example, we might have a robot where there is a cost 1o move, but zero cost to rotate 90°; the algorithms in this chapter can handle this as long as no more than three consecutive 90° turns are allowed. There is also a complication with problems that have an infinite number of arbitrarily small action costs. Consi ion of Zeno's paradox where there is an action to move half way to the goal, at a cost of half of the previous move. This problem has no solution with a finite number of actions, but to prevent a search from taking an unbounded number of actions without quite reaching the goal, we can require that all action costs be at least , for some small positive value c.
66
Chapter 3 Solving Problems by Searching 3.1.2
Formulating problems
Our formulation of the problem of getting to Bucharest is a model—an abstract mathematical
description—and not the real thing. Compare the simple atomic state description Arad to an actual cross-country trip, where the state of the world includes so many things: the traveling companions, the current radio program, the scenery out of the window, the proximity of law enforcement officers, the distance to the next rest stop, the condition of the road, the weather,
the traffic, and so on.
Abstraction
All these considerations are left out of our model because they are
irrelevant to the problem of finding a route to Bucharest.
The process of removing detail from a representation is called abstraction.
A good
problem formulation has the right level of detail. If the actions were at the level of “move the
right foot forward a centimeter” or “turn the steering wheel one degree left,” the agent would
Level of abstraction
probably never find its way out of the parking lot, let alone to Bucharest.
Can we be more precise about the appropriate level of abstraction? Think of the abstract
states and actions we have chosen as corresponding to large sets of detailed world states and
detailed action sequences.
Now consider a solution to the abstract problem:
for example,
the path from Arad to Sibiu to Rimnicu Vilcea to Pitesti to Bucharest. This abstract solution
corresponds to a large number of more detailed paths. For example, we could drive with the radio on between Sibiu and Rimnicu Vilcea, and then switch it off for the rest of the trip.
The abstraction is valid if we can elaborate any abstract solution into a solution in the
more detailed world; a sufficient condition
is that for every detailed state that is “in Arad,”
there is a detailed path to some state that is “in Sibiu,” and so on.* The abstraction is useful if
carrying out each of the actions in the solution is easier than the original problem; in our case,
the action “drive from Arad to Sibiu™ can be carried out without further search or planning by a driver with average skill. The choice ofa good abstraction thus involves removing as much
detail as possible while retaining validity and ensuring that the abstract actions are easy to
carry out. Were it not for the ability to construct useful abstractions, intelligent agents would be completely swamped by the real world. 3.2
Example
Problems
The problem-solving approach has been applied to a vast array of task environments. We list Standardized problem Real-world problem
some of the best known here, distinguishing between standardized and real-world problems.
A standardized problem is intended to illustrate or exercise various problem-solving meth-
ods.
It can be given a concise, exact description and hence is suitable as a benchmark for
researchers to compare the performance of algorithms. A real-world problem, such as robot
navigation, is one whose solutions people actually use, and whose formulation is idiosyn-
cratic, not standardized, because, for example, each robot has different sensors that produce different data.
3.2.1 Grid world
Standardized problems
A grid world problem is a two-dimensional rectangular array of square cells in which agents can move from cell to cell. Typically the agent can move to any obstacle-free adjacent cell— horizontally or vertically and in some problems diagonally. Cells can contain objects, which
Section 3.2 Example Problems
67
Figure 3.2 The state-space graph for the two-cell vacuum world. There are 8 states and three actions for each state: L = Lefr, R = Right, S = Suck.
the agent can pick up, push, or otherwise act upon; a wall or other impassible obstacle in a cell prevents an agent from moving into that cell. The vacuum world from Section 2.1 can be formulated as a grid world problem as follows: e States:
A state of the world says which objects are in which cells.
For the vacuum
world, the objects are the agent and any dirt. In the simple two-cell version, the agent
can be in either of the two cells, and each call can either contain dirt or not, so there are 2-2-2 = 8 states (see Figure 3.2). In general, a vacuum environment with n cells has
n-2" states.
o Initial state: Any state can be designated as the initial state. ® Actions: In the two-cell world we defined three actions: Suck, move Lefr, and move Right. Tn a two-dimensional multi-cell world we need more movement actions. We
could add Upward and Downward, giving us four absolute movement actions, or we could switch to egocentric actions, defined relative to the viewpoint of the agent—for example, Forward, Backward, TurnRight, and TurnLefi. o Transition model: Suck removes any dirt from the agent’s cell; Forward moves the agent ahead one cell in the direction it is facing, unless it hits a wall, in which case
the action has no effect.
Backward moves the agent in the opposite direction, while
TurnRight and TurnLeft change the direction it is facing by 90°. * Goal states: The states in which every cell is clean. ® Action cost: Each action costs 1.
Another type of grid world is the sokoban puzzle, in which the agent’s goal is to push a
number of boxes, scattered about the grid, to designated storage locations. There can be at
most one box per cell. When an agent moves forward into a cell containing a box and there is an empty cell on the other side of the box, then both the box and the agent move forward.
Sckoban puzzle
68
Chapter 3 Solving Problems by Searching
Start State
Goal State
Figure 3.3 A typical instance of the 8-puzzle. The agent can’t push a box into another box or a wall. For a world with 2 non-obstacle cells
and b boxes, there are n x n!/(b!(n — b)!) states; for example on an 8 x 8 grid with a dozen
Sliding-tile puzzle
boxes, there are over 200 trillion states.
In a sliding-tile puzzle, a number of tiles (sometimes called blocks or pieces) are ar-
ranged in a grid with one or more blank spaces so that some of the tiles can slide into the blank space.
8-puzzle 15-puzzle
One variant is the Rush Hour puzzle, in which cars and trucks slide around a
6% 6 grid in an attempt to free a car from the traffic jam. Perhaps the best-known variant is the 8-puzzle (see Figure 3.3), which consists of a 3 x 3 grid with eight numbered tiles and
one blank space, and the 15-puzzle on a 4 x 4 grid. The object is to reach a specified goal state, such as the one shown on the right of the figure.
puzzle is as follows:
The standard formulation of the 8
o States: A state description specifies the location of each of the tiles.
o Initial state: Any state can be designated as the initial state. Note that a parity prop-
erty partitions the state space—any given goal can be reached from exactly half of the possible initial states (see Exercise 3.PART).
e Actions: While in the physical world it is a tile that slides, the simplest way of describ-
ing an action is to think of the blank space moving Left, Right, Up, or Down. If the blank is at an edge or corner then not all actions will be applicable.
« Transition model: Maps a state and action to a resulting state; for example, if we apply Left to the start state in Figure 3.3, the resulting state has the 5 and the blank switched.
Goal state: Although any state could be the goal, we typically specify a state with the numbers in order, as in Figure 3.3.
® Action cost: Each action costs 1.
Note that every problem formulation involves abstractions. The 8-puzzle actions are ab-
stracted to their beginning and final states, ignoring the intermediate locations where the tile
is sliding. We have abstracted away actions such as shaking the board when tiles get stuck
and ruled out extracting the tiles with a knife and putting them back again. We are left with a
description of the rules, avoiding all the details of physical manipulations.
Our final standardized problem was devised by Donald Knuth (1964) and illustrates how
infinite state spaces can arise. Knuth conjectured that starting with the number 4, a sequence
Section 3.2 Example Problems
of square root, floor, and factorial operations can reach any desired positive integer. For example, we can reach 5 from 4 as follows:
|
@y =s
The problem definition is simple: States: Positive real numbers. o Initial state: 4.
o Actions: Apply square root, floor, or factorial operation (factorial for integers only). o Transition model: As given by the mathematical definitions of the operations.
o Goal state: The desired positive integer. ® Action cost: Each action costs 1.
The state space for this problem is infinite: for any integer greater than 2 the factorial oper-
ator will always yield a larger integer. The problem is interesting because it explores very large numbers: the shortest path to 5 goes through (4!)! = 620,448,401,733,239,439,360,000.
Infinite state spaces arise frequently in tasks involving the generation of mathematical expressions, circuits, proofs, programs, and other recursively defined objects. 3.2.2
Real-world problems
We have already seen how the route-finding problem is defined in terms of specified locations and transitions along edges between them. Route-finding algorithms are used in a variety of applications.
Some, such as Web sites and in-car systems that provide driving
directions, are relatively straightforward extensions
of the Romania example.
(The main
complications are varying costs due to traffic-dependent delays, and rerouting due to road closures.) Others, such as routing video streams in computer networks, military operations planning, and airline travel-planning systems, involve much more complex specifications. Consider the airline travel problems that must be solved by a travel-planning Web site: o States: Each state obviously includes a location (e.g., an airport) and the current time. Furthermore, because the cost of an action (a flight segment) may depend on previous segments, their fare bases, and their status as domestic or international, the state must record extra information about these “historical” aspects. o Initial state: The user’s home airport.
® Actions: Take any flight from the current location, in any seat class, leaving after the
current time, leaving enough time for within-airport transfer if needed.
o Transition model: The state resulting from taking a flight will have the flight’s destination as the new location and the flight’s arrival time as the new time.
o Goal state: A destination city. Sometimes the goal can be more complex, such as “arrive at the destination on a nonstop flight.”
* Action cost: A combination of monetary cost, waiting time, flight time, customs and immigration procedures, seat quality, time of day, type of airplane, frequent-flyer re-
ward points, and 50 on.
69
70
Chapter 3 Solving Problems by Searching Commercial travel advice systems use a problem formulation of this kind, with many addi-
tional complications to handle the airlines’ byzantine fare structures. Any seasoned traveler
Touring problem Traveling salesperson problem (TSP)
knows, however, that not all air travel goes according to plan. A really good system should include contingency plans—what happens if this flight is delayed and the connection is missed? Touring problems describe a set of locations that must be visited, rather than a single
goal destination. The traveling salesperson problem (TSP) is a touring problem in which every city on a map must be visited. optimization
The aim is to find a tour with cost < C (or in the
version, to find a tour with the lowest cost possible).
An enormous
amount
of effort has been expended to improve the capabilities of TSP algorithms. The algorithms
can also be extended to handle fleets of vehicles. For example, a search and optimization algorithm for routing school buses in Boston saved $5 million, cut traffic and air pollution, and saved time for drivers and students (Bertsimas ez al., 2019). In addition to planning trips,
search algorithms have been used for tasks such as planning the movements of automatic
VLSI layout
circuit-board drills and of stocking machines on shop floors.
A VLSI layout problem requires positioning millions of components and connections on a chip to minimize area, minimize circuit delays, minimize stray capacitances, and maximize manufacturing yield. The layout problem comes after the logical design phase and is usually split into two parts: cell layout and channel routing. In cell layout, the primitive components of the circuit are grouped into cells, each of which performs some recognized function. Each cell has a fixed footprint (size and shape) and requires a certain number of connections to each of the other cells. The aim is to place the cells on the chip so that they do not overlap and so that there is room for the connecting wires to be placed between the cells. Channel
routing finds a specific route for each wire through the gaps between the cells. These search
Robot navigation
problems are extremely complex, but definitely worth solving.
Robot navigation is a generalization of the route-finding problem described earlier. Rather than following distinct paths (such as the roads in Romania), a robot can roam around,
in effect making its own paths. For a circular robot moving on a flat surface, the space is
essentially two-dimensional. When the robot has arms and legs that must also be controlled,
the search space becomes many-dimensional—one dimension for each joint angle. Advanced
techniques are required just to make the essentially continuous search space finite (see Chap-
ter 26). In addition to the complexity of the problem, real robots must also deal with errors
in their sensor readings and motor controls, with partial observability, and with other agents
Automatic assembly sequencing
that might alter the environment.
Automatic assembly sequencing of complex objects (such as electric motors) by a robot
has been standard industry practice since the 1970s. Algorithms first find a feasible assembly
sequence and then work to optimize the process. Minimizing the amount of manual human
labor on the assembly line can produce significant savings in time and cost. In assembly problems, the aim is to find an order in which to assemble the parts of some object.
If the
wrong order is chosen, there will be no way to add some part later in the sequence without
undoing some of the work already done. Checking an action in the sequence for feasibility is a difficult geometrical search problem closely related to robot navigation. Thus, the generation Protein design
of legal actions is the expensive part of assembly sequencing. Any practical algorithm must avoid exploring all but a tiny fraction of the state space. One important assembly problem is
protein design, in which the goal is to find a sequence of amino acids that will fold into a three-dimensional protein with the right properties to cure some disease.
Section 33 Search Algorithms 3.3 Search Algorithms A search algorithm takes a search problem as input and returns a solution, or an indication of ~Search algorithm failure. In this chapter we consider algorithms that superimpose a search tree over the state-
space graph, forming various paths from the initial state, trying to find a path that reaches a goal state. Each node in the search tree corresponds to a state in the state space and the edges
in the search tree correspond to actions. The root of the tree corresponds to the initial state of
Node
the problem.
It is important to understand the distinction between the state space and the search tree.
The state space describes the (possibly infinite) set of states in the world, and the actions that allow transitions from one state to another.
The search tree describes paths between
these states, reaching towards the goal. The search tree may have multiple paths to (and thus
‘multiple nodes for) any given state, but each node in the tree has a unique path back to the root (as in all trees). Figure 3.4 shows the first few steps in finding a path from Arad to Bucharest. The root
node of the search tree is at the initial state, Arad. We can expand the node, by considering
Figure 3.4 Three partial search trees for finding a route from Arad to Bucharest. Nodes that have been expanded are lavender with bold letters; nodes on the frontier that have been generated but not yet expanded are in green; the set of states corresponding to these two
types of nodes are said to have been reached. Nodes that could be generated next are shown
in faint dashed lines. Notice in the bottom tree there is a cycle from Arad to Sibiu to Arad; that can’t be an optimal path, so search should not continue from there.
Expand
72
Chapter 3 Solving Problems by Searching
Figure 3.5 A sequence of search trees generated by a graph search on the Romania problem of Figure 3.1. At each stage, we have expanded every node on the frontier, extending every path with all applicable actions that don’t result in a state that has already been reached. Notice that at the third stage, the topmost city (Oradea) has two successors, both of which have already been reached by other paths, so no paths are extended from Oradea.
(a)
(b)
©
Figure 3.6 The separation property of graph search, illustrated on a rectangular-grid problem. The frontier (green) separates the interior (lavender) from the exterior (faint dashed).
The frontier is the set of nodes (and corresponding states) that have been reached but not yet expanded; the interior s the set of nodes (and corresponding states) that have been expanded; and the exterior is the set of states that have not been reached. In (a), just the root has been
expanded. In (b), the top frontier node is expanded. In (c), the remaining successors of the root are expanded in clockwise order.
Generating Child node Successor node
the available ACTIONS
for that state, using the RESULT function to see where those actions
lead to, and generating a new node (called a child node or successor node) for each of the resulting states. Each child node has Arad as its parent node. Now we must choose which of these three child nodes to consider next.
This is the
Parent node
essence of search—following up one option now and putting the others aside for later. Sup-
Frontier Reached
panded nodes (outlined in bold). We call this the frontier of the search tree. We say that any state that has had a node generated for it has been reached (whether or not that node has been
Separator
pose we choose to expand Sibiu first. Figure 3.4 (bottom) shows the result: a set of 6 unex-
expanded). Figure 3.5 shows the search tree superimposed on the state-space graph.
Note that the frontier separates two regions of the state-space graph: an interior region
where every state has been expanded, and an exterior region of states that have not yet been
reached. This property is illustrated in Figure 3.6. 5 Some authors call the frontier the open list, which s both geographically less evocative and computationally less appropriate, because a queue is more efficient than a list here. Those authors use the term elosed list to refer 10 the set of previously expanded nodes, which in our terminology would be the reached nodes minus the frontier.
Section 33 Search Algorithms
73
function BEST-FIRST-SEARCH(problem, ) returns a solution node or failure node « NODE(STATE=problem.INITIAL)
frontier ® B D
c
G F G E D F G E Figure 3.8 Breadth-first scarch on a simple binary tree. At each stage, the node to be expanded next s indicated by the triangular marker.
Section 3.4
Uninformed Search Strategies
77
function BREADTH-FIRST-SEARCH(problem) returns a solution node or failure
node < NODE(problem.INITIAL) if problem.1s-GOAL(node. STATE) then return node frontier 4+ b = 0(b%) All the nodes remain in memory, so both time and space complexity are O(b). Exponential bounds like that are scary. As a typical real-world example, consider a problem with branching factor b = 10, processing speed 1 million nodes/second, and memory requirements of 1 A search to depth d = 10 would take less than 3 hours, but would require 10
terabytes of memory. The memory requirements are a bigger problem for breadth-first search than the execution time.
But time is still an important factor.
At depth d = 14, even with
infinite memory, the search would take 3.5 years. In general, exponential-complexity search problems cannot be solved by uninformed search for any but the smallest instances. 3.4.2
A A
Kbyte/node.
Dijkstra’s algorithm or uniform-cost search
‘When actions have different costs, an obvious choice is to use best-first search where the
evaluation function is the cost of the path from the root to the current node. This is called Di-
jkstra’s algorithm by the theoretical computer science community, and uniform-cost search
by the AT community. The idea is that while breadth-first search spreads out in waves of uni-
form depth—first depth 1, then depth 2, and so on—uniform-cost search spreads out in waves of uniform path-cost. The algorithm can be implemented as a call to BEST-FIRST-SEARCH with PATH-COST as the evaluation function, as shown in Figure 3.9.
Uniform-cost search
78
Chapter 3 Solving Problems by Searching
Bucharest Figure 3.10 Part of the Romania state space, selected to illustrate uniform-cost search.
Consider Figure 3.10, where the problem is to get from Sibiu to Bucharest. The succes-
sors of Sibiu are Rimnicu Vilcea and Fagaras, with costs 80 and 99, respectively. The least-
cost node, Rimnicu Vilcea, is expanded next, adding Pitesti with cost 80 +97=177.
The
least-cost node is now Fagaras, so it is expanded, adding Bucharest with cost 99 +211=310.
Bucharest is the goal, but the algorithm tests for goals only when it expands a node, not when it generates a node, so it has not yet detected that this is a path to the goal.
The algorithm continues on, choosing Pitesti for expansion next and adding a second path
to Bucharest with cost 80 +97 + 101 =278.
path in reached and is added to the frontier.
It has a lower cost, so it replaces the previous
It turns out this node now has the lowest cost,
50 it is considered next, found to be a goal, and returned. Note that if we had checked for a
goal upon generating a node rather than when expanding the lowest-cost node, then we would have returned a higher-cost path (the one through Fagaras).
The complexity of uniform-cost search is characterized in terms of C*, the cost of the
optimal solution,® and ¢, a lower bound on the cost of each action, with ¢ > 0.
Then the
algorithm’s worst-case time and space complexity is O(b'*1€"/ ¢>0;? cost-optimal if action costs are all identical; * if both directions are breadth-first
or uniform-cost. 3.5
Informed (Heuristic) Search Strategies
Informed search
This section shows how an informed search strategy—one that uses domain-specific hints
Heuristic function
The hints come in the form of a heuristic function, denoted /(n):'*
about the location of goals—can find solutions more efficiently than an uninformed strategy. h(n) = estimated cost of the cheapest path from the state at node n to a goal state.
For example, in route-finding problems, we can estimate the distance from the current state to a goal by computing the straight-line distance on the map between the two points. We study heuristics
and where they come from in more detail in Section 3.6.
10 It may seem odd that the heuristic function operates on a node, when all it really needs is the node’s state. It is traditional t0 use /(n) rather than h(s) to be consistent with the evaluation function J (n) and the path cost g(n).
Section 3.5 Arad Bucharest Craiova
366 0 160 242 161 176 77 151 226 244
Informed (Heuristic) Search Strategies
Mehadia Neamt Oradea Pitesti Rimnicu Vilcea Sibiu Timisoara Urziceni Vaslui Zerind
241 234 380 100 193 253 329 80 199 374
Figure 3.16 Values of is;p—straight-line distances to Bucharest. 3.5.1
Greedy best-first search
Greedy best-first search is a form of best-first search that expands first the node with the §eedy best-irst lowest /(n) value—the node that appears to be closest to the goal—on the grounds that this is likely to lead to a solution quickly. So the evaluation function f(n) = h(n).
Let us see how this works for route-finding problems in Romania; we use the straight-
line distance heuristic, which we will call hg;p. If the goal is Bucharest, we need to know the straight-line distances to Bucharest, which are shown in Figure 3.16. For example, hsip(Arad)=366. Notice that the values of hg;p cannot be computed from the problem description itself (that is, the ACTIONS
and RESULT functions).
Straight-ine distance
Moreover, it takes a certain
amount of world knowledge to know that g p is correlated with actual road distances and is, therefore, a useful heuristic.
Figure 3.17 shows the progress of a greedy best-first search using hg;p to find a path
from Arad to Bucharest.
The first node to be expanded from Arad will be Sibiu because the
heuristic says it is closer to Bucharest than is either Zerind or Timisoara. The next node to be
expanded will be Fagaras because it is now closest according to the heuristic. Fagaras in turn generates Bucharest, which is the goal. For this
particular problem, greedy best-first search
using hg;p finds a solution without ever expanding a node that is not on the solution path.
The solution it found does not have optimal cost, however: the path via Sibiu and Fagaras to Bucharest is 32 miles longer than the path through Rimnicu Vilcea and Pitesti. This is why
the algorithm is called “greedy”—on each iteration it tries to get as close to a goal as it can, but greediness can lead to worse results than being careful.
Greedy best-first graph search is complete in finite state spaces, but not in infinite ones. The worst-case time and space complexity is O(|V'|). With a good heuristic function, however, the complexity can be reduced substantially, on certain problems reaching O(bm). 3.5.2
A" search
The most common
informed search algorithm is A* search (pronounced “A-star search”), a
best-first search that uses the evaluation function
F(n) = g(n) +h(n) where g(n) is the path cost from the initial state to node n, and h(n) is the estimated cost of the shortest path from 7 to a goal state, so we have f(n) = estimated cost of the best path that continues from 7 to a goal.
A" search
86
Chapter 3 Solving Problems by Searching (a) The initial state
366
(b) After expanding Arad
Chmd>
b >
isoad>
253
CZerind>
329
374
(¢) After expanding Sibiu
(d) After expanding Fagaras
Figure 3.17 Stages in a greedy best-first tree-like search for Bucharest with the straight-line distance heuristic /iszp. Nodes are labeled with their A-values. In Figure 3.18, we show the progress of an A* search with the goal of reaching Bucharest.
The values of g are computed from the action costs in Figure 3.1, and the values of hgs.p are
given in Figure 3.16. Notice that Bucharest first appears on the frontier at step (e), but it is not selected for expansion (and thus not detected as a solution) because at f =450 it is not the lowest-cost node on the frontier—that would be Pitesti, at f=417.
Another way to say this
is that there might be a solution through Pitesti whose cost is as low as 417, so the algorithm will not settle for a solution that costs
450.
At step (f), a different path to Bucharest is now
the lowest-cost node, at f =418, so it is selected and detected as the optimal solution.
Admissible heuristic
A search is complete.!! Whether A* is cost-optimal depends on certain properties of
the heuristic.
A key property is admissibil
an admissible heuristic is one that never
overestimates the cost to reach a goal. (An admissible heuristic is therefore optimistic.) With
11 Again, assuming all action costs are > ¢ > 0, and the state space either has a solution or is finite.
Section 3.5
(a) The initial state
Informed (Heuristic) Search Strategies
366-0+366
(b) After expanding Arad 449754374
646=280+366 415-239+176 671-291+380 413-220+193
(d) After expanding Rimnicu Vilcea
SOI=3384253 4S0450+0
S26=366+160 4173174100 §53-300-253
(f) After expanding Pitesti 449754374
41841840 615-455+160 607=414+193
Figure 3.18 Stages in an A" search for Bucharest. Nodes are labeled with f = g+ /. The h values are the straight-line distances to Bucharest taken from Figure 3.16.
87
88
Chapter 3 Solving Problems by Searching
Figure 3.19 Triangle inequality: If the heuristic / is consistent, then the single number k()
will be less than the sum of the cost ¢(n,a,d’) of the action from n to n plus the heuristic
estimate h(n').
an admissible heuristic, A* is cost-optimal, which we can show with a proof by contradiction. Suppose the optimal path has cost C*, but the algorithm returns a path with cost C > C*. Then there must be some node 7 which is on the optimal path and is unexpanded (because if all the nodes on the optimal path had been expanded, then we would have returned that optimal solution). So then, using the notation g*(n) to mean the cost of the optimal path from the start to n, and h*(n) to mean the cost of the optimal path from # to the nearest goal, we have:
f(n) > € (otherwise n would have been expanded) f(n) = g(n)+h(n) f(n)
= g"(n)+h(n)
(by definition)
(because n is on an optimal path)
f(n) < g"(n)+"(n) (because of admissibility, h(n) < h*(n)) f(n) < € (by definition,C* = g*(n) + h*(n)) The first and last lines form a contradiction, so the supposition that the algorithm could return
Consistency
a suboptimal path must be wrong—it must be that A* returns only cost-optimal paths.
A slightly stronger property is called consistency.
A heuristic i(n) is consistent if, for
every node n and every successor n’ of n generated by an action a, we have:
h(n) < e(nan’)+h(n').
Triangle inequality
This is a form of the triangle inequality, which stipulates that a side of a triangle cannot
be longer than the sum of the other two sides (see Figure 3.19). An example of a consistent
heuristic is the straight-line distance /s> that we used in getting to Bucharest.
Every consistent heuristic is admissible (but not vice versa), so with a consistent heuristic, A" is cost-optimal. In addition, with a consistent heuristic, the first time we reach a state it will be on an optimal path, so we never have to re-add a state to the frontier, and never have to
change an entry in reached. But with an inconsistent heuristic, we may end up with multiple paths reaching the same state, and if each new path has a lower path cost than the previous
one, then we will end up with multiple nodes for that state in the frontier, costing us both
time and space. Because of that, some implementations of A* take care to only enter a state into the frontier once, and if a better path to the state is found, all the successors of the state
are updated (which requires that nodes have child pointers as well as parent pointers). These complications have led many implementers to avoid inconsistent heuristics, but Felner et al. (2011) argues that the worst effects rarely happen in practice, and one shouldn’t be afraid of inconsistent heuristics.
Section 3.5
Informed (Heuristic) Search Strategies
89
Figure 3.20 Map of Romania showing contours at f = 380, f = 400, and f = 420, with Arad as the start state. Nodes inside a given contour have f = g+ costs less than or equal to the contour value.
With an inadmissible heuristic, A* may or may not be cost-optimal. Here are two cases where it is: First, if there is even one cost-optimal path on which h(n) is admissible for all nodes n on the path, then that path will be found, no matter what the heuristic says for states off the path. Second, if the optimal solution has cost C*, and the second-best has cost C,, and
if h(n) overestimates some costs, but never by more than C; — C*, then A is guaranteed to
return cost-optimal solutions.
3.5.3 Search contours A useful way to visualize a search is to draw contours in the state space, just like the contours
in a topographic map. Figure 3.20 shows an example. Inside the contour labeled 400, all nodes have f(n) = g(n) +h(n) < 400, and so on. Then, because A" expands the frontier node
Contour
of lowest f-cost, we can see that an A” search fans out from the start node, adding nodes in concentric bands of increasing f-cost. ‘With uniform-cost search, we also have contours, but of g-cost, not g + . The contours
with uniform-cost search will be “circular” around the start state, spreading out equally in all
directions with no preference towards the goal. With A* search using a good heuristic, the g+ h bands will stretch toward a goal state (as in Figure 3.20) and become more narrowly focused around an optimal path. It should be clear that as you extend a path, the g costs are monotonic:
the path cost
always increases as you go along a path, because action costs are always positive.'? Therefore
you get concentric contour lines that don’t cross each other, and if you choose to draw the
lines fine enough, you can put a line between any two nodes on any path.
12 Technically, we say decrease, but might rem:
ctly monotonic™ for costs that always increase, and “monotonic” for cost the same.
that never
Monotonic
Chapter 3 Solving Problems by Searching But it is not obvious whether the f = g+ h cost will monotonically increase. As you ex-
tend a path from 7 to 1, the cost goes from g(n) +h(n) to g(n) +c(n,a,n’) +h(n’). Canceling
out the g(n) term, we see that the path’s cost will be monotonically increasing if and only if h(n) < c(n,a,n") + h(n'); in other words if and only if the heuristic is consistent.'> But note that a path might contribute several nodes in a row with the same g(n) + (n) score; this will
happen whenever the decrease in / is exactly equal to the action cost just taken (for example,
in a grid problem, when n is in the same row as the goal and you take a step towards the goal,
g s increased by 1 and h is decreased by 1). If C* is the cost of the optimal solution path, then we can say the following:
Surely expanded nodes
+ A expands all nodes that can be reached from the initial state on a path where every
node on the path has f(n) < C*. We say these are surely expanded nodes. A" might then expand some of the nodes right on the “goal contour” (where f(n) = C*) before selecting a goal node.
« A" expands no nodes with f(n) > C*.
Optimally efficient
We say that A* with a consistent heuristic is optimally efficient in the sense that any algorithm
that extends search paths from the initial state, and uses the same heuristic information, must
expand all nodes that are surely expanded by A* (because any one of them could have been
part of an optimal solution). Among the nodes with f(n)=C", one algorithm could get lucky
and choose the optimal one first while another algorithm is unlucky; we don’t consider this
Pruning
difference in defining optimal efficiency.
A" is efficient because it prunes away search tree nodes that are not necessary for finding an optimal solution. Tn Figure 3.18(b) we see that Timisoara has f = 447 and Zerind has f 449. Even though they are children of the root and would be among the first nodes expanded by uniform-cost or breadth-first search, they are never expanded by A* search because the solution with f = 418 is found first. The concept of pruning—eliminating possibilities from consideration without having to examine them—is important for many areas of Al
That A* search is complete, cost-optimal, and optimally efficient among all such algo-
rithms is rather satisfying.
Unfortunately, it does not mean that A* is the answer to all our
searching needs. The catch is that for many problems, the number of nodes expanded can be exponential in the length of the solution. For example, consider a version of the vacuum world with a super-powerful vacuum that can clean up any one square at a cost of 1 unit, without even having to visit the square; in that scenario, squares can be cleaned in any order.
With N initially dirty squares, there are 2V states where some subset has been cleaned; all of those states are on an optimal solution path, and hence satisfy f(n) < C*, so all of them would be visited by A*.
3.5.4
Inadmissible heuristic
Satisficing search:
Inadmissible heuristics and weighted A*
A" search has many good qualities, but it expands a lot of nodes. We can explore fewer nodes (taking less time and space) if we are willing to accept solutions that are suboptimal, but are “good enough”—what we call satisficing solutions. If we allow A® search to use an inadmissible heuristic—one that may overestimate—then we risk missing the optimal solution, but the heuristic can potentially be more accurate, thereby reducing the number of 13 In fact, the term * Jonotonic heuristic” is a synonym for “consistent heurist . The two ideas were developed independently, and ther was proved that they are equivalent (Pearl, 1984).
Section 3.5
Informed (Heuristic) Search Strategies
91
(b)
(a)
Figure 3.21 Two searches on the same grid: (a) an A" search and (b) a weighted A" search with weight W = 2. The gray bars are obstacles, the purple line is the path from the green start to red goal, and the small dots are states that were reached by each search. On this particular problem, weighted A” explores 7 times fewer states and finds a path that is 5% more costly.
nodes expanded. For example, road engineers know the concept of a detour index, which is Detour index a multiplier applied to the straight-line distance to account for the typical curvature of roads. A detour index of 1.3 means that if two cities are 10 miles apart in straight-line distance, a good estimate of the best path between them is 13 miles. For most localities, the detour index
ranges between 1.2 and 1.6. We can apply this idea to any problem, not just ones involving roads, with an approach
called weighted A* search where we weight the heuristic value more heavily, giving us the
evaluation function £(n) = g(n) +W x h(n), for some W > 1. Figure 3.21 shows a search problem on a grid world. In (), an A" search finds the optimal solution, but has to explore a large portion of the state space to find it. In (b), a weighted A*
search finds a solution that is slightly costlier, but the search time is much faster. We see that the weighted search focuses the contour of reached states towards a goal.
That means that
fewer states are explored, but if the optimal path ever strays outside of the weighted search’s
contour (as it does in this case), then the optimal path will not be found. In general, if the optimal solution costs C*, a weighted A* search will find a solution that costs somewhere
between C* and W x C*; but in practice we usually get results much closer to C* than W x C*.
We have considered searches that evaluate states by combining g and / in various ways; weighted A* can be seen as a generalization of the others:
At search:
Uniform-cost search:
Greedy best-first search: Weighted A" search:
g(n)+h(n) 2(n) h(n)
g(n) +W x h(n)
(1 0 then current
next
else current —next only with probability e~2£/T
Figure 4.5 The simulated annealing algorithm, a version of stochastic hill climbing where some downhill moves are allowed. The schedule input determines the value of the “tempera-
ture” 7 as a function of time.
all the probability is concentrated on the global maxima, which the algorithm will find with probability approaching 1. Simulated annealing was used to solve VLSI layout problems beginning in the 1980s. Tt has been applied widely to factory scheduling and other large-scale optimization tasks. 4.1.3
Local beam search
Keeping just one node in memory might seem to be an extreme reaction to the problem of
memory limitations.
The local beam search algorithm keeps track of k states rather than
just one. Tt begins with k randomly generated states. At each step, all the successors of all k states are generated. If any one is a goal, the algorithm halts. Otherwise, it selects the k best
Local beam search
successors from the complete list and repeats. At first sight, a local beam search with k states might seem to be nothing more than
running k random restarts in parallel instead of in sequence. are quite different.
In fact, the two algorithms
In a random-restart search, each search process runs independently of
the others. In a local beam search, useful information is passed among the parallel search threads.
In effect, the states that generate the best successors say to the others, “Come over
2, which occurs
only rarely in nature but is easy enough to simulate on computers.
Selection
« The selection process for selecting the individuals who will become the parents of the next generation: one possibility is to select from all individuals with probability proportional to their fitness score. Another possibility is to randomly select n individuals (n > p), and then select the p most fit ones as parents.
Crossover point
« The recombination procedure. One common approach (assuming p = 2), is to randomly select a crossover point to split each of the parent strings, and recombine the parts to form two children, one with the first part of parent 1 and the second part of parent 2; the other with the second part of parent 1 and the first part of parent 2.
Mutation rate
+ The mutation rate, which determines how often offspring have random mutations to
their representation. Once an offspring has been generated, every bit in its composition is flipped with probability equal to the mutation rate.
« The makeup of the next generation. This can be just the newly formed offspring, or it Elitism
can include a few top-scoring parents from the previous generation (a practice called
elitism, which guarantees that overall fitness will never decrease over time). The practice of culling, in which all individuals below a given threshold are discarded, can lead
to a speedup (Baum e al., 1995). Figure 4.6(a) shows a population of four 8-digit strings, each representing a state of the 8-
queens puzzle: the c-th digit represents the row number of the queen in column c. In (b), each state is rated by the fitness function.
Higher fitness values are better, so for the 8-
Section 4.1
Local Search and Optimization Problems
17
Figure 4.7 The 8-queens states corresponding to the first two parents in Figure 4.6(c) and the first offspring in Figure 4.6(d). The green columns are lost in the crossover step and the red columns are retained. (To interpret the numbers in Figure 4.6: row 1 is the bottom row, and 8 is the top row.) queens problem we use the number of nonattacking pairs of queens, which has a value of 8 x 7/2 =28 for a solution. The values of the four states in (b) are 24, 23, 20, and 11. The
fitness scores are then normalized to probabilities, and the resulting values are shown next to
the fitness values in (b).
In (c), two pairs of parents are selected, in accordance with the probabilities in (b). Notice that one individual is selected twice and one not at all. For each selected pair, a crossover point (dotted line) is chosen randomly. In (d), we cross over the parent strings at the crossover
points, yielding new offspring. For example, the first child of the first pair gets the first three digits (327) from the first parent and the remaining digits (48552) from the second parent.
The 8-queens states involved in this recombination step are shown in Figure 4.7.
Finally, in (e), each location in each string is subject to random mutation with a small independent probability. One digit was mutated in the first, third, and fourth offspring. In the
8-queens problem, this corresponds to choosing a queen at random and moving it to a random
square in its column. It is often the case that the population is diverse early on in the process,
so crossover frequently takes large steps in the state space early in the search process (as in simulated annealing). After many generations of selection towards higher fitness, the popu-
lation becomes less diverse, and smaller steps are typical. Figure 4.8 describes an algorithm
that implements all these steps.
Genetic algorithms are similar to stochastic beam search, but with the addition of the
crossover operation. This is advantageous if there are blocks that perform useful functions.
For example, it could be that putting the first three queens in positions 2, 4, and 6 (where they do not attack each other) constitutes a useful block that can be combined with other useful blocks that appear in other individuals to construct a solution. It can be shown mathematically
that, if the blocks do not serve a purpose—for example if the positions of the genetic code
are randomly permuted—then crossover conveys no advantage.
The theory of genetic algorithms explains how this works using the idea of a schema,
which is a substring in which some of the positions can be left unspecified. For example, the schema 246#*###
Schema
describes all 8-queens states in which the first three queens are in
positions 2, 4, and 6, respectively. Strings that match the schema (such as 24613578) are
called instances of the schema. It can be shown that if the average fitness of the instances of a schema is above the mean, then the number of instances of the schema will grow over time.
Instance
118
Chapter 4 Search in Complex Environments Evolution and Search The theory of evolution was developed by Charles Darwin in On the Origin of Species by Means of Natural Selection (1859) and independently by Alfred Russel ‘Wallace (1858).
The central idea is simple:
variations occur in reproduction and
will be preserved in successive generations approximately in proportion to their effect on reproductive fitness.
Darwin’s theory was developed with no knowledge of how the traits of organisms can be inherited and modified. The probabilistic laws governing these processes were first identified by Gregor Mendel (1866), a monk who experimented
with sweet peas. Much later, Watson and Crick (1953) identified the structure of the
DNA molecule and its alphabet, AGTC (adenine, guanine, thymine, cytosine). In the standard model, variation occurs both by point mutations in the letter sequence and by “crossover” (in which the DNA of an offspring is generated by combining long sections of DNA from each parent). The analogy to local search algorithms has already been described; the prin-
cipal difference between stochastic beam search and evolution is the use of sexual
reproduction, wherein successors are generated from multiple individuals rather than just one.
The actual mechanisms of evolution are, however, far richer than
most genetic algorithms allow. For example, mutations can involve reversals, duplications, and movement of large chunks of DNA;
some viruses borrow DNA
from one organism and insert it into another; and there are transposable genes that
do nothing but copy themselves many thousands of times within the genome.
There are even genes that poison cells from potential mates that do not carry
the gene, thereby increasing their own chances of replication. Most important is the fact that the genes themselves encode the mechanisms whereby the genome is reproduced and translated into an organism. In genetic algorithms, those mechanisms
are a separate program that is not represented within the strings being manipulated. Darwinian evolution may appear inefficient, having generated blindly some 10" or so organisms without improving its search heuristics one iota. But learning does play a role in evolution.
Although the otherwise great French naturalist
Jean Lamarck (1809) was wrong to propose that traits ing an organism’s lifetime would be passed on to its (1896) superficially similar theory is correct: learning ness landscape, leading to an acceleration in the rate of
acquired by adaptation duroffspring, James Baldwin’s can effectively relax the fitevolution. An organism that
has a trait that is not quite adaptive for its environment will pass on the trait if it also
has enough plasticity to learn to adapt to the environment in a way that is beneficial. Computer simulations (Hinton and Nowlan, 1987) confirm that this Baldwin
effect is real, and that a consequence is that things that are hard to learn end up in the genome, but things that are easy to learn need not reside there (Morgan and
Griffiths, 2015).
Section 4.2
Local Search in Continuous Spaces
function GENETIC-ALGORITHM(population, fitness) returns an individual
repeat weights— WEIGHTED-BY (population, fitness) population2 +—empty list for i = 1 to SIZE(population) do
parent], parent2< WEIGHTED-R ANDOM-CHOICES (population, weights, 2) child REPRODUCE (parent ], parent2) if (small random probability) then child < MUTATE(child) add child to population2 population < population2 until some individual is fit enough, or enough time has elapsed return the best individual in population, according to fitness function REPRODUCE(parent], parent2) returns an individual
14 LENGTH(parent])
¢ +random number from 1 to n
return APPEND(SUBSTRING(parentl, 1,c), SUBSTRING (parent2, ¢ + 1,m))
Figure 4.8 A genetic algorithm. Within the function, population is an ordered list of individuals, weights is a list of corresponding fitness values for each individual, and fitess is a function to compute these values.
Clearly, this effect is unlikely to be significant if adjacent bits are totally unrelated to
each other, because then there will be few contiguous blocks that provide a consistent benefit. Genetic algorithms work best when schemas correspond to meaningful components of a
solution. For example, if the string is a representation of an antenna, then the schemas may represent components of the antenna, such as reflectors and deflectors. A good component is likely to be good in a variety of different designs. This suggests that successful use of genetic
algorithms requires careful engineering of the representation. In practice, genetic algorithms have their place within the broad landscape of optimization methods (Marler and Arora, 2004), particularly for complex structured problems such as circuit layout or job-shop scheduling, and more recently for evolving the architecture of deep neural networks (Miikkulainen ef al., 2019). It is not clear how much of the appeal of genetic algorithms arises from their superiority on specific tasks, and how much from the appealing metaphor of evolution.
Local Search in Continuous Spaces In Chapter 2, we explained the distinction between discrete and continuous environments,
pointing out that most real-world environments are continuous.
A continuous action space
has an infinite branching factor, and thus can’t be handled by most of the algorithms we have covered so far (with the exception of first-choice hill climbing and simulated annealing). This section provides a very brief introduction to some local search techniques for continuous spaces. The literature on this topic is vast; many of the basic techniques originated
119
120
Chapter 4 Search in Complex Environments in the 17th century, after the development of calculus by Newton and Leibniz? We find uses for these techniques in several places in this book, including the chapters on learning, vision, and robotics.
We begin with an example. Suppose we want to place three new airports anywhere in
Romania, such that the sum of squared straight-line distances from each city on the map
Variable
to its nearest airport is minimized. (See Figure 3.1 for the map of Romania.) The state space is then defined by the coordinates of the three airports: (xy,y;), (x2,y2), and (x3,y3).
This is a six-dimensional space; we also say that states are defined by six variables. In general, states are defined by an n-dimensional vector of variables, x. Moving around in this space corresponds to moving one or more of the airports on the map. The objective function f(X) = f(x1,y1.X2,y2,x3,y3) is relatively easy to compute for any particular state once we
compute the closest cities. Let C; be the set of cities whose closest airport (in the state x) is airport i. Then, we have
3
F) = fiynx e ys) = Y Z(X:*X‘)2+(Yi*n)2»
@
This equation is correct not only for the state x but also for states in the local neighborhood
of x. However, it is not correct globally; if we stray too far from x (by altering the location
of one or more of the airports by a large amount) then the set of closest cities for that airport
Discretization
changes, and we need to recompute C;.
One way to deal with a continuous state space is to discretize it. For example, instead of
allowing the (x;,y;) locations to be any point in continuous two-dimensional space, we could
limit them to fixed points on a rectangular grid with spacing of size § (delta). Then instead of
having an infinite number of successors, each state in the space would have only 12 successors, corresponding to incrementing one of the 6 variables by 4. We can then apply any of our local search algorithms to this discrete space. Alternatively, we could make the branching factor finite by sampling successor states randomly, moving in a random direction by a small
Empirical gradient
amount, 5. Methods that measure progress by the change in the value of the objective function between two nearby points are called empirical gradient methods.
Empirical gradient
search is the same as steepest-ascent hill climbing in a discretized version of the state space.
Reducing the value of § over time can give a more accurate solution, but does not necessarily
converge to a global optimum in the limit.
Often we have an objective function expressed in a mathematical form such that we can
Gradient
use calculus to solve the problem analytically rather than empirically. Many methods attempt
to use the gradient of the landscape to find a maximum. The gradient of the objective function
is a vector V£ that gives the magnitude and direction of the steepest slope. For our problem, we have
vio (9F
9f
9f
9f
9f
of
1= (S A 2 5)
In some cases, we can find a maximum by solving the equation V £ =0. (This could be done,
for example, if we were placing just one airport; the solution is the arithmetic mean of all the
cities’ coordinates.) In many cases, however, this equation cannot be solved in closed form.
For example, with three airports, the expression for the gradient depends on what cities are
Knowledge of vectors, mat ces, and derivatives s useful for this
(see Appendix A).
Section 4.2
Local Search in Continuous Spaces
121
closest to each airport in the current state. This means we can compute the gradient locally (but not globally); for example, of 22 ¥ (x—x0). “2) Given alocally correct expression for the gradient, we can perform steepest-ascent hill climbing by updating the current state according to the formula X
x+aVf(x),
where a (alpha) is a small constant often called the step size. There exist a huge variety Step size of methods for adjusting a.. The basic problem is that if « is too small, too many steps are needed; if a is too large, the search could overshoot the maximum.
The technique of line
search tries to overcome this dilemma by extending the current gradient direction—usually by repeatedly doubling a—until f starts to decrease again. becomes
the new current state.
The point at which this occurs
There are several schools of thought about how the new
direction should be chosen at this point. For many problems, the most effective algorithm is the venerable Newton-Raphson
method. This is a general technique for finding roots of functions—that is, solving equations
of the form g(x)=0. Newton’s formula
Line search
Newton-Raphson
It works by computing a new estimate for the root x according to
x e x—g(x)/g'(x).
To find a maximum or minimum of f, we need to find x such that the gradient is a zero vector
(i.e., V(x)=0). Thus, g(x) in Newton’s formula becomes V £(x), and the update equation can be written in matrix-vector form as
x4 x—H
(x)Vf(x),
where Hy(x) is the Hessian matrix of second derivatives, whose elements H;; are given by
92f/9x;dx;. For our airport example, we can see from Equation (4.2) that Hy(x) is particularly simple: the off-diagonal elements are zero and the diagonal elements for airport are just
Hessian
twice the number of cities in C;. A moment’s calculation shows that one step of the update
moves airport i directly to the centroid of C;, which is the minimum of the local expression
for f from Equation (4.1).3 For high-dimensional problems, however, computing the n? en-
tries of the Hessian and inverting it may be expensive, so many approximate versions of the
Newton-Raphson method have been developed.
Local search methods suffer from local maxima, ridges, and plateaus in continuous state
spaces just as much as in discrete spaces. Random restarts and simulated annealing are often helpful. High-dimensional continuous spaces are, however, big places in which it is very easy to get lost. Constrained A final topic is constrained optimization. An optimization problem is constrained if optimization solutions must satisfy some hard constraints on the values of the variables.
For example, in
our airport-siting problem, we might constrain sites to be inside Romania and on dry land (rather than in the middle of lakes).
The difficulty of constrained
optimization problems
depends on the nature of the constraints and the objective function. The best-known category is that of linear programming problems, in which constraints must be linear inequalities
3 In general, the Newton-Raphson update can be seen s fitting a quadratic surface to / at x and then moving direetly to the minimum of that surface—which is also the minimum of f if / is quadrati.
Linear programming
122
Convex set Convex optimization
Chapter 4 Search in Complex Environments forming a convex set* and the objective function is also linear. The time complexity of linear programming is polynomial in the number of variables. Linear programming is probably the most widely studied and broadly useful method for optimization. It is a special case of the more general problem of convex optimization, which
allows the constraint region to be any convex region and the objective to be any function that is
convex within the constraint region. Under certain conditions, convex optimization problems are also polynomially solvable and may be feasible in practice with thousands of variables. Several important problems in machine learning and control theory can be formulated as convex optimization problems (see Chapter 20). 4.3
Search with Nondeterministic Actions
In Chapter 3, we assumed a fully observable, deterministic, known environment.
Therefore,
an agent can observe the initial state, calculate a sequence of actions that reach the goal, and
execute the actions with its “eyes closed,” never having to use its percepts. When the environment is partially observable, however, the agent doesn’t know for sure what state it is in; and when the environment is nondeterministic, the agent doesn’t know
what state it transitions to after taking an action. That means that rather than thinking “I'm in state 51 and ifI do action a I'll end up in state s»,” an agent will now be thinking “I'm either
Belief state Conditional plan
in state s or 53, and if I do action a 'l end up in state s,,54 or s5.” We call a set of physical states that the agent believes are possible a belief state.
In partially observable and nondeterministic environments, the solution to a problem is
no longer a sequence, but rather a conditional plan (sometimes called a contingency plan or a
strategy) that specifies what to do depending on what percepts agent receives while executing the plan. We examine nondeterminism in this section and partial observability in the next.
4.3.1
The erratic vacuum world
The vacuum world from Chapter 2 has eight states, as shown in Figure 4.9. There are three
actions—Right, Left, and Suck—and the goal is to clean up all the dirt (states 7 and 8). If the
environment is fully observable, deterministic,
and completely known, then the problem is
easy to solve with any of the algorithms in Chapter 3, and the solution is an action sequence.
For example, if the initial state is 1, then the action sequence [Suck, Right, Suck] will reach a
goal state, 8.
Now suppose that we introduce nondeterminism in the form of a powerful but erratic
vacuum cleaner. In the erratic vacuum world, the Suck action works as follows:
+ When applied to a dirty square the action cleans the square and sometimes cleans up
dirt in an adjacent square, too.
« When applied to a clean square the action sometimes deposits dirt on the carpet.”
To provide a precise formulation of this problem, we need to generalize the notion of a transition model from Chapter 3. Instead of defining the transition model by a RESULT function
4 A set of points is convex if the line joining any two points in & is also contained in S. A convex function is one for which the space “above” it forms a convex set; by definition, convex functions have no local (as opposed 10 global) minima. 5 We assume that most readers milar problems and can sympathize with our agent. We a owners of modern, efficient cleaning appliances who cannot take advantage of this pedagogical de
Section 4.3
Search with Nondeterministic Actions
123
=0 -
Figure 4.9 The eight possible states of the vacuum world; states 7 and 8 are goal states. that returns a single outcome state, we use a RESULTS function that returns a set of possible
outcome states. For example, in the erratic vacuum world, the Suck action in state 1 cleans
up either just the current location, or both locations RESULTS (1, Suck)= {5,7}
If we start in state 1, no single sequence of actions solves the problem, but the following
conditional plan does:
[Suck,if State =5 then [Right, Suck] else []] .
4.3)
Here we see that a conditional plan can contain if-then—else steps; this means that solutions are frees rather than sequences. Here the conditional in the if statement tests to see what the current state is; this is something the agent will be able to observe at runtime, but doesn’t
know at planning time. Alternatively, we could have had a formulation that tests the percept
rather than the state. Many problems in the real, physical world are contingency problems, because exact prediction of the future is impossible. For this reason, many people keep their eyes open while walking around. 4.3.2
AND-OR
search trees
How do we find these contingent solutions to nondeterministic problems?
As in Chapter 3,
we begin by constructing search trees, but here the trees have a different character. In a de-
terministic environment, the only branching is introduced by the agent’s own choices in each
state: I can do this action or that action. We call these nodes OR nodes. In the vacuum world,
Or node
environment, branching is also introduced by the environment’s choice of outcome for each action. We call these nodes AND nodes. For example, the Suck action in state 1 results in the
And node
two kinds of nodes alternate, leading to an AND-OR tree as illustrated in Figure 4.10.
And-or tree
for example, at an OR node the agent chooses Left or Right or Suck. In a nondeterministic
belief state {5,7}, so the agent would need to find a plan for state 5 and for state 7. These
124
Chapter 4 Search in Complex Environments
L
GoaL
=
Suck,
el
el
LOOP
Loop
=
GOAL
Light
Left
Fle]
[ S
EME
L
Loor
Suck
17
o]
GOAL
LooP
Figure 4.10 The first two levels of the search tree for the erratic vacuum world. State nodes are OR nodes where some action must be chosen. At the AND nodes, shown as circles, every ‘outcome must be handled, as indicated by the arc linking the outgoing branches. The solution found is shown in bold lines.
A solution for an AND-OR search problem is a subtree of the complete search tree that
(1) has a goal node at every leaf, (2) specifies one action at each of its OR nodes, and (3) includes every outcome branch at each of its AND nodes. The solution is shown in bold lines
in the figure; it corresponds to the plan given in Equation (4.3).
Figure 4.11 gives a recursive, depth-first algorithm for AND-OR graph search. One key aspect of the algorithm is the way in which it deals with cycles, which often arise in nonde-
terministic problems (e.g., if an action sometimes has no effect or if an unintended effect can be corrected). If the current state is identical to a state on the path from the root, then it returns with failure. This doesn’t mean that there is no solution from the current state; it simply
means that if there is a noncyclic solution, it must be reachable from the earlier incarnation of
the current state, so the new incarnation can be discarded. With this check, we ensure that the
algorithm terminates in every finite state space, because every path must reach a goal, a dead
end, or a repeated state. Notice that the algorithm does not check whether the current state is
a repetition of a state on some other path from the root, which is important for efficiency.
AND-OR graphs can be explored either breadth-first or best-first. The concept of a heuris-
tic function must be modified to estimate the cost of a contingent solution rather than a se-
quence, but the notion of admissibility carries over and there is an analog of the A* algorithm for finding optimal solutions. (See the bibliographical notes at the end of the chapter.)
Section 4.3
Search with Nondeterministic Actions
125
function AND-OR-SEARCH(problem) returns a conditional plan, or failure
return OR-SEARCH(problem, problem.INITIAL, [])
function OR-SEARCH(problem, state, path) returns a conditional plan, or failure if problem.1s-GOAL(state) then return the empty plan if Is-CYCLE(path) then return failure
for each action in problem. ACTIONS state) do
plan < AND-SEARCH(problem, RESULTS(state, action), [state] + path])
if plan # failure then return [action] + plan] return failure
function AND-SEARCH(problem, states, path) returns a conditional plan, or failure for each s; in states do plan; < OR-SEARCH(problem, s;. path)
if plan; = failure then return failure return [ifs, then plan, else if s, then plan, else ...if s,
then plan,,_, else plan,]
Figure 4.11 An algorithm for searching AND-OR graphs generated by nondeterministic environments. A solution is a conditional plan that considers every nondeterministic outcome and makes a plan for each one. 4.3.3
Try, try ag
Consider a slippery vacuum world, which is identical to the ordinary (non-erratic) vacuum
world except that movement actions sometimes fail, leaving the agent in the same location.
For example, moving Right in state 1 leads to the belief state {1,2}. Figure 4.12 shows part of the search graph; clearly, there are no longer any acyclic solutions from state 1, and
AND-OR-SEARCH would return with failure. There is, however, a cyclic solution, which is to keep trying Right until it works. We can express this with a new while construct:
[Suck, while State =S5 do Right, Suck] or by adding a label to denote some portion of the plan and referring to that label later: [Suck,Ly : Right,if State=S5 then L; else Suck].
When is a cyclic plan a solution? A minimum condition is that every leaf is a goal state and
that a leaf is reachable from every point in the plan. In addition to that, we need to consider the cause of the nondeterminism. If it is really the case that the vacuum robot’s drive mechanism
works some of the time, but randomly and independently slips on other occasions, then the
agent can be confident that if the action is repeated enough times, eventually it will work and the plan will succeed.
But if the nondeterminism is due to some unobserved fact about the
robot or environment—perhaps a drive belt has snapped and the robot will never move—then repeating the action will not help. One way to understand this decision is to say that the initial problem formulation (fully observable, nondeterministic) is abandoned in favor of a different formulation (partially observable, deterministic) where the failure of the cyclic plan is attributed to an unobserved
property of the drive belt. In Chapter 12 we discuss how to decide which of several uncertain possibilities is more likely.
Cyelic solution
126
Chapter 4 Search in Complex Environments
Figure 4.12 Part of the search graph for a slippery vacuum world, where we have shown (some) cycles explicitly. All solutions for this problem are cyclic plans because there is no way to move reliably. 4.4
Search in Partially Observable
Environments
‘We now turn to the problem of partial observability, where the agent’s percepts are not enough to pin down the exact state.
That means that some of the agent’s actions will be aimed at
reducing uncertainty about the current state.
4.4.1
Sensorless Conformant
Searching with no observation
‘When the agent’s percepts provide no information at all, we have what is called a sensorless
problem (or a conformant problem). At first, you might think the sensorless agent has no hope of solving a problem if it has no idea what state it starts in, but sensorless solutions are
surprisingly common and useful, primarily because they don’t rely on sensors working properly. In manufacturing systems, for example, many ingenious methods have been developed for orienting parts correctly from an unknown initial position by using a sequence of actions
with no sensing at all. Sometimes a sensorless plan is better even when a conditional plan
with sensing is available. For example, doctors often prescribe a broad-spectrum antibiotic
rather than using the conditional plan of doing a blood test, then waiting for the results to
come back, and then prescribing a more specific antibiotic. The sensorless plan saves time and money, and avoids the risk of the infection worsening before the test results are available.
Consider a sensorless version of the (deterministic) vacuum world. Assume that the agent
knows the geography of its world, but not its own location or the distribution of dirt. In that
case, its initial belief state is {1,2,3,4,5,6,7,8} (see Figure 4.9).
Now, if the agent moves
Right it will be in one of the states {2,4,6,8}—the agent has gained information without
Coercion
perceiving anything! After [Right,Suck] the agent will always end up in one of the states {4,8}. Finally, after [Right,Suck,Lefr,Suck] the agent is guaranteed to reach the goal state 7, no matter what the start state. We say that the agent can coerce the world into state 7.
Section 4.4
Search in Partially Observable Environments
The solution to a sensorless problem is a sequence of actions, not a conditional plan
(because there is no perceiving).
But we search in the space of belief states rather than
physical states.S In belief-state space, the problem is fully observable because the agent always knows its own belief state. Furthermore, the solution (if any) for a sensorless problem
is always a sequence of actions. This is because, as in the ordinary problems of Chapter 3, the percepts received after each action are completely predictable—they’re always empty! So there are no contingencies to plan for. This is true even if the environment is nondeterministic.
‘We could introduce new algorithms for sensorless search problems. But instead, we can
use the existing algorithms from Chapter 3 if we transform the underlying physical problem
into a belief-state problem, in which we search over belief states rather than physical states.
The original problem, P, has components Actionsp, Resultp etc., and the belief-state problem has the following components:
o States: The belief-state space contains every possible subset of the physical states. If P
has N states, then the belief-state problem has 2V belief states, although many of those may be unreachable from the initial state.
o Initial state: Typically the belief state consisting of all states in P, although in some cases the agent will have more knowledge than this.
o Actions: This is slightly tricky. Suppose the agent is in belief state b={s;,s2}, but
ACTIONSp(s1) # ACTIONSp(s>): then the agent is unsure of which actions are legal. If we assume that illegal actions have no effect on the environment, then it is safe to take the union of all the actions in any of the physical states in the current belief state b:
AcTIONS(b) = [J ACTIONSp(s). seb
On the other hand, if an illegal action might lead to catastrophe, it is safer to allow only the intersection, that is, the set of actions legal in all the states. For the vacuum world,
every state has the same legal actions, so both methods give the same result.
o Transition model: For deterministic actions, the new belief state has one result state for each of the current possible states (although some result states may be the same):
b =RESULT(b,a) = {s" : '
=RESULTp(s,a) and 5 € b}.
(4.4)
‘With nondeterminism, the new belief state consists of all the possible results of applying the action to any of the
b =ResuLT(b,a)
states in the current belief state:
= {s':5' € RESULTS p(s,a) and s € b}
= [JREsuLtsp(s,a),
seb
The size of b’ will be the same or smaller than b for deterministic actions, but may be larger than b with nondeterministic actions (see Figure 4.13).
o Goal test: The agent possibly achieves the goal if any state s in the belief state satisfies
the goal test of the underlying problem, Is-GOALp(s). The agent necessarily achieves the goal if every state satisfies IS-GOALp(s). We aim to necessarily achieve the goal.
® Action cost:
This is also tricky.
If the same action can have different costs in dif-
ferent states, then the cost of taking an action in a given belief state could be one of
© Ina fully observable environment, each belief state contai ins one physical state. Thus, we can view the algorithms in Chapter 3 as searching in a belief-state space of leton belief stat
127
128
Chapter 4 Search in Complex Environments
(@
(b)
Figure 4.13 (a) Predicting the next belief state for the sensorless vacuum world with the deterministic action, Right. (b) Prediction for the same belief state and action in the slippery version of the sensorless vacuum world.
several values. (This gives rise to a new class of problems, which we explore in Exercise 4.MVAL.) For now we assume that the cost of an action is the same in all states and 5o can be transferred directly from the underlying physical problem. Figure 4.14 shows the reachable belief-state space for the deterministic, sensorless vacuum world. There are only 12 reachable belief states out of 28 =256 possible belief states. The preceding definitions enable the automatic construction of the belief-state problem
formulation from the definition of the underlying physical problem. Once this is done, we can solve sensorless problems with any of the ordinary search algorithms of Chapter 3. In ordinary graph search, newly reached states are tested to see if they were previously reached. This works for belief states, too; for example, in Figure 4.14, the action sequence [Suck.Left,Suck] starting at the initial state reaches the same belief state as [Right,Left,Suck), namely, {5,7}. Now, consider the belief state reached by [Lefr], namely, {1,3,5,7}. Obviously, this is not identical to {5,7}, but it is a superser. We can discard (prune) any such superset belief state. Why? Because a solution from {1,3,5,7} must be a solution for each
of the individual states 1, 3, 5, and 7, and thus it is a solution for any combination of these
individual states, such as {5,7}; therefore we don’t need to try to solve {1,3,5,7}, we can
concentrate on trying to solve the strictly easier belief state {5,7}.
Conversely, if {1,3,5,7} has already been generated and found to be solvable, then any
subset, such as {5,7}, is guaranteed to be solvable. (If I have a solution that works when I'm very confused about what state I'm in, it will still work when I'm less confused.) This extra
level of pruning may dramatically improve the efficiency of sensorless problem solving. Even with this improvement, however, sensorless problem-solving as we have described
itis seldom feasible in practice. One issue is the vastness of the belief-state space—we saw in
the previous chapter that often a search space of size N is too large, and now we have search
spaces of size 2V. Furthermore, each element of the search space is a set of up to N elements.
For large N, we won’t be able to represent even a single belief state without running out of
memory space.
One solution is to represent the belief state by some more compact description.
In En-
glish, we could say the agent knows “Nothing” in the initial state; after moving Left, we could
Section 4.4
Search in Partially Observable Environments
129
L L1 3 [=] 5
1[ 54~7
L
| =]
—
2| e [=A] |55
sfA] %=
4y [
5-4“
o
o[
o[
1[5
s
1 [
=]
&
R
¢
|
|
|s [ [ 5[ 7[=a] ] o[
S
EL|
s [\EL
——
7[=] R
¥
o[ T s
7[=]
L
—
[~ lk
L
L R
S
e[ T
A A
TL
|
=
[F
s[4 Ll
= s |
[~
L
3[=
R
7~
l———l
TR
"=
Figure 4.14 The reachable portion of the belief-state space for the deterministic, sensorless
vacuum world. Each rectangular box corresponds to a single belief state. At any given point,
the agent has a belief state but does not know which physical state it is in. The initial belief state (complete ignorance) is the top center box.
say, “Not in the rightmost column,” and so on. Chapter 7 explains how to do this in a formal representation scheme.
Another approach is to avoid the standard search algorithms, which treat belief states as
black boxes just like any other problem state. Instead, we can look inside the belief states
and develop incremental belief-state search algorithms that build up the solution one phys- Incremental belief-state search ical state at a time.
For example, in the sensorless vacuum world, the initial belief state is
{1,2,3,4,5,6,7,8}, and we have to find an action sequence that works in all 8 states. We can
do this by first finding a solution that works for state 1; then we check if it works for state 2; if not, go back and find a different solution for state 1, and so on.
Just as an AND-OR search has to find a solution for every branch at an AND node, this
algorithm has to find a solution for every state in the belief state; the difference is that AND—
OR search can find a different solution for each branch, whereas search has to find one solution that works for all the states.
an incremental belief-state
The main advantage of the incremental approach is that it is typically able to detect failure
quickly—when a belief state is unsolvable, it is usually the case that a small subset of the
130
Chapter 4 Search in Complex Environments belief state, consisting of the first few states examined, is also unsolvable. In some cases, this
leads to a speedup proportional to the size of the belief states, which may themselves be as
large as the physical state space itself. 4.4.2
Searching in partially observable environments
Many problems cannot be solved without sensing. For example, the sensorless 8-puzzle is impossible. On the other hand, a little bit of sensing can go a long way: we can solve 8puzzles if we can see just the upper-left corner square.
The solution involves moving each
tile in turn into the observable square and keeping track of its location from then on .
For a partially observable problem, the problem specification will specify a PERCEPT (s)
function that returns the percept received by the agent in a given state. If sensing is non-
deterministic, then we can use a PERCEPTS function that returns a set of possible percepts. For fully observable problems, PERCEPT (s) = s for every state s, and for sensorless problems PERCEPT (s) = null.
Consider a local-sensing vacuum world, in which the agent has a position sensor that
yields the percept L in the left square, and R in the right square, and a dirt sensor that yields
Dirty when the current square is dirty and Clean when it is clean. Thus, the PERCEPT in state
Lis [L, Dirty]. With partial observability, it will usually be the case that several states produce
the same percept; state 3 will also produce [L, Dirfy]. Hence, given this
initial percept, the
initial belief state will be {1,3}. We can think of the transition model between belief states
for partially observable problems as occurring in three stages, as shown in Figure 4.15:
+ The prediction stage computes the belief state resulting from the action, RESULT (b, a),
exactly as we did with sensorless problems. To emphasize that this is a prediction, we
use the notation h=RESULT(b,a), where the “hat” over the b means “estimated,” and we also use PREDICT(b,a) as a synonym for RESULT (b, a). + The possible percepts stage computes the set of percepts that could be observed in the predicted belief state (using the letter o for observation):
POSSIBLE-PERCEPTS (b) = {0 : 0=PERCEPT(s) and s € b} . « The update stage computes, for each possible percept, the belief state that would result from the percept.
The updated belief state b, is the set of states in b that could have
produced the percept:
b, = UPDATE(b,0) = {s : 0=PERCEPT(s) and 5 € b}. The agent needs to deal with possible percepts at planning time, because it won’t know
the actual percepts until it executes the plan. Notice that nondeterminism in the phys-
ical environment can enlarge the belief state in the prediction stage, but each updated belief state b, can be no larger than the predicted belief state b; observations can only
help reduce uncertainty. Moreover, for deterministic sensing, the belief states for the
different possible percepts will be disjoint, forming a partition of the original predicted belief state.
Putting these three stages together, we obtain the possible belief states resulting from a given action and the subsequent possible percepts:
RESULTS (b,a) = {b, : b, = UPDATE(PREDICT(b,a),0) and 0 € POSSIBLE-PERCEPTS (PREDICT(b,a))} .
(4.5)
Section 4.4
Search in Partially Observable Environments
(a)
B.Dirty]
(b)
=
i1/
o=
‘
[ o[z
)
1B Clean] Figure 4.15 Two examples of transitions in local-sensing vacuum worlds. (a) In the deterministic world, Right is applied in the initial belief state, resulting in a new predicted belief
state with two possible physical states; for those states, the possible percepts are [R, Dirty]
and [R, Clean], leading to two belief states, each of which is a singleton. (b) In the slippery world, Right i applied in the initial belief state, giving a new belief state with four physical states; for those states, the possible percepts are [L, Dirty], [R, Dirty], and [R, Clean], leading to three belief states as shown.
[4,Clean]
Figure 4.16 The first level of the AND-OR
[8.0iry) L
[B.Clean]
search tree for a problem in the local-sensing
vacuum world; Suck is the first action in the solution.
131
132
Chapter 4 Search in Complex Environments 4.4.3
Solving partially observable problems
The preceding section showed how to derive the RESULTS function for a nondeterministic
belief-state problem from an underlying physical problem, given the PERCEPT function. With this formulation, the AND—OR search algorithm of Figure 4.11 can be applied directly to
derive a solution.
Figure 4.16 shows part of the search tree for the local-sensing vacuum
world, assuming an initial percept [A, Dirty]. The solution is the conditional plan [Suck, Right, if Bstate={6} then Suck else []]. Notice that, because we supplied a belief-state problem to the AND-OR
search algorithm, it
returned a conditional plan that tests the belief state rather than the actual state. This is as it
should be:
in a partially observable environment the agent won’t know the actual state.
As in the case of standard search algorithms applied to sensorless problems, the AND—
OR search algorithm treats belief states as black boxes, just like any other states. One can improve on this by checking for previously generated belief states that are subsets or supersets
of the current state, just as for sensorless problems.
One can also derive incremental search
algorithms, analogous to those described for sensorless problems, that provide substantial
speedups over the black-box approach.
4.4.4
An agent for partially observable environments
An agent for partially observable environments formulates a problem, calls a search algorithm (such as AND-OR-SEARCH)
to solve it, and executes the solution. There are two main
differences between this agent and the one for fully observable deterministic environments.
First, the solution will be a conditional plan rather than a sequence; to execute an if-then—else
expression, the agent will need to test the condition and execute the appropriate branch of the conditional.
Second, the agent will need to maintain its belief state as it performs actions
and receives percepts. This process resembles the prediction-observation-update process in
Equation (4.5) but is actually simpler because the percept is given by the environment rather
than calculated by the agent. Given an initial belief state b, an action a, and a percept o, the new belief state is:
b’ = UPDATE(PREDICT(b,a),0).
(4.6)
Consider a kindergarten vacuum world wherein agents sense only the state of their current
square, and any square may become dirty at any time unless the agent is actively cleaning it at that moment.” Figure 4.17 shows the belief state being maintained in this environment. In partially observable environments—which include the vast majority of real-world Monitoring Filtering
State estimation
environments—maintaining one’s belief state is a core function of any intelligent system.
This function goes under various names, including monitoring, filtering, and state estima-
tion. Equation (4.6) is called a recursive state estimator because it computes the new belief
state from the previous one rather than by examining the entire percept sequence. If the agent is not to “fall behind,” the computation has to happen as fast as percepts are coming in. As
the environment becomes more complex, the agent will only have time to compute an ap-
proximate belief state, perhaps focusing on the implications of the percept for the aspects of the environment that are of current interest. Most work on this problem has been done for
7 The usual apologies to those who are unfamiliar with the effect of small children on the environment.
Section 4.4
[A.Clean)
1B.Diny]
4l
Suck
133
Search in Partially Observable Environments
Figure 4.17 Two prediction-update cycles of belief-state maintenance in the kindergarten vacuum world with local sensing. stochastic, continuous-state environments with the tools of probability theory, as explained in
Chapter 14.
In this section we will show an example in a discrete environment with deterministic
sensors and nondeterministic actions. The example concerns a robot with a particular state estimation task called localization: working out where it is, given a map of the world and
a sequence of percepts and actions. Our robot is placed in the maze-like environment of Figure 4.18. The robot is equipped with four sonar sensors that tell whether there is an obstacle—the outer wall or a dark shaded square in the figure—in each of the four compass directions. The percept is in the form of a bt vector, one bit for each of the directions north,
east, south, and west in that order, so 1011 means there are obstacles to the north, south, and
west, but not east. ‘We assume that the sensors give perfectly correct data, and that the robot has a correct
map of the environment.
But unfortunately, the robot’s navigational system is broken, so
when it executes a Right action, it moves randomly to one of the adjacent squares. robot’s task is to determine its current location.
The
Suppose the robot has just been switched on, and it does not know where it is—its initial
belief state b consists of the set of all locations.
The robot then receives the percept 1011
and does an update using the equation b, =UPDATE(1011), yielding the 4 locations shown
in Figure 4.18(a). You can inspect the maze to see that those are the only four locations that yield the percept 1011.
Next the robot executes a Right action, but the result is nondeterministic. The new belief
state, b,
=PREDICT (b, Right), contains all the locations that are one step away from the lo-
cations in b,. When the second percept, 1010, arrives, the robot does UPDATE (b,, 1010) and finds that the belief state has collapsed down to the single location shown in Figure 4.18(b). That’s the only location that could be the result of
UPDATE (PREDICT (UPDATE(b, 1011),Right),1010) . ‘With nondeterministic actions the PREDICT step grows the belief state, but the UPDATE step
shrinks it back down—as long as the percepts provide some useful identifying information.
Sometimes the percepts don’t help much for localization: If there were one or more long east-
west corridors, then a robot could receive a long sequence of 1010 percepts, but never know
Localization
134
Chapter 4 Search in Complex Environments
(b) Possible locations of robot after E; = 1011, E,=1010
Figure 4.18 Possible positions of the robot, ©, (a) after one observation, £; = 1011, and (b) after moving one square and making a second observation, E> = 1010. When sensors are noiseless and the transition model is accurate, there is only one possible location for the robot consistent with this sequence of two observations. where in the corridor(s) it was. But for environments with reasonable variation in geography, localization often converges quickly to a single point, even when actions are nondeterministic.
What happens if the sensors are faulty? If we can reason only with Boolean logic, then we.
have to treat every sensor bit as being either correct or incorrect, which is the same as having
no perceptual information at all. But we will see that probabilistic reasoning (Chapter 12), allows us to extract useful information from a faulty sensor as long as it is wrong less than half the time.
Online Search Agents and Unknown Offline search Online search
Environments
So far we have concentrated on agents that use offfine search algorithms. They compute
a complete solution before taking their first action.
In contrast, an online search® agent
interleaves computation and action: first it takes an action, then it observes the environment
and computes the next action. Online search is a good idea in dynamic or semi-dynamic environments, where there is a penalty for sitting around and computing too long. Online 8 The term “onl i¢” here refers to algorithms that must proc nput as it is received rather than waiting for the entire input ata set to become available. This usage of “online’ unrelated to the concept of “having an Internet
connecti
"
Section 4.5
Online Search Agents and Unknown Environments
135
search is also helpful in nondeterministic domains because it allows the agent to focus its
computational efforts on the contingencies that actually arise rather than those that might
happen but probably won’t.
Of course, there is a tradeoff:
the more an agent plans ahead, the less often it will find
itself up the creek without a paddle. In unknown environments, where the agent does not
know what states exist or what its actions do, the agent must use its actions as experiments in order to learn about the environment.
A canonical example of online search is the mapping problem: a robot is placed in an Mapping problem
unknown building and must explore to build a map that can later be used for getting from
A to B. Methods for escaping from labyrinths—required knowledge for aspiring heroes of antiquity—are also examples of online search algorithms. Spatial exploration is not the only
form of online exploration, however. Consider a newborn baby: it has many possible actions but knows the outcomes of none of them, and it has experienced only a few of the possible states that it can reach.
4.5.1
Online search problems
An online search problem is solved by interleaving computation, sensing, and acting. We’ll start by assuming a deterministic and fully observable environment (Chapter 17 relaxes these assumptions) and stipulate that the agent knows only the following: * ACTIONS(s), the legal actions in state s;
* ¢(s,a,s'), the cost of applying action a in state s to arrive at state s'. Note that this cannot be used until the agent knows that s” is the outcome. * Is-GOAL(s), the goal test. Note in particular that the agent cannot determine RESULT (s, @) except by actually being in s
and doing a. For example, in the maze problem shown in Figure 4.19, the agent does not know that going Up from (1,1) leads to (1,2); nor, having done that, does it know that going Down will take it back to (1,1). This degree of ignorance can be reduced in some applications—for example, a robot explorer might know how its movement actions work and be ignorant only of the locations of obstacles.
Finally, the agent might have access to an admissible heuristic function /(s) that estimates
the distance from the current state to a goal state. For example, in Figure 4.19, the agent might know the location of the goal and be able to use the Manhattan-distance heuristic (page 97).
Typically, the agent’s objective is to reach a goal state while minimizing cost. (Another
possible objective is simply to explore the entire environment.) The cost is the total path
cost that the agent incurs as it travels. It is common to compare this cost with the path cost
the agent would incur if it knew the search space in advance—that is, the optimal path in the known environment. In the language of online algorithms, this comparison is called the competitive ratio; we would like it to be as small as possible.
Online explorers are vulnerable to dead ends: states from which no goal state is reach-
able. If the agent doesn’t know what each action does, it might execute the “jump into bottomless pit” action, and thus never reach the goal. In general, no algorithm can avoid dead ends in all state spaces. Consider the two dead-end state spaces in Figure 4.20(a). An on-
line search algorithm that has visited states S and A cannot tell if it is in the top state or the bottom one; the two look identical based on what the agent has seen. Therefore, there is no
Competitive ratio Dead end
n, then the constraint cannot be satisfied.
Section 6.2 Constraint Propagation: Inference in CSPs
189
This leads to the following simple algorithm: First, remove any variable in the constraint
that has a singleton domain, and delete that variable’s value from the domains of the re-
maining variables. Repeat as long as there are singleton variables. If at any point an empty
domain is produced or there are more variables than domain values left, then an inconsistency has been detected. This method can detect the inconsistency in the assignment {WA = red, NSW =red} for Figure 6.1. Notice that the variables SA, NT, and Q are effectively connected by an Alldiff
constraint because each pair must have two different colors. After applying AC-3 with the partial assignment, the domains of SA, NT, and Q are all reduced to {green, blue}. That is, we have three variables and only two colors, so the Alldiff constraint is violated.
Thus, a
simple consistency procedure for a higher-order constraint is sometimes more effective than
applying arc consistency to an equivalent set of binary constraints.
Another important higher-order constraint is the resource constraint, sometimes called
the Atmost constraint. For example, in a scheduling problem, let P, ..., P; denote the numbers.
Resource constraint
of personnel assigned to each of four tasks. The constraint that no more than 10 personnel
are assigned in total is written as Ammost(10, Py, Py, Py, Py). We can detect an inconsistency simply by checking the sum of the minimum values of the current domains; for example, if each variable has the domain {3,4,5,6}, the Armost constraint cannot be satisfied. We can
also enforce consistency by deleting the maximum value of any domain if it is not consistent
with the minimum values of the other domains. Thus, if each variable in our example has the domain {2,3,4,5,6}, the values 5 and 6 can be deleted from each domain.
For large resource-limited problems with integer values—such as logistical problems involving moving thousands of people in hundreds of vehicles—it is usually not possible to represent the domain of each variable as a large set of integers and gradually reduce that set by consistency-checking methods. Instead, domains are represented by upper and lower bounds and are managed by bounds propagation. For example, in an airline-scheduling
problem, let’s suppose there are two flights, i and F, for which the planes have capacities 165 and 385, respectively. The initial domains for the numbers of passengers on flights Fi and F; are then
D;=[0,165]
and
Bounds propagation
D;=[0,385]
Now suppose we have the additional constraint that the two flights together must carry 420
people: Fj + Fy = 420. Propagating bounds constraints, we reduce the domains to
Dy =[35,165]
and D, =|[255,385].
‘We say that a CSP is bounds-consistent if for every variable X, and for both the lower-bound and upper-bound values of X, there exists some value of ¥ that satisfies the constraint between
Bounds-consistent
X and Y for every variable Y. This kind of bounds propagation is widely used in practical constraint problems. 6.2.6 Sudoku
The popular Sudoku puzzle has introduced millions of people to constraint satisfaction problems, although they may not realize it. A Sudoku board consists of 81 squares, some of which are initially filled with digits from 1 to 9. The puzzle is to fill in all the remaining squares such that no digit appears twice in any row, column, or 3x 3 box (see Figure 6.4). A row, column, or box is called a unit.
Sudoku
(@)
R [o|r|unfo|a|w|—| i.
To solve a tree-structured CSP, first pick any variable to be the root of the tree, and choose
an ordering of the variables such that each variable appears after its parent in the tree. Such an ordering is called a topological sort. Figure 6.10(a) shows a sample tree and (b) shows
one possible ordering. Any tree with n nodes has n— 1 edges, so we can make this graph directed arc-consistent in O(n) steps, each of which must compare up to d possible domain
values for two variables, for a total time of O(nd?). Once we have a directed arc-consistent
graph, we can just march down the list of variables and choose any remaining value. Since each edge from a parent to its child is arc-consistent, we know that for any value we choose
for the parent, there will be a valid value left to choose for the child. That means we won’t
3 A careful cartographer or patriotic Tasmanian might object that Tasmania should not be colored the same as its nearest mainland neighbor, to avoid the impression that it might be part of that state. 4 Sadly, very few regions of the world have tree-structured maps, although Sulawesi comes close.
Topological sort
200
Chapter 6
Constraint Satisfaction Problems
ze ez @
®
Figure 6.10 () The constraint graph of a tree-structured CSP. (b) A linear ordering of the variables consistent with the tree with A as the root. This is known as a topological sort of the variables. function TREE-CSP-SOLVER(csp) returns a solution, or failure inputs: csp, a CSP with components X, D, C n < number of variables in X
assignment < an empty assignment root any variable in X X < TOPOLOGICALSORT(X, roof) for j=n down to 2 do
MAKE-ARC-CONSISTENT(PARENT(X)), X))
if it cannot be made consistent then return failure fori=1tondo
assignment[X;]
any consistent value from D;
if there is no consistent value then return failure
return assignment
Figure 6.11 The TREE-CSP-SOLVER algorithm for solving tree-structured CSPs. If the (CSP has a solution, we will find it in linear time; if not, we will detect a contradiction.
have to backtrack; we can move linearly through the variables. The complete algorithm is shown in Figure 6.11.
Now that we have an efficient algorithm for trees, we can consider whether more general constraint graphs can be reduced to trees somehow. There are two ways to do this: by
removing nodes (Section 6.5.1) or by collapsing nodes together (Section 6.5.2). 6.5.1
Cutset conditioning
The first way to reduce a constraint graph to a tree involves assigning values to some variables so that the remaining variables form a tree. Consider the constraint graph for Australia, shown
again in Figure 6.12(a). Without South Australia, the graph would become a tree, as in (b). Fortunately, we can delete South Australia (in the graph, not the country) by fixing a value for SA and deleting from the domains of the other variables any values that are inconsistent with the value chosen for SA.
Now, any solution for the CSP after SA and its constraints are removed will be consistent
with the value chosen for SA. (This works for binary CSPs; the situation is more complicated with higher-order constraints.) Therefore, we can solve the remaining tree with the algorithm
Section 6.5
(@)
201
The Structure of Problems
(®)
Figure 6.12 (a) The original constraint graph from Figure 6.1. (b) After the removal of SA,
the constraint graph becomes a forest of two trees.
given above and thus solve the whole problem. OF course, in the general case (as opposed o map coloring), the value chosen for SA could be the wrong one, so we would need to try each possible value. The general algorithm is as follows: 1. Choose a subset S of the CSP’s variables such that the constraint graph becomes a tree after removal of S. is called a cycle cutset.
Cycle cutset
2. For each possible assignment to the variables in S that satisfies all constraints on S,
(a) remove from the domains of the remaining variables any values that are inconsis-
tent with the assignment for S, and (b) if the remaining CSP has a solution, return it together with the assignment for S.
If the cycle cutset has size c, then the total run time is O(d® - (n — ¢)d?): we have to try each of
the d° combinations of values for the variables in §, and for each combination we must solve
atree problem of size n — c. If the graph is “nearly a tree,” then ¢ will be small and the savings over straight backtracking will be huge—for our 100-Boolean-variable example, if we could
find a cutset of size ¢ =20, this would get us down from the lifetime of the Universe to a few
minutes. In the worst case, however, ¢ can be as large as (n — 2). Finding the smallest cycle cutset is NP-hard, but several efficient approximation algorithms are known.
The overall
algorithmic approach is called cutset conditioning; it comes up again in Chapter 13, where
itis used for reasoning about probabilities.
6.5.2
Cutset conditioning
Tree decomposition
The second way to reduce a constraint graph to a tree is based on constructing a tree decom-
position of the constraint graph: a transformation of the original graph into a tree where each node in the tree consists of a set of variables, as in Figure 6.13. A tree decomposition must
satisfy these three requirements: + Every variable in the original problem appears in at least one of the tree nodes. If two variables are connected by a constraint in the original problem, they must appear together (along with the constraint) in at least one of the tree nodes.
« If a variable appears in two nodes in the tree, it must appear in every node along the
path connecting those nodes.
Tree decomposition
202
Chapter 6
Constraint Satisfaction Problems
Figure 6.13 A tree decomposition of the constraint graph in Figure 6.12(a).
The first two conditions ensure that all the variables and constraints are represented in the tree decomposition.
The third condition seems rather technical, but allows us to say that
any variable from the original problem must have the same value wherever it appears: the
constraints in the tree say that a variable in one node of the tree must have the same value as
the corresponding variable in the adjacent node in the tree. For example, SA appears in all four of the connected nodes in Figure 6.13, so each edge in the tree decomposition therefore
includes the constraint that the value of SA in one node must be the same as the value of SA
in the next. You can verify from Figure 6.12 that this decomposition makes sense.
Once we have a tree-structured graph, we can apply TREE-CSP-SOLVER to get a solution
in O(nd?) time, where n is the number of tree nodes and d is the size of the largest domain. But note that in the tree, a domain is a set of tuples of values, not just individual values.
For example, the top left node in Figure 6.13 represents, at the level of the original prob-
lem, a subproblem with variables {WA,NT,SA}, domain {red, green, blue}, and constraints WA # NT,SA # NT,WA # SA. At the level of the tree, the node represents a single variable, which we can call SANTWA,
whose value must be a three-tuple of colors,
such as
(red, green,blue), but not (red,red,blue), because that would violate the constraint SA # NT
from the original problem. We can then move from that node to the adjacent one, with the variable we can call SANTQ, and find that there is only one tuple, (red, green, blue), that is
consistent with the choice for SANTWA.
The exact same process is repeated for the next two
nodes, and independently we can make any choice for T.
We can solve any tree decomposition problem in O(nd?) time with TREE-CSP-SOLVER,
which will be efficient as long as d remains small. Going back to our example with 100 Boolean variables, if each node has 10 variables, then d=2'° and we should be able to solve
the problem in seconds. But if there is a node with 30 variables, it would take centuries.
Tree width
A given graph admits many tree decompositions; in choosing a decomposition, the aim is to make the subproblems as small as possible. (Putting all the variables into one node is technically a tree, but is not helpful.) The tree width of a tree decomposition of a graph is
Summary
203
one less than the size of the largest node; the tree width of the graph itself is defined to be
the minimum width among all its tree decompositions.
If a graph has tree width w then the
problem can be solved in O(nd"*!) time given the corresponding tree decomposition. Hence, CSPs with constraint graphs of bounded tree width are solvable in polynomial time.
Unfortunately, finding the decomposition with minimal tree width is NP-hard, but there
are heuristic methods that work well in practice.
Which is better: the cutset decomposition
with time O(d°- (n — ¢)d?), or the tree decomposition with time O(nd"*')? Whenever you
have a cycle-cutset of size c, there is also a tree width of size w < ¢+ 1, and it may be far
smaller in some cases. So time consideration favors tree decomposition, but the advantage of
the cycle-cutset approach is that it can be executed in linear memory, while tree decomposition requires memory exponential in w.
6.5.3
Value symmetry
So far, we have looked at the structure of the constraint graph. There can also be important structure in the values of variables, or in the structure of the constraint relations themselves.
Consider the map-coloring problem with d colors.
For every consistent solution, there is
actually a set of d! solutions formed by permuting the color names. For example, on the
Australia map we know that WA, NT, and SA must all have different colors, but there are
31
=6 ways to assign three colors to three regions. This is called value symmetry. We would Value symmetry
like to reduce the search space by a factor of d! by breaking the symmetry in assignments.
Symmetry-breaking We do this by introducing a symmetry-breaking constraint. For our example, we might cons traint impose an arbitrary ordering constraint, NT < SA < WA, that requires the three values to be in alphabetical order.
This constraint ensures that only one of the d! solutions is possible:
{NT = blue,SA = green, WA = red}.
For map coloring, it was easy to find a constraint that eliminates the symmetry. In general it is NP-hard to eliminate all symmetry, but breaking value symmetry has proved to be
important and effective on a wide range of problems.
Summary + Constraint satisfaction problems (CSPs) represent a state with a set of variable/value
pairs and represent the conditions for a solution by a set of constraints on the variables. Many important real-world problems can be described as CSPs.
+ A number of inference techniques use the constraints to rule out certain variable as-
signments. These include node, arc, path, and k-consistency.
+ Backtracking search, a form of depth-first search, is commonly used for solving CSPs. Inference can be interwoven with search.
« The minimum-remaining-values and degree heuristics are domain-independent methods for deciding which variable to choose next in a backtracking search.
The least-
constraining-value heuristic helps in deciding which value to try first for a given variable. Backtracking occurs when no legal assignment can be found for a variable.
Conflict-directed backjumping backtracks directly to the source of the problem. Constraint learning records the conflicts as they are encountered during search in order to avoid the same conflict later in the search.
204
Chapter 6
Constraint Satisfaction Problems
+ Local search using the min-conflicts heuristic has also been applied to constraint satis-
faction problems with great success. + The complexity of solving a CSP is strongly related to the structure of its constraint
graph. Tree-structured problems can be solved in linear time. Cutset conditioning can
reduce a general CSP to a tree-structured one and is quite efficient (requiring only lin-
ear memory) if a small cutset can be found. Tree decomposition techniques transform
the CSP into a tree of subproblems and are efficient if the tree width of the constraint
graph is small; however they need memory exponential in the tree width of the con-
straint graph. Combining cutset conditioning with tree decomposition can allow a better
tradeoff of memory versus time.
Bibliographical and Historical Notes
Diophantine equations
The Greek mathematician Diophantus (c. 200-284) presented and solved problems involving algebraic constraints on equations, although he didn’t develop a generalized methodology. ‘We now call equations over integer domains Diophantine equations.
The Indian mathe-
matician Brahmagupta (c. 650) was the first to show a general solution over the domain of integers for the equation ax+ by = c. Systematic methods for solving linear equations by variable elimination were studied by Gauss (1829); the solution of linear inequality constraints
goes back to Fourier (1827). Finite-domain constraint satisfaction problems also have a long history. For example, graph coloring (of which map coloring is a special case) is an old problem in mathematics.
The four-color conjecture (that every planar graph can be colored with four or fewer colors) was first made by Francis Guthrie, a student of De Morgan, in 1852.
It resisted solution—
despite several published claims to the contrary—until a proof was devised by Appel and Haken (1977) (see the book Four Colors Suffice (Wilson, 2004)). Purists were disappointed that part of the proof relied on a computer,
so Georges Gonthier (2008), using the COQ
theorem prover, derived a formal proof that Appel and Haken’s proof program was correct.
Specific classes of constraint satisfaction problems occur throughout the history of com-
puter science. One of the most influential early examples was SKETCHPAD (Sutherland, 1963), which solved geometric constraints in diagrams and was the forerunner of modern
drawing programs and CAD tools. The identification of CSPs as a general class is due to Ugo Montanari (1974). The reduction of higher-order CSPs to purely binary CSPs with auxiliary variables (see Exercise 6.NARY) is due originally to the 19th-century logician Charles Sanders Peirce. It was introduced into the CSP literature by Dechter (1990b) and was elaborated by Bacchus and van Beek (1998). CSPs with preferences among solutions are studied
widely in the optimization literature; see Bistarelli er al. (1997) for a generalization of the
CSP framework to allow for preferences.
Constraint propagation methods were popularized by Waltz’s (1975) success on polyhedral line-labeling problems for computer vision. Waltz showed that in many problems, propagation completely eliminates the need for backtracking. Montanari (1974) introduced the notion of constraint graphs and propagation by path consistency. Alan Mackworth (1977)
proposed the AC-3 algorithm for enforcing arc consistency as well as the general idea of combining backtracking with some degree of consistency enforcement. AC-4, a more efficient
Bibliographical and Historical Notes
205
arc-consistency algorithm developed by Mohr and Henderson (1986), runs in O(cd?) worstcase time but can be slower than AC-3 on average cases. The PC-2 algorithm (Mackworth,
1977) achieves path consistency in much the same way that AC-3 achieves arc consistency. Soon after Mackworth’s paper appeared, researchers began experimenting with the trade-
off between the cost of consistency enforcement and the benefits in terms of search reduction. Haralick and Elliott (1980) favored the minimal forward-checking algorithm described
by McGregor (1979), whereas Gaschnig (1979) suggested full arc-consistency checking after each variable assignment—an algorithm later called MAC
by Sabin and Freuder (1994). The
latter paper provides somewhat convincing evidence that on harder CSPs, full arc-consistency
checking pays off. Freuder (1978, 1982) investigated the notion of k-consistency and its relationship to the complexity of solving CSPs. Dechter and Dechter (1987) introduced directional arc consistency. Apt (1999) describes a generic algorithmic framework within which consistency propagation algorithms can be analyzed, and surveys are given by Bessiére (2006) and Bartdk et al. (2010). Special methods for handling higher-order or global constraints were developed first logic within the context of constraint logic programming. Marriott and Stuckey (1998) pro- Constraint programming vide excellent coverage of research in this area. The Alldiff constraint was studied by Regin (1994), Stergiou and Walsh (1999), and van Hoeve (2001). There are more complex inference algorithms for Alldiff (see van Hoeve and Katriel, 2006) that propagate more constraints but are more computationally expensive to run. Bounds constraints were incorporated into con-
straint logic programming by Van Hentenryck et al. (1998). A survey of global constraints is provided by van Hoeve and Katriel (2006). Sudoku has become the most widely known CSP and was described as such by Simonis (2005). Agerbeck and Hansen (2008) describe some of the strategies and show that Sudoku
on an n? x n? board is in the class of NP-hard problems. In 1850, C. F. Gauss described
a recursive backtracking algorithm
for solving the 8-
queens problem, which had been published in the German chess magazine Schachzeirung in
1848. Gauss called his method Tatonniren, derived from the French word taronner—to grope around, as if in the dark. According to Donald Knuth (personal communication), R. J. Walker introduced the term backtrack in the 1950s. Walker (1960) described the basic backtracking algorithm and used it to find all solutions to the 13-queens problem. Golomb and Baumert (1965) formulated, with
examples, the general class of combinatorial problems to which backtracking can be applied, and introduced what we call the MRV
heuristic.
Bitner and Reingold (1975) provided an
influential survey of backtracking techniques. Brelaz (1979) used the degree heuristic as a
tiebreaker after applying the MRV heuristic. The resulting algorithm, despite its simplicity,
is still the best method for k-coloring arbitrary graphs. Haralick and Elliott (1980) proposed
the least-constraining-value heuristic. The basic backjumping
method is due to John Gaschnig (1977,
1979).
Kondrak and
van Beek (1997) showed that this algorithm is essentially subsumed by forward checking. Conflict-directed backjumping was devised by Prosser (1993).
Dechter (1990a) introduced
graph-based backjumping, which bounds the complexity of backjumping-based algorithms. as a function of the constraint graph (Dechter and Frost, 2002).
A very general form of intelligent backtracking was developed early on by Stallman and Sussman (1977). Their technique of dependency-directed backtracking combines bacl k. Dependency-directed backtracking
206
Chapter 6
Constraint Satisfaction Problems
jumping with no-good learning (McAllester, 1990) and led to the development of truth maintenance systems (Doyle, 1979), which we discuss in Section 10.6.2. The connection between
Constraint learning
the two areas is analyzed by de Kleer (1989).
The work of Stallman and Sussman also introduced the idea of constraint learning, in which partial results obtained by search can be saved and reused later in the search. The
idea was formalized by Dechter (1990a). Backmarking (Gaschnig, 1979) is a particularly
simple method in which consistent and inconsistent pairwise assignments are saved and used to avoid rechecking constraints. Backmarking can be combined with conflict-directed back-
jumping; Kondrak and van Beek (1997) present a hybrid algorithm that provably subsumes either method taken separately. The method of dynamic backtracking (Ginsberg,
1993) retains successful partial as-
signments from later subsets of variables when backtracking over an earlier choice that does
not invalidate the later success.
Moskewicz er al. (2001) show how these techniques and
others are used to create an efficient SAT solver. Empirical studies of several randomized backtracking methods were done by Gomes ez al. (2000) and Gomes and Selman (2001).
Van Beek (2006) surveys backtracking. Local search in constraint satisfaction problems was popularized by the work of Kirkpatrick et al. (1983) on simulated annealing (see Chapter 4), which is widely used for VLSI
layout and scheduling problems. Beck ef al. (2011) give an overview of recent work on jobshop scheduling. The min-conflicts heuristic was first proposed by Gu (1989) and was devel-
oped independently by Minton ez al. (1992).
Sosic and Gu (1994) showed how it could be
applied to solve the 3,000,000 queens problem in less than a minute. The astounding success
of local search using min-conflicts on the n-queens problem led to a reappraisal of the nature and prevalence of “easy” and “hard” problems.
Peter Cheeseman ez al. (1991) explored the
difficulty of randomly generated CSPs and discovered that almost all such problems either are trivially easy or have no solutions. Only if the parameters of the problem generator are
set in a certain narrow range, within which roughly half of the problems are solvable, do we. find “hard” problem instances. We discuss this phenomenon further in Chapter 7. Konolige (1994) showed that local search is inferior to backtracking search on problems
with a certain degree of local structure; this led to work that combined
local search and
inference, such as that by Pinkas and Dechter (1995). Hoos and Tsang (2006) provide a survey of local search techniques, and textbooks are offered by Hoos and Stiitzle (2004) and
Aarts and Lenstra (2003). Work relating the structure and complexity of CSPs originates with Freuder (1985) and
Mackworth and Freuder (1985), who showed that search on arc-consistent trees works with-
out any backtracking. A similar result, with extensions to acyclic hypergraphs, was developed in the database community (Beeri ef al., 1983). Bayardo and Miranker (1994) present an algorithm for tree-structured CSPs that runs in linear time without any preprocessing. Dechter
(1990a) describes the cycle-cutset approach. Since those papers were published, there has been a great deal of progress in developing more general results relating the complexity of solving a CSP to the structure of its constraint
graph.
The notion of tree width was introduced by the graph theorists Robertson and Sey-
mour (1986). Dechter and Pearl (1987, 1989), building on the work of Freuder, applied a related notion (which they called induced width but is identical to tree width) to constraint
satisfaction problems and developed the tree decomposition approach sketched in Section 6.5.
Bibliographical and Historical Notes Drawing on this work and on results from database theory, Gottlob ez al. (1999a, 1999b)
developed a notion, hypertree width, that is based on the characterization of the CSP as a
hypergraph. In addition to showing that any CSP with hypertree width w can be solved in time O(n"*+!logn), they also showed that hypertree width subsumes all previously defined measures of “width” in the sense that there are cases where the hypertree width is bounded
and the other measures are unbounded.
The RELSAT algorithm of Bayardo and Schrag (1997) combined constraint learning and
backjumping and was shown to outperform many other algorithms of the time. This led to AND-OR search algorithms applicable to both CSPs and probabilistic reasoning (Dechter and Mateescu, 2007). Brown et al. (1988) introduce the idea of symmetry breaking in CSPs,
and Gent et al. (2006) give a survey.
The field of distributed constraint satisfaction looks at solving CSPs when there is a
collection of agents, each of which controls a subset of the constraint variables. There have
been annual workshops on this problem since 2000, and good coverage elsewhere (Collin et al., 1999; Pearce et al., 2008).
Comparing CSP algorithms is mostly an empirical science: few theoretical results show that one algorithm dominates another on all problems; instead, we need to run experiments to see which algorithms perform better on typical instances of problems. As Hooker (1995)
points out, we need to be careful to distinguish between competitive testing—as occurs in
competitions among algorithms based on run time—and scientific testing, whose goal is to identify the properties of an algorithm that determine its efficacy on a class of problems. The textbooks by Apt (2003), Dechter (2003), Tsang (1993), and Lecoutre (2009), and the collection by Rossi er al. (2006), are excellent resources on constraint processing. There
are several good survey articles, including those by Dechter and Frost (2002), and Bartdk et al. (2010). Carbonnel and Cooper (2016) survey tractable classes of CSPs.
Kondrak and
van Beek (1997) give an analytical survey of backtracking search algorithms, and Bacchus and van Run (1995) give a more empirical survey. Constraint programming is covered in the books by Apt (2003) and Fruhwirth and Abdennadher (2003). Papers on constraint satisfac-
tion appear regularly in Artificial Intelligence and in the specialist journal Constraints; the latest SAT solvers are described in the annual International SAT Competition. The primary
conference venue is the International Conference on Principles and Practice of Constraint
Programming, often called CP.
207
TR
[
LOGICAL AGENTS In which we design agents that can form representations ofa complex world, use a process. of inference to derive new representations about the world, and use these new representa-
tions to deduce what to do.
Knowledge-based agents.
Reasoning Representation
Humans, it seems, know things; and what they know helps them do things. In AI knowledge-
based agents use a process of reasoning over an internal representation of knowledge to
decide what actions to take.
The problem-solving agents of Chapters 3 and 4 know things, but only in a very limited, inflexible sense. They know what actions are available and what the result of performing a specific action from a specific state will be, but they don’t know general facts. A route-finding agent doesn’t know that it is impossible for a road to be a negative number of kilometers long.
An 8-puzzle agent doesn’t know that two tiles cannot occupy the same space. The knowledge they have is very useful for finding a path from the start to a goal, but not for anything else. The atomic representations used by problem-solving agents are also very limiting. In a partially observable environment, for example, a problem-solving agent’s only choice for representing what it knows about the current state is to list all possible concrete states. I could
give a human the goal of driving to a U.S. town with population less than 10,000, but to say
that to a problem-solving agent, I could formally describe the goal only as an explicit set of the 16,000 or so towns that satisfy the description.
Chapter 6 introduced our first factored representation, whereby states are represented as
assignments of values to variables; this is a step in the right direction, enabling some parts of
the agent to work in a domain-independent way and allowing for more efficient algorithms.
In this chapter, we take this step to its logical conclusion, so to speak—we develop logic as a
general class of representations to support knowledge-based agents. These agents can com-
bine and recombine information to suit myriad purposes. This can be far removed from the needs of the moment—as
when a mathematician proves a theorem or an astronomer calcu-
lates the Earth’s life expectancy.
Knowledge-based agents can accept new tasks in the form
new environment, the wumpus
world, and illustrates the operation of a knowledge-based
of explicitly described goals; they can achieve competence quickly by being told or learning new knowledge about the environment; and they can adapt to changes in the environment by updating the relevant knowledge. We begin in Section 7.1 with the overall agent design. Section 7.2 introduces a simple agent without going into any technical detail. Then we explain the general principles of logic
in Section 7.3 and the specifics of propositional logic in Section 7.4. Propositional logic is
a factored representation; while less expressive than first-order logic (Chapter 8), which is the canonical structured representation, propositional logic illustrates all the basic concepts
Section 7.1
Knowledge-Based Agents
209
function KB-AGENT(percept) returns an action
persistent: KB, a knowledge base 1, acounter, initially 0, indicating time TELL(KB, MAKE-PERCEPT-SENTENCE(percept, 1))
action ASK(KB, MAKE-ACTION-QUERY(1)) TELL(KB, MAKE-ACTION-SENTENCE(action, 1)) tertl return action
Figure 7.1 A generic knowledge-based agent. Given a percept, the agent adds the percept o0 its knowledge base, asks the knowledge base for the best action, and tells the knowledge base that it has in fact taken that action.
of logic. It also comes with well-developed inference technologies, which we describe in sections 7.5 and 7.6. Finally, Section 7.7 combines the concept of knowledge-based agents with the technology of propositional logic to build some simple agents for the wumpus world. 7.1
Knowledge-Based
Agents
The central component of a knowledge-based agent is its knowledge base, or KB. A knowledge base
is a set of sentences.
(Here “sentence” is used as a technical term.
It is related
but not identical to the sentences of English and other natural languages.) Each sentence is
Knowledge base Sentence
expressed in a language called a knowledge representation language and represents some
Knowledge representation language
from other sentences, we call it an axiom.
Axiom
assertion about the world. When the sentence is taken as being given without being derived
There must be a way to add new sentences to the knowledge base and a way to query
what is known.
The standard names for these operations are TELL and ASK, respectively.
Both operations may involve inference—that is, deriving new sentences from old. Inference Inference must obey the requirement that when one ASKs a question of the knowledge base, the answer
should follow from what has been told (or TELLed) to the knowledge base previously. Later in this chapter, we will be more precise about the crucial word “follow.” For now, take it to
mean that the inference process should not make things up as it goes along. Figure 7.1 shows the outline of a knowledge-based agent program. Like all our agents, it takes a percept as input and returns an action. The agent maintains a knowledge base, KB,
which may initially contain some background knowledge.
Each time the agent program is called, it does three things. First, it TELLS the knowledge
base what it perceives. Second, it ASKs the knowledge base what action it should perform. In
the process of answering this query, extensive reasoning may be done about the current state of the world, about the outcomes of possible action sequences, and so on. Third, the agent
program TELLs the knowledge base which action was chosen, and returns the action so that it can be executed.
The details of the representation language are hidden inside three functions that imple-
ment the interface between the sensors and actuators on one side and the core representation and reasoning system on the other.
MAKE-PERCEPT-SENTENCE
constructs a sentence as-
Background knowledge
210
Chapter 7 Logical Agents serting that the agent perceived the given percept at the given time. MAKE-ACTION-QUERY constructs a sentence that asks what action should be done at the current time. Finally, MAKE-ACTION-SENTENCE constructs a sentence asserting that the chosen action was executed. The details of the inference mechanisms are hidden inside TELL and ASK. Later
sections will reveal these details.
The agent in Figure 7.1 appears quite similar to the agents with internal state described
in Chapter 2. Because of the definitions of TELL and ASK, however, the knowledge-based
Knowledge level
agent iis not an arbitrary program for calculating actions. It is amenable to a description at the knowledge level, where we need specify only what the agent knows and what its goals are, in order to determine its behavior.
For example, an automated taxi might have the goal of taking a passenger from San
Francisco to Marin County and might know that the Golden Gate Bridge is the only link
between the two locations. Then we can expect it to cross the Golden Gate Bridge because it
Implementation level
knows that that will achieve its goal. Notice that this analysis is independent of how the taxi
works at the implementation level. It doesn’t matter whether its geographical knowledge is
implemented as linked lists or pixel maps, or whether it reasons by manipulating strings of symbols stored in registers or by propagating noisy signals in a network of neurons.
A knowledge-based agent can be built simply by TELLing it what it needs to know. Start-
Declarative Procedural
ing with an empty knowledge base, the agent designer can TELL sentences one by one until
the agent knows how to operate in its environment. This is called the declarative approach to system building. In contrast, the procedural approach encodes desired behaviors directly as program code. In the 1970s and 1980s, advocates of the two approaches engaged in heated debates.
We now understand that a successful agent often combines both declarative and
procedural elements in its design, and that declarative knowledge can often be compiled into more efficient procedural code.
We can also provide a knowledge-based agent with mechanisms that allow it to learn for
itself. These mechanisms, which are discussed in Chapter 19, create general knowledge about
the environment from a series of percepts. A learning agent can be fully autonomous.
7.2 Wumpus world
The Wumpus
World
In this section we describe an environment in which knowledge-based agents can show their
worth. The wumpus world is a cave consisting of rooms connected by passageways. Lurking
somewhere in the cave is the terrible wumpus, a beast that eats anyone who enters its room.
The wumpus can be shot by an agent, but the agent has only one arrow. Some rooms contain bottomless pits that will trap anyone who wanders into these rooms (except for the wumpus, which
is too big to fall in).
The only redeeming feature of this bleak environment is the
possibility of finding a heap of gold. Although the wumpus world is rather tame by modern computer game standards, it illustrates some important points about intelligence. A sample wumpus world is shown in Figure 7.2. The precise definition of the task environment is given, as suggested in Section 2.3, by the PEAS description: o Performance measure:
+1000 for climbing out of the cave with the gold, 1000
for
falling into a pit or being eaten by the wumpus, —1 for each action taken, and ~10 for
using up the arrow. The game ends either when the agent dies or when the agent climbs out of the cave.
Section7.2
4
| S8
L0
Zhreozs-
W= o
|sss START. 1
The Wumpus World
@)= o PIT
2
3
4
Figure 7.2 A typical wumpus world. The agent is in the botiom left corner, facing east (rightward). o Environment: A 4 x4 grid of rooms, with walls surrounding the grid. The agent always starts in the square labeled [1,1], facing to the east. The locations of the gold and the wumpus are chosen randomly, with a uniform distribution, from the squares other
than the start square. In addition, each square other than the start can be a pit, with probability 0.2. o Actuators: The agent can move Forward, TurnLeft by 90°, or TurnRight by 90°. The agent dies a miserable death if it enters a square containing a pit or a live wumpus. (It is safe, albeit smelly, to enter a square with a dead wumpus.) If an agent tries to move
forward and bumps into a wall, then the agent does not move. The action Grab can be
used to pick up the gold if it is in the same square as the agent. The action Shoot can
be used to fire an arrow in a straight line in the direction the agent is facing. The arrow continues until it either hits (and hence kills) the wumpus or hits a wall. The agent has
only one arrow, so only the first Shoot action has any effect. Finally, the action Climb can be used to climb out of the cave, but only from square [1,1].
o Sensors: The agent has five sensors, each of which gives a single bit of information: — In the squares directly (not diagonally) adjacent to the wumpus, the agent will — — — ~
perceive a Stench.!
In the In the When When where
squares directly adjacent to a pit, the agent will perceive a Breeze. square where the gold is, the agent will perceive a Glitter. an agent walks into a wall, it will perceive a Bump. the wumpus is killed, it emits a woeful Scream that can be perceived anyin the cave.
The percepts will be given to the agent program in the form of a list of five symbols;
for example, if there is a stench and a breeze, but no glitter, bump, or scream, the agent program will get [Stench, Breeze, None, None, None]. ! Presumably the square containing the wumpus also has a stench, but any agent entering that square is eaten before being able to perceive anything.
211
212
Chapter 7 Logical Agents 34
33
A
B G o » s v W
2 31
=Wumpus
14
24
34
44
13
23
33
43
12
22 P
32
42
11
0K v oK.
@
31 5
[A1
()
Figure 7.3 The first step taken by the agent in the wumpus world. (a) The initial situation, after percept [None, None, None, None, None]. (b) After moving to [2,1] and perceiving [None, Breeze, None, None, None]. We can characterize the wumpus environment along the various dimensions given in Chapter 2. Clearly, it is deterministic, discrete, static, and single-agent.
(The wumpus doesn’t
move, fortunately.) It is sequential, because rewards may come only after many actions are taken. Itis partially observable, because some aspects of the state are not directly perceivable:
the agent’s location, the wumpus'’s state of health, and the availability of an arrow. As for the
locations of the pits and the wumpus: we could treat them as unobserved parts of the state— in which case, the transition model for the environment is completely known, and finding the
locations of pits completes the agent’s knowledge of the state. Alternatively, we could say that the transition model itself is unknown because the agent doesn’t know which Forward
actions are fatal—in which case, discovering the locations of pits and wumpus completes the agent’s knowledge of the transition model.
For an agent in the environment, the main challenge is its initial ignorance of the configuration of the environment; overcoming this ignorance seems to require logical reasoning. In most instances of the wumpus world, it is possible for the agent to retrieve the gold safely.
Occasionally, the agent must choose between going home empty-handed and risking death to
find the gold. About 21% of the environments are utterly unfair, because the gold is in a pit or surrounded by pits. Let us watch a knowledge-based wumpus agent exploring the environment shown in Figure 7.2. We use an informal knowledge representation language consisting of writing down symbols in a grid (as in Figures 7.3 and 7.4). The agent’s initial knowledge base contains the rules of the environment, as described
previously; in particular, it knows that it is in [1,1] and that [1,1] is a safe square; we denote that with an “A” and “OK,” respectively, in square [1,1]. The first percept is [None, None, None, None, None], from which the agent can conclude that its neighboring squares, [1,2] and [2,1], are free of dangers—they are OK. Figure 7.3(a) shows the agent’s state of knowledge at this point.
Section7.2
14
24
3.4
44
13w
|28
33
4.3
22
3.2
4.2
14
3.4
44
13 1
33 p,
[43
12
3.2
4.2
i OK
B
3.1
P
4.1
v OK
1.1
v oK 2,1
W OK
(@)
»
s
v oK
0K 2.1
2.4
The Wumpus World
B
3,1
Pt
4.1
WL OK
®)
Figure 7.4 Two later stages in the progress of the agent. (a) After moving to [1.1] and then [1.2), and perceiving [Stench, None, None, None, None]. (b) After moving to [2.2] and then [2.3), and perceiving [Stench, Breeze, Glitter, None, None]. A cautious agent will move only into a square that it knows to be OK. Let us suppose
the agent decides to move forward to [2,1]. The agent perceives a breeze (denoted by “B”) in [2,1], 50 there must be a pit in a neighboring square. The pit cannot be in [1,1], by the rules of the game, so there must be a pit in [2,2] or [3,1] or both. The notation “P?” in Figure 7.3(b) indicates a possible pit in those squares. At this point, there is only one known square that is
OK and that has not yet been visited. So the prudent agent will turn around, go back to [1,1], and then proceed to [1,2]. The agent perceives a stench in [1,2], resulting in the state of knowledge shown in Figure 7.4(a). The stench in [1,2] means that there must be a wumpus nearby. But the wumpus cannot be in [1,1], by the rules of the game, and it cannot be in [2,2] (or the agent would have detected a stench when it was in [2,1]). Therefore, the agent can infer that the wumpus
is in [1,3]. The notation W' indicates this inference. Moreover, the lack of a breeze in [1,2] implies that there is no pit in [2,2]. Yet the agent has already inferred that there must be a pit in either [2,2] or [3,1], so this means it must be in [3,1]. This is a fairly difficult inference, because it combines knowledge gained at different times in different places and relies on the
lack of a percept to make one crucial step.
The agent has now proved to itself that there is neither a pit nor a wumpus in [2,2], so it
is OK to move there. We do not show the agent’s state of knowledge at [2,2]; we just assume
that the agent turns and moves to [2,3], giving us Figure 7.4(b). In [2,3], the agent detects a glitter, so it should grab the gold and then return home.
Note that in each case for which the agent draws a conclusion from the available infor-
mation, that conclusion is guaranteed to be correct if the available information is correct.
This is a fundamental property of logical reasoning. In the rest of this chapter, we describe
how to build logical agents that can represent information and draw conclusions such as those
described in the preceding paragraphs.
213
214
Chapter 7 Logical Agents 7.3
Logic
This section summarizes the fundamental concepts of logical representation and reasoning. These beautiful ideas are independent of any of logic’s particular forms. We therefore post-
pone the technical details of those forms until the next section, using instead the familiar
example of ordinary arithmetic. Syntax Semantics Truth Possible world Model
In Section 7.1, we said that knowledge bases consist of sentences. These sentences are expressed according to the syntax of the representation language, which specifies all the
sentences that are well formed. The notion of syntax is clear enough in ordinary arithmetic: “x+y=4"is a well-formed sentence, whereas “x4y+ =" is not.
A logic must also define the semantics, or meaning, of sentences. The semantics defines
the truth of each sentence with respect to each possible world. For example, the semantics
for arithmetic specifies that the sentence “x+y=4"is true in a world where x is 2 and y is 2, but false in a world where x is 1 and y is 1. In standard logics, every sentence must be either true or false in each possible world—there is no “in between.”
When we need to be precise, we use the term model in place of “possible world.” Whereas possible worlds might be thought of as (potentially) real environments that the agent
might or might not be in, models are mathematical abstractions, each of which has a fixed
truth value (true or false) for every relevant sentence. Informally, we may think of a possible world as, for example, having x men and y women sitting at a table playing bridge, and the sentence x + y=4 is true when there are four people in total. Formally, the possible models are just all possible assignments of nonnegative integers to the variables x and y. Each such
Satisfaction Entailment
assignment determines the truth of any sentence of arithmetic whose variables are x and y. If a sentence « is true in model m, we say that m satisfies o or sometimes m is a model of o
‘We use the notation M(«) to mean the set of all models of a.
Now that we have a notion of truth, we are ready to talk about logical reasoning. This in-
volves the relation of logical entailment between sentences—the idea that a sentence follows
logically from another sentence. In mathematical notation, we write
aks
to mean that the sentence « entails the sentence 3. The formal definition of entailment is this:
«a k= Bif and only if, in every model in which a is true, 4 s also true. Using the notation just introduced, we can write
a k= Bifand only if M(a) C M(B). (Note the direction of the C here: if o |= 3, then « is a stronger assertion than 3 it rules out more possible worlds.) The relation of entailment is familiar from arithmetic; we are happy with the idea that the sentence x = 0 entails the sentence xy = 0. Obviously, in any model
where x i zero, it is the case that xy is zero (regardless of the value of y). We can apply the same kind of analysis to the wumpus-world reasoning example given in the preceding section.
Consider the situation in Figure 7.3(b):
the agent has detected
nothing in [1,1] and a breeze in [2,1]. These percepts, combined with the agent’s knowledge
of the rules of the wumpus world, constitute the KB. The agent is interested in whether the
adjacent squares [1,2], [2,2], and [3,1] contain pits. Each of the three squares might or might
2 Fuzzy logic,
cussed in Chapter 13, allows for degrees of truth.
Section7.3
Logic
215
(@
Figure 7.5 Possible models for the presence of pits in squares [1,2], [2:2], and [3.1]. The KB corresponding to the observations of nothing in [1.1] and a breeze in [2,1]is shown by the solid line. (a) Dotted line shows models of a1 (no pit in [1.2]). (b) Dotted line shows models of a3 (no pit in [2.2]). not contain a pit, so (ignoring other aspects of the world for now) there are 23=8 possible models. These eight models are shown in Figure 7.5.3
The KB can be thought of as a set of sentences or as a single sentence that asserts all the individual sentences. The KB is false in models that contradict what the agent knows—
for example, the KB is false in any model in which [1,2] contains a pit, because there is no breeze in [1,1]. There are in fact just three models in which the KB is true, and these are shown surrounded by a solid line in Figure 7.5. Now let us consider two possible conclusions: ay = “There is no pit in [1,2].”
=
“There is no pit in [2,2]."
We have surrounded the models of a; and a2 with dotted lines in Figures 7.5(a) and 7.5(b), respectively. By inspection, we see the following: in every model in which KB s true, v is also true. Hence, KB = a: there is no pit in [1,2]. We can also see that in some models in which KB is true, a is false.
Hence, KB does not entail a: the agent cannot conclude that there is no pit in [2,2]. (Nor can it conclude that there is a pit in [2,2].)*
The preceding example not only illustrates entailment but also shows how the definition
of entailment can be applied to derive conclusions—that is, to carry out logical inference.
The inference algorithm illustrated in Figure 7.5 is called model checking, because it enu‘merates all possible models to check that a is true in all models in which KB is true, that is, that M(KB) C M(a). 3 Although the figure shows the models as partial wumpus worlds, they are really nothing more than assignments of true and false 1o the sentences “there s a pit in [1,2]” etc. Models, in the mathematical sense, do not need to have “orrible "airy wumpuses in them. 4 The agent can calculate the probability that there is a pit in [2.2]; Chapter 12 shows how.
Logical inference Model checking
Chapter 7 Logical Agents
216
Aspects of the ~~=~Follows =~ > Aspect of the " veaiworld atwond '
Figure 7.6 Sentences are physical configurations of the agent, and reasoning is a process of constructing new physical configurations from old ones. Logical reasoning should ensure that the new configurations represent aspects of the world that actually follow from the aspects that the old configurations represent.
In understanding entailment and inference, it might help to think of the set of all consequences of KB as a haystack and of a as a needle. Entailment is like the needle being in the haystack; inference is like finding it. This distinction is embodied in some formal notation: if an inference algorithm i can derive from KB, we write KBt a, Sound
Truth-preserving
which is pronounced “a is derived from KB by i or “i derives a from KB.” An inference algorithm that derives only entailed sentences is called sound or truth-
preserving. Soundness is a highly desirable property. An unsound inference procedure essentially makes things up as it goes along—it announces the discovery of nonexistent needles.
Itis easy to see that model checking, when it is applicable,’ is a sound procedure.
Completeness
The property of completeness is also desirable:
an inference algorithm is complete if
it can derive any sentence that is entailed. For real haystacks, which are finite in extent, it seems obvious that a systematic examination can always decide whether the needle is in
the haystack. For many knowledge bases, however, the haystack of consequences is infinite, and completeness becomes an important issue.® Fortunately, there are complete inference procedures for logics that are sufficiently expressive to handle many knowledge bases. We have described a reasoning process whose conclusions are guaranteed to be true in any world in which the premises are true; in particular, if KB is true in the real world, then any sentence o derived from KB by a sound inference procedure is also true in the real world. So, while an inference process operates on “syntax™—internal physical configurations such as bits in registers or patterns of electrical blips in brains—the process corresponds to the realworld relationship whereby some aspect of the real world is the case by virtue of other aspects
of the real world being the case.” This correspondence between world and representation is Grounding
illustrated in Figure 7.6.
|
The final
issue to consider is grounding—the connection between logical reasoning pro-
cesses and the real environment in which the agent exists. In particular, how do we know that
5 Model checking works if the space of models is finite—for example, in wumpus worlds of fixed size. For arithmetic, on the other hand, the space of models is infinite: even if we restrict ourselves to the integers, there are infinitely many pairs of values forx and y in the sentence x-+y = 4. 6 Compare with the case of infinite search spaces in Chapter 3, where depth-first search is not complete. 7 As Wittgenstein (1922) put it in his famous Tractatus: “The world is everything that is the case.”
Section7.4
Propositional Logic: A Very Simple Logic
217
KB is true in the real world? (After all, KB is just “syntax” inside the agent’s head.) This is a
philosophical question about which many, many books have been written. (See Chapter 27.) A simple answer is that the agent’s sensors create the connection. For example, our wumpusworld agent has a smell sensor. The agent program creates a suitable sentence whenever there is a smell. Then, whenever that sentence is in the knowledge base, it is true in the real world.
Thus, the meaning and truth of percept sentences are defined by the processes of sensing and sentence construction that produce them. What about the rest of the agent’s knowledge, such
as its belief that wumpuses cause smells in adjacent squares? This is not a direct representation of a single percept, but a general rule—derived, perhaps, from perceptual experience but not identical to a statement of that experience.
General rules like this are produced by
a sentence construction process called learning, which is the subject of Part V. Learning is
fallible. Tt could be the case that wumpuses cause smells except on February 29 in leap years, which is when they take their baths.
Thus, KB may not be true in the real world, but with
good learning procedures, there is reason for optimism. Propo:
nal Logic:
A Very
ple Log
We now present propositional logic. We describe its syntax (the structure of sentences) and Propositional logic its semantics
(the way in which the truth of sentences is determined). From these, we derive
a simple, syntactic algorithm for logical inference that implements the semantic notion of
entailment. Everything takes place, of course, in the wumpus world. 7.4.1
Syntax
The syntax of propositional logic defines the allowable sentences. The atomic sentences Atomic sentences consist of a single proposition symbol. Each such symbol stands for a proposition that can be true or false.
Proposition symbol
We use symbols that start with an uppercase letter and may contain other
letters or subscripts, for example: P, Q, R, Wy3 and FacingEast. The names are arbitrary but are often chosen to have some mnemonic value—we use W 3 to stand for the proposition that the wumpus is in [1,3]. (Remember that symbols such as W;3 are atomic, i.e., W, 1, and 3 are not meaningful parts of the symbol.) There are two proposition symbols with fixed meanings: True is the always-true proposition and False is the always-false proposition. Complex sentences are constructed from simpler sentences, using parentheses and operators Complex sentences
called logical connectives. There are five connectives in common use:
Logical connectives
~ (not). A sentence such as =W, 3 is called the negation of W,5. A literal is cither an Negation Literal atomic sentence (a positive literal) or a negated atomic sentence (a negative literal). A (and). A sentence whose main connective is A, such as Wy 3 A Py 1, is called a conjunction; its parts are the conjuncts. (The A looks like an “A” for “And.”)
V (or). A sentence whose main connective is V, such as (Wi3 A Ps,1) V Wao, is a disjunc-
Conjunction
Disjunction tion; its parts are disjuncts—in this example, (Wj3 A Py;) and Wa. = (implies). A sentence such as (W3 APs;) = —Way is called an implication (or con- Implication ditional). Its premise or antecedent is (Wi3 A Ps.1), and its conclusion or consequent is =Wa.
Implications are also known as rules or if-then statements. The implication
symbol is sometimes written in other books as D or —.
< (if and only if). The sentence W; 3 —W, is a biconditional.
Premise Conclusion Rules Biconditional
218
Chapter 7 Logical Agents Sentence
—
AtomicSentence ComplexSentence
— —
True| False| P| Q| R| ... ( Sentence)
|
- Sentence
|
Sentence A Sentence
|
Sentence V Sentence
|
Sentence
=
Sentence
|
Sentence
=8 possible models—exactly those depicted in Figure 7.5. Notice, however, that the models are purely mathematical objects with no
necessary connection to wumpus worlds. P> is just a symbol; it might mean “there is a pit in [1,2]” or “I'm in Paris today and tomorrow.”
The semantics for propositional logic must specify how to compute the truth value of any
sentence, given a model. This is done recursively. All sentences are constructed from atomic
sentences and the five connectives; therefore, we need to specify how to compute the truth
of atomic sentences and how to compute the truth of sentences formed with each of the five
connectives. Atomic sentences are easy:
« True is true in every model and False is false in every model.
* The truth value of every other proposition symbol must be specified directly in the
model. For example, in the model m, given earlier, Py ; is false.
Section7.4 Propositional Logic: A Very Simple Logic
219
Figure 7.8 Truth tables for the five logical connectives. To use the table to compute, for example, the value of PVQ when P is true and Q is false, first look on the left for the row where P is true and Q is false (the third row). Then look in that row under the PV Q column
to see the result: true.
For complex sentences, we have five rules, which hold for any subsentences P and Q (atomic or complex) in any model m (here “iff” means “if and only if”): « =Pis true iffP s false in m. « PAQis true iff both P and Q are true in m. « PV Qis true iff either P or Q is truc in m. « P= Qis true unless P is true and Q is false in m. « P& Qis true iff P and Q are both true or both false in m. The rules can also be expressed with truth tables that specify the truth value of a complex
sentence for each possible assignment of truth values to its components. Truth tables for the
five connectives are given in Figure 7.8. From these tables, the truth value of any sentence s can be computed with respect to any model m by a simple recursive evaluation.
For ex-
ample, the sentence —P3 A (P2 V Py 1), evaluated in my, gives true A (false\/ true) = trueA true =true. Exercise 7.TRUV asks you to write the algorithm PL-TRUE?(s, m), which com-
putes the truth value of a propositional logic sentence s in a model 1.
The truth tables for “and,” “or,” and “not” are in close accord with our intuitions about
the English words. The main point of possible confusion is that P or Q is true or both.
Q is true when P is true
A different connective, called “exclusive or” (“xor” for short), yields
false when both disjuncts are true.® There is no consensus on the symbol for exclusive or;
some choices are \/ or # or &
The truth table for = may not quite fit one’s intuitive understanding of “P implies Q” or “if P then Q. For one thing, propositional logic does not require any relation of causation or relevance between P and Q. The sentence “5 is odd implies Tokyo is the capital of Japan” is a true sentence of propositional logic (under the normal interpretation), even though it is a decidedly odd sentence of English. Another point of confusion is that any implication is true whenever its antecedent is false. For example, “5 is even implies Sam is smart” is true, regardless of whether Sam is smart. This seems bizarre, but it makes sense if you think of
“P = Q" as saying, “If P is true, then I am claiming that Q is true; otherwise I am making no claim.” The only way for this sentence to be false is if P is true but Q is false. The biconditional, P < @, is true whenever both P => Q and Q = P are true. In English,
this is often written as “P if and only if Q.” Many of the rules of the wumpus world are best
8 Latin u;
two separate words:
el is inclusive or and “aut” is exclusive or.
Truth table
220
Chapter 7 Logical Agents
written using
For this reason, we have used a special notation—the double-boxed link—in Figure 10.4.
This link asserts that
Vx x€Persons
=
[Vy HasMother(x,y)
= y& FemalePersons).
‘We might also want to assert that persons have two legs—that is, Vx x€Persons
= Legs(x,2).
As before, we need to be careful not to assert that a category has legs; the single-boxed link
in Figure 10.4 is used to assert properties of every member of a category.
The semantic network notation makes it convenient to perform inheritance reasoning of
the kind introduced in Section 10.2. For example, by virtue of being a person, Mary inherits the property of having two legs. Thus, to find out how many legs Mary has, the inheritance
algorithm follows the MemberOf link from Mary to the category she belongs to, and then S Several early systems failed to distinguish between properties of members of a category and properties of the category as a whole. Thi istencies, as pointed out by Drew McDermott (1976) in his article “Artificial Intelligence Meets Natural Stupidity.” Another common problem was the use of /s links for both subset and membership relations, in correspondence with English usage: “a cat isa mammal” and “Fif is 4 cat” See Exercise 10.NATS for more on these is ues.
330
Chapter 10 Knowledge Representation
Tastioter
&5 Subser0y
ez
Persons )—=2 \uhm(l/ Subser0) Memberf
sisier0f
Menberof
Figure 10.4 A semantic network with four objects (John, Mary, 1, and 2) and four categories. Relations are denoted by labeled links.
Menberof Agent
During
Figure 10.5 A fragment of a semantic network showing the representation of the logical assertion Fly(Shankar, NewYork, NewDelhi, Yesterday). follows SubsetOf links up the hierarchy until it finds a category for which there is a boxed Legs link—in this case, the Persons category. The simplicity and efficiency of this inference mechanism, compared with semidecidable logical theorem proving, has been one of the main attractions of semantic networks.
Inheritance becomes complicated when an object can belong to more than one category
Multiple inheritance
or when a category can be a subset of more than one other category; this is called multiple inheritance. In such cases, the inheritance algorithm might find two or more conflicting values
answering the query. For this reason, multiple inheritance is banned in some object-oriented
programming (OOP) languages, such as Java, that use inheritance in a class hierarchy. It is usually allowed in semantic networks, but we defer discussion of that until Section 10.6.
The reader might have noticed an obvious drawback of semantic network notation, com-
pared to first-order logic: the fact that links between bubbles represent only binary relations.
For example, the sentence Fly(Shankar,NewYork, NewDelhi, Yesterday) cannot be asserted
directly in a semantic network.
Nonetheless, we can obtain the effect of n-ary assertions
by reifying the proposition itself as an event belonging to an appropriate event category.
Figure 10.5 shows the semantic network structure for this particular event.
Notice that the
restriction to binary relations forces the creation of a rich ontology of reified concepts.
Section 105
Reasoning Systems for Categories
331
Reification of propositions makes it possible to represent every ground, function-free
atomic sentence of first-order logic in the semantic network notation. Certain kinds of univer-
sally quantified sentences can be asserted using inverse links and the singly boxed and doubly boxed arrows applied to categories, but that still leaves us a long way short of full first-order logic.
Negation, disjunction, nested function symbols, and existential quantification are all
missing. Now it is possible to extend the notation to make it equivalent to first-order logic—as
in Peirce’s existential graphs—but doing so negates one of the main advantages of semantic
networks, which s the simplicity and transparency of the inference processes. Designers can build a large network and still have a good idea about what queries will be efficient, because (a) it is easy to visualize the steps that the inference procedure will go through and (b) in some cases the query language is so simple that difficult queries cannot be posed.
In cases where the expressive power proves to be too limiting, many semantic network
systems provide for procedural attachment to fill in the gaps.
Procedural attachment is a
technique whereby a query about (or sometimes an assertion of) a certain relation results in a
Procedural attachment
call to a special procedure designed for that relation rather than a general inference algorithm.
One of the most important aspects of semantic networks is their ability to represent default values for categories. Examining Figure 10.4 carefully, one notices that John has one Default value leg, despite the fact that he is a person and all persons have two legs. In a strictly logical KB, this would be a contradiction, but in a semantic network, the assertion that all persons have
two legs has only default status; that is, a person is assumed to have two legs unless this is contradicted by more specific information. The default semantics is enforced naturally by the inheritance algorithm, because it follows links upwards from the object itself (John in this
case) and stops as soon as it finds a value. We say that the default is overridden by the more
specific value. Notice that we could also override the default number of legs by creating a category of OneLeggedPersons, a subset of Persons of which John is a member. ‘We can retain a strictly logical semantics for the network if we say that the Legs
Overriding
rtion
for Persons includes an exception for John:
Vx x€Persons Ax # John = Legs(x,2). For a fixed network, this is semantically adequate but will be much less concise than the network notation itself if there are lots of exceptions. For a network that will be updated with
more assertions, however, such an approach fails—we really want to say that any persons as
yet unknown with one leg are exceptions t0o. Section 10.6 goes into more depth on this issue and on default reasoning in general. 10.5.2
Description logics
The syntax of first-order logic is designed to make it easy to say things about objects. Description logics are notations that are designed to make it easier to describe definitions
and
properties of categories. Description logic systems evolved from semantic networks in re-
Description logic
sponse to pressure to formalize what the networks mean while retaining the emphasis on
taxonomic structure as an organizing principle. The principal inference tasks for description logics are subsumption (checking if one Subsumption category is a subset of another by comparing their definitions) and classification (checking Classification whether an object belongs to a category). Some systems also include consistency of a cate-
gory definition—whether the membership criteria are logically satisfiable.
Consistency
332
Chapter 10 Knowledge Representation Concept
—
Thing| ConceptName
| And(Concept,...) | All(RoleName, Concept) | AtLeast(Integer, RoleName) | AtMost(Integer,RoleName)
| Fills(RoleName, IndividualName, . |
| Path
ConceptName RoleName
—
SameAs(Path, Path)
OneOf(IndividualName, . [RoleName, ...
— Adult| Female| Male| — Spouse| Daughter | Son| ...
Figure 10.6 The syntax of descriptions in a subset of the CLASSIC language.
The CLASSIC language (Borgida er al., 1989) is a typical description logic. The syntax
of CLASSIC descriptions is shown in Figure 10.6.% unmarried adult males we would write
For example, to say that bachelors are
Bachelor = And(Unmarried, Adult, Male) The equivalent in first-order logic would be Bachelor(x) < Unmarried(x) AAdult(x) A Male(x). Notice that the description logic has an algebra of operations on predicates, which of course
we can’t do in first-order logic. Any description in CLASSIC can be translated into an equiv-
alent first-order sentence, but some descriptions are more straightforward in CLASSIC.
For
example, to describe the set of men with at least three sons who are all unemployed and married to doctors, and at most two daughters who are all professors in physics or math
departments, we would use
And(Man, AtLeast(3, Son), AtMost(2, Daughter), All(Son, And(Unemployed, Married All(Spouse, Doctor))), All(Daughter, And(Professor, Fills(Department, Physics, Math)))) .
‘We leave it as an exercise to translate this into first-order logic.
Perhaps the most important aspect of description logics s their emphasis on tractability of inference. A problem instance is solved by describing it and then asking if it is subsumed by one of several possible solution categories. In standard first-order logic systems, predicting the solution time is often impossible. It is frequently left to the user to engineer the represen-
tation to detour around sets of sentences that seem to be causing the system to take several
© Notice that the language does not allow one to simply state that one concept, or category, is a subset of another. This is a deliberate policy: subsumption between categories must be derivable from some aspects of the descriptions of the categories. If not, then something is missing from the descriptions.
Section 10.6
Reasoning with Default Information
333
weeks to solve a problem. The thrust in description logics, on the other hand, is to ensure that
subsumption-testing can be solved in time polynomial in the size of the descriptions.”
This sounds wonderful in principle, until one realizes that it can only have one of two
consequences: either hard problems cannot be stated at all, or they require exponentially
large descriptions! However, the tractability results do shed light on what sorts of constructs
cause problems and thus help the user to understand how different representations behave. For example, description logics usually lack negation and disjunction. Each forces firstorder logical systems to go through a potentially exponential case analysis in order to ensure completeness.
CLASSIC allows only a limited form of disjunction in the Fills and OneOf
constructs, which permit disjunction over explicitly enumerated individuals but not over descriptions. With disjunctive descriptions, nested definitions can lead easily to an exponential
number of alternative routes by which one category can subsume another. 10.6
Reasoning with Default Information
In the preceding section, we saw a simple example of an assertion with default status: people have two legs. This default can be overridden by more specific information, such as that Long John Silver has one leg. We saw that the inheritance mechanism in semantic networks
implements the overriding of defaults in a simple and natural way. In this section, we study
defaults more generally, with a view toward understanding the semantics of defaults rather than just providing a procedural mechanism. 10.6.1
Circumscription and default logic
‘We have seen two examples of reasoning processes that violate the monotonicity property of
logic that was proved in Chapter 7.5 In this chapter we saw that a property inherited by all members of a category in a semantic network could be overridden by more specific informa-
Monotonicity
tion for a subcategory. In Section 9.4.4, we saw that under the closed-world assumption, ifa
proposition « is not mentioned in KB then KB }= —a, but KBA o |= ..
Simple introspection suggests that these failures of monotonicity are widespread in com-
monsense reasoning. It seems that humans often “jump to conclusions.” For example, when
one sees a car parked on the street, one is normally willing to believe that it has four wheels
even though only three are visible. Now, probability theory can certainly provide a conclusion
that the fourth wheel exists with high probability; yet, for most people, the possibility that the
car does not have four wheels will not arise unless some new evidence presents itself. Thus,
it seems that the four-wheel conclusion is reached by default, in the absence of any reason to
doubt it. If new evidence arrives—for example, if one sees the owner carrying a wheel and notices that the car is jacked up—then the conclusion can be retracted. This kind of reasoning
is said to exhibit nonmonotonicity, because the set of beliefs does not grow monotonically over time as new evidence arrives. Nonmonotonic logics have been devised with modified notions of truth and entailment in order to capture such behavior. We will look at two such
logics that have been studied extensively: circumscription and default logic.
Circumscription can be seen as a more powerful and precise version of the closed-world
7 CLASSIC provides efficient subsumption testing in practice, but the worst-case run time is exponential. 8 Recall that monotonicity requires all entailed sentences to remain entailed after new sentences are added to the KB. Thatis, if KB |= o then KBA 5 = .
Nonmonotonicity
Nonmonotonic logic Circumscription
334
Chapter 10 Knowledge Representation assumption. The idea is to specify particular predicates that are assumed to be “as false as possible”—that is, false for every object except those for which they are known to be true.
For example, suppose we want to assert the default rule that birds fly. We would introduce a predicate, say Abnormal, (x), and write Bird(x) A —Abnormaly (x) = Flies(x). If we say that Abnormal;
is to be circumscribed, a circumscriptive reasoner is entitled to
assume —Abnormal, (x) unless Abnormal; (x) is known to be true. This allows the conclusion
Model preference
Flies(Tweety) to be drawn from the premise Bird(Tweety), but the conclusion no longer holds if Abnormal, (Tweety) is asserted. Circumscription can be viewed as an example of a model preference logic.
In such
logics, a sentence is entailed (with default status) if it is true in all preferred models of the KB,
as opposed to the requirement of truth in all models in classical logic. For circumscription,
one model is preferred to another if it has fewer abnormal objects.” Let us see how this idea
works in the context of multiple inheritance in semantic networks. The standard example for
which multiple inheritance is problematic is called the “Nixon diamond.” It arises from the observation that Richard Nixon was both a Quaker (and hence by default a pacifist) and a
Republican (and hence by default not a pacifist). We can write this as follows: Republican(Nixon) A Quaker(Nixon). Republican(x) A ~Abnormaly(x) = —Pacifist(x).
Quaker(x) A —~Abnormals(x) = Pacifist(x).
If we circumscribe Abnormaly and Abnormals, there are two preferred models: one in which
Abnormaly(Nixon) and Pacifist(Nixon) are true and one in which Abnormals(Nixon) and —Pacifist(Nixon) are true. Thus, the circumscriptive reasoner remains properly agnostic as Prioritized circumscription
Default logic Default rules
to whether Nixon was a pacifist. If we wish, in addition, to assert that religious beliefs
take
precedence over political beliefs, we can use a formalism called prioritized circumscription to give preference to models where Abnormals is minimized.
Default logic is a formalism in which default rules can be written to generate contingent,
nonmonotonic conclusions. A default rule looks like this:
Bird(x) : Flies(x)/Flies(x). This rule means that if Bird(x) is true, and if Flies(x) is consistent with the knowledge base, then Flies(x) may be concluded by default. In general, a default rule has the form P:li,...,Jn/C
where P s called the prerequisite, C is the conclusion, and J; are the justifications—if any one
of them can be proven false, then the conclusion cannot be drawn. Any variable that appears
in J; or C must also appear in P. The Nixon-diamond example can be represented in default logic with one fact and two default rules:
Republican(Nixon) A Quaker(Nixon).
Republican(x) : ~Pacifist(x) | ~Pacifist(x) . Quaker(x) : Pacifist(x) /Pacifist(x) mption, one model s preferred to another if it has fewer true atoms—that i, preferred models are minimal models. There is a natural connection between the closed-world assumption and definiteclause KBs, because the fixed point reached by forward chaining on definite-clause KBs is the unique minimal model. See page 231 for more on this 9 Forthe closed-world
Section 10.6
Reasoning with Default Information
To interpret what the default rules mean, we define the notion of an extension of a default
335 Extension
theory to be a maximal set of consequences of the theory. That is, an extension S consists
of the original known facts and a set of conclusions from the default rules, such that no additional conclusions can be drawn from S, and the justifications of every default conclusion
in S are consistent with S. As in the case of the preferred models in circumscription, we have
two possible extensions for the Nixon diamond: one wherein he is a pacifist and one wherein he is not. Prioritized schemes exist in which some default rules can be given precedence over
others, allowing some ambiguities to be resolved. Since 1980, when nonmonotonic logics were first proposed, a great deal of progress
has been made in understanding their mathematical properties. There are still unresolved questions, however. For example, if “Cars have four wheels” is false, what does it mean to have it in one’s knowledge base? What is a good set of default rules to have? If we cannot
decide, for each rule separately, whether it belongs in our knowledge base, then we have a serious problem of nonmodularity.
Finally, how can beliefs that have default status be used
to make decisions? This is probably the hardest issue for default reasoning.
Decisions often involve tradeoffs, and one therefore needs to compare the strengths of be-
lief in the outcomes of different actions, and the costs of making a wrong decision. In cases
where the same kinds of decisions are being made repeatedly, it is possible to interpret default rules as “threshold probability” statements. For example, the default rule “My brakes are always OK” really means “The probability that my brakes are OK, given no other information, is sufficiently high that the optimal decision is for me to drive without checking them.” When
the decision context changes—for example, when one is driving a heavily laden truck down a steep mountain road—the default rule suddenly becomes inappropriate, even though there is no new evidence of faulty brakes. These considerations have led researchers to consider how
to embed default reasoning within probability theory or utility theory.
10.6.2
Truth maintenance systems
‘We have seen that many of the inferences drawn by a knowledge representation system will
have only default status, rather than being absolutely certain. Inevitably, some of these inferred facts will turn out to be wrong and will have to be retracted in the face of new infor-
mation. This process is called belief revision.!® Suppose that a knowledge base KB contains a sentence P—perhaps a default conclusion recorded by a forward-chaining algorithm, or perhaps just an incorrect assertion—and we want to execute TELL(KB, —P).
ating a contradiction, we must first execute RETRACT(KB, P).
To avoid cre-
This sounds easy enough.
Problems arise, however, if any additional sentences were inferred from P and asserted in the KB. For example, the implication P = Q might have been used to add Q. The obvious “solution”—retracting all sentences inferred from P—fails because such sentences may have other justifications besides P. For example, ifR and R = Q are also in the KB, then Q does not have to be removed after all. Truth maintenance systems, or TMSs, are designed to
handle exactly these kinds of complications. One simple approach to truth maintenance is to keep track of the order in which sentences are told to the knowledge
base by numbering
Belief revision
them from P; to P,.
When
the call
10 Belief revision is often contrasted with belief update, which occurs when a knowledge base is revised to reflect a change in the world rather than new information about a fixed world. Belief update combines belief revis with reasoning about time and change; it is also related to the process of filtering described in Chapter 14.
Truth maintenance system
336
Chapter 10 Knowledge Representation RETRACT(KB,
JT™S Justification
P,) is made, the system reverts to the state just before P; was added, thereby
removing both P, and any inferences that were derived from P, The sentences P through P, can then be added again. This is simple, and it guarantees that the knowledge base will be consistent, but retracting P; requires retracting and reasserting n — i sentences as well as undoing and redoing all the inferences drawn from those sentences. For systems to which many facts are being added—such as large commercial databas is impractical. A more efficient approach is the justification-based truth maintenance system, or JTMS.
In a JTMS, each sentence in the knowledge base is annotated with a justification consisting
of the set of sentences from which it was inferred. For example, if the knowledge base already contains P = Q, then TELL(P) will cause Q to be added with the justification {P, P = Q}.
In general, a sentence can have any number of justifications. Justifications make retraction efficient. Given the call RETRACT(P), the JTMS will delete exactly those sentences for which P is a member of every justification. So, if a sentence Q had the single justification {P. P =
Q}, it would be removed; if it had the additional justification {P, PVR
would still be removed; but if it also had the justification {R, PVR
=
Q}, it
= Q}, then it would
be spared. In this way, the time required for retraction of P depends only on the number of sentences derived from P rather than on the number of sentences added after P.
The JTMS assumes that sentences that are considered once will probably be considered
again, so rather than deleting a sentence from the knowledge base entirely when it loses
all justifications, we merely mark the sentence as being our of the knowledge base. If a subsequent assertion restores one of the justifications, then we mark the sentence as being back in.
In this way, the JTMS
retains all the inference chains that it uses and need not
rederive sentences when a justification becomes valid again.
In addition to handling the retraction of incorrect information, TMSs
can be used to
speed up the analysis of multiple hypothetical situations. Suppose, for example, that the Romanian Olympic Committee is choosing sites for the swimming, athletics, and equestrian
events at the 2048 Games to be held in Romania.
For example, let the first hypothesis be
Site(Swimming, Pitesti), Site(Athletics, Bucharest), and Site(Equestrian,Arad). A great deal of reasoning must then be done to work out the logistical consequences and hence
the desirability of this selection.
If we
want to consider Site(Athletics, Sibiu)
instead, the TMS avoids the need to start again from scratch. Instead, we simply retract Site(Athletics, Bucharest) and as ert Site(Athletics, Sibiu) and the TMS takes care of the necessary revisions. Inference chains generated from the choice of Bucharest can be reused with
ATMS
Sibiu, provided that the conclusions are the same.
An assumption-based truth maintenance system, or ATMS,
makes this type of context-
switching between hypothetical worlds particularly efficient. In a JTMS, the maintenance of
justifications allows you to move quickly from one state to another by making a few retrac-
tions and assertions, but at any time only one state is represented. An ATMS represents all the states that have ever been considered at the same time. Whereas a JTMS simply labels each
sentence as being in or out, an ATMS keeps track, for each sentence, of which assumptions
would cause the sentence to be true. In other words, each sentence has a label that consists of
a set of assumption sets. The sentence is true just in those cases in which all the assumptions
Explanation
in one of the assumption sets are true. Truth maintenance systems also provide a mechanism for generating explanations. Tech-
nically, an explanation of a sentence P is a set of sentences E such that E entails P. If the
Summary sentences in E are already known to be true, then E simply provides a sufficient basis for proving that P must be the case. But explanations can also include assumptions—sentences
that are not known to be true, but would suffice to prove P if they were true. For example, if your car won’t start, you probably don’t have enough information to definitively prove the
reason for the problem. But a reasonable explanation might include the assumption that the
battery is dead. This, combined with knowledge of how cars operate, explains the observed nonbehavior.
In most cases, we will prefer an explanation E that is minimal, meaning that
there is no proper subset of E that is also an explanation. An ATMS can generate explanations for the “car won’t start” problem by making assumptions (such as “no gas in car” or “battery
dead”) in any order we like, even if some assumptions are contradictory. Then we look at the
label for the sentence “car won’t start™ to read off the sets of assumptions that would justify the sentence. The exact algorithms used to implement truth maintenance systems are a little compli-
cated, and we do not cover them here. The computational complexity of the truth maintenance. problem is at least as great as that of propositional inference—that is, NP-hard.
Therefore,
you should not expect truth maintenance to be a panacea. When used carefully, however, a TMS can provide a substantial increase in the ability of a logical system to handle complex environments and hypotheses.
Summary
By delving into the details of how one represents a variety of knowledge, we hope we have given the reader a sense of how real knowledge bases are constructed and a feeling for the interesting philosophical issues that arise. The major points are as follows: + Large-scale knowledge representation requires a general-purpose ontology to organize
and tie together the various specific domains of knowledge. « A general-purpose ontology needs to cover a wide variety of knowledge and should be capable, in principle, of handling any domain. « Building a large, general-purpose ontology is a significant challenge that has yet to be fully realized, although current frameworks seem to be quite robust.
« We presented an upper ontology based on categories and the event calculus. We covered categories, subcategories, parts, structured objects, measurements, substances,
events, time and space, change, and beliefs. « Natural kinds cannot be defined completely in logic, but properties of natural kinds can be represented. * Actions, events, and time can be represented with the event calculus.
Such represen-
tations enable an agent to construct sequences of actions and make logical inferences
about what will be true when these actions happen.
* Special-purpose representation systems, such as semantic networks and description logics, have been devised to help in organizing a hierarchy of categories. Inheritance is an important form of inference, allowing the properties of objects to be deduced from
their membership in categories.
337
Assumption
338
Chapter 10 Knowledge Representation « The closed-world assumption, as implemented in logic programs, provides a simple way to avoid having to specify lots of negative information. default that can be overridden by additional information.
It is best interpreted as a
+ Nonmonotonic logics, such as circumscription and default logic, are intended to cap-
ture default reasoning in general.
+ Truth maintenance systems handle knowledge updates and revisions efficiently.
« It is difficult to construct large ontologies by hand; extracting knowledge from text makes the job easier. Bibliographical and Historical Notes Briggs (1985) claims that knowledge representation research began with first millennium BCE
Indian theorizing about the grammar of Shastric Sanskrit. Western philosophers trace their work on the subject back to c. 300 BCE in Aristotle’s Metaphysics (literally, what comes after
the book on physics). The development of technical terminology in any field can be regarded as a form of knowledge representation.
Early discussions of representation in Al tended to focus on “problem representation”
rather than “knowledge representation.” (See, for example, Amarel’s (1968) discussion of the “Mi naries and Cannibals” problem.) In the 1970s, Al emphasized the development of
“expert systems” (also called “knowledge-based systems”) that could, if given the appropriate domain knowledge, match or exceed the performance of human experts on narrowly defined tasks. For example, the first expert system, DENDRAL (Feigenbaum ef al., 1971; Lindsay
et al., 1980), interpreted the output of a mass spectrometer (a type of instrument used to ana-
Iyze the structure of organic chemical compounds) as accurately as expert chemists. Although
the success of DENDRAL was instrumental in convincing the Al research community of the importance of knowledge representation, the representational formalisms used in DENDRAL
are highly specific to the domain of chemistry.
Over time, researchers became interested in standardized knowledge representation for-
malisms and ontologies that could assist in the creation of new expert systems. This brought
them into territory previously explored by philosophers of science and of language. The discipline imposed in Al by the need for one’s theories to “work” has led to more rapid and deeper progress than when these problems were the exclusive domain of philosophy (although it has at times also led to the repeated reinvention of the wheel). But to what extent can we trust expert knowledge?
As far back as 1955, Paul Meehl
(see also Grove and Meehl, 1996) studied the decision-making processes of trained experts at subjective tasks such as predicting the success of a student in a training program or the recidivism of a criminal.
In 19 out of the 20 studies he looked at, Meehl found that simple
statistical learning algorithms (such as linear regression or naive Bayes) predict better than the experts. Tetlock (2017) also studies expert knowledge and finds it lacking in difficult cases. The Educational Testing Service has used an automated program to grade millions of essay questions on the GMAT exam since 1999. The program agrees with human graders 97%
of the time, about the same level that two human graders agree (Burstein ez al., 2001).
(This does not mean the program understands essays, just that it can distinguish good ones from bad ones about as well as human graders can.)
Bibliographical and Historical Notes The creation of comprehensive taxonomies or classifications dates back to ancient times.
Aristotle (384-322 BCE) strongly emphasized classification and categorization schemes. His
Organon, a collection of works on logic assembled by his students after his death, included a
treatise called Categories in which he attempted to construct what we would now call an upper ontology. He also introduced the notions of genus and species for lower-level classification. Our present system of biological classification, including the use of “binomial nomenclature” (classification via genus and species in the technical sense), was invented by the Swedish biologist Carolus Linnaeus, or Carl von Linne (1707-1778).
The problems associated with
natural kinds and inexact category boundaries have been addressed by Wittgenstein (1953), Quine (1953), Lakoff (1987), and Schwartz (1977), among others. See Chapter 24 for a discussion of deep neural network representations of words and concepts that escape some of the problems ofa strict ontology, but also sacrifice some of the precision.
We still don’t know the best way to combine the advantages of neural networks
and logical semantics for representation. Interest in larger-scale ontologies is increasing, as documented by the Handbook on Ontologies (Staab, 2004).
The OPENCYC
project (Lenat and Guha,
1990; Matuszek er al.,
2006) has released a 150,000-concept ontology, with an upper ontology similar to the one in Figure 10.1 as well as specific concepts like “OLED Display” and “iPhone,” which is a type of “cellular phone,” which in turn is a type of “consumer electronics,” “phone,” “wireless communication device,” and other concepts. The NEXTKB project extends CYC and other resources including FrameNet and WordNet into a knowledge base with almost 3 million
facts, and provides a reasoning engine, FIRE to go with it (Forbus et al., 2010). The DBPEDIA
project extracts structured data from Wikipedia,
specifically from In-
foboxes: the attribute/value pairs that accompany many Wikipedia articles (Wu and Weld, 2008; Bizer et al., 2007).
As of 2015, DBPEDIA
contained 400 million facts about 4 mil-
lion objects in the English version alone; counting all 110 languages yields 1.5 billion facts (Lehmann et al., 2015). The IEEE working group P1600.1 created SUMO, the Suggested Upper Merged Ontology (Niles and Pease, 2001; Pease and Niles, 2002), with about 1000 terms in the upper ontology and links to over 20,000 domain-specific terms. Stoffel ef al. (1997) describe algorithms for efficiently managing a very large ontology. A survey of techniques for extracting knowledge from Web pages is given by Etzioni ef al. (2008). On the Web, representation languages are emerging. RDF (Brickley and Guha, 2004)
allows for assertions to be made in the form of relational triples and provides some means for
evolving the meaning of names over time. OWL (Smith ef al., 2004) is a description logic that supports inferences over these triples. So far, usage seems to be inversely proportional
to representational complexity: the traditional HTML and CSS formats account for over 99% of Web content, followed by the simplest representation schemes, such as RDFa (Adida and Birbeck, 2008), and microformats (Khare, 2006; Patel-Schneider, 2014) which use HTML
and XHTML markup to add attributes to text on web pages. Usage of sophisticated RDF and OWL ontologies is not yet widespread, and the full vision of the Semantic Web (BernersLee et al., 2001) has not been realized. The conferences on Formal Ontology in Information
Systems (FOIS) covers both general and domain-specific ontologies.
The taxonomy used in this chapter was developed by the authors and is based in part
on their experience in the CYC project and in part on work by Hwang and Schubert (1993)
339
Chapter 10 Knowledge Representation and Davis (1990, 2005). An inspirational discussion of the general project of commonsense
knowledge representation appears in Hayes’s (1978, 1985b) “Naive Physics Manifesto.” Successful deep ontologies within a specific field include the Gene Ontology project (Gene Ontology Consortium, 2008) and the Chemical Markup Language (Murray-Rust ef al., 2003). Doubts about the feasibility of a single ontology for all knowledge are expressed by Doctorow (2001), Gruber (2004), Halevy et al. (2009), and Smith (2004).
The event calculus was introduced by Kowalski and Sergot (1986) to handle continuous time, and there have been several variations (Sadri and Kowalski, 1995; Shanahan, 1997) and overviews (Shanahan, 1999; Mueller, 2006). James Allen introduced time intervals for the same reason (Allen, 1984), arguing that intervals were much more natural than situations for reasoning about extended and concurrent events. In van Lambalgen and Hamm (2005) we see how the logic of events maps onto the language we use to talk about events. An alternative to the event and situation calculi is the fluent calculus (Thielscher, 1999), which reifies the facts
out of which states are composed.
Peter Ladkin (1986a, 1986b) introduced “concave” time intervals (intervals with gaps—
essentially, unions of ordinary “convex” time intervals) and applied the techniques of mathematical abstract algebra to time representation. Allen (1991) systematically investigates the wide variety of techniques available for time representation; van Beek and Manchak (1996)
analyze algorithms for temporal reasoning. There are significant commonalities between the
event-based ontology given in this chapter and an analysis of events due to the philosopher
Donald Davidson (1980). The histories in Pat Hayes’s (1985a) ontology of liquids and the chronicles in McDermott’s (1985) theory of plans were also important influences on the field
and on this chapter. The question of the ontological status of substances has a long history. Plato proposed
that substances were abstract entities entirely distinct from physical objects; he would say MadeOf (Butters, Butter) rather than Butters € Butter. This leads to a substance hierarchy in which, for example, UnsaltedButter is a more specific substance than Butter.
The position
adopted in this chapter, in which substances are categories of objects, was championed by
Richard Montague (1973). It has also been adopted in the CYC project. Copeland (1993) mounts a serious, but not invincible, attack.
The alternative approach mentioned in the chapter, in which butter is one object consisting of all buttery objects in the universe, was proposed originally by the Polish logician Lesniewski (1916).
His mereology (the name is derived from the Greek word for “part”)
used the part-whole relation as a substitute for mathematical set theory, with the aim of elim-
inating abstract entities such as sets. A more readable exposition of these ideas is given by Leonard and Goodman (1940), and Goodman’s The Structure of Appearance (1977) applies the ideas to various problems in knowledge representation. While some aspects of the mereological approach are awkward—for example, the need for a separate inheritance mechanism based on part-whole relations—the approach gained the support of Quine (1960). Harry Bunt (1985) has provided an extensive analysis of its use in knowledge representation. Casati and Varzi (1999) cover parts, wholes, and a general theory of spatial locations.
There are three main approaches to the study of mental objects. The one taken in this chapter, based on modal logic and possible worlds, is the classical approach from philosophy (Hintikka,
1962; Kripke,
1963; Hughes and Cresswell,
1996).
The book Reasoning about
Bibliographical and Historical Notes Knowledge (Fagin et al., 1995) provides a thorough introduction, and Gordon and Hobbs
(2017) provide A Formal Theory of Commonsense Psychology. The second approach is a first-order theory in which mental objects are fluents. Davis
(2005) and Davis and Morgenstern (2005) describe this approach. It relies on the possibleworlds formalism, and builds on work by Robert Moore (1980, 1985).
The third approach is a syntactic theory, in which mental objects are represented by
character strings. A string is just a complex term denoting a list of symbols, so CanFly(Clark)
can be represented by the list of symbols [C,a,n,F,1,y,(,C,l,a,r,k,)]. The syntactic theory of mental objects was first studied in depth by Kaplan and Montague (1960), who showed
that it led to paradoxes if not handled carefully. Ernie Davis (1990) provides an excellent comparison of the syntactic and modal theories of knowledge. Pnueli (1977) describes a
temporal logic used to reason about programs, work that won him the Turing Award and which was expanded upon by Vardi (1996). Littman e al. (2017) show that a temporal logic can be a good language for specifying goals to a reinforcement learning robot in a way that is easy for a human to specify, and generalizes well to different environments.
The Greek philosopher Porphyry (c. 234-305 CE), commenting on Aristotle’s Caregories, drew what might qualify as the first semantic network. Charles S. Peirce (1909) developed existential graphs as the first semantic network formalism using modern logic. Ross Quillian (1961), driven by an interest in human memory and language processing, initiated work on semantic networks within Al An influential paper by Marvin Minsky (1975)
presented a version of semantic networks called frames; a frame was a representation of an object or category, with attributes and relations to other objects or categories.
The question of semantics arose quite acutely with respect to Quillian’s semantic net-
works (and those of others who followed his approach), with their ubiquitous and very vague
“IS-A links” Bill Woods’s (1975) famous article “What's In a Link?” drew the attention of AT
researchers to the need for precise semantics in knowledge representation formalisms. Brachman (1979) elaborated on this point and proposed solutions.
Ron
Patrick Hayes’s (1979)
“The Logic of Frames” cut even deeper, claiming that “Most of ‘frames’ is just a new syntax for parts of first-order logic.” Drew McDermott’s (1978b) “Tarskian Semantics, or, No No-
tation without Denotation!”
argued that the model-theoretic approach to semantics used in
first-order logic should be applied to all knowledge representation formalisms. This remains
a controversial idea; notably, McDermott himself has reversed his position in “A Critique of Pure Reason” (McDermott, 1987). Selman and Levesque (1993) discuss the complexity of
inheritance with exceptions, showing that in most formulations it is NP-complete.
Description logics were developed as a useful subset of first-order logic for which infer-
ence is computationally tractable.
Hector Levesque and Ron Brachman (1987) showed that
certain uses of disjunction and negation were primarily responsible for the intractability of logical inference. This led to a better understanding of the interaction between complexity
and expressiveness in reasoning systems. Calvanese ef al. (1999) summarize the state of the art, and Baader et al. (2007) present a comprehensive handbook of description logic. The three main formalisms for dealing with nonmonotonic inference—circumscription
(McCarthy, 1980), default logic (Reiter, 1980), and modal nonmonotonic logic (McDermott
and Doyle, 1980)—were all introduced in one special issue of the Al Journal. Delgrande and Schaub (2003) discuss the merits of the variants, given 25 years of hindsight. Answer
set programming can be seen as an extension of negation as failure or as a refinement of
341
342
Chapter 10 Knowledge Representation circumscription; the underlying theory of stable model semantics was introduced by Gelfond
and Lifschitz (1988), and the leading answer set programming systems are DLV (Eiter ef al.,
1998) and SMODELS (Niemelii et al., 2000). Lifschitz (2001) discusses the use of answer set programming for planning. Brewka er al. (1997) give a good overview of the various approaches to nonmonotonic logic. Clark (1978) covers the negation-as-failure approach to logic programming and Clark completion. Lifschitz (2001) discusses the application of answer set programming to planning. A variety of nonmonotonic reasoning systems based on
logic programming are documented in the proceedings of the conferences on Logic Program-
ming and Nonmonotonic Reasoning (LPNMR).
The study of truth maintenance systems began with the TMS (Doyle, 1979) and RUP
(McAllester,
1980) systems, both of which were essentially JTMSs.
Forbus and de Kleer
(1993) explain in depth how TMSs can be used in Al applications. Nayak and Williams
(1997) show how an efficient incremental TMS called an ITMS makes it feasible to plan the
operations of a NASA spacecraft in real time.
This chapter could not cover every area of knowledge representation in depth. The three
principal topics omitted are the following: Qualitative physics.
Qualitative physics: Qualitative physics is a subfield of knowledge representation concerned specifically with constructing a logical, nonnumeric theory of physical objects and processes. The term was coined
by Johan de Kleer (1975), although the enterprise could be said to
have started in Fahlman’s (1974) BUILD, a sophisticated planner for constructing complex towers of blocks.
Fahlman discovered in the process of designing it that most of the effort
(80%, by his estimate) went into modeling the physics of the blocks world to calculate the
stability of various subassemblies of blocks, rather than into planning per se. He sketches a hypothetical naive-physics-like process to explain why young children can solve BUILD-like
problems without access to the high-speed floating-point arithmetic used in BUILD’s physical
modeling. Hayes (1985a) uses “histories"—four-dimensional slices of space-time similar to
Davidson’s events—to construct a fairly complex naive physics of liquids. Davis (2008) gives an update to the ontology of liquids that describes the pouring of liquids into containers.
De Kleer and Brown (1985), Ken Forbus (1985), and Benjamin Kuipers (1985) independently and almost simultaneously developed systems that can reason about a physical system based on qualitative abstractions of the underlying equations. Qualitative physics soon developed to the point where it became possible to analyze an impressive variety of complex phys-
ical systems (Yip, 1991). Qualitative techniques have been used to construct novel designs
for clocks, windshield wipers, and six-legged walkers (Subramanian and Wang, 1994). The collection Readings in Qualitative Reasoning about Physical Systems (Weld and de Kleer, 1990), an encyclopedia article by Kuipers (2001), and a handbook article by Davis (2007)
provide good introductions to the field.
Spatial reasoning
Spatial reasoning: The reasoning necessary to navigate in the wumpus world is trivial in comparison to the rich spatial structure of the real world.
The earliest serious attempt to
capture commonsense reasoning about space appears in the work of Ernest Davis (1986, 1990). The region connection calculus of Cohn ef al. (1997) supports a form of qualitative spatial reasoning and has led to new kinds of geographical information systems; see also (Davis, 2006). As with qualitative physics, an agent can go a long way, so to speak, without
resorting to a full metric representation.
Bibliographical and Historical Notes Psychological reasoning: Psychological reasoning involves the development of a working
psychology for artificial agents to use in reasoning about themselves and other agents. This
is often based on so-called folk psychology, the theory that humans in general are believed
to use in reasoning about themselves and other humans.
When Al researchers provide their
artificial agents with psychological theories for reasoning about other agents, the theories are
frequently based on the researchers” description of the logical agents” own design. Psychological reasoning is currently most useful within the context of natural language understanding,
where divining the speaker’s intentions is of paramount importance. Minker (2001) collects papers by leading researchers in knowledge representation, summarizing 40 years of work in the field. The proceedings of the international conferences on Principles of Knowledge Representation and Reasoning provide the most up-to-date sources. for work in this area. Readings in Knowledge Representation (Brachman and Levesque, 1985) and Formal Theories of the Commonsense
World (Hobbs and Moore,
1985) are ex-
cellent anthologies on knowledge representation; the former focuses more on historically
important papers in representation languages and formalisms, the latter on the accumulation of the knowledge itself. Davis (1990), Stefik (1995), and Sowa (1999) provide textbook introductions to knowledge representation, van Harmelen et al. (2007) contributes a handbook, and Davis and Morgenstern (2004) edited a special issue of the Al Journal on the topic. Davis
(2017) gives a survey of logic for commonsense reasoning. The biennial conference on Theoretical Aspects of Reasoning About Knowledge (TARK) covers applications of the theory of knowledge in Al, economics, and distributed systems.
Psychological reasoning
TR
1 1
AUTOMATED PLANNING In which we see how an agent can take advantage of the structure ofa problem to efficiently construct complex plans of action.
Planning a course of action is a key requirement for an intelligent agent. The right representation for actions and states and the right algorithms can make this easier.
In Section 11.1
we introduce a general factored representation language for planning problems that can naturally and succinctly represent a wide variety of domains, can efficiently scale up to large
problems, and does not require ad hoc heuristics for a new domain. Section 11.4 extends the
representation language to allow for hierarchical actions, allowing us to tackle more complex
problems. We cover efficient algorithms for planning in Section 11.2, and heuristics for them in Section
11.3.
In Section
11.5 we account for partially observable and nondeterministic
domains, and in Section 11.6 we extend the language once again to cover scheduling problems with resource constraints. This gets us closer to planners that are used in the real world
for planning and scheduling the operations of spacecraft, factories, and military campaigns. Section 11.7 analyzes the effectiveness of these techniques.
11.1 Classical planning
Definition of Classical Planning
Classical planning is defined as the task of finding a sequence of actions to accomplish a
goal in a discrete, deterministic, static, fully observable environment. We have seen two approaches to this task: the problem-solving agent of Chapter 3 and the hybrid propositional
logical agent of Chapter 7. Both share two limitations. First, they both require ad hoc heuristics for each new domai
heuristic evaluation function for search, and hand-written code
for the hybrid wumpus agent. Second, they both need to explicitly represent an exponentially large state space.
For example, in the propositional logic model of the wumpus world, the
axiom for moving a step forward had to be repeated for all four agent orientations, T time. steps, and n* current locations. PDDL
In response to these limitations, planning researchers have invested in a factored repre-
sentation using a family of languages called PDDL, the Planning Domain Definition Lan-
guage (Ghallab ef al., 1998), which allows us to express all 47n? actions with a single action schema, and does not need domain-specific knowledge.
Basic PDDL can handle classical
planning domains, and extensions can handle non-classical domains that are continuous, partially observable, concurrent, and multi-agent. The syntax of PDDL is based on Lisp, but we. State
will translate it into a form that matches the notation used in this book.
In PDDL, a state is represented as a conjunction of ground atomic fluents. Recall that
“ground” means no variables, “fluent” means an aspect of the world that changes over time,
Section 11.1
Definition of Classical Planning
and “ground atomic” means there is a single predicate, and if there are any arguments, they must be constants. For example, Poor A Unknown might represent the state of a hapless agent,
and Ar(Truck,, Melbourne) AAt(Trucks, Sydney) could represent a state in a package delivery problem. PDDL uses database semantics: the closed-world assumption means that any fluents that are not mentioned are false, and the unique names
assumption means that Truck;
and Truck; are distinct. The following fluents are nor allowed in a state: Ar(x,y) (because it has variables), ~Poor (because it is a negation), and Ar(Spouse(Ali), Sydney) (because it uses a function symbol, Spouse). When convenient, we can think of the conjunction of fluents as a set of fluents.
An action schema represents a family of ground actions. For example, here is an action Action schema
schema for flying a plane from one location to another:
Action(Fly(p. from,t0),
PRECOND:At(p. from) A Plane(p) A Airport(from) A Airport(to) EFFECT:—A!(p. from) A At(p, t0))
The schema consists of the action name,
a list of all the variables used in the schema,
a
precondition and an effect. The precondition and the effect are each conjunctions of literals (positive or negated atomic sentences).
yielding a ground (variable-free) action:
We can choose constants to instantiate the variables,
Precondition Effect
Action(Fly(P,,SFO,JFK),
PRECOND:At(Py,SFO) A Plane(Py) AAirport(SFO) AAirport(JFK) EFFECT:—AI(P1,SFO) AA1(Py,JFK))
A ground action a is applicable in state s if s entails the precondition of a; that is, if every
positive literal in the precondition is in s and every negated literal is not.
The result of executing applicable action a in state s is defined as a state s’ which is
represented by the set of fluents formed by starting with s, removing the fluents that appear as negative literals in the action’s effects (what we call the delete list or DEL(a)), and adding the fluents that are positive literals in the action’s effects (what we call the add list or ADD(a)):
RESULT(s,a) = (s — DEL(a)) UADD(a) .
(11.1)
For example, with the action Fiy(Py,SFO,JFK), we would remove the fluent Ar(Py,SFO) and add the fluent At(Py,JFK). A set of action schemas serves as a definition of a planning domain. A specific problem within the domain is defined with the addition of an initial state and a goal. The initial state is a conjunction of ground fluents (introduced with the keyword Inir in Figure 11.1). As with all states, the closed-world assumption is used, which means that any atoms that are not mentioned are false. The goal (introduced with Goal) is just like a precondition: a conjunction of literals (positive or negative) that may contain variables. For example, the goal At(C1,SFO) A=At(C2,SFO) AAt(p, SFO), refers to any state in which cargo C} is at SFO but C is not, and in which there is a plane at SFO. 11.1.1
Example domain:
Air cargo transport
Figure 11.1 shows an air cargo transport problem involving loading and unloading cargo and flying it from place to place. The problem can be defined with three actions: Load, Unload,
and Fly. The actions affect two predicates: In(c, p) means that cargo c is inside plane p, and Ar(x,a) means that object x (either plane or cargo) is at airport a. Note that some care
Delete list Add list
Chapter 11
Automated Planning
Init(AK(Cy, SFO) A At(Cy, JFK) A At(Py, SFO) A At(Py, JFK) A Cargo(Cy) A Cargo(Cs) A Plane(P;) A Plane(Ps) AAirport(JFK) A Airport(SFO)) Goal(A1(Cy, JFK) A At(Ca, SFO)) Action(Load(c, p, @), PRECOND: At(c, a) A At(p, a) A Cargo(c)
EFFECT: = At(c, a) A In(c, p))
Action(Unload(c, p, a),
PRECOND: In(c, p) A At(p, a) A Cargo(c) EFFECT: At(c, a) A —In(c, p))
A Plane(p) A Airport(a)
A Plane(p) A Airport(a)
Action(Fly(p, from, t0),
PRECOND: At(p, from) A Plane(p) A Airport(from) A Airport(to) EFFECT: = At(p, from) A\ At(p, t0))
Figure 1.1 A PDDL description of an air cargo transportation planning problem. must be taken to make sure the Az predicates are maintained properly.
When a plane flies
from one airport to another, all the cargo inside the plane goes with it. Tn first-order logic it would be easy to quantify over all objects that are inside the plane. But PDDL does not have
a universal quantifier, so we need a different solution. The approach we use is to say that a piece of cargo ceases to be Ar anywhere when it is /n a plane; the cargo only becomes At the new airport when it is unloaded.
So At really means “available for use at a given location.”
The following plan is a solution to the problem:
[Load(Cy, Py, SFO). Fly(Py,SFO.JFK), Unload(C\,Pi,JFK). Load(Ca. P2, JFK), Fly(P2,JFK SFO), Unload(Ca, P2, SFO)]..
11.1.2
Example domain:
The spare tire problem
Consider the problem of changing a flat tire (Figure 11.2). The goal is to have a good spare tire properly mounted onto the car’s axle, where the initial state has a flat tire on the axle and a good spare tire in the trunk.
To keep it simple, our version of the problem is
an abstract
one, with no sticky lug nuts or other complications. There are just four actions: removing the spare from the trunk, removing the flat tire from the axle, putting the spare on the axle, and leaving the car unattended overnight.
We assume that the car is parked in a particu-
larly bad neighborhood, so that the effect of leaving it overnight is that the tires disappear. [Remove(Flat,Axle),Remove(Spare, Trunk), PutOn(Spare,Axle)] is a solution to the problem. 11.1.3
Example domain:
The blocks world
One of the most famous planning domains is the blocks world. This domain consists of a set
of cube-shaped blocks sitting on an arbitrarily-large table.! The blocks can be stacked, but
only one block can fit directly on top of another. A robot arm can pick up a block and move it
to another position, either on the table or on top of another block. The arm can pick up only
one block at a time, so it cannot pick up a block that has another one on top of it. A typical
goal to get block A on B and block B on C (see Figure 11.3). 1" The blocks world commonly used in planning research is much simpler than SHRDLUs version (page 20).
Section 11.1
Definition of Classical Planning
Init(Tire(Flat) A Tire(Spare) A At(Flat,Axle) A At(Spare, Trunk)) Goal(At(Spare,Axle))
Action(Remove(obj. loc),
PRECOND: At(obj. loc)
EFFECT: At(obj,loc) A A(obj, Ground))
Action(PutOn(t, Axle),
PRECOND: Tire(t) A At(t,Ground) A — At(Flat,Axle) N\ — At(Spare,Axle) EFFECT: = At(1,Ground) A At(t,Axle))
Action(LeaveOvernight, PRECOND: EFFECT: — At(Spare, Ground) A — At(Spare, Axle) A = At(Spare, Trunk) A = At(Flat,Ground) A = At(Flat,Axle) A = At(Flat, Trunk))
Figure 11.2 The simple spare tire problem.
BN Start State
A
Goal State
Figure 11.3 Diagram of the blocks-world problem in Figure 11.4.
Init(On(A, Table) A On(B,Table) A On(C,A) A Block(A) A Block(B) A Block(C) A Clear(B) A Clear(C) A Clear(Table)) Goal(On(A,B) A On(B,C))
Action(Move(b.x.y),
PRECOND: On(b,x) A Clear(b) A Clear(y) A Block(b) A Block(y) N
(b#x) A (b#y) A (x#),
EFFECT: On(b,y) A Clear(x) A —On(b,x) A —Clear(y))
Action(MoveToTable(b,x),
PRECOND: On(b,x) A Clear(b) A Block(b) A Block(x), EFFECT: On(b, Table) A Clear(x) A —On(b,x))
Figure 11.4 A planning problem in the blocks world: building a three-block tower. One solution is the sequence [MoveToTuble(C.A), Move(B, Table.C), Move(A, Table, B)).
347
Chapter 11
Automated Planning
We use On(b,x) to indicate that block b is on x, where x is either another block or the
table. The action for moving block b from the top of x to the top of y will be Move(b,x,y).
Now, one of the preconditions on moving b is that no other block be on it. In first-order logic,
this would be —3x On(x,b) or, alternatively, Vx —On(x,b). Basic PDDL does not allow
quantifiers, so instead we introduce a predicate Clear(x) that is true when nothing is on x.
(The complete problem description is in Figure 11.4.)
The action Move moves a block b from x to y if both b and y are clear. After the move is made, b is still clear but y is not. A first attempt at the Move schema is
Action(Move(b, x.y),
PRECOND:On(b.x) A Clear(b) A Clear(y),
EFFECT:On(b,y) A Clear(x) A =On(b,x) A—Clear(y)).
Unfortunately, this does not maintain Clear properly when x or y is the table. When x is
the Table, this action has the effect Clear(Table), but the table should not become clear; and
when y=Table, it has the precondition Clear(Table), but the table does not have to be clear for us to move a block onto it. To fix this, we do two things.
action to move a block b from x to the table:
First, we introduce another
Action(MoveToTable(b,x),
PRECOND:On(b,x) A Clear(b),
EFFECT: On(b, Table) A Clear(x) A—On(b,x)) .
Second, we take the interpretation of Clear(x) to be “there is a clear space on x to hold a block.” Under this interpretation, Clear(Table) will always be true. The only problem is that nothing prevents the planner from using Move(b,x, Table) instead of MoveToTable(b,x). We could live with this problem—it will lead to a larger-than-necessary scarch space, but will not lead to incorrect answers—or we could introduce the predicate Block and add Block(b) A Block(y) to the precondition of Move, as shown in Figure 11.4. 11.2
Algorithms for Classical Planning
The description of a planning problem provides an obvious way to search from the initial state through the space of states, looking for a goal. A nice advantage of the declarative representation of action schemas is that we can also search backward from the goal, looking
for the initial state (Figure 11.5 compares forward and backward searches). A third possibility
is to translate the problem description into a set of logic sentences, to which we can apply a
logical inference algorithm to find a solution. 11.2.1
Forward state-space search for planning
We can solve planning problems by applying any of the heuristic search algorithms from Chapter 3 or Chapter 4. The states in this search state space are ground states, where every fluent is either true or not.
The goal is a state that has all the positive fluents in the prob-
lem’s goal and none of the negative fluents. The applicable actions in a state, Actions(s), are
grounded instantiations of the action schemas—that is, actions where the variables have all
been replaced by constant values.
To determine the applicable actions we unify the current state against the preconditions
of each action schema.
For each unification that successfully results in a substitution, we
Section 112 Algorithms for Classical Planning apply the substitution to the action schema to yield a ground action with no variables. (It is a requirement of action schemas that any variable in the effect must also appear in the
precondition; that way, we are guaranteed that no variables remain after the substitution.) Each schema may unify in multiple ways. In the spare tire example (page 346), the Remove action has the precondition Ar(obj, loc), which matches against the initial state in two
ways, resulting in the two substitutions {obj/Flat,loc/Axle} and {obj/Spare,loc/Trunk}; applying these substitutions yields two ground actions.
If an action has multiple literals in
the precondition, then each of them can potentially be matched against the current state in multiple ways.
At first, it seems that the state space might be too big for many problems. Consider an
air cargo problem with 10 airports, where each airport initially has 5 planes and 20 pieces of cargo. The goal is to move all the cargo at airport A to airport B. There is a 41-step solution
to the problem: load the 20 pieces of cargo into one of the planes at A, fly the plane to B, and
unload the 20 pieces. Finding this apparently straightforward solution can be difficult because the average branching factor is huge: each of the 50 planes can fly to 9 other airports, and each of the 200
packages can be either unloaded (if it is loaded) or loaded into any plane at its airport (if it is unloaded).
So in any state there is a minimum of 450 actions (when all the packages are
at airports with no planes) and a maximum of 10,450 (when all packages and planes are at
Fly(P,, A B) Fly(P,, A B)
Figure 11.5 Two approaches to searching for a plan. (a) Forward (progression) search through the space of ground states, starting in the initial state and using the problem’s actions to search forward for a member of the set of goal states.
(b) Backward (regression)
search through state descriptions, starting at the goal and using the inverse of the actions to search backward for the initial state.
349
350
Chapter 11
Automated Planning
the same airport). On average, let’s say there are about 2000 possible actions per state, so the search graph up to the depth of the 41-step solution has about 2000*' nodes.
Clearly, even this relatively small problem instance is hopeless without an accurate heuristic. Although many real-world applications of planning have relied on domain-specific heuris-
tics, it turns out (as we see in Section 11.3) that strong domain-independent heuristics can be
derived automatically; that is what makes forward search feasible. 11.2.2 Regression search Relevant action
Backward search for planning
In backward search (also called regression search) we start at the goal and apply the actions backward until we find a sequence of steps that reaches the initial state. At each step we
consider relevant actions (in contrast to forward search, which considers actions that are
applicable). This reduces the branching factor significantly, particularly in domains with many possible actions.
A relevant action is one with an effect that unifies with one of the goal literals, but with
no effect that negates any part of the goal. For example, with the goal ~PoorA Famous, an action with the sole effect Famous would be relevant, but one with the effect Poor A Famous
is not considered relevant: even though that action might be used at some point in the plan (to
establish Famous), it cannot appear at rhis point in the plan because then Poor would appear in the final state.
Regression
What does it mean to apply an action in the backward direction?
Given a goal g and
an action a, the regression from g over a gives us a state description g’ whose positive and negative literals are given by Pos(¢') = (Pos(g) — ADD(a)) UPOS(Precond(a))
NEG(g') = (NEG(g) — DEL(a)) UNEG(Precond(a)).
That is, the preconditions
must have held before, or else the action could not have been
executed, but the positive/negative literals that were added/deleted by the action need not have been true before.
These equations are straightforward for ground literals, but some care is required when there are variables in g and a. For example, suppose the goal is to deliver a specific piece of cargo to SFO: A1(Ca,SFO). The Unload action schema has the effect Ar(c,a). When we
unify that with the goal, we get the substitution {c/C,a/SFO}; applying that substitution to the schema gives us a new schema which captures the idea of using any plane that is at SFO:
Action(Unload(Cs, p/,SFO),
PRECOND:In(Cy, p') AAt(p',SFO) A Cargo(C,) A Plane(p') A Airport(SFO) EFFECT:A1(C2,SFO) A —In(Ca, p')) .
Here we replaced p with a new variable named p'. This is an instance of standardizing apart
variable names so there will be no conflict between different variables that happen to have the
same name (see page 284). The regressed state description gives us a new goal: & = In(Ca, p') AAI(p',SFO) A Cargo(C2) A Plane(p') A Airport(SFO) As another example, consider the goal of owning a book with a specific ISBN number: Own(9780134610993). Given a trillion 13-digit ISBNs and the single action schema A = Action(Buy(i), PRECOND:ISBN (i), EFFECT: Own(i)) .
a forward search without a heuristic would have to start enumerating the 10 billion ground
Buy actions. But with backward search, we would unify the goal Own(9780134610993 ) with
Section 112
Algorithms for Classical Planning
the effect Own(i"), yielding the substitution 6 = {i /9780134610993 }. Then we would regress over the action Subst(6,A) to yield the predecessor state description ISBN (9780134610993).
This is part of the initial state, so we have a solution and we are done, having considered just one action, not a trillion.
More formally, assume a goal description g that contains a goal literal g; and an action
schema A. If A has an effect literal ¢); where Unify(g;.¢})=0 and where we define A’
SUBST(0,A) and if there is no effect in A’ that is the negation of a literal in g, then A’ is a
relevant action towards g.
For most problem domains backward search keeps the branching factor lower than for-
ward search.
However, the fact that backward search uses states with variables rather than
ground states makes it harder to come up with good heuristics. That is the main reason why
the majority of current systems favor forward search. 11.2.3
Planning as Boolean satisfiability
In Section 7.7.4 we showed how some clever axiom-rewriting could turn a wumpus world problem into a propositional logic satisfiability problem that could be handed to an efficient satisfiability solver. SAT-based planners such as SATPLAN operate by translating a PDDL problem description into propositional form. The translation involves a series of steps: « Propositionalize the actions: for each action schema, form ground propositions by substituting constants for each of the variables. So instead of a single Unload(c,p,a) schema, we would have separate action propositions for each combination of cargo,
plane, and airport (here written with subscripts), and for each time step (here written as a superscript). + Add action exclusion axioms saying that no two actions can occur at the same time, e.g.
—(FlyP;SFOJFK' A FlyP,;SFOBUH").
+ Add precondition axioms:
For each ground action A, add the axiom A’ =
PRE(A)',
that is, if an action is taken at time 7, then the preconditions must have been true. For
example, FIyP;SFOJFK' = At(P,,SFO) A Plane(P\) A Airport(SFO) AAirport (JFK).
« Define the initial state: assert FO for every fluent F in the problem’s initial state, and
—F" for every fluent not mentioned in the initial state.
« Propositionalize the goal: the goal becomes a disjunction over all of its ground instances, where variables are replaced by constants. For example, the goal of having block A on another block, On(A,x) ABlock(x) in a world with objects A, B and C, would
be replaced by the goal
(On(A,A) ABlock(A)) V (On(A,B) ABlock(B)) V (On(A,C) A Block(C)). + Add successor-state axioms: For each fluent ', add an axiom of the form
F'*' & ActionCausesF' V (F' A=ActionCausesNotF"), where ActionCausesF stands for a disjunction of all the ground actions that add F, and ActionCausesNotF stands for a disjunction of all the ground actions that delete F.
The resulting translation is typically much larger than the original PDDL, but modern the efficiency of modern SAT solvers often more than makes up for this.
351
352
Chapter 11 11.2.4
Planning graph
Automated Planning
Other classical planning approaches
The three approaches we covered above are not the only ones tried in the 50-year history of automated planning. We briefly describe some others here. An approach called Graphplan uses a specialized data structure, a planning graph, to
encode constraints on how actions are related to their preconditions and effects, and on which
Situation calculus
things are mutually exclusive. Situation calculus is a method of describing planning problems in first-order logic. It uses successor-state axioms just as SATPLAN
does, but first-order logic allows for more
flexibility and more succinct axioms. Overall the approach has contributed to our theoretical
understanding of planning, but has not made a big impact in practical applications, perhaps
because first-order provers are not as well developed as propositional satisfiability programs. It is possible to encode a bounded planning problem (i.c., the problem of finding a plan
of length k) as a constraint satisfaction problem
(CSP). The encoding
is similar to the
encoding to a SAT problem (Section 11.2.3), with one important simplification: at each time
step we need only a single variable, Action', whose domain s the set of possible actions. We. no longer need one variable for every action, and we don’t need the action exclusion axioms.
Partial-order planning
All the approaches we have seen so far construct fotally ordered plans consisting of strictly linear sequences of actions. But if an air cargo problem has 30 packages being loaded onto one plane and 50 packages being loaded onto another, it seems pointless to decree a specific linear ordering of the 80 load actions. An alternative called partial-order planning represents a plan as a graph rather than a linear sequence: each action is a node in the graph, and for each precondition of the action there is an edge from another action (or from the initial state) that indicates that the predeces-
sor action establishes the precondition. So we could have a partial-order plan that says that ac-
tions Remove(Spare, Trunk) and Remove(Flat, Axle) must come before PutOn(Spare,Axle),
but without saying which of the two Remove actions should come first. We search in the space
of plans rather than world-states, inserting actions to satisfy conditions.
In the 1980s and 1990s, partial-order planning was seen as the best way to handle planning problems with independent subproblems. By 2000, forward-search planners had developed excellent heuristics that allowed them to efficiently discover the independent subprob-
lems that partial-order planning was designed for. Moreover, SATPLAN was able to take ad-
vantage of Moore’s law: a propositionalization that was hopelessly large in 1980 now looks tiny, because computers have 10,000 times more memory today.
As a result, partial-order
planners are not competitive on fully automated classical planning problems.
Nonetheless, partial-order planning remains an important part of the field.
For some
specific tasks, such as operations scheduling, partial-order planning with domain-specific heuristics is the technology of choice. Many of these systems use libraries of high-level plans, as described in Section 11.4.
Partial-order planning is also often used in domains where it is important for humans
to understand the plans. For example, operational plans for spacecraft and Mars rovers are generated by partial-order planners and are then checked by human operators before being uploaded to the vehicles for execution. The plan refinement approach makes it easier for the humans to understand what the planning algorithms are doing and to verify that the plans are
correct before they are executed.
Section 11.3 11.3
Heuristics for Planning
353
Heuristics for Planning
Neither forward nor backward search is efficient without a good heuristic function.
Recall
from Chapter 3 that a heuristic function h(s) estimates the distance from a state s to the goal, and that if we can derive an admissible heuristic for this distance—one that does not
overestimate—then we can use A* search to find optimal solutions.
By definition, there is no way to analyze an atomic state, and thus it requires some ingenuity by an analyst (usually human) to define good domain-specific heuristics for search problems with atomic states. But planning uses a factored representation for states and actions, which makes it possible to define good domain-independent heuristics.
Recall that an admissible heuristic can be derived by defining a relaxed problem that is
easier to solve. The exact cost of a solution to this easier problem then becomes the heuristic
for the original problem. A search problem is a graph where the nodes are states and the edges are actions.
The problem is to find a path connecting the initial state to a goal state.
There are two main ways we can relax this problem to make it easier: by adding more edges to the graph, making it strictly easier to find a path, or by grouping multiple nodes together, forming an abstraction of the state space that has fewer states, and thus is easier to search.
We look first at heuristics that add edges to the graph. Perhaps the simplest is the ignore-
preconditions heuristic, which drops all preconditions from actions. Every action becomes
applicable in every state, and any single goal fluent can be achieved in one step (if there are any applicable actions—if not, the problem is impossible). This almost implies that the number of steps required to solve the relaxed problem is the number of unsatisfied goals—
Ignore-preconditions heuristic
almost but not quite, because (1) some action may achieve multiple goals and (2) some actions
may undo the effects of others. For many problems an accurate heuristic is obtained by considering (1) and ignoring (2). First, we relax the actions by removing all preconditions and all effects except those that are
literals in the goal. Then, we count the minimum number of actions required such that the union of those actions’ effects satisfies the goal. This is an instance of the set-cover problem.
There is one minor irritation: the set-cover problem is NP-hard. Fortunately a simple greedy algorithm is guaranteed to return a set covering whose size is within a factor of logn of the true minimum covering, where n is the number of literals in the goal.
Unfortunately, the
greedy algorithm loses the guarantee of admissibility. It is also possible to ignore only selected preconditions of actions. Consider the slidingtile puzzle (8-puzzle or 15-puzzle) from Section 3.2. We could encode this as a planning problem involving tiles with a single schema Slide: Action(Slide(t,s1,52),
PRECOND:On(t,sy) ATile(t) A Blank(sy) A Adjacent (s, s2) EFFECT:On(t,s2) A Blank(s;) A =On(t,s1) A —Blank(s;))
As we saw in Section 3.6, if we remove the preconditions Blank(s) A Adjacent(sy ,s2) then any tile can move in one action to any space and we get the number-of-misplaced-tiles heuris-
tic. If we remove only the Blank(s,) precondition then we get the Manhattan-distance heuris
tic. It is easy to see how these heuristics could be derived automatically from the action schema description. The ease of manipulating the action schemas is the great advantage of the factored representation of planning problems, as compared with the atomic representation
of search problems.
Set-cover problem
354
Chapter 11
Automated Planning
Figure 11.6 Two state spaces from planning problems with the ignore-delete-lists heuristic. The height above the bottom plane is the heuristic score of a state; states on the bottom plane are goals. There are no local minima, so search for the goal is straightforward. From Hoffmann (2005). Ignore-delete-lists heuristic
Another possibility is the ignore-delete-lists heuristic.
Assume for a moment that all
goals and preconditions contain only positive literals.> We want to create a relaxed version
of the original problem that will be easier to solve, and where the length of the solution will serve as a good heuristic.
We can do that by removing the delete lists from all actions
(i.e., removing all negative literals from effects). That makes it possible to make monotonic
progress towards the goal—no action will ever undo progress made by another action. It turns
out it is still NP-hard to find the optimal solution to this relaxed problem, but an approximate
solution can be found in polynomial time by hill climbing.
Figure 11.6 diagrams part of the state space for two planning problems using the ignore-
delete-lists heuristic. The dots represent states and the edges actions, and the height of each
dot above the bottom plane represents the heuristic value. States on the bottom plane are solutions. Tn both of these problems, there is a wide path to the goal. There are no dead ends, 50 no need for backtracking; a simple hill-climbing search will easily find a solution to these
problems (although it may not be an optimal solution). 11.3.1
Dom:
ndependent
pruning
Factored representations make it obvious that many states are just variants of other states. For
example, suppose we have a dozen blocks on a table, and the goal is to have block A on top
of a three-block tower. The first step in a solution is to place some block x on top of block y
(where x, y, and A are all different). After that, place A on top ofx and we’re done. There are 11 choices for x, and given x, 10 choices for y, and thus 110 states to consider. But all these
Symmetry reduction
states are symmetric: choosing one over another makes no difference, and thus a planner should only consider one of them. This is the process of symmetry reduction: we prune out aren’t, replace every negative literal ~P in tate and the action effects accordingly.
Section 11.3
Heuristics for Planning
355
of consideration all symmetric branches of the search tree except for one. For many domains, this makes the difference between intractable and efficient solving.
Another possibility is to do forward pruning, accepting the risk that we might prune away an optimal solution, in order to focus the search on promising branches. We can define
a preferred action as follows: First, define a relaxed version of the problem, and solve it to
Preferred action
teractions can be ruled out. We say that a problem has serializable subgoals if there exists
Serializable subgoals
get a relaxed plan. Then a preferred action s either a step of the relaxed plan, or it achieves some precondition of the relaxed plan. Sometimes it is possible to solve a problem efficiently by recognizing that negative in-
an order of subgoals such that the planner can achieve them in that order without having to undo any of the previously achieved subgoals. For example, in the blocks world, if the goal is to build a tower (e.g., A on B, which in turn is on C, which in turn is on the Table, as in
Figure 11.3 on page 347), then the subgoals are serializable bottom to top: if we first achieve C on Table, we will never have to undo it while we are achieving the other subgoals.
A
planner that uses the bottom-to-top trick can solve any problem in the blocks world without backtracking (although it might not always find the shortest plan). As another example, if there is a room with n light switches, each controlling a separate light, and the goal is to have them all on, then we don’t have to consider permutations of the order; we could arbitrarily
restrict ourselves to plans that flip switches in, say, ascending order.
For the Remote Agent planner that commanded NASA’s Deep Space One spacecraft, it was determined that the propositions involved in commanding a spacecraft are serializable. This is perhaps not too surprising, because a spacecraft is designed by its engineers to be as easy as possible to control (subject to other constraints). Taking advantage of the serialized
ordering of goals, the Remote Agent planner was able to eliminate most of the search. This meant that it was fast enough to control the spacecraft in real time, something previously
considered impossible. 11.3.2
State abstraction in planning
A relaxed problem leaves us with a simplified planning problem just to calculate the value of the heuristic function. Many planning problems have 10'% states or more, and relaxing
the actions does nothing to reduce the number of states, which means that it may still be
expensive to compute the heuristic. Therefore, we now look at relaxations that decrease the
number of states by forming a state abstraction—a many-to-one mapping from states in the State abstraction ground representation of the problem to the abstract representation.
The easiest form of state abstraction is to ignore some fluents. For example, consider an
air cargo problem with 10 airports, 50 planes, and 200 pieces of cargo. Each plane can be at one of 10 airports and each package can be either in one of the planes or unloaded at one of the airports. So there are 10% x (50 + 10)** ~ 10**5 states. Now consider a particular
problem in that domain in which it happens that all the packages are at just 5 of the airports,
and all packages at a given airport have the same destination. Then a useful abstraction of the
problem is to drop all the Ar fluents except for the ones involving one plane and one package
at each of the 5 airports. Now there are only 10% x (5+ 10)° ~ 10'! states. A solution in this abstract state space will be shorter than a solution in the original space (and thus will be an admissible heuristic), and the abstract solution is easy to extend to a solution to the original
problem (by adding additional Load and Unload actions).
356 Decomposition B ence
Chapter 11
Automated Planning
A key idea in defining heuristics is decomposition: dividing a problem into parts, solving each part independently, and then combining the parts. The subgoal independence assumption is that the cost of solving a conjunction of subgoals is approximated by the sum of the costs of solving each subgoal independently. The subgoal independence assumption can be optimistic or pessimistic.
It is optimistic when there are negative interactions between
the subplans for each subgoal—for example, when an action in one subplan deletes a goal achieved by another subplan. It is pessimistic, and therefore inadmissible, when subplans contain redundant actions—for instance, two actions that could be replaced by a single action in the merged plan. Suppose the goal is a set of fluents G, which we divide into disjoint subsets Gy, . ., G. ‘We then find optimal plans P, ..., P, that solve the respective subgoals. What is an estimate
of the cost of the plan for achieving all of G? We can think of each COST(P;) as a heuristic estimate, and we know that if we combine estimates by taking their maximum
value, we
always get an admissible heuristic. So max;COST(P,) is admissible, and sometimes it is exactly correct: it could be that Py serendipitously achieves all the G;. But usually the estimate is too low. Could we sum the costs instead? For many problems that is a reasonable estimate, but it is not admissible. The best case is when G; and G; are independent, in the sense that plans for one cannot reduce the cost of plans for the other. In that case, the estimate
COST(P,) + CosT(P;) is admissible, and more accurate than the max estimate.
It is clear that there is great potential for cutting down the search space by forming abstractions. The trick is choosing the right abstractions and using them in a way that makes the total cost—defining an abstraction, doing an abstract search, and mapping the abstraction back to the original problem—Iless than the cost of solving the original problem. The techniques of pattern databases from Section 3.6.3 can be useful, because the cost of creating the pattern database can be amortized over multiple problem instances. A system that makes use of effective heuristics is FF, or FASTFORWARD
(Hoffmann,
2005), a forward state-space searcher that uses the ignore-delete-lists heuristic, estimating
the heuristic with the help of a planning graph. FF then uses hill climbing search (modified
to keep track of the plan) with the heuristic to find a solution. FF’s hill climbing algorithm is
nonstandard:
it avoids local maxima by running a breadth-first search from the current state
until a better one is found. If this fails, FF switches to a greedy best-first search instead.
11.4
Hierarchical Planning
The problem-solving and planning methods of the preceding chapters all operate with a fixed set of atomic actions.
Actions can be strung together, and state-of-the-art algorithms can
generate solutions containing thousands of actions. That’s fine if we are planning a vacation and the actions are at the level of “fly from San Francisco to Honolulu,” but at the motor-
control level of “bend the left knee by 5 degrees” we would need to string together millions or billions of actions, not thousands.
Bridging this gap requires planning at higher levels of abstraction. A high-level plan for
a Hawaii vacation might be “Go to San Francisco airport; take flight HA 11 to Honolulu; do vacation stuff for two weeks; take HA 12 back to San Francisco; go home.” Given such
a plan, the action “Go to San Francisco airport™ can be viewed as a planning task in itself,
with a solution such as “Choose a ride-hailing service; order a car; ride to airport.” Each of
Section 11.4
Hierarchical Planning
357
these actions, in turn, can be decomposed further, until we reach the low-level motor control
actions like a button-press.
In this example, planning and acting are interleaved; for example, one would defer the problem of planning the walk from the curb to the gate until after being dropped off. Thus, that particular action will remain at an abstract level prior to the execution phase. discussion of this topic until Section 11.5.
We defer
Here, we concentrate on the idea of hierarchi-
cal decomposition, an idea that pervades almost all attempts to manage complexity.
For
example, complex software is created from a hierarchy of subroutines and classes; armies, governments and corporations have organizational hierarchies. The key benefit of hierarchi-
Hierarchical decomposition
cal structure is that at each level of the hierarchy, a computational task, military mission, or
administrative function is reduced to a small number of activities at the next lower level, so
the computational cost of finding the correct way to arrange those activities for the current problem is small. 11.4.1
High-level actions
The basic formalism we adopt to understand hierarchical decomposition comes from the area
of hierarchical task networks or HTN planning. For now we assume full observability and determinism and a set of actions, now called primitive actions, with standard precondition—
effect schemas.
The key additional concept is the high-level action or HLA—for example,
the action “Go to San Francisco airport.” Each HLA has one or more possible refinements,
into a sequence of actions, each of which may be an HLA or a primitive action. For example, the action “Go to San Francisco airport,” represented formally as Go(Home,SFO), might have two possible refinements, as shown in Figure 11.7. The same figure shows a recursive
Hierarchical task network Primitive action High-level action Refinement
refinement for navigation in the vacuum world: to get to a destination, take a step, and then
20 to the destination.
These examples show that high-level actions and their refinements embody knowledge
about how to do things. For instance, the refinements for Go(Home,SFO) say that to get to
the airport you can drive or take a ride-hailing service; buying milk, sitting down, and moving the knight to e4 are not to be considered.
An HLA refinement that contains only primitive actions is called an implementation
of the HLA. In a grid world, the sequences [Right,Right, Down] and [Down, Right,Right] both implement the HLA Navigate([1,3],[3,2]). An implementation of a high-level plan (a sequence of HLAS) is the concatenation of implementations of each HLA in the sequence. Given the precondition—effect definitions of each primitive action, it is straightforward to
determine whether any given implementation of a high-level plan achieves the goal. We can say, then, that a high-level plan achieves the goal from a given state if at least one of its implementations achieves the goal from that state. ~ The “at least one” in this
definition is crucial—not all implementations need to achieve the goal, because the agent gets
to decide which implementation it will execute. Thus, the set of possible implementations in
HTN planning—each of which may have a different outcome—is not the same as the set of
possible outcomes in nondeterministic planning. There, we required that a plan work for all outcomes because the agent doesn’t get to choose the outcome; nature does.
The simplest case is an HLA that has exactly one implementation. In that case, we can
compute the preconditions and effects of the HLA from those of the implementation (see
Exercise 11.HLAU) and then treat the HLA exactly as if it were a primitive action itself. It
Implementation
358
Chapter 11
Automated Planning
Refinement(Go(Home,SFO), STEPS: [Drive(Home, SFOLongTermParking), Shuttle(SFOLongTermParking, SFO)] ) Refinement(Go(Home,SFO), STEPS: [Taxi(Home,SFO)] )
Refinement(Navigate([a,b], [x,]), PRECOND: a=x A b=y
STEPS: ] ) Refinement(Navigate([a.b), x.
PRECOND: Connected([ab], a— 1.b]) STEPS: [Left, Navigate([a— 1,b].[x.])] ) Refinement(Navigate([a,b], [ PRECOND: Connected ([a,b], a + 1,5])
STEPS: [Right. Navigate([a+ 1,5],[x.])]) Figure 11.7 Definitions of possible refinements for two high-level actions: going to San Francisco airport and navigating in the vacuum world. In the latter case, note the recursive nature of the refinements and the use of preconditions.
can be shown that the right collection of HLAs can result in the time complexity of blind
search dropping from exponential in the solution depth to linear in the solution depth, although devising such a collection of HLAs may be a nontrivial task in itself. When HLAs
have multiple possible implementations, there are two options: one is to search among the implementations for one that works, as in Section 11.4.2; the other is to reason directly about
the HLAs—despite the multiplicity of implementations—as explained in Section 11.4.3. The latter method enables the derivation of provably correct abstract plans, without the need to
consider their implementations.
11.4.2
Searching for primitive solutions
HTN planning is often formulated with a single “top level” action called Act, where the aim is to find an implementation of Act that achieves the goal. This approach is entirely general. For example, classical planning problems can be defined as follows: for each primitive action
a;, provide one refinement of Act with steps [a;, Act]. That creates a recursive definition of Act
that lets us add actions. But we need some way to stop the recursion; we do that by providing
one more refinement for Act, one with an empty list of steps and with a precondition equal to the goal of the problem.
This says that if the goal is already achieved, then the right
implementation is to do nothing. The approach leads to a simple algorithm: repeatedly choose an HLA in the current plan and replace it with one of its refinements, until the plan achieves the goal. One possible implementation based on breadth-first tree search is shown in Figure 11.8. Plans are considered
in order of depth of nesting of the refinements, rather than number of primitive steps. It is straightforward to design a graph-search version of the algorithm as well as depth-first and iterative deepening versions.
Section 11.4
Hierarchical Planning
function HIERARCHICAL-SEARCH(problem, hierarchy) returns a solution or failure
frontier —a FIFO queue with [Aci] as the only element while rrue do if IS-EMPTY( frontier) then return failure
plan — POP(frontier)
1/ chooses the shallowest plan in frontier
hla + the first HLA in plan, or null if none
prefix,suffix — the action subsequences before and after hla in plan outcome«— RESULT(problem.INITIAL, prefix) if hla is null then
// so plan is primitive and outcome is its result
if problem.1s-GOAL(outcome) then return plan else for each sequence in REFINEMENTS (hla, outcome, hierarchy) do add APPEND(prefix, sequence, suffix) to frontier
Figure 11.8 A breadth-first implementation of hierarchical forward planning search. The
initial plan supplied to the algorithm is [Acr]. The REFINEMENTS function returns a set of
action sequences, one for each refinement of the HLA whose preconditions are satisfied by the specified state, outcome.
In essence, this form of hierarchical search explores the space of sequences that conform to the knowledge contained in the HLA library about how things are to be done. A great deal of knowledge can be encoded, not just in the action sequences specified in each refinement but also in the preconditions for the refinements. For some domains, HTN planners have been able to generate huge plans with very little search. For example, O-PLAN (Bell and Tate, 1985), which combines HTN planning with scheduling, has been used to develop production plans for Hitachi. A typical problem involves a product line of 350 different products, 35 assembly machines, and over 2000 different operations. The planner generates a 30-day schedule with three 8-hour shifts a day, involving tens of millions of steps. Another important
aspect of HTN plans is that they are, by definition, hierarchically structured; usually this makes them easy for humans to understand.
The computational benefits of hierarchical search can be seen by examining an ideal-
ized case.
Suppose that a planning problem has a solution with d primitive actions.
a nonhierarchical,
For
forward state-space planner with b allowable actions at each state, the
cost is O(b?), as explained in Chapter 3. For an HTN planner, let us suppose a very regular refinement structure:
each nonprimitive action has r possible refinements,
k actions at the next lower level. there are with this structure.
Now,
each into
We want to know how many different refinement trees
if there are d actions at the primitive level, then the
number of levels below the root is log,d, so the number of internal refinement nodes is
14 k+ K24+ K24~
= (¢ — 1) /(k— 1). Each internal node has r possible refinements,
s0 rld=1/(=1) possible decomposition trees could be constructed.
Examining this formula, we see that keeping r small and k large can result in huge sav-
ings: we are taking the kth root of the nonhierarchical cost, if » and r are comparable. Small r
and large k means a library of HLAs with a small number of refinements each yielding a long
action sequence. This is not always possible: long action sequences that are usable across a
wide range of problems are extremely rare.
359
360
Chapter 11
Automated Planning
The key to HTN planning is a plan library containing known methods for implementing
complex, high-level actions.
One way to construct the library is to learn the methods from
problem-solving experience. After the excruciating experience of constructing a plan from scratch, the agent can save the plan in the library as a method for implementing the high-level
action defined by the task. In this way, the agent can become more and more competent over
time as new methods are built on top of old methods. One important aspect of this learning process is the ability to generalize the methods that are constructed, eliminating detail that is specific to the problem instance (e.g., the name of the builder or the address of the plot of land) and keeping just the key elements of the plan. It seems to us inconceivable that humans could be as competent as they are without some such mechanism. 11.4.3
Searching for abstract solutions
The hierarchical search algorithm in the preceding section refines HLAs all the way to primitive action sequences to determine if a plan is workable. This contradicts common sense: one should be able to determine that the two-HLA high-level plan
[Drive(Home,SFOLongTermParking), Shuttle(SFOLongTermParking, SFO)]
gets one to the airport without having to determine a precise route, choice of parking spot, and so on. The solution is to write precondition—effect descriptions of the HLAs, just as we
do for primitive actions. From the descriptions, it ought to be easy to prove that the high-level
plan achieves the goal. This is the holy grail, so to speak, of hierarchical planning, because if we derive a high-level plan that provably achieves the goal, working in a small search space of high-level actions, then we can commit to that plan and work on the problem of refining each step of the plan. This gives us the exponential reduction we seek.
For this to work, it has to be the case that every high-level plan that “claims” to achieve
Downward refinement property
the goal (by virtue of the descriptions of its steps) does in fact achieve the goal in the sense defined earlier: it must have at least one implementation that does achieve the goal. This
property has been called the downward refinement property for HLA descriptions.
Writing HLA descriptions that satisfy the downward refinement property is, in principle,
easy: as long as the descriptions are frue, then any high-level plan that claims to achieve
the goal must in fact do so—otherwise, the descriptions are making some false claim about
what the HLAs do. We have already seen how to write true descriptions for HLAs that have
exactly one implementation (Exercise 11.HLAU); a problem arises when the HLA has multiple implementations.
How can we describe the effects of an action that can be implemented in
many different ways? One safe answer (at least for problems where all preconditions and goals are positive) is
to include only the positive effects that are achieved by every implementation of the HLA and
the negative effects of any implementation. Then the downward refinement property would be satisfied. Unfortunately, this semantics for HLAs is much too conservative.
Consider again the HLA Go(Home, SFO), which has two refinements, and suppose, for the sake of argument, a simple world in which one can always drive to the airport and park, but taking a taxi requires Cash as a precondition. In that case, Go(Home, SFO) doesn’t always get you to the airport.
In particular, it fails if Cash is false, and so we cannot assert
At(Agent,SFO) as an effect of the HLA. This makes no sense, however; if the agent didn’t
have Cash, it would drive itself. Requiring that an effect hold for every implementation is
equivalent to assuming that someone else—an adversary—will choose the implementation.
Section 11.4
(@)
Hierarchical Planning
361
(b)
Figure 11.9 Schematic examples of reachable sets. The set of goal states is shaded in purple. Black and gray arrows indicate possible implementations of Ay and ha, respectively. (a) The reachable set of an HLA hy in a state s. (b) The reachable set for the sequence [k, ha]. Because this intersects the goal set, the sequence achieves the goal. It treats the HLA’s multiple outcomes exactly as if the HLA were a nondeterministic action,
as in Section 4.3. For our case, the agent itself will choose the implementation.
The programming languages community has coined the term demonic nondeterminism Demeric
for the case where an adversary makes the choices, contrasting this with angelic nondeterminism, where the agent itself makes the choices. We borrow this term to define angelic semantics for HLA descriptions. The basic concept required for understanding angelic se-
mantics is the reachable set of an HLA: given a state s, the reachable set for an HLA h,
written as REACH (s, h), is the set of states reachable by any of the HLA’s implementations.
The key idea is that the agent can choose which element of the reachable set it ends up in when it executes the HLA; thus, an HLA with multiple refinements is more “powerful” than the same HLA with fewer refinements. We can also define the reachable set of a sequence of HLAs. For example, the reachable set of a sequence [hy, /] is the union of all the reachable
sets obtained by applying h, in each state in the reachable set of h;: REACH(s, [, ha]) = U REACH(S, ). #eREACH(s.In) Given these definitions, a high-level plan—a sequence of HLAs—achieves the goal if its reachable set intersects the set of goal states. (Compare this to the much stronger condition for demonic semantics,
where every member of the reachable set has to be a goal state.)
Conversely, if the reachable set doesn’t intersect the goal, then the plan definitely doesn’t work. Figure 11.9 illustrates these ideas.
The notion of reachable sets yields a straightforward algorithm: search among highlevel plans, looking for one whose reachable set intersects the goal; once that happens, the
algorithm can commit to that abstract plan, knowing that it works, and focus on refining the
plan further. We will return to the algorithmic issues later; for now consider how the effects
:;g;l‘; rmminism Angelic semantics Reachable set
362
Chapter 11
Automated Planning
of an HLA—the reachable set for each possible initial state—are represented.
action can set a fluent to true or false or leave it unchanged.
A primitive
(With conditional effects (see
Section 11.5.1) there is a fourth possibility: flipping a variable to its opposite.)
An HLA under angelic semantics can do more: it can control the value of a fluent, setting
it to true or false depending on which implementation is chosen. That means that an HLA can
have nine different effects on a fluent: if the variable starts out true, it can always keep it true,
always make it false, or have a choice; if the fluent starts out false, it can always keep it false, always make it true, or have a choice; and the three choices for both cases can be combined
arbitrarily, making nine. Notationally, this is a bit challenging. We’ll use the language of add lists and delete lists (rather than true/false fluents) along with the ~ symbol to mean “possibly, if the agent
so chooses.” Thus, the effect -4 means “possibly add A,” that is, either leave A unchanged
or make it true. or delete A
Similarly, —A means “possibly delete A” and FA means “possibly add
For example, the HLA
Go(Home,SFO),
with the two refinements shown in
Figure 11.7, possibly deletes Cash (if the agent decides to take a taxi), so it should have the
effect ~Cash. Thus, we see that the descriptions of HLAs are derivable from the descriptions
of their refinements. Now, suppose we have the following schemas for the HLAs /; and h»: Action(hy, PRECOND: A, EFFECT:A A ~B)
Action(hy, PRECOND: ~B, EFFECT: +A A £C) That is, /; adds A and possibly deletes B, while &, possibly adds A and has full control over
C. Now, if only B is true in the initial state and the goal is A A C then the sequence [hy,hs]
achieves the goal: we choose an implementation of /; that makes B false, then choose an implementation of /2, that leaves A true and makes C true. The preceding discussion assumes that the effects of an HLA—the reachable set for any
given initial state—can be described exactly by describing the effect on each fluent. It would
be nice if this cause an HLA gly reachable page 243. For
were always true, but in many cases we can only approximate may have infinitely many implementations and may produce sets—rather like the wiggly-belief-state problem illustrated in example, we said that Go(Home, SFO) possibly deletes Cash;
the effects bearbitrarily wigFigure 7.21 on it also possibly
adds At(Car,SFOLongTermParking); but it cannot do both—in fact, it must do exactly one.
Optimistic lescription essimistic description
As with belief states, we may need to write approximate descriptions. We will use two kinds of approximation: an optimistic description REACH* (s, ) of an HLA h may overstate the reachable set, while a pessimistic description REACH ™ (s, 1) may understate the reachable set. Thus, we have
REACH ™ (s,h) C REACH(s, h) C REACHT (s, h)
For example, an optimistic description of Go(Home, SFO) says that it possibly deletes Cash and possibly adds At(Car,SFOLongTermParking). Another good example arises in the 8puzzle, half of whose states are unreachable from any given state (see Exercise 11.PART): the optimistic description of Act might well include the whole state space, since the exact reachable set is quite wiggly.
With approximate descriptions, the test for whether a plan achieves the goal needs to be modified slightly. If the optimistic reachable set for the plan doesn’t intersect the goal, then the plan doesn’t work; if the pessimistic reachable set intersects the goal, then the plan
does work (Figure 11.10(a)). With exact descriptions, a plan either works or it doesn’t, but
Section 11.4
Hierarchical Planning
Figure 11.10 Goal achievement for high-level plans with approximate descriptions. The set of goal states is shaded in purple. For each plan, the pessimistic (solid lines, light blue) and optimistic (dashed lines, light green) reachable sets are shown. (a) The plan indicated by the black arrow definitely achieves the goal, while the plan indicated by the gray arrow definitely doesn’t. (b) A plan that possibly achieves the goal (the optimistic reachable set intersects the goal) but does not necessarily achieve the goal (the pessimistic reachable set does not intersect the goal). The plan would need to be refined further to determine if it really does achieve the goal. with approximate descriptions, there is a middle ground: if the optimistic set intersects the goal but the pessimistic set doesn’t, then we cannot tell if the plan works (Figure 11.10(b)).
When this circumstance arises, the uncertainty can be resolved by refining the plan. This is a very common situation in human reasoning. For example, in planning the aforementioned
two-week Hawaii vacation, one might propose to spend two days on each of seven islands.
Prudence would indicate that this ambitious plan needs to be refined by adding details of inter-island transportation.
An algorithm for hierarchical planning with approximate angelic descriptions is shown in Figure 11.11. For simplicity, we have kept to the same overall scheme used previously in Figure 11.8, that is, a breadth-first search in the space of refinements. As just explained,
the algorithm can detect plans that will and won’t work by checking the intersections of the optimistic and pessimistic reachable sets with the goal. (The details of how to compute
the reachable sets of a plan, given approximate descriptions of each step, are covered in Exercise 1 1.LHLAP.)
When a workable abstract plan is found, the algorithm decomposes the original problem into subproblems, one for each step of the plan. The initial state and goal for each subproblem are obtained by regressing a guaranteed-reachable goal state through the action schemas for each step of the plan. (See Section 11.2.2 for a discussion of how regression works.) Figure 11.9(b) illustrates the basic idea: the right-hand circled state is the guaranteed-reachable
goal state, and the left-hand circled state is the intermediate goal obtained by regressing the
goal through the final action.
363
364
Chapter 11
Automated Planning
function ANGELIC-SEARCH(problem, hierarchy, initialPlan) returns solution or fail Jfrontier +a FIFO queue with initialPlan as the only element while rrue do if EMPTY?(frontier) then return fail
plan & POP(frontier) /1 chooses the shallowest node in frontier if REACH* (problem.INITIAL, plan) intersects problem.GOAL then if plan is primitive then return plan
// REACH" is exact for primitive plans
guaranteed— REACH™ (problem.INITIAL, plan) N problem.GOAL
if guaranteed#{ } and MAKING-PROGRESS(plan, initialPlan) then
JfinalState —any element of guaranteed return DECOMPOSE (hierarchy, problem.INITIAL, plan, finalState)
hla+some HLA in plan
prefix,suffix < the action subsequences before and after hla in plan outcome < RESULT(problem.INITIAL, prefix) for each sequence in REFINEMENTS(hla, outcome, hierarchy) do Sfrontier «— Insert( APPEND(prefix, sequence, suffix), frontier)
function DECOMPOSE hierarchy, so. plan, sy) returns a solution solution «an empty plan while plan is not empty do action < REMOVE-LAST(plan)
si¢—astate in REACH™ (sp, plan) such that s;€REACH™ (s;, action)
problem «—a problem with INITIAL = s; and GOAL = sy solution « APPEND(ANGELIC-SEARCH(problem, hierarchy, action), solution) Spesi
return solution
Figure 11.11 A hierarchical planning algorithm that uses angelic semantics to identify and commit to high-level plans that work while avoiding high-level plans that don’t. The predicate MAKING-PROGRESS
checks to make sure that we aren’t stuck in an infinite regression
of refinements. At top level, call ANGELIC-SEARCH with [Act] as the initialPlan.
The ability to commit to or reject high-level plans can give ANGELIC-SEARCH
a sig-
nificant computational advantage over HIERARCHICAL-SEARCH, which in turn may have a
large advantage over plain old BREADTH-FIRST-SEARCH.
Consider, for example, cleaning
up a large vacuum world consisting of an arrangement of rooms connected by narrow corridors, where each room is a w x h rectangle of squares. It makes sense to have an HLA for Navigate (as shown in Figure 11.7) and one for CleanWholeRoom. (Cleaning the room could
be implemented with the repeated application of another HLA to clean each row.) Since there
are five primitive actions, the cost for BREADTH-FIRST-SEARCH grows as 59, where d is the
length of the shortest solution (roughly twice the total number of squares); the algorithm
cannot manage even two 3 x 3 rooms.
HIERARCHICAL-SEARCH
is more efficient, but still
suffers from exponential growth because it tries all ways of cleaning that are consistent with the hierarchy. ANGELIC-SEARCH scales approximately linearly in the number of squares— it commits to a good high-level sequence of room-cleaning and navigation steps and prunes
away the other options.
Section 1.5
Planning and Acting in Nondeterministic Domains
365
Cleaning a set of rooms by cleaning each room in turn is hardly rocket science: it is
easy for humans because of the hierarchical structure of the task. When we consider how
difficult humans find it to solve small puzzles such as the 8-puzzle, it seems likely that the
human capacity for solving complex problems derives not from considering combinatorics, but rather from skill in abstracting and decomposing problems to eliminate combinatorics.
The angelic approach can be extended to find least-cost solutions by generalizing the
notion of reachable set. Instead of a state being reachable or not, each state will have a cost for the most efficient way to get there.
(The cost is
infinite for unreachable states.)
optimistic and pessimistic descriptions bound these costs.
The
In this way, angelic search can
find provably optimal abstract plans without having to consider their implementations. The
same approach can be used to obtain effective hierarchical look-ahead algorithms for online
search, in the style of LRTA* (page 140). In some ways, such algorithms mirror aspects of human deliberation in tasks such as planning a vacation to Hawaii—consideration of alternatives is done initially at an abstract
level over long time scales; some parts of the plan are left quite abstract until execution time,
such as how to spend two lazy days on Moloka'i, while others parts are planned in detail, such as the flights to be taken and lodging to be reserved—without these latter refinements,
there is no guarantee that the plan would be feasible. 11.5
Planning and Acting in Nondeterministic
Domains
In this section we extend planning to handle partially observable, nondeterministic, and un-
known environments. The basic concepts mirror those in Chapter 4, but there are differences
arising from the use of factored representations rather than atomic representations. This affects the way we represent the agent’s capability for action and observation and the way
we represent belief states—the sets of possible physical states the agent might be in—for partially observable environments.
We can also take advantage of many of the domain-
independent methods given in Section 11.3 for calculating search heuristics.
‘We will cover sensorless planning (also known as conformant planning) for environ-
ments with no observations; contingency planning for partially observable and nondeterministic environments; and online planning and replanning for unknown environments. This
will allow us to tackle sizable real-world problems.
Consider this problem:
given a chair and a table, the goal is to have them match—have
the same color. In the initial state we have two cans
of paint, but the colors of the paint and
the furniture are unknown. Only the table is initially in the agent’s field of view:
Init(Object(Tuble) A Object(Chair) A Can(Cy) A Can(Cz) A InView(Tuble)) Goal(Color(Chair,c) A Color(Table,c))
There are two actions:
removing the lid from a paint can and painting an object using the
paint from an open can. Action(RemoveLid(can), PRECOND:Can(can) EFFECT: Open(can))
Action(Paint(x,can),
PRECOND:Object(x) A Can(can) A Color(can,c) A Open(can) EFFECT: Color(x,c))
Hierarchical look-ahead
366
Chapter 11
Automated Planning
The action schemas are straightforward, with one exception: preconditions and effects now may contain variables that are not part of the action’s variable list.
That is, Paint(x,can)
does not mention the variable ¢, representing the color of the paint in the can. In the fully
observable case, this is not allowed—we would have to name the action Paint(x,can,c). But
in the partially observable case, we might or might not know what color is in the can.
To solve a partially observable problem, the agent will have to reason about the percepts
it will obtain when it is executing the plan.
The percept will be supplied by the agent’s
sensors when it is actually acting, but when it is planning it will need a model of its sensors. Percept schema
In Chapter 4, this model was given by a function, PERCEPT(s).
PDDL with a new type of schema, the percept schema: Percept(Color(x.c),
For planning, we augment
PRECOND: Object(x) A InView(x)
Percept(Color(can,c),
PRECOND:Can(can) A InView(can) A Open(can)
The first schema says that whenever an object is in view, the agent will perceive the color of the object (that is, for the object x, the agent will learn the truth value of Color(x,c) for all ¢). The second schema says that if an open can is in view, then the agent perceives the color of the paint in the can. Because there are no exogenous events in this world, the color
of an object will remain the same, even if it is not being perceived, until the agent performs
an action to change the object’s color. Of course, the agent will need an action that causes objects (one at a time) to come into view:
Action(LookAt(x),
PRECOND:InView(y) A (x # y) EFFECT: InView(x) A —~InView(y))
For a fully observable environment, we would have a Percept schema with no preconditions
for each fluent. A sensorless agent, on the other hand, has no Percept schemas at all. Note
that even a sensorless agent can solve the painting problem. One solution is to open any can of paint and apply it to both chair and table, thus coercing them to be the same color (even
though the agent doesn’t know what the color is). A contingent planning agent with sensors can generate a better plan. First, look at the
table and chair to obtain their colors; if they are already the same then the plan is done. If
not, look at the paint cans; if the paint in a can is the same color as one piece of furniture,
then apply that paint to the other piece. Otherwise, paint both pieces with any color.
Finally, an online planning agent might generate a contingent plan with fewer branches
at first—perhaps ignoring the possibility that no cans match any of the furniture—and deal
with problems when they arise by replanning. It could also deal with incorrectness of its
action schemas.
Whereas a contingent planner simply assumes that the effects of an action
always succeed—that painting the chair does the job—a replanning agent would check the result and make an additional plan to fix any unexpected failure, such as an unpainted area or the original color showing through. In the real world, agents use a combination of approaches. Car manufacturers sell spare tires and air bags, which are physical embodiments of contingent plan branches designed to handle punctures or crashes.
On the other hand, most car drivers never consider these
possibilities; when a problem arises they respond as replanning agents. In general, agents
Section 1.5
Planning and Acting in Nondeterministic Domains
plan only for contingencies that have important consequences and a nonnegligible chance of happening. Thus, a car driver contemplating a trip across the Sahara desert should make explicit contingency plans for breakdowns, whereas a trip o the supermarket requires less advance planning. We next look at cach of the three approaches in more detail. 11.5.1
Sensorless planning
Section 4.4.1 (page 126) introduced the basic idea of searching in belief-state space to find
a solution for sensorless problems. Conversion of a sensorless planning problem to a beliefstate planning problem works much the same way as it did in Section 4.4.1; the main dif-
ferences are that the underlying physical transition model is represented by a collection of
action schemas, and the belief state can be represented by a logical formula instead of by
an explicitly enumerated set of states. We assume that the underlying planning problem is deterministic.
The initial belief state for the sensorless painting problem can ignore InView fluents
because the agent has no sensors. Furthermore, we take as given the unchanging facts Object(Table) A Object(Chair) A Can(Cy) A Can(Cs) because these hold in every belief state. The agent doesn’t know the colors of the cans or the objects, or whether the cans are open or closed, but it does know that objects and cans have colors:
Skolemizing (see Section 9.5.1), we obtain the initial belief state:
Vx
3¢
Color(x,c).
After
bo = Color(x,C(x)).
In classical planning, where the closed-world assumption is made, we would assume that any fluent not mentioned in a state is false, but in sensorless (and partially observable) plan-
ning we have to switch to an open-world assumption in which states contain both positive and negative fluents, and if a fluent does not appear, its value is unknown.
Thus, the belief
state corresponds exactly to the set of possible worlds that satisfy the formula. Given this initial belief state, the following action sequence is a solution:
[RemoveLid(Cany ), Paint(Chair, Cany ), Paint(Table, Cany ).
‘We now show how to progress the belief state through the action sequence to show that the
final belief state satisfies the goal.
First, note that in a given belief state b, the agent can consider any action whose preconditions are satisfied by b. (The other actions cannot be used because the transition model doesn’t define the effects of actions whose preconditions might be unsatisfied.) According
to Equation (4.4) (page 127), the general formula for updating the belief state b given an applicable action a in a deterministic world is as follows:
b =RESULT(b,a) = {s' : ' =RESULTp(s,a) and s € b} where RESULTp defines the physical transition model. For the time being, we assume that the
initial belief state is always a conjunction of literals, that is, a 1-CNF formula. To construct the new belief state 5, we must consider what happens to each literal £ in each physical state s in b when action a is applied. For literals whose truth value is already known in b, the truth value in b is computed from the current value and the add list and delete list of the action. (For example, if £ is in the delete list of the action, then —¢ is added to &'.) What about a literal whose truth value is unknown in b? There are three cas 1. If the action adds ¢, then ¢ will be true in b’ regardless of its initial value.
367
368
Chapter 11
Automated Planning
2. If the action deletes £, then ¢ will be false in b’ regardless of its initial value.
3. If the action does not affect £, then ¢ will retain its initial value (which is unknown) and will not appear in 4. Hence, we see that the calculation of 4/ is almost identical to the observable case, which was specified by Equation (11.1) on page 345: b =ResuLT(b,a) = (b— DEL(a)) UADD(a). ‘We cannot quite use the set semantics because (1) we must make sure that b’ does not contain both ¢ and —¢, and (2) atoms may contain unbound variables.
But it is still the case
that RESULT (b, a) is computed by starting with b, setting any atom that appears in DEL(a) to false, and setting any atom that appears in ADD(a) to true. For example, if we apply
RemoveLid(Cany) to the initial belief state by, we get
by = Color(x,C(x)) A Open(Can,). When we apply the action Paint(Chair,Cany), the precondition Color(Cany,c) is satisfied by the literal Color(x,C(x)) with binding {x/Cany,c/C(Can;)} and the new belief state is by = Color(x,C(x)) A Open(Cany) A Color(Chair,C(Cany ). Finally, we apply the action Paint(Table, Cany) to obtain by = Color(x,C(x)) A Open(Can,) A Color(Chair,C(Cany ) A Color(Table,C(Cany)).
The final belief state satisfies the goal, Color(Table,c) A Color(Chair,c), with the variable ¢ bound to C(Cany). The preceding analysis of the update rule has
shown a very important fact: the family
of belief states defined as conjunctions of literals is closed under updates defined by PDDL action schemas.
That is, if the belief state starts as a conjunction of literals, then any update
will yield a conjunction of literals.
That means that in a world with n fluents, any belief
state can be represented by a conjunction of size O(n).
This is a very comforting result,
considering that there are 2" states in the world. It says we can compactly represent all the
subsets of those 2" states that we will ever need. Moreover, the process of checking for belief
states that are subsets or supersets of previously visited belief states is also easy, at least in
the propositional case.
The fly in the ointment of this pleasant picture is that it only works for action schemas
that have the same effects for all states in which their preconditions are satisfied.
It is this
property that enables the preservation of the 1-CNF belief-state representation. As soon as
the effect can depend on the state, dependencies are introduced between fluents, and the 1-
CNF property is lost.
Consider, for example, the simple vacuum world defined in Section 3.2.1. Let the fluents be AzL and AtR for the location of the robot and CleanL and CleanR for the state of the
squares. According to the definition of the problem, the Suck action has no precondition—it
Conditional effect
can always be done. The difficulty is that its effect depends on the robot’s location: when the robot is ArL, the result is CleanL, but when it is AzR, the result is CleanR. For such actions, our action schemas will need something new: a conditional effect. These have the syntax
Section 1.5
Planning and Acting in Nondeterministic Domains
“when condition: effect,” where condition is a logical formula to be compared against the current state, and effect is a formula describing the resulting state. For the vacuum world:
Action(Suck,
EFFECT:when ArL: CleanL \ when AtR: CleanR) .
When applied to the initial belief state True, the resulting belief state is (AL A CleanL) V' (ARA CleanR), which is no longer in 1-CNF. (This transition can be seen in Figure 4.14
on page 129.) In general, conditional effects can induce arbitrary dependencies among the fluents in a belief state, leading to belief states of exponential size in the worst case.
Itis important to understand the difference between preconditions and conditional effects.
All conditional effects whose conditions are satisfied have their effects applied to generate the resulting belief state; if none are satisfied, then the resulting state is unchanged. On the other hand, if a precondition is unsatisfied, then the action is inapplicable and the resulting state is undefined.
From the point of view of sensorless planning, it is better to have conditional
effects than an inapplicable action. unconditional effects as follows:
Action(SuckL,
PRECOND:AL;
Action(SuckR,
For example, we could split Suck into two actions with
EFFECT: CleanL)
PRECOND:AtR; EFFECT: CleanR).
Now we have only unconditional schemas, so the belief states all remain in 1-CNF; unfortu-
nately, we cannot determine the applicability of SuckL and SuckR in the initial belief state.
It seems inevitable, then, that nontrivial problems will involve wiggly belief states, just
like those encountered when we considered the problem of state estimation for the wumpus
world (see Figure 7.21 on page 243). The solution suggested then was to use a conservative
approximation to the exact belief state; for example, the belief state can remain in 1-CNF
if it contains all literals whose truth values can be determined and treats all other literals as
unknown.
While this approach is sound, in that it never generates an incorrect plan, it is
incomplete because it may be unable to find solutions to problems that necessarily involve interactions among literals.
To give a trivial example, if the goal is for the robot to be on
a clean square, then [Suck] is a solution but a sensorless agent that insists on 1-CNF belief
states will not find it.
Perhaps a better solution is to look for action sequences that keep the belief state as simple as possible. In the sensorless vacuum world, the action sequence [Right, Suck, Left, Suck] generates the following sequence of belief states: by
=
True
by
=
AR
by = ARACleanR by = AILACleanR by
= AtL A CleanR N CleanL
That is, the agent can solve the problem while retaining a 1-CNF belief state, even though
some sequences (e.g., those beginning with Suck) go outside 1-CNE. The general lesson is not lost on humans: we are always performing little actions (checking the time, patting our
369
370
Chapter 11
Automated Planning
pockets to make sure we have the car keys, reading street signs as we navigate through a city) to eliminate uncertainty and keep our belief state manageable.
There is another, quite different approach to the problem of unmanageably wiggly belief states: don’t bother computing them at all. Suppose the initial belief state is by and we would like to know the belief state resulting from the action sequence [a1, ...,
). Instead of com-
puting it explicitly, just represent it as “by then [ar, ..., a,,].” This is a lazy but unambiguous
representation of the belief state, and it’s quite concise—O(n +m) where n is the size of the initial belief state (assumed to be in 1-CNF) and m is the maximum
length of an action se-
quence. As a belief-state representation, it suffers from one drawback, however: determining
whether the goal is satisfied, or an action is applicable, may require a lot of computation.
The computation can be implemented as an entailment test: if A,, represents the collec-
tion of successor-state axioms required to define occurrences of the actions aj....,a,—as
explained for SATPLAN in Section 11.2.3—and G, asserts that the goal is true after m steps,
then the plan achieves the goal if by A A,, = G—that is, if by A Ay A G, is unsatisfiable. Given a modern SAT solver, it may be possible to do this much more quickly than computing the full belief state. For example, if none of the actions in the sequence has a particular goal fluent in its add list, the solver will detect this immediately.
It also helps if partial results
about the belief state—for example, fluents known to be true or false—are cached to simplify
subsequent computations.
The final piece of the sensorless planning puzzle is a heuristic function to guide the
search. The meaning of the heuristic function is the same as for classical planning: an esti-
mate (perhaps admissible) of the cost of achieving the goal from the given belief state. With
belief states, we have one additional fact: solving any subset of a belief state is necessarily easier than solving the belief state:
if by C by then h*(by) < h*(b2). Hence, any admissible heuristic computed for a subset is admissible for the belief state itself. The most obvious candidates are the singleton subsets, that is, individual physical states. We
can take any random collection of states s admissible heuristic /, and return
sy that are in the belief state b, apply any
H(b) =max{h(s1),....h(sn)}
as the heuristic estimate for solving b. We can also use inadmissible heuristics such as the ignore-delete-lists heuristic (page 354), which seems to work quite well in practice. 11.5.2
Contingent planning
We saw in Chapter 4 that contingency planning—the generation of plans with conditional
branching based on percepts—is appropriate for environments with partial observability, non-
determinism, or both. For the partially observable painting problem with the percept schemas given earlier, one possible conditional solution is as follows:
[LookAt(Table), LookAt(Chair), if Color(Tuble,c) A Color(Chair, c) then NoOp else [RemoveLid(Can, ), LookAt(Can, ), RemoveLid(Cans), LookAt(Cans), if Color(Table,c) A Color(can, c) then Paint(Chair, can) else if Color(Chair,c) A Color(can,c) then Paint(Table,can) else [Paint(Chair, Can), Paint (Table, Can, )|]]
Section 1.5
Planning and Acting in Nondeterministic Domains
Variables in this plan should be considered existentially quantified;
the second
line says
that if there exists some color ¢ that is the color of the table and the chair, then the agent
need not do anything to achieve the goal. When executing this plan, a contingent-planning
agent can maintain its belief state as a logical formula and evaluate each branch condition
by determining if the belief state entails the condition formula or its negation.
(It is up to
the contingent-planning algorithm to make sure that the agent will never end up in a belief
state where the condition formula’s truth value is unknown.)
Note that with first-order
conditions, the formula may be satisfied in more than one way; for example, the condition Color(Table, c) A Color(can, ¢) might be satisfied by {can/Cany } and by {can/Cans} if both cans are the same color as the table. In that case, the agent can choose any satisfying substitution to apply to the rest of the plan.
As shown in Section 4.4.2, calculating the new belief state b after an action a and subse-
quent percept is done in two stages. The first stage calculates the belief state after the action,
just as for the sensorless agent: b= (b—DEL(a))UADD(a)
where, as before, we have assumed a belief state represented as a conjunction of literals. The
second stageis a little trickier. Suppose that percept literals p;., ..., py are received. One might
think that we simply need to add these into the belief state; in fact, we can also infer that the
preconditions for sensing are satisfied. Now, if a percept p has exactly one percept schema, Percept(p, PRECOND:c), where c is a conjunction of literals, then those literals can be thrown into the belief state along with p. On the other hand, if p has more than one percept schema
whose preconditions might hold according to the predicted belief state b, then we have to add in the disjunction of the preconditions. Obviously, this takes the belief state outside 1-CNF
and brings up the same complications as conditional effects, with much the same classes of
solutions.
Given a mechanism for computing exact or approximate belief states, we can generate
contingent plans with an extension of the AND—OR
forward search over belief states used
in Section 4.4. Actions with nondeterministic effects—which are defined simply by using a
disjunction in the EFFECT of the action schema—can be accommodated with minor changes
to the belief-state update calculation and no change to the search algorithm.® For the heuristic
function, many of the methods suggested for sensorless planning are also applicable in the partially observable, nondeterministic case. 11.5.3
Online planning
Imagine watching a spot-welding robot in a car plant. The robot’s fast, accurate motions are
repeated over and over again as each car passes down the line. Although technically impressive, the robot probably does not seem at all intelligent because the motion is a fixed, preprogrammed sequence; the robot obviously doesn’t “know what it’s doing” in any meaningful sense.
Now suppose that a poorly attached door falls off the car just as the robot is
about to apply a spot-weld. The robot quickly replaces its welding actuator with a gripper,
picks up the door, checks it for scratches, reattaches it to the car, sends an email to the floor supervisor, switches back to the welding actuator, and resumes its work.
All of a sudden,
3 If cyclic solutions are required for a nondeterministic problem, AND—OR search must be generalized to a loopy version such as LAO" (Hansen and Zilberstein, 2001).
371
372
Chapter 11
Automated Planning
the robot’s behavior seems purposive rather than rote; we assume it results not from a vast,
precomputed contingent plan but from an online replanning process—which means that the
Execution monitoring
robot does need to know what it’s trying to do. Replanning presupposes some form of execution monitoring to determine the need for a new plan. One such need arises when a contingent planning agent gets tired of planning
for every little contingency, such as whether the sky might fall on its head.* This means that the contingent plan is left in an incomplete form.
For example, Some branches of a
partially constructed contingent plan can simply say Replan; if such a branch is reached
during execution, the agent reverts to planning mode. As we mentioned earlier, the decision
as to how much of the problem to solve in advance and how much to leave to replanning
is one that involves tradeoffs among possible events with different costs and probabilities of
occurring.
Nobody wants to have a car break down in the middle of the Sahara desert and
only then think about having enough water.
Missing precondition Missing effect
Missing fluent Exogenous event
Replanning may be needed if the agent’s model of the world is incorrect. The model
for an action may have a missing precondition—for example, the agent may not know that
removing the lid of a paint can often requires a screwdriver. The model may have a missing effect—painting an object may get paint on the floor as well.
Or the model may have a
missing fluent that is simply absent from the representation altogether—for example, the model given earlier has no notion of the amount of paint in a can, of how its actions affect
this amount, or of the need for the amount to be nonzero. The model may also lack provision
for exogenous events such as someone knocking over the paint can. Exogenous events can
also include changes in the goal, such as the addition of the requirement that the table and
chair not be painted black. Without the ability to monitor and replan, an agent’s behavior is likely to be fragile if it relies on absolute correctness of its model.
The online agent has a choice of (at least) three different approaches for monitoring the environment during plan execution:
Action monitoring
+ Action monitoring: before executing an action, the agent verifies that all the precondi-
Plan monitoring
+ Plan monitoring: before executing an action, the agent verifies that the remaining plan
Goal monitoring
* Goal monitoring: before executing an action, the agent checks to see if there is a better
tions still hold.
will still succeed.
set of goals it could be trying to achieve.
In Figure 11.12 we see a schematic of action monitoring.
The agent keeps track of both its
original plan, whole plan, and the part of the plan that has not been executed yet, which is denoted by plan.
After executing the first few steps of the plan, the agent expects to be in
state E. But the agent observes that it is actually in state O. It then needs to repair the plan by
finding some point P on the original plan that it can get back to. (It may be that P is the goal
state, G.) The agent tries to minimize the total cost of the plan: the repair part (from O to P)
plus the continuation (from P to G).
# In 1954, a Mrs. Hodges of Alabama was hit by meteorite that crashed through her roof. In 1992, a piece of the Mbale metcorite hit a small boy on the head; fortunately. its descent was slowed by banana leaves (Jenniskens etal., 1994). Andiin 2009, a German boy claimed to have been injuries resulted from any of these incidents, suggesting that the need for preplanning st such contingencies is sometimes overstated.
Section 1.5
Planning and Acting in Nondeterministic Domains whole plan
Figure 11.12 At first, the sequence “whole plan” is expected to get the agent from S to G. The agent executes steps of the plan until it expects to be in state £, but observes that it is actually in O. The agent then replans for the minimal repair plus continuation to reach G. Now let’s return to the example problem of achieving a chair and table of matching color.
Suppose the agent comes up with this plan:
[LookAt(Table), LookAt(Chair), if Color(Tuble, ) A Color(Chair, ) then NoOp else [RemoveLid(Cany ), LookAt(Cany),
if Color(Tuble, c) A Color(Can) ) then Paint(Chair, Cany) else REPLAN]].
Now the agent is ready to execute the plan. The agent observes that the table and can of paint are white and the chair is black.
It then executes Paint(Chair,Can,).
At this point a
classical planner would declare victory; the plan has been executed. But an online execution monitoring agent needs to check that the action succeeded.
Suppose the agent perceives that the chair is a mottled gray because the black paint is
showing through. The agent then needs to figure out a recovery position in the plan to aim for and a repair action sequence to get there. The agent notices that the current state is identical to the precondition before the Paint(Chair,Can;) action, so the agent chooses the empty
sequence for repair and makes its plan be the same [Paint] sequence that it just attempted.
With this new plan in place, execution monitoring resumes, and the Paint action is retried.
This behavior will loop until the chair is perceived to be completely painted. But notice that
the loop is created by a process of plan-execute-replan, rather than by an explicit loop in a plan. Note also that the original plan need not cover every contingency. If the agent reaches
the step marked REPLAN, it can then generate a new plan (perhaps involving Cany).
Action monitoring is a simple method of execution monitoring, but it can sometimes lead
to less than intelligent behavior. For example, suppose there is no black or white paint, and
the agent constructs a plan to solve the painting problem by painting both the chair and table
red. Suppose that there is only enough red paint for the chair. With action monitoring, the agent would go ahead and paint the chair red, then notice that it is out of paint and cannot
paint the table, at which point it would replan a repair—perhaps painting both chair and table
green. A plan-monitoring agent can detect failure whenever the current state is such that the
remaining plan no longer works. Thus, it would not waste time painting the chair red.
373
374
Chapter 11
Automated Planning
Plan monitoring achieves this by checking the preconditions for success of the entire
remaining plan—that is, the preconditions of each step in the plan, except those preconditions
that are achieved by another step in the remaining plan. Plan monitoring cuts off execution of a doomed plan as soon as possible, rather than continuing until the failure actually occurs.3 Plan monitoring also allows for serendipity—accidental success.
If someone comes along
and paints the table red at the same time that the agent is painting the chair red, then the final plan preconditions are satisfied (the goal has been achieved), and the agent can go home early. It is straightforward to modify a planning algorithm so that each action in the plan is annotated with the action’s preconditions, thus enabling action monitoring.
It is slightly more
complex to enable plan monitoring. Partial-order planners have the advantage that they have already built up structures that contain the relations necessary for plan monitoring. Augment-
ing state-space planners with the necessary annotations can be done by careful bookkeeping as the goal fluents are regressed through the plan. Now that we have described a method for monitoring and replanning, we need to ask, “Does it work?” This is a surprisingly tricky question. If we mean, “Can we guarantee that the agent will always achieve the goal?” then the answer is no, because the agent could
inadvertently arrive at a dead end from which there is no repair. For example, the vacuum
agent might have a faulty model of itself and not know that its batteries can run out. Once
they do, it cannot repair any plans. If we rule out dead ends—assume that there exists a plan to reach the goal from any state in the environment—and
assume that the environment is
really nondeterministic, in the sense that such a plan always has some chance of success on
any given execution attempt, then the agent will eventually reach the goal.
Trouble occurs when a seemingly-nondeterministic action is not actually random, but
rather depends on some precondition that the agent does not know about. For example, sometimes a paint can may be empty, so painting from that can has no effect.
No amount
of retrying is going to change this.® One solution is to choose randomly from among the set of possible repair plans, rather than to try the same one each time. In this case, the repair plan of opening another can might work. A better approach is to learn a better model. Every
prediction failure is an opportunity for learning; an agent should be able to modify its model
of the world to accord with its percepts. From then on, the replanner will be able to come up with a repair that gets at the root problem, rather than relying on luck to choose a good repair.
11.6_Time, Schedules, and Resources Classical planning talks about what to do, in what order, but does not talk about time: how
Scheduling Resource constraint
long an action takes and when it occurs. For example, in the airport domain we could produce aplan saying what planes go where, carrying what, but could not specify departure and arrival
times. This is the subject matter of scheduling. The real world also imposes resource constraints: an airline has a limited number of staff, and staff who are on one flight cannot be on another at the same time. This section
introduces techniques for planning and scheduling problems with resource constraints.
5 Plan monitoring means that finally, after 374 pages, we have an agent that is smarter than a dung beetle (see page 41). A plan-monitoring agent would notice that the dung ball was missing from its grasp and would replan 10 get another ball and plug its hole. 6 Futile repetition of a plan repair is exactly the behavior exhibited by the sphex wasp (page 41).
Section 11.6
Time, Schedules, and Resources
375
Jobs({AddEnginel < AddWheels] < Inspectl}, {AddEngine2 < AddWheels2 < Inspeci2}) Resources(EngineHoists(1), WheelStations(1), Inspectors(e2), LugNuts(500)) Action(AddEnginel, DURATION:30,
USE:EngineHoists(1)) Action(AddEngine2, DURATION:60,
USE:EngineHoists(1)) Action(AddWheels], DURATION:30,
CONSUME: LugNuts(20), USE: WheelStations(1))
Action(AddWheels2, DURATION:15,
CONSUME: LugNuts(20), USE:WheelStations(1))
Action(Inspect;, DURATION: 10,
UsE:Inspectors(1))
Figure 11.13 A job-shop scheduling problem for assembling two cars, with resource constraints. The notation A < B means that action A must precede action B. The approach we take is “plan first, schedule later”: divide the overall problem into a planning phase in which actions are selected, with some ordering constraints, to meet the goals of the problem, and a later scheduling phase, in which temporal information is added to the plan to ensure that it meets resource and deadline constraints. This approach is common in real-world manufacturing and logistical settings, where the planning phase is sometimes
automated, and sometimes performed by human experts. 11.6.1
Representing temporal and resource constraints
A typical job-shop scheduling problem (see Section 6.1.2), consists of a set of jobs, each
of which has a collection of actions with ordering constraints among them. Each action has
a duration and a set of resource constraints required by the action.
A constraint specifies
a fype of resource (e.g., bolts, wrenches, or pilots), the number of that resource required,
and whether that resource is consumable (e.g., the bolts are no longer available for use) or
reusable (e.g., a pilot is occupied during a flight but is available again when the flight is over). Actions can also produce resources (¢.g., manufacturing and resupply actions). A solution to a job-shop scheduling problem specifies the start times for each action and must satisfy all the temporal ordering constraints and resource constraints.
Job-shop scheduling problem Job Duration Consumable Reusable
As with search
and planning problems, solutions can be evaluated according to a cost function; this can be quite complicated, with nonlinear resource costs, time-dependent delay costs, and so on. For simplicity, we assume that the cost function is just the total duration of the plan, which is called the makespan.
Figure 11.13 shows a simple example: a problem involving the assembly of two cars. The problem consists of two jobs, each of the form [AddEngine,AddWheels, Inspect]. Then
the Resources statement declares that there are four types of resources, and gives the number of each type available at the start: 1 engine hoist, 1 wheel station, 2 inspectors, and 500 lug nuts. The action schemas give the duration and resource needs of each action. The lug nuts
Makespan
376
Chapter 11
Automated Planning
are consumed as wheels are added to the car, whereas the other resources are “borrowed” at
the start of an action and released at the action’s end.
Aggregation
The representation of resources as numerical quantities, such as Inspectors(2), rather than as named entities, such as Inspector (1) and Inspector (L), is an example of a technique called aggregation:
grouping individual objects into quantities when the objects are all in-
distinguishable. In our assembly problem, it does not matter which inspector inspects the car, so there is no need to make the distinction. Aggregation is essential for reducing complexity.
Consider what happens when a proposed schedule has 10 concurrent Inspect actions but only
9 inspectors are available. With inspectors represented as quantities, a failure is detected im-
mediately and the algorithm backtracks to try another schedule. With inspectors represented as individuals, the algorithm would try all 9! ways of assigning inspectors to actions before noticing that none of them work. 11.6.2
Solving scheduling problems
We begin by considering just the temporal scheduling problem, ignoring resource constraints.
To minimize makespan (plan duration), we must find the earliest start times for all the actions
consistent with the ordering constraints supplied with the problem. It is helpful to view these Critical path method
ordering constraints as a directed graph relating the actions, as shown in Figure 11.14. We can
apply the critical path method (CPM) to this graph to determine the possible start and end
times of each action. A path through a graph representing a partial-order plan is a linearly
Critical path
ordered sequence of actions beginning with Start and ending with Finish. (For example, there are two paths in the partial-order plan in Figure 11.14.) The critical path is that path whose total duration is longest; the path is “critical” because it determines the duration of the entire plan—shortening other paths doesn’t shorten the plan
as a whole, but delaying the start of any action on the critical path slows down the whole plan. Actions that are off the critical path have a window of time in which they can be executed.
Slack Schedule
The window is specified in terms of an earliest possible start time, ES, and a latest possible
start time, LS.
The quantity LS — ES is known as the slack of an action.
We can see in
Figure 11.14 that the whole plan will take 85 minutes, that each action in the top job has 15 minutes of slack, and that each action on the critical path has no slack (by definition). Together the ES and LS times for all the actions constitute a schedule for the problem.
The following formulas define ES and LS and constitute a dynamic-programming algo-
rithm to compute them. A and B are actions, and A < B means that A precedes B:
ES(Start) =0
ES(B) = maxs < ES(A) + Duration(A)
LS(Finish) = ES(Finish) LS(A) = ming,. o LS(B) — Duration(A) .
The idea is that we start by assigning ES(Start) to be 0. Then, as soon as we get an action
B such that all the actions that come immediately before B have ES values assigned, we
set ES(B) to be the maximum of the earliest finish times of those immediately preceding
actions, where the earliest finish time of an action is defined as the earliest start time plus
the duration. This process repeats until every action has been assigned an ES value. The LS values are computed in a similar manner, working backward from the Finish action.
The complexity of the critical path algorithm is just O(Nb), where N is the number of
actions and b is the maximum
branching factor into or out of an action.
(To see this, note
Section 11.6
Start
o
0
o g 30
[EX5] 07 nsawiess =] tmpect 30 10
o7 g2 &
[@er ] niawteci 15
2
o
“
Time, Schedules, and Resources
[85.85]
Finish
[FEE ! o o
0
@
o
"
»
Figure 11.14 Top: a representation of the temporal constraints for the job-shop scheduling problem of Figure 11.13. The duration of each action is given at the bottom of each rectangle. In solving the problem, we compute the earliest and latest start times as the pair [ES, LS],
displayed in the upper left. The difference between these two numbers is the slack of an
action; actions with zero slack are on the critical path, shown with bold arrows. Bottom: the
same solution shown as a timeline. Grey rectangles represent time intervals during which an action may be executed, provided that the ordering constraints are respected. The unoccupied portion ofa gray rectangle indicates the slack.
that the LS and ES computations are done once for each action, and each computation iterates
over at most b other actions.) Therefore, finding a minimum-duration schedule, given a partial
ordering on the actions and no resource constraints, is quite easy.
Mathematically speaking, critical-path problems are easy to solve because they are de-
fined as a conjunction of linear inequalities on the start and end times. When we introduce
resource constraints, the resulting constraints on start and end times become more complicated. For example, the AddEngine actions, which begin at the same time in Figure 11.14, require the same EngineHoist and so cannot overlap. The “cannot overlap” constraint is a disjunction of two linear inequalities, one for each possible ordering. The introduction of disjunctions turns out to make scheduling with resource constraints NP-hard.
Figure 11.15 shows the solution with the fastest completion time, 115 minutes. This is
30 minutes longer than the 85 minutes required for a schedule without resource constraints.
Notice that there is no time at which both inspectors are required, so we can immediately
‘move one of our two inspectors to a more productive position. There is a long history of work on optimal scheduling. A challenge problem posed in 1963—to find the optimal schedule for a problem involving just 10 machines and 10 jobs of 100 actions each—went unsolved for 23 years (Lawler et al., 1993). Many approaches have been tried, including branch-and-bound, simulated annealing, tabu search, and constraint sat-
377
378
Chapter 11
Automated Planning
——&
Engincoist)
o
T
oo
T
—
T
T w0
T 0
T @
W
T s»
T %
T w
T
Figure 11.15 A solution to the job-shop scheduling problem from Figure 11.13, taking into account resource constraints. The left-hand margin lists the three reusable resources, and actions are shown aligned horizontally with the resources they use. There are two pos ble schedules, depending on which assembly uses the engine hoist first; we’ve shown the shortest-duration solution, which takes 115 minutes. Minimum slack
isfaction. One popular approach is the minimum slack heuristic:
on each iteration, schedule
for the earliest possible start whichever unscheduled action has all its predecessors scheduled and has the least slack; then update the ES and LS times for each affected action and
repeat. This greedy heuristic resembles the minimum-remaining-values (MRV) heuristic in constraint satisfaction. It often works well in practice, but for our assembly problem it yields a 130-minute solution, not the 115-inute solution of Figure 11.15.
Up to this point, we have assumed that the set of actions and ordering constraints is fixed. Under these assumptions, every scheduling problem can be solved by a nonoverlapping sequence that avoids all resource conflicts, provided that each action is feasible by itself.
However if a scheduling problem is proving very difficult, it may not be a good idea to solve
it this way—it may be better to reconsider the actions and constraints, in case that leads to a
much easier scheduling problem. Thus, it makes sense to infegrate planning and scheduling by taking into account durations and overlaps during the construction of a plan. Several of
the planning algorithms in Section 11.2 can be augmented to handle this information.
11.7
Analysis of Planning Approaches
Planning combines the two major areas of Al we have covered so far: search and logic. A planner can be seen either as a program that searches for a solution or as one that (constructively) proves the existence of a solution. The cross-fertilization of ideas from the two areas
has allowed planners to scale up from toy problems where the number of actions and states was limited to around a dozen, to real-world industrial applications with millions of states and thousands of actions.
Planning is foremost an exercise in controlling combinatorial explosion. If there are n propositions in a domain, then there are 2" states. Against such pessimism, the identification of independent subproblems can be a powerful weapon. In the best case—full decomposability of the problem—we get an exponential speedup. Decomposability is destroyed, however, by negative interactions between actions.
SATPLAN can encode logical relations between
subproblems. Forward search addresses the problem heuristically by trying to find patterns (subsets of propositions) that cover the independent subproblems.
Since this approach is
heuristic, it can work even when the subproblems are not completely independent.
Summary Unfortunately, we do not yet have a clear understanding of which techniques work best on which kinds of problems. Quite possibly, new techniques will emerge, perhaps providing a synthesis of highly expressive first-order and hierarchical representations with the highly efficient factored and propositional representations that dominate today. We are seeing exam-
ples of portfolio planning systems, where a collection of algorithms are available to apply to
any given problem. This can be done selectively (the system classifies each new problem to choose the best algorithm for it), or in parallel (all the algorithms run concurrently, each on a different CPU), or by interleaving the algorithms according to a schedule.
Summary
In this chapter, we described the PDDL representation for both classical and extended planning problems, and presented several algorithmic approaches for finding solutions. The points to remember: + Planning systems are problem-solving algorithms that operate on explicit factored rep-
resentations of states and actions. These representations make possible the derivation of
efffective domain-independent heuristics and the development of powerful and flexible algorithms for solving problems. « PDDL, the Planning Domain Definition Language, describes the initial and goal states as conjunctions of literals, and actions in terms of their preconditions and effects. Ex-
tensions represent time, resources, percepts, contingent plans, and hierarchical plans
+ State-space search can operate in the forward direction (progression) or the backward
direction (regression). Effective heuristics can be derived by subgoal independence assumptions and by various relaxations of the planning problem.
« Other approaches include encoding a planning problem as a Boolean satisfiability problem or as a constraint satisfaction problem; and explicitly searching through the space of partially ordered plans. + Hierarchical task network (HTN) planning allows the agent to take advice from the
domain designer in the form of high-level actions (HLAs) that can be implemented in
various ways by lower-level action sequences. The effects of HLAs can be defined with angelic semantics, allowing provably correct high-level plans to be derived without consideration of lower-level implementations. HTN methods can create the very large plans required by many real-world applications. + Contingent plans allow the agent to sense the world during execution to decide what
branch of the plan to follow. In some cases, sensorless or conformant planning can be used to construct a plan that works without the need for perception. Both conformant and contingent plans can be constructed by search in the space of belief states. Efficient representation or computation of belief states is a key problem.
« An online planning agent uses execution monitoring and splices in repairs as needed to recover from unexpected situations, which can be due to nondeterministic
exogenous events, or incorrect models of the environment.
actions,
* Many actions consume resources, such as money, gas, or raw materials. It is convenient
to treat these resources as numeric measures in a pool rather than try to reason about,
379
Portfolio
380
Chapter 11
Automated Planning
say, each individual coin and bill in the world.
Time is one of the most important
resources. It can be handled by specialized scheduling algorithms, or scheduling can be integrated with planning. « This chapter extends classical planning to cover nondeterministic environments (where
outcomes of actions are uncertain), but it is not the last word on planning. Chapter 17 describes techniques for stochastic environments (in which outcomes of actions have
probabilities associated with them): Markov decision processes, partially observable Markov decision processes, and game theory. In Chapter 22 we show that reinforcement learning allows an agent to learn how to behave from past successes and failures. Bibliographical and Historical Notes
AT planning arose from investigations into state-space search, theorem proving, and control theory.
STRIPS (Fikes and Nilsson, 1971, 1993), the first major planning system, was de-
signed as the planner for the Shakey robot at SRL The first version of the program ran on a
computer with only 192 KB of memory.
Its overall control structure was modeled on GPS,
the General Problem Solver (Newell and Simon, 1961), a state-space search system that used
means—ends analysis.
The STRIPS representation language evolved into the Action Description Language, or ADL (Pednault, 1986), and then the Problem Domain Description Language, or PDDL (Ghallab et al., 1998), which has been used for the International Planning Competition since 1998. The most recent version is PDDL 3.1 (Kovacs, 2011).
Linear planning
Planners in the early 1970s decomposed problems by computing a subplan for each subgoal and then stringing the subplans together in some order. This approach, called linear planning by Sacerdoti (1975), was soon discovered to be incomplete. It cannot solve some very simple problems, such as the Sussman anomaly (see Exercise 11.5Uss), found by Allen Brown during experimentation with the HACKER system (Sussman, 1975). A complete plan-
ner must allow for interleaving of actions from different subplans within a single sequence.
‘Warren’s (1974) WARPLAN system achieved that, and demonstrated how the logic programming language Prolog can produce concise programs; WARPLAN is only 100 lines of code.
Partial-order planning dominated the next 20 years of research, with theoretical work
describing the detection of conflicts (Tate, 1975a) and the protection of achieved conditions (Sussman, 1975), and implementations including NOAH (Sacerdoti, 1977) and NONLIN (Tate, 1977). That led to formal models (Chapman, 1987; McAllester and Rosenblitt, 1991)
that allowed for theoretical analysis of various algorithms and planning problems, and to a widely distributed system, UCPOP (Penberthy and Weld, 1992).
Drew McDermott suspected that the emphasis on partial-order planning was crowding out
other techniques that should perhaps be reconsidered now that computers had 100 times the
memory of Shakey’s day. His UNPOP (McDermott, 1996) was a state-space planning program employing the ignore-delete-list heuristic. HSP, the Heuristic Search Planner (Bonet and Geffner, 1999; Haslum, 2006) made state-space search practical for large planning problems. The FF or Fast Forward planner (Hoffmann, 2001; Hoffmann and Nebel, 2001; Hoffmann, 2005) and the FASTDOWNWARD variant (Helmert, 2006) won international planning
competitions in the 2000s.
Bibliographical and Historical Notes Bidirectional search (see Section 3.4.5) has also been known
381
to suffer from a lack of
heuristics, but some success has been obtained by using backward search to create a perimeter around the goal, and then refining a heuristic to search forward towards that perimeter (Torralba et al., 2016). The SYMBA*
bidirectional search planner (Torralba ez al., 2016)
won the 2016 competition.
Researchers turned to PDDL and the planning paradigm so that they could use domain independent heuristics. Hoffmann (2005) analyzes the search space of the ignore-deletelist heuristic.
Edelkamp (2009) and Haslum er al. (2007) describe how to construct pattern
databases for planning heuristics. Felner ef al. (2004) show encouraging results using pattern databases for sliding-tile puzzles, which can be thought of as a planning domain, but Hoffmann ef al. (2006) show some limitations of abstraction for classical planning problems. (Rintanen, 2012) discusses planning-specific variable-selection heuristics for SAT solving. Helmert et al. (2011) describe the Fast Downward Stone Soup (FDSS) system, a portfolio
planner that, as in the fable of stone soup, invites us to throw in as many planning algorithms. as possible. The system maintains a set of training problems, and for each problem and each algorithm records the run time and resulting plan cost of the problem’s solution. Then when
faced with a new problem, it uses the past experience to decide which algorithm(s) to try, with what time limits, and takes the solution with minimal cost. FDSS
was a winner in the 2018
International Planning Competition (Seipp and Rger, 2018). Seipp ef al. (2015) describe a machine learning approach to automatically learn a good portfolio, given a new problem. Vallati ef al. (2015) give an overview of portfolio planning. The idea of algorithm portfolios for combinatorial search problems goes back to Gomes and Selman (2001). Sistla and Godefroid (2004) cover symmetry reduction, and Godefroid (1990) covers
heuristics for partial ordering. Richter and Helmert (2009) demonstrate the efficiency gains
of forward pruning using preferred actions.
Blum and Furst (1997) revitalized the field of planning with their Graphplan system, which was orders of magnitude faster than the partial-order planners of the time. Bryce and Kambhampati (2007) give an overview of planning graphs. The use of situation calculus for planning was introduced by John McCarthy (1963) and refined by Ray Reiter (2001). Kautz et al. (1996) investigated various ways to propositionalize action schemas, finding
that the most compact forms did not necessarily lead to the fastest solution times. A systematic analysis was carried out by Emst er al. (1997), who also developed an automatic “compiler” for generating propositional representations from PDDL problems. The BLACKBOX
planner, which combines ideas from Graphplan and SATPLAN, was developed by Kautz and Selman (1998). Planners based on constraint satisfaction include CPLAN van Beek and Chen
(1999) and GP-CSP (Do and Kambhampati, 2003). There has also been interest in the representation of a plan as a binary decision diagram
(BDD), a compact data structure for Boolean expressions widely studied in the hardware
verification community (Clarke and Grumberg, 1987; McMillan, 1993). There are techniques
for proving properties of binary decision diagrams, including the property of being a solution to a planning problem. Cimatti ef al. (1998) present a planner based on this approach. Other representations have also been used, such as integer programming (Vossen et al., 2001).
There are some interesting comparisons of the various approaches to planning. Helmert (2001) analyzes several classes of planning problems, and shows that constraint-based approaches such as Graphplan and SATPLAN are best for NP-hard domains, while search-based
Binary decision diagram (BDD)
382
Chapter 11
Automated Planning
approaches do better in domains where feasible solutions can be found without backtracking.
Graphplan and SATPLAN have trouble in domains with many objects because that means
they must create many actions. In some cases the problem can be delayed or avoided by generating the propositionalized actions dynamically, only as needed, rather than instantiating them all before the search begins.
Macrops Abstraction hierarchy
The first mechanism for hierarchical planning was a facility in the STRIPS program for
learning macrops—‘macro-operators”
consisting of a sequence of primitive steps (Fikes
et al., 1972). The ABSTRIPS system (Sacerdoti, 1974) introduced the idea of an abstraction
hierarchy, whereby planning at higher levels was permitted to ignore lower-level precon-
ditions of actions in order to derive the general structure of a working plan. Austin Tate’s Ph.D. thesis (1975b) and work by Earl Sacerdoti (1977) developed the basic ideas of HTN
planning. Erol, Hendler, and Nau (1994, 1996) present a complete hierarchical decomposi-
tion planner as well as a range of complexity results for pure HTN planners. Our presentation of HLAs and angelic semantics is due to Marthi ez al. (2007, 2008).
One of the goals of hierarchical planning has been the reuse of previous planning experience in the form of generalized plans. The technique of explanation-based learning has been used as a means of generalizing previously computed plans in systems such as SOAR (Laird et al., 1986) and PRODIGY
(Carbonell ef al., 1989). An alternative approach is
to store previously computed plans in their original form and then reuse them to solve new,
Case-based planning
similar problems by analogy to the original problem. This is the approach taken by the field called case-based planning (Carbonell, 1983; Alterman, 1988). Kambhampati (1994) argues that case-based planning should be analyzed as a form of refinement planning and provides a formal foundation for case-based partial-order planning. Early planners lacked conditionals and loops, but some could use coercion to form conformant plans. Sacerdoti’s NOAH solved the “keys and boxes” problem (in which the planner knows little about the initial state) using coercion.
Mason (1993) argued that sensing often
can and should be dispensed with in robotic planning, and described a sensorless plan that can move a tool into a specific position on a table by a sequence of tilting actions, regardless
of the initial position.
Goldman and Boddy (1996) introduced the term conformant planning, noting that sen-
sorless plans are often effective even if the agent has sensors. The first moderately efficient conformant planner was Smith and Weld's (1998) Conformant Graphplan (CGP). Ferraris and Giunchiglia (2000) and Rintanen (1999) independently developed SATPLAN-based conformant planners. Bonet and Geffner (2000) describe a conformant planner based on heuristic
search in the space of belief states, drawing on ideas first developed in the 1960s for partially
observable Markov decision processes, or POMDPs (see Chapter 17).
Currently, there are three main approaches to conformant planning. The first two use
heuristic search in belief-state space:
HSCP (Bertoli ef al., 2001a) uses binary decision di-
agrams (BDDs) to represent belief states, whereas Hoffmann and Brafman (2006) adopt the lazy approach of computing precondition and goal tests on demand using a SAT solver. The third approach, championed primarily by Jussi Rintanen (2007), formulates the entire sensorless planning problem as a quantified Boolean formula (QBF) and solves it using a
general-purpose QBF solver. Current conformant planners are five orders of magnitude faster than CGP. The winner of the 2006 conformant-planning track at the International Planning
Competition was Tj (Palacios and Geffner, 2007), which uses heuristic search in belief-state
Bibliographical and Historical Notes
383
space while keeping the belief-state representation simple by defining derived literals that cover conditional effects. Bryce and Kambhampati (2007) discuss how a planning graph can be generalized to generate good heuristics for conformant and contingent planning. The contingent-planning approach described in the chapter is based on Hoffmann and Brafman (2005), and was influenced by the efficient search algorithms for cyclic AND—OR graphs developed by Jimenez and Torras (2000) and Hansen and Zilberstein (2001).
The
problem of contingent planning received more attention after the publication of Drew Mc-
Dermott’s (1978a) influential article, Planning and Acting. Bertoli et al. (2001b) describe MBP
(Model-Based
Planner), which uses binary decision diagrams to do conformant and
contingent planning. Some authors use “conditional planning” and “contingent planning” as synonyms; others make the distinction that “conditional” refers to actions with nondetermin-
istic effects, and “contingent” means using sensing to overcome partial observability. In retrospect, it is now possible to see how the major classical planning algorithms led to
extended versions for uncertain domains. Fast-forward heuristic search through state space led to forward search in belief space (Bonet and Geffner, 2000; Hoffmann and Brafman, 2005); SATPLAN led to stochastic SATPLAN (Majercik and Littman, 2003) and to planning
with quantified Boolean logic (Rintanen, 2007); partial order planning led to UWL (Etzioni et al., 1992) and CNLP (Peot and Smith, 1992); Graphplan led to Sensory Graphplan or SGP
(Weld et al., 1998).
The first online planner with execution
monitoring
was
PLANEX
(Fikes ef al.,
1972),
which worked with the STRIPS planner to control the robot Shakey. SIPE (System for Interactive Planning and Execution monitoring) (Wilkins, 1988) was the first planner to deal
systematically with the problem of replanning. It has been used in demonstration projects in
several domains, including planning operations on the flight deck of an aircraft carrier, jobshop scheduling for an Australian beer factory, and planning the construction of multistory buildings (Kartam and Levitt, 1990).
In the mid-1980s, pessimism about the slow run times of planning systems led to the pro-
posal of reflex agents called reactive planning systems (Brooks, 1986; Agre and Chapman,
1987). “Universal plans” (Schoppers, 1989) were developed as a lookup-table method for
reactive planning, but turned out to be a rediscovery of the idea of policies that had long been
used in Markov decision processes (see Chapter 17). Koenig (2001) surveys online planning techniques, under the name Agent-Centered Search.
Planning with time constraints was first dealt with by DEVISER (Vere, 1983). The representation of time in plans was addressed by Allen (1984) and by Dean et al. (1990) in the FORBIN
system.
NONLIN+
(Tate and Whiter,
1984) and SiPE (Wilkins,
1990) could rea-
son about the allocation of limited resources to various plan steps. O-PLAN (Bell and Tate,
1985) has been applied to resource problems such as software procurement planning at Price Waterhouse and back-axle assembly planning at Jaguar Cars. The two planners SAPA
(Do and Kambhampati,
2001) and T4 (Haslum and Geffner,
2001) both used forward state-space search with sophisticated heuristics to handle actions
with durations and resources. An alternative is to use very expressive action languages, but guide them by human-written, domain-specific heuristics, as is done by ASPEN (Fukunaga et al., 1997), HSTS (Jonsson et al., 2000), and IxTeT (Ghallab and Laruelle, 1994). A number of hybrid planning-and-scheduling systems have been deployed: Isis (Fox et al., 1982; Fox, 1990) has been used for job-shop scheduling at Westinghouse, GARI (De-
Reactive planning
384
Chapter 11
Automated Planning
scotte and Latombe,
1985) planned
the machining and construction
of mechanical
parts,
FORBIN was used for factory control, and NONLIN+ was used for naval logistics planning.
We chose to present planning and scheduling as two separate problems; Cushing et al. (2007) show that this can lead to incompleteness on certain problems. There is a long history of scheduling in aerospace. T-SCHED (Drabble, 1990) was used
to schedule mission-command sequences for the UOSAT-II satellite. OPTIMUM-AIV (Aarup et al., 1994) and PLAN-ERS1 (Fuchs ez al., 1990), both based on O-PLAN, were used for
spacecraft assembly and observation planning, respectively, at the European Space Agency.
SPIKE (Johnston and Adorf, 1992) was used for observation planning at NASA for the Hub-
ble Space Telescope, while the Space Shuttle Ground Processing Scheduling System (Deale et al., 1994) does job-shop scheduling of up to 16,000 worker-shifts. Remote Agent (Muscettola et al., 1998) became the first autonomous planner—scheduler to control a spacecraft, when
it flew onboard the Deep Space One probe in 1999. Space applications have driven the development of algorithms for resource allocation; see Laborie (2003) and Muscettola (2002). The literature on scheduling is presented in a classic survey article (Lawler ef al., 1993), a book (Pinedo, 2008), and an edited handbook (Blazewicz et al., 2007).
The computational complexity of of planning has been analyzed by several authors (By-
lander,
1994; Ghallab et al., 2004; Rintanen, 2016).
There are two main tasks:
PlanSAT
is the question of whether there exists any plan that solves a planning problem.
Bounded
PlanSAT asks whether there is a solution of length k or less; this can be used to find an optimal plan. Both are decidable for classical planning (because the number of states is finite). But if we add function symbols to the language, then the number of states becomes infinite,
and PlanSAT becomes only semidecidable. For propositionalized problems both are in the
complexity class PSPACE, a class that is larger (and hence more difficult) than NP and refers
to problems that can be solved by a deterministic Turing machine with a polynomial amount of space. These theoretical results are discouraging, but in practice, the problems we want to solve tend to be not so bad.
The true advantage of the classical planning formalism is
that it has facilitated the development of very accurate domain-independent heuristics; other
approaches have not been as fruitful. Readings in Planning (Allen et al., 1990) is a comprehensive anthology of early work in the field. Weld (1994, 1999) provides two excellent surveys of planning algorithms of the
1990s. It is interesting to see the change in the five years between the two surveys: the first
concentrates on partial-order planning, and the second introduces Graphplan and SATPLAN.
Automated Planning and Acting (Ghallab ez al., 2016) is an excellent textbook on all aspects of the field. LaValle’s text Planning Algorithms (2006) covers both classical and stochastic
planning, with extensive coverage of robot motion planning.
Planning research has been central to Al since its inception, and papers on planning are a staple of mainstream Al journals and conferences. There are also specialized conferences such as the International Conference on Automated Planning and Scheduling and the Inter-
national Workshop on Planning and Scheduling for Space.
TS
12
QUANTIFYING UNCERTAINTY In which we see how to tame uncertainty with numeric degrees of belief.
12.1
Acting under Uncertainty
Agents in the real world need to handle uncertainty, whether due to partial observability,
nondeterminism, or adversaries. An agent may never know for sure what state it is in now or
Uncertainty
where it will end up after a sequence of actions.
We have seen problem-solving and logical agents handle uncertainty by keeping track of
a belief state—a representation of the set of all possible world states that it might be in—and
generating a contingency plan that handles every possible eventuality that its sensors may report during execution. This approach works on simple problems, but it has drawbacks: « The agent must consider every possible explanation for its sensor observations, no matter how unlikely. This leads to a large belief-state full of unlikely possibilities. « A correct contingent plan that handles every eventuality can grow arbitrarily large and must consider arbitrarily unlikely contingencies.
« Sometimes there is no plan that is guaranteed to achieve the goal—yet the agent must act. It must have some way to compare the merits of plans that are not guaranteed.
Suppose, for example, that an automated taxi has the goal of delivering a passenger to the
airport on time. The taxi forms a plan, A, that involves leaving home 90 minutes before the
flight departs and driving at a reasonable speed. Even though the airport is only 5 miles away,
alogical agent will not be able to conclude with absolute certainty that “Plan Agy will get us to the airport in time.” Instead, it reaches the weaker conclusion “Plan Agg will get us to the
airport in time, as long as the car doesn’t break down, and I don’t get into an accident, and
the road isn’t closed, and no meteorite hits the car, and ... .” None of these conditions can be
deduced for sure, so we can’t infer that the plan succeeds. This is the logical qualification
problem (page 241), for which we so far have seen no real solution.
Nonetheless, in some sense Ag is in fact the right thing to do. What do we mean by this?
As we discussed in Chapter 2, we mean that out of all the plans that could be executed, Agy
is expected to maximize the agent’s performance measure (where the expectation is relative
to the agent’s knowledge about the environment). The performance measure includes getting to the airport in time for the flight, avoiding a long, unproductive wait at the airport, and
avoiding speeding tickets along the way. The agent’s knowledge cannot guarantee any of these outcomes for Agg, but it can provide some degree of belief that they will be achieved. Other plans, such as Ago, might increase the agent’s belief that it will get to the airport on time, but also increase the likelihood of a long, boring wait. The right thing to do—the
rational decision—therefore depends on both the relative importance of various goals and
=28,224 entries—still a manageable number.
If we add the possibility of dirt in each of the 42 squares, the number of states is multiplied
by 2*2 and the transition matrix has more than 102 entries—no longer a manageable number.
In general, if the state is composed of n discrete variables with at most d values each, the
corresponding HMM transition matrix will have size O(d2") and the per-update computation time will also be O(d?").
For these reasons, although HMMs have many uses in areas ranging from speech recogni-
tion to molecular biology, they are fundamentally limited in their ability to represent complex
processes.
In the terminology introduced in Chapter 2, HMMs are an atomic representation:
states of the world have no internal structure and are simply labeled by integers. Section 14.5
shows how to use dynamic Bayesian networks—a factored representation—to model domains
with many state variables. The next section shows how to handle domains with continuous state variables, which of course lead to an infinite state space.
Section 14.4
14.4
Kalman
Kalman Filters
Filters
Imagine watching a small bird flying through dense jungle foliage at dusk: you glimpse
brief, intermittent flashes of motion; you try hard to guess where the bird is and where it will
appear next so that you don’t lose it. Or imagine that you are a World War II radar operator
peering at a faint, wandering blip that appears once every 10 seconds on the screen. Or, going
back further still, imagine you are Kepler trying to reconstruct the motions of the planets
from a collection of highly inaccurate angular observations taken at irregular and imprecisely measured intervals.
In all these cases, you are doing filtering: estimating state variables (here, the position and velocity of a moving object) from noisy observations over time. If the variables were discrete, we could model the system with a hidden Markov model.
This section examines
methods for handling continuous variables, using an algorithm called Kalman filtering, after Kalman filtering
one of its inventors, Rudolf Kalman.
The bird’s flight might be specified by six continuous variables at each time point; three for position (X;,Y;,Z) and three for velocity (X;.,¥;,Z). We will need suitable conditional
densities to represent the transition and sensor models; as in Chapter 13, we will use linear— Gaussian distributions. This means that the next state X,
must be a linear function of the
current state X, plus some Gaussian noise, a condition that turns out to be quite reasonable in
practice. Consider, for example, the X-coordinate of the bird, ignoring the other coordinates
for now. Let the time interval between observations be A, and assume constant velocity during
the interval; then the position update is given by X;;a = X, + X A. Adding Gaussian noise (to account for wind variation, etc.), we obtain a linear-Gaussian transition model:
P(Xrra=3eal X =2, X =5) = N (0a3% + 5% A,0%).
The Bayesian network structure for a system with position vector X, and velocity X, is shown
in Figure 14.9. Note that this is a very specific form of linear-Gaussian model: the general
form will be described later in this
section and covers a vast array of applications beyond the
simple motion examples of the first paragraph. The reader might wish to consult Appendix A for some of the mathematical properties of Gaussian distributions; for our immediate purposes, the most important is that a multivariate Gaussian
distribution for ¢ variables is
specified by a d-element mean 12 and a d x d covariance matrix . 14.4.1
Updating
Gaussian distributions
In Chapter 13 on page 423, we alluded to a key property of the lincar-Gaussian family of distributions: it remains closed under Bayesian updating.
(That is, given any evidence, the
posterior is still in the linear-Gaussian family.) Here we make this claim precise in the context
of filtering in a temporal probability model. The required properties correspond to the twostep filtering calculation in Equation (14.5):
1. If the current distribution P(X, |ey) is Gaussian and the transition model P(X,+|x,) is linear-Gaussian, then the one-step predicted distribution given by
P(Xpiifer) = |Jx, P(Xoior[%)P(x |er)dx is also a Gaussian distribution.
(14.17)
479
Chapter 14 Probabilistic Reasoning over Time
P
480
Figure 14.9 Bayesian network structure for a linear dynamical system with position X,. velocity X,. and position measurement Z 2. If the prediction P(X,; |e|,) is Gaussian and the sensor model P(e,1| X, 1) is linear— Gaussian, then, after conditioning on the new evidence, the updated distribution
P(Xpiieri1) = aP(erst [ X
is also a Gaussian distribution.
) P(Xrii | err)
(14.18)
Thus, the FORWARD operator for Kalman filtering takes a Gaussian forward message f.;,
specified by a mean z, and covariance ¥, and produces a new multivariate Gaussian forward
message f1... 1, specified by a mean i,
and covariance ¥, . So if we start with a Gaussian
prior f1.0=P(Xo) = (119, o), filtering with a linear-Gaussian model produces a Gaussian state distribution for all time.
This seems to be a nice, elegant result, but why is it so important? The reason is that except for a few special cases such as this, filtering with continuous or hybrid (discrete and continuous) networks generates state distributions whose representation grows without bound
over time. This statement is not easy to prove in general, but Exercise 14.KFSW shows what
happens for a simple example. 14.4.2
A simple one-dimensional example
‘We have said that the FORWARD operator for the Kalman filter maps a Gaussian into a new
Gaussian. This translates into computing a new mean and covariance from the previous mean
and covariance. Deriving the update rule in the general (multivariate) case requires rather a
lot of linear algebra, so we will stick to a very simple univariate case for now, and later give
the results for the general case. Even for the univariate case, the calculations are somewhat
tedious, but we feel that they are worth seeing because the usefulness of the Kalman filter is
tied so intimately to the mathematical properties of Gaussian distributions.
The temporal model we consider describes a random walk of a single continuous state
variable X, with a noisy observation Z. An example might be the “consumer confidence” in-
dex, which can be modeled as undergoing a random Gaussian-distributed change each month
and is measured by a random consumer survey that also introduces Gaussian sampling noise.
The prior distribution is assumed to be Gaussian with variance a‘%:
P(xo) =ae °
Section 14.4
Kalman Filters
(For simplicity, we use the same symbol o for all normalizing constants in this section.) The transition model adds a Gaussian perturbation of constant variance o2 to the current state: e
Pl
= e ().
The sensor model assumes Gaussian noise with variance o2:2
Pla
= ae
(),
Now, given the prior P(Xo), the one-step predicted distribution comes from Equation (14.17):
p) = [ Plalsopain=a [~ Al ‘This integral looks rather complicated. The key to progress is to notice that the exponent is the sum of two expressions that are quadratic in xo and hence
is itself a quadratic in xo. A simple
trick known as completing the square allows the rewriting of any quadratic ax} +bxo+cas Sempletie the the sum of a squared term a(xo — 52)? and a residual term ¢ — % that is independent of xo. In this case, we have a= (03 +02)/(0302), b=—2(03x) + 2 p0)/(03072). and c= (o} + o2u3)/(0302). The residual term can be taken outside the integral, giving us
P() = ae HEE) /"“ e Hatw-2) gy
Now the integral is just the integral of a Gaussian over its full range, which is simply 1. Thus,
we are left with only the residual term from the quadratic. Plugging back in the expressions for a, b, and ¢ and simplifying, we obtain -4
)
That is, the one-step predicted distribution is a Gaussian with the same mean 19 and a variance
equal to the sum of the original variance o7 and the transition variance 2.
To complete the update step, we need to condition on the observation at the first time
step, namely, z;. From Equation (14.18), this is given by
P(xi|z1) = aP(a|x)P(x) Once again, we combine the exponents and complete the square (Exercise 14.KALM), obtain-
ing the following expression for the posterior:
(14.19)
P(x|z1) =ae *
Thus, after one update cycle, we have a new Gaussian distribution for the state variable. From the Gaussian formula in Equation (14.19), we see that the new mean and standard deviation can be calculated from the old mean and standard deviation as follows: M1
=
+ o)z (2
+otm
Ftoit ol
and
(07 +a})e? 2 _ _
Fraitol
1420 14.20,
481
Chapter 14 Probabilistic Reasoning over Time 0.45 0.4 0.35 0.3
P(x)
482
0.25 0.2
0.15 0.1 0.05
xposition Figure 14.10 Stages in the Kalman filter update cycle for a random walk with a prior given by f1o=0.0 and o= 1.5, transition noise given by o, =2.0, sensor noise given by o= 1.0, and a first observation
.5 (marked on the x-axis).
flattened out, relative to P(xg), by the transition noise.
Notice how the prediction P(xi) is
Notice also that the mean of the
posterior distribution P(x; |2y is slightly to the left of the observation z; because the mean is a weighted average of the prediction and the observation.
Figure 14.10 shows one update cycle of the Kalman filter in the one-dimensional case for particular values of the transition and sensor models. Equation (14.20) plays exactly the same role as the general filtering equation (14.5) or the HMM filtering equation (14.12). Because of the special nature of Gaussian distributions, however, the equations have some interesting additional properties. First, we can interpret the calculation for the new mean y, 1| as a weighted mean of the
new observation z; and the old mean ;. If the observation is unreliable, then o2 is large
and we pay more attention to the old mean; if the old mean is unreliable (o7 is large) or the
process is highly unpredictable (o7 is large), then we pay more attention to the observation.
Second, notice that the update for the variance o7, | is independent of the observation. We
can therefore compute in advance what the sequence of variance values will be. Third, the sequence of variance values converges quickly to a fixed value that depends only on o2 and
2, thereby substantially simplifying the subsequent calculations. (See Exercise 14.VARL) 14.4.3
The general case
The preceding derivation illustrates the key property of Gaussian distributions that allows
Kalman filtering to work: the fact that the exponent is a quadratic form. This is true not just
for the univariate case; the full multivariate Gaussian distribution has the form L
N(x:p,E) =ae’5((x’m
)
B
(x’m)
Multiplying out the terms in the exponent, we see that the exponent is also a quadratic func-
tion of the values x; in . Thus, filtering preserves the Gaussian nature of the state distribution.
Section 14.4
Kalman Filters
483
Let us first define the general temporal model used with Kalman filtering. Both the tran-
sition model and the sensor model are required to be a linear transformation with additive Gaussian noise. Thus, we have
P(%1|x) = N(x1:F%, %) P(z|x) = N(z:Hx,.Z.),
(14.21)
where F and Z, are matrices describing the linear transition model and transition noise co-
variance, and H and X are the corresponding matrices for the sensor model. Now the update equations for the mean and covariance, in their full, hairy horribleness, are
Hrir = Fpy+ Ko (241 — HFpy,) S where K,
(14.22)
= (=K H)(FEFT 1),
=
(FL,F +X)H
(H(FZ,F' +X,)H" +X;)"! is the Kalman gain matrix. Be-
lieve it or not, these equations make some intuitive sense. For example, consider the up-
date for the mean state estimate j.. The term Fy, is the predicted state at 1 +
Kalman gain matrix
1, so HFpy, is
the predicted observation. Therefore, the term 7| — HFyy, represents the error in the pre-
dicted observation. This is multiplied by K, to correct the predicted state; hence, K, is a measure of how seriously to take the new observation relative to the prediction. As in
Equation (14.20), we also have the property that the variance update is independent of the observations.
The sequence of values for £, and K, can therefore be computed offline, and
the actual calculations required during online tracking are quite modest.
To illustrate these equations at work, we have applied them to the problem of tracking an object moving on the X—Y plane. The state variables are X = (X,Y,X,¥)7, so F, £,, H, and X are 4 x 4 matrices. Figure 14.11(a) shows the true trajectory, a series of noisy observations,
and the trajectory estimated by Kalman filtering, along with the covariances indicated by the one-standard-deviation contours. The filtering process does a good job of tracking the actual
motion, and, as expected, the variance quickly reaches a fixed point. We can also derive equations for smoothing as well as filtering with linear-Gaussian
models. The smoothing results are shown in Figure 14.11(b). Notice how the variance in the
position estimate is sharply reduced, except at the ends of the trajectory (why?), and that the
estimated trajectory is much smoother. 14.4.4
Applicability of Kalman
filtering
The Kalman filter and its elaborations are used in a vast array of applications. The “classical”
application is in radar tracking of aircraft and missiles. Related applications include acoustic tracking of submarines and ground vehicles and visual tracking of vehicles and people. In a slightly more esoteric vein, Kalman filters are used to reconstruct particle trajectories from
bubble-chamber photographs and ocean currents from satellite surface measurements. The range of application is much larger than just the tracking of motion: any system characterized
by continuous state variables and noisy measurements will do. Such systems include pulp mills, chemical plants, nuclear reactors, plant ecosystems, and national economies.
The fact that Kalman filtering can be applied to a system does not mean that the re-
sults will be valid or useful. The assumptions made—linear-Gaussian transition and sensor
Kalman models—are very strong. The extended Kalman filter (EKF) attempts to overcome nonlin- Extended filter (EKF)
earities in the system being modeled.
A system is nonlinear if the transition model cannot
be described as a matrix multiplication of the state vector, as in Equation (14.21). The EKF
Nonlinear
484
Chapter 14 Probabilistic Reasoning over Time 2D filiering iT
2D smoothing, e hered
(a)
Served Smoathed
(b)
Figure 14.11 (a) Results of Kalman filtering for an object moving on the XY plane, showing the true trajectory (left to right), a series of noisy observations, and the trajectory estimated by Kalman filtering. Variance in the position estimate is indicated by the ovals. (b) The results of Kalman smoothing for the same observation sequence. works by modeling the system as locally linear in x, in the region of X, = j1,, the mean of the
current state distribution. This works well for smooth, well-behaved systems and allows the
tracker to maintain and update a Gaussian state distribution that is a reasonable approximation to the true posterior. A detailed example is given in Chapter 26.
What does it mean for a system to be “unsmooth” or “poorly behaved”? Technically,
it means that there is significant nonlinearity in system response within the region that is “close” (according to the covariance X;) to the current mean f,.
To understand this idea
in nontechnical terms, consider the example of trying to track a bird as it flies through the
jungle. The bird appears to be heading at high speed straight for a tree trunk. The Kalman
filter, whether regular or extended, can make only a Gaussian prediction of the location of the
bird, and the mean of this Gaussian will be centered on the trunk, as shown in Figure 14.12(a). A reasonable model of the bird, on the other hand, would predict evasive action to one side or the other, as shown in Figure 14.12(b). Such a model is highly nonlinear, because the bird’s
decision varies sharply depending on its precise location relative to the trunk.
To handle examples like these, we clearly need a more expressive language for repre-
senting the behavior of the system being modeled. Within the control theory community, for
Switching Kalman filter
which problems such as evasive maneuvering by aircraft raise the same kinds of difficulties,
the standard solution is the switching Kalman filter. In this approach, multiple Kalman filters run in parallel, each using a different model of the system—for example, one for straight
flight, one for sharp left turns, and one for sharp right turns. A weighted sum of predictions
is used, where the weight depends on how well each filter fits the current data. We will see
in the next section that this is simply a special case of the general dynamic Bayesian net-
work model, obtained by adding a discrete “maneuver” state variable to the network shown in Figure 14.9. Switching Kalman filters are discussed further in Exercise 14.KFSW.
Section 14.5
Dynamic Bayesian Networks
485
Figure 14.12 A bird flying toward a tree (top views). (a) A Kalman filter will predict the location of the bird using a single Gaussian centered on the obstacle. (b) A more realistic model allows for the bird’s evasive action, predicting that it will fly to one side or the other. 14.5
Dyna
Bayesian
Networks
Bayesian Dynamic Bayesian networks, or DBNs, extend the semantics of standard Bayesian networks Dynamic network to handle temporal probability models of the kind described in Section 14.1. We have already seen examples of DBN:
the umbrella network in Figure 14.2 and the Kalman filter network
in Figure 14.9. In general, each slice of a DBN can have any number of state variables X,
and evidence variables E,. For simplicity, we assume that the variables, their links, and their
conditional distributions are exactly replicated from slice to slice and that the DBN represents
a first-order Markov process, so that each variable can have parents only in its own slice or the immediately preceding slice. In this way, the DBN corresponds to a Bayesian network with infinitely many variables.
It should be clear that every hidden Markov model can be represented as a DBN
with
a single state variable and a single evidence variable. It is also the case that every discrete-
variable DBN can be represented as an HMM; as explained in Section 14.3, we can combine all the state variables in the DBN into a single state variable whose values are all possible tuples of values of the individual state variables. Now, if every HMM is a DBN and every DBN can be translated into an HMM, what's the difference? The difference is that, by decomposing the state of a complex system into its constituent variables, we can take advantage
of sparseness in the temporal probability model.
To see what this means in practice, remember that in Section 14.3 we said that an HMM
representation for a temporal process with n discrete variables, each with up to d values,
needs a transition matrix of size 0({]2”). ‘The DBN representation, on the other hand, has size
0O(nd") if the number of parents of each variable is bounded by . In other words, the DBN
representation is linear rather than exponential in the number of variables.
For the vacuum
robot with 42 possibly dirty locations, the number of probabilities required is reduced from 5% 10% to a few thousand.
We have already explained that every Kalman filter model can be represented in a DBN
with continuous
variables and linear-Gaussian conditional distributions (Figure
14.9).
It
should be clear from the discussion at the end of the preceding section that not every DBN
can be represented by a Kalman filter model. In a Kalman filter, the current state distribution
486
Chapter 14 Probabilistic Reasoning over Time
o
o)
07!
Ry [PRilRo) S|
07
03
B
(P(UIIRY)| 0.9 0.2
Figure 14.13 Left: Specification of the prior, transition model. and sensor model for the umbrella DBN. Subsequent slices are copies of slice 1. Right: A simple DBN for robot motion in the X-Y plane. is always a single multivariate Gaussian distribution—that s, a single “bump” in a particular location. DBNSs, on the other hand, can model arbitrary distributions.
For many real-world applications, this flexibility is essential. Consider, for example, the current location of my keys. They might be in my pocket, on the bedside table, on the kitchen counter, dangling from the front door, or locked in the car. A single Gaussian bump that included all these places would have to allocate significant probability to the keys being in mid-air above the front garden. Aspects of the real world such as purposive agents, obstacles, and pockets introduce “nonlinearities” that require combinations of discrete and continuous variables in order to get reasonable models.
14.5.1
Constructing
DBNs
To construct a DBN, one must specify three kinds of information: the prior distribution over
the state variables, P(Xo); the transition model P(X,+ 1| X;); and the sensor model P(E, |X,).
To specify the transition and sensor models, one must also specify the topology of the connections between successive slices and between the state and evidence variables. Because the transition and sensor models are assumed to be time-homogeneous—the same for all r—
it is most convenient simply to specify them for the first slice. For example, the complete DBN specification for the umbrella world is given by the three-node network shown in Figure 14.13(a). From this specification, the complete DBN with an unbounded number of time
slices can be constructed as needed by copying the first slice. Let us now consider a more interesting example: monitoring a battery-powered robot moving in the X-Y plane, as introduced at the end of Section 14.1.
First, we need state
variables, which will include both X, = (X,,¥;) for position and X, = (X;.;) for velocity. We assume some method of measuring position—perhaps a fixed camera or onboard GPS (Global Positioning System)—yielding measurements Z,. The position at the next time step depends
on the current position and velocity, as in the standard Kalman filter model. The velocity at the next step depends on the current velocity and the state of the battery. We add Battery; to
Section 14.5
Dynamic Bayesian Networks
487
represent the actual battery charge level, which has as parents the previous battery level and the velocity, and we add BMeter,, which measures the battery charge level. This gives us the basic model shown in Figure 14.13(b). It is worth looking in more depth at the nature of the sensor model for BMeter;.
Let us
suppose, for simplicity, that both Battery, and BMeter, can take on discrete values 0 through 5. (Exercise
model.)
14.BATT asks you to relate this discrete model to a corresponding continuous
If the meter is always accurate, then the CPT P(BMeter, | Battery,) should have
probabilities of 1.0 “along the diagonal” and probabilities of 0.0 elsewhere.
always creeps into measurements.
In reality, noise
For continuous measurements, a Gaussian distribution
with a small variance might be used.”
For our discrete variables, we can approximate a
Gaussian using a distribution in which the probability of error drops off in the appropriate
way, so that the probability of a large error is very small. We use the term Gaussian error model to cover both the continuous and discrete versions.
Anyone with hands-on experience of robotics, computerized process control, or other
Gaussian error model
forms of automatic sensing will readily testify to the fact that small amounts of measurement noise are often the least of one’s
problems.
Real sensors fail.
When a sensor fails, it does
not necessarily send a signal saying, “Oh, by the way, the data I'm about to send you is a load of nonsense.”
Instead, it simply sends the nonsense.
The simplest kind of failure is
called a transient failure, where the sensor occasionally decides to send some nonsense. For
example, the battery level sensor might have a habit of sending a reading of 0 when someone bumps the robot, even if the battery is fully charged.
Let’s see what happens when a transient failure occurs with a Gaussian error model that
doesn’t accommodate such failures.
Suppose, for example, that the robot is sitting quietly
and observes 20 consecutive battery readings of 5. Then the battery meter has a temporary seizure and the next reading is BMetery;
=0. What will the simple Gaussian error model lead
us to believe about Barteryy? According to Bayes’ rule, the answer depends on both the sensor model P(BMeter =0| Batterys) and the prediction P(Batterys) | BMetery). If the
probability of a large sensor error is significantly less than the probability of a transition to
Battery,; =0, even if the latter is very unlikely, then the posterior distribution will assign a high probability to the battery’s being empty.
A second reading of 0 at 7 =22 will make this conclusion almost certain. If the transient
failure then disappears and the reading returns to 5 from 7 =23 onwards, the estimate for the
battery level will quickly return to 5. (This does not mean the algorithm thinks the battery magically recharged itself, which may be physically impossible; instead, the algorithm now
believes that the battery was never low and the extremely unlikely hypothesis that the battery
meter had two consecutive huge errors must be the right explanation.) This course of events
is illustrated in the upper curve of Figure 14.14(a), which shows the expected value (see
Appendix A) of Battery, over time, using a discrete Gaussian error model.
Despite the recovery, there is a time (r=22) when the robot is convinced that its battery is empty; presumably, then, it should send out a mayday signal and shut down. Alas, its oversimplified sensor model has led it astray. The moral of the story is simple: for the system 10 handle sensor failure properly, the sensor model must include the possibility of failure. 7 Strictly speaking, a Gaussia ution is problematic because it assigns nonzero probability to large negative charge levels. The beta distribution s a better choice for a variable whose range is restricted.
Transient failure
488
Chapter 14 Probabilistic Reasoning over Time E(Battery, ..5555005555...)
5
5
4
4
S 2 4 1
S 2 = 1
T
\
3
£
o
El
E(Battery, ..5555005555...
KM e K E(Battery, |...5555000000...) 15
20
25
30
Time step ¢
0
Bl
¥\
\
\\
k\'—n—n—n—n—n—n—x E(Bartery, |...5555000000...) 15
20
Time step
()
25
30
(®)
Figure 14.14 (a) Upper curve: trajectory of the expected value of Battery, for an observation sequence consisting of all Ss except for Os at =21 and =22, using a simple Gaussian error model. Lower curve: trajectory when the observation remains at 0 from =21 onwards. (b) The same experiment run with the transient failure model. The transient failure is handled well, but the persistent failure results in excessive pessimism about the battery charge. The simplest kind of failure model for a sensor allows a certain probability that the sensor will return some completely incorrect value, regardless of the true state of the world. For
example, if the battery meter fails by returning 0, we might say that
P(BMeter,=0| Battery, =5)=0.03, Transient failure
which is presumably much larger than the probability assigned by the simple Gaussian error
model. Let's call this the transient failure model. How does it help when we are faced
with a reading of 0? Provided that the predicted probability of an empty battery, according
to the readings so far, is much less than 0.03, then the best explanation of the observation
BMeter;; =0 is that the sensor has temporarily failed. Intuitively, we can think of the belief
about the battery level as having a certain amount of “inertia” that helps to overcome tempo-
rary blips in the meter reading. The upper curve in Figure 14.14(b) shows that the transient failure model can handle transient failures without a catastrophic change in beliefs.
So much for temporary blips. What about a persistent sensor failure? Sadly, failures of
this kind are all too common.
If the sensor returns 20 readings of 5 followed by 20 readings
of 0, then the transient sensor failure model described in the preceding paragraph will result in the robot gradually coming to believe that its battery is empty when in fact it may be that the meter has failed.
The lower curve in Figure
14.14(b) shows the belief “trajectory” for
this case. By r=25—five readings of O—the robot is convinced that its battery is empty.
Obviously, we would prefer the robot to believe that its battery meter is broken—if indeed
Pt
fallure
this is the more likely event.
Unsurprisingly, to handle persistent failure, we need a persistent failure model that
describes how the sensor behaves
under normal conditions
and after failure.
To do this,
we need to augment the state of the system with an additional variable, say, BMBroken, that
describes the status of the battery meter. The persistence of failure must be modeled by an
Section 14.5
Dynamic Bayesian Networks E(Battery,|...5555005555...)
B, | P(B)
7 [ 1.000
/|
489
ey
0.001
E(Battery,|..5555000000...)
BMBrokeny
P(BMBroken, |..5555000000...) Roe e eea
BMeter,
P(BMBroken, ...5555005555..) 15
(a)
20
“Time step
25
30
(b)
Figure 14.15 (a) A DBN fragment showing the sensor status variable required for modeling persistent failure of the battery sensor. (b) Upper curves: trajectories of the expected value of Battery, for the “transient failure” and “permanent failure” observations sequences. Lower curves: probability trajectories for BMBroken given the two observation sequences. arc linking BMBrokeny to BMBroken,.
This persistence arc has a CPT that gives a small
probability of failure in any given time step, say, 0.001, but specifies that the sensor stays
broken once it breaks. When the sensor is OK, the sensor model for BMeter is identical to the transient failure model; when the sensor is broken, it says BMeter is always 0, regardless
of the actual battery charge. The persistent failure model for the battery sensor is shown in Figure 14.15(a). Its performance on the two data sequences (temporary blip and persistent failure) is shown in Figure 14.15(b).
There are several things to notice about these curves.
First, in the case of the
temporary blip, the probability that the sensor is broken rises significantly after the second
0 reading, but immediately drops back to zero once a 5 is observed. Second, in the case of persistent failure, the probability that the sensor is broken rises quickly to almost 1 and stays there.
Finally, once the sensor is known to be broken, the robot can only assume that its
battery discharges at the “normal” rate. This is shown by the gradually descending level of E(Battery,|...).
So far, we have merely scraiched the surface of the problem of representing complex
processes.
The variety of transition models is huge, encompassing
topics as disparate as
‘modeling the human endocrine system and modeling multiple vehicles driving on a freeway. Sensor modeling is also a vast subfield in itself. But dynamic Bayesian networks can model
even subtle phenomena, such as sensor drift, sudden decalibration, and the effects of exoge-
nous conditions (such as weather) on sensor readings. 14.5.2
Exact inference in DBNs
Having sketched some ideas for representing complex processes
as DBNs, we now turn to the
question of inference. In a sense, this question has already been answered: dynamic Bayesian networks are Bayesian networks, and we already have algorithms for inference in Bayesian networks. Given a sequence of observations, one can construct the full Bayesian network rep-
Persistence arc
490
Chapter 14 Probabilistic Reasoning over Time
BR[O Ry|P(R)|R,
07
03
I
07
P(R)|R,
R,
[P(R IR,
Ry
|P(RyR;
f1
03
f1
03
f1
03
@@@@
T
(R, | P(UIR,)|
AN
f1l
02
T e
Ry |
@@
Tl T Tl (R |P(UIR)|
o[
fl
09
02
Ry [ PCUAR,)| fl
[
o9 | 02
(R; [ PCUSR,)
[t]
£l
o9
02
Figure 14.16 Unrolling a dynamic Bayesian network: slices are replicated to accommodate the observation sequence Umbrellay;s. Further slices have no effect on inferences within the observation period. resentation of a DBN by replicating slices until the network is large enough to accommodate the observations, as in Figure 14.16. This technique is called unrolling. (Technically, the DBN
is equivalent to the semi-infinite network obtained by unrolling forever.
Slices added
beyond the last observation have no effect on inferences within the observation period and can be omitted.) Once the DBN is unrolled, one can use any of the inference algorithms— variable elimination, clustering methods, and so on—described in Chapter 13.
Unfortunately, a naive application of unrolling would not be particularly efficient. If
we want to perform filtering or smoothing with a long sequence of observations ey, the
unrolled network would require O(r) space and would thus grow without bound as more observations were added. Moreover, if we simply run the inference algorithm anew each time an observation is added, the inference time per update will also increase as O(f). Looking back to Section 14.2.1, we see that constant time and space per filtering update
can be achieved if the computation can be done recursively. Essentially, the filtering update in Equation (14.5) works by summing out the state variables of the previous time step to get the distribution for the new time step. Summing out variables is exactly what the variable
elimination (Figure 13.13) algorithm does, and it turns out that running variable elimination
with the variables in temporal order exactly mimics the operation of the recursive filtering
update in Equation (14.5). The modified algorithm keeps at most two slices in memory at any one time: starting with slice 0, we add slice 1, then sum out slice 0, then add slice 2, then
sum out slice 1, and so on. In this way, we can achieve constant space and time per filtering update.
(The same performance can be achieved by suitable modifications to the clustering
algorithm.) Exercise 14.DBNE asks you to verify this fact for the umbrella network.
So much for the good news; now for the bad news: It turns out that the “constant” for the
per-update time and space complexity is, in almost all cases, exponential in the number of state variables. What happens is that, as the variable elimination proceeds, the factors grow
to include all the state variables (or, more precisely, all those state variables that have parents
in the previous time slice). The maximum factor size is O(d"**) and the total update cost per
step is O(nd"**), where d is the domain size of the variables and k is the maximum number
of parents of any state variable.
Of course, this is much less than the cost of HMM updating, which is O(d?"), but it s still infeasible for large numbers of variables. This grim fact means is that even though we can use
DBN s to represent very complex temporal processes with many sparsely connected variables,
Section 14.5
Dynamic Bayesian Networks
491
we cannot reason efficiently and exactly about those processes. The DBN model itself, which
represents the prior joint distribution over all the variables, is factorable into its constituent CPTs, but the posterior joint distribution conditioned on an observation sequence—that
is,
the forward message—is generally not factorable. The problem is intractable in general, so we must fall back on approximate methods.
14.5.3
Approximate inference in DBNs
Section 13.4 described two approximation algorithms: likelihood weighting (Figure 13.18)
and Markov chain Monte Carlo (MCMC, Figure 13.20). Of the two, the former is most easily adapted to the DBN context. (An MCMC filtering algorithm is described briefly in the notes
at the end of this chapter.) We will see, however, that several improvements are required over
the standard likelihood weighting algorithm before a practical method emerges. Recall that likelihood weighting works by sampling the nonevidence nodes of the network in topological order, weighting each sample by the likelihood it accords to the observed evidence variables. As with the exact algorithms, we could apply likelihood weighting directly to an unrolled DBN, but this would suffer from the same problems of increasing time
and space requirements per update as the observation sequence grows. The problem is that
the standard algorithm runs each sample in turn, all the way through the network.
Instead, we can simply run all N samples together through the DBN, one slice at a time.
The modified algorithm fits the general pattern of filtering algorithms, with the set of N samples as the forward message. The first key innovation, then, is to use the samples themselves as an approximate representation of the current state distribution.
This meets the require-
For a single customer Cj recommending a single book B1, the Bayes net might look like the one shown in Figure 15.2(a). (Just as in Section 9.1, expressions with parentheses such as Honest(Cy) are just fancy symbols—in this case, fancy names for random variables.) With 1" The name relational probability model was given by Pfeffer (2000) to a slightly different representation, but the underlying ideas are the same. 2 A game theorist would advise a dishones! customer to avoid detection by occasionally recommending a good book from a competitor. See Chapter 18.
Section 15.1
503
Relational Probability Models
two customers and two books, the Bayes net looks like the one in Figure 15.2(b). For larger
numbers of books and customers, it is quite impractical to specify a Bayes net by hand.
Fortunately, the network has a lot of repeated structure. Each Recommendation(c,b) vari-
able has as its parents the variables Honest(c), Kindness(c), and Quality(b). Moreover, the conditional probability tables (CPTs) for all the Recommendation(c,b) variables are identical, as are those for all the Honest(c) variables, and so on. The situation seems tailor-made
for a first-order language. We would like to say something like Recommendation(c,b) ~ RecCPT (Honest(c), Kindness c), Quality(b))
which means that a customer’s recommendation for a book depends probabilistically on the
customer’s honesty and kindness and the book’s quality according to a fixed CPT.
Like first-order logic, RPMs have constant, function, and predicate symbols. We will also
assume a type signature for each function—that is, a specification of the type of each argu-
ment and the function’s value. (If the type of each object is known, many spurious possible worlds are eliminated by this mechanism; for example, we need not worry about the kindness of each book, books recommending customers, and so on.)
Type signature
For the book-recommendation
domain, the types are Customer and Book, and the type signatures for the functions and predicates are as follows:
Honest : Customer — {true false}
Kindness : Customer — {1,2,3,4,5} Quality : Book — {1,2,3,4,5} Recommendation : Customer x Book — {1,2,3,4,5} The constant symbols will be whatever customer and book names appear in the retailer’s data
set. In the example given in Figure 15.2(b), these were Cy, C> and B1, B. Given the constants and their types, together with the functions and their type signatures,
the basic random variables of the RPM are obtained by instantiating each function with each possible combination of objects. For the book recommendation model, the basic random
variables include Honest(Cy), Quality(Bs), Recommendation(C\,B3), and so on. These are exactly the variables appearing in Figure 15.2(b). Because each type has only finitely many instances (thanks to the domain closure assumption), the number of basic random variables is also finite.
To complete the RPM, we have to write the dependencies that govern these random vari-
ables. There function is a For example, of honesty is
is one dependency statement for each function, where each argument of the logical variable (i.c., a variable that ranges over objects, as in first-order logic). the following dependency states that, for every customer ¢, the prior probability 0.9 true and 0.01 false:
Honest(c) ~ (0.99,0.01)
Similarly, we can state prior probabilities for the kindness value of each customer and the quality of each book, each on the 1-5 scale: Kindness(c) ~ (0.1,0.1,0.2,0.3,0.3)
Quality(h) ~ (0.05,0.2,0.4,0.2,0.15) Finally, we need the dependency for recommendations: for any customer ¢ and book b, the score depends on the honesty and kindness of the customer and the quality of the book:
Recommendation(c,b) ~ RecCPT(Honest(c), Kindness(c), Quality (b))
Basic random variable
504
Chapter 15 Probabilistic Programming where RecCPT is a separately defined conditional probability table with 2 x 5 x 5=50 rows,
each with 5 entries. For the purposes of illustration, we’ll assume that an honest recommen-
dation for a book of quality ¢ from a person of kindness is uniformly distributed in the range 1155, 145411 The semantics of the RPM can be obtained by instantiating these dependencies for all known constants, giving a Bayesian network (as in Figure 15.2(b)) that defines a joint distribution over the RPM’s random variables.>
The set of possible worlds is the Cartesian product of the ranges of all the basic random variables, and, as with Bayesian networks, the probability for each possible world is the product of the relevant conditional probabilities from the model. With C customers and B books, there are C Honest variables, C Kindness variables, B Quality variables, and BC Recommendation variables, leading to 265*3+5C possible worlds. With ten million books
and a billion customers, that’s about 107> 1" worlds. Thanks to the expressive power of
RPMs, the complete probability model still has only fewer than 300 parameters—most of them in the RecCPT table. We can refine the model by asserting a context-specific independence (see page 420) to reflect the fact that dishonest customers ignore quality when giving a recommendation; more-
over, kindness plays no role in their decisions. Thus, Recommendation(c,b) is independent of Kindness(c) and Quality(b) when Honest(c) = false:
Recommendation(c,b) ~
if Honest(c) then HonestRecCPT (Kindness(c), Quality(b)) else (0.4,0.1,0.0,0.1,0.4).
This kind of dependency may look like an ordinary if-then—else statement in a programming language, but there is a key difference: the inference engine doesn’t necessarily know the value of the conditional test because Honest(c) is a random variable.
We can elaborate this model in endless ways to make it more realistic.
For example,
suppose that an honest customer who is a fan of a book’s author always gives the book a 5, regardless of quality:
Recommendation(c,b) ~
if Honest(c) then if Fan(c, Author (b)) then Exactly(5) else HonestRecCPT (Kindness(c), Quality(b)) else (0.4,0.1,0.0,0.1,0.4)
Again, the conditional test Fan(c,Author(b)) is unknown, but if a customer gives only 5s to a
particular author’s books and is not otherwise especially kind, then the posterior probability
that the customer is a fan of that author will be high. Furthermore, the posterior distribution will tend to discount the customer’s 5s in evaluating the quality of that author’s books.
In this example, we implicitly assumed that the value of Author(b) is known for every b, but this may not be the case. How can the system reason about whether, say, C) is a fan of Author(B,) when Author(Bs) is unknown?
The answer is that the system may have to
reason about all possible authors. Suppose (to keep things simple) 3 Some technical conditions are required for an RPM to define a proper distribution. be acyclic; otherwise the resulting Bayesian network will have cycles. Second, the be well-founded: there can be no infinite ancestor chains, such as might arise from Exercise 15.HAMD for an exception to this rule.
that there are just two First, the dependencies must dependencies must (usually) recursive dependen
Section 15.1
Recommendation(Cy, B>
Relational Probability Models
505
Recommendation(Cy, By)
Figure 15.3 Fragment of the equivalent Bayes net for the book recommendation RPM when Author(Bz) is unknown. authors, A; and Ay, Then Author(Ba) is a random variable with two possible values, A; and Ay, and it is a parent of Recommendation(Cy, By). The variables Fan(Cy,A;) and Fan(C),As)
are parents too. The conditional distribution for Recommendation(C),B,) is then essentially a
‘multiplexer in which the Author(B,) parent acts as a selector to choose which of Fan(Cy,A;)
and Fan(C,A,) actually gets to influence the recommendation. A fragment of the equivalent
Multiplexer
Bayes net is shown in Figure 15.3. Uncertainty in the value of Author(B,), which affects the dependency structure of the network, is an instance of relational uncertainty. e In case you are wondering how the system can possibly work out who the author of B, is: consider the possibility that three other customers are fans of A} (and have no other favorite authors in common) and all three have given B; a 5, even though most other customers find
it quite dismal. In that case, it is extremely likely that A, is the author of B,. The emergence of sophisticated reasoning like this from an RPM model of just a few lines is an intriguing
example of how probabilistic influences spread through the web of interconnections among objects in the model. As more dependencies and more objects are added, the picture conveyed by the posterior distribution often becomes clearer and clearer.
15.1.2
Example:
Rating player skill levels
Many competitive games have a numerical measure of players’ skill levels, sometimes called arating. Perhaps the best-known is the Elo rating for chess players, which rates a typical be- Rating ginner at around 800 and the world champion usually somewhere above 2800. Although Elo ratings have a statistical basis, they have some ad hoc elements. We can develop a Bayesian rating scheme as follows: each player i has an underlying skill level Skill(i); in each game g, Ps actual performance is Performance(i,g), which may vary from the underlying skill level; and the winner of g is the player whose performance in g is better. As an RPM, the model looks like this: Skill(i) ~ N (11, 0?) Performance(i,g) ~ N (Skill(i),%)
Win(i.j,g) = if Game(g,i, j) then (Performance(i,g) > Performance(j,g))
where /32 is the variance ofa player’s actual performance in any specific game relative to the
player’s underlying skill level. Given a set of players and games, as well as outcomes for
some of the games, an RPM inference engine can compute a posterior distribution over the
skill of each player and the probable outcome of any additional game that might be played.
506
Chapter 15 Probabilistic Programming For team games, we’ll assume, as a first approximation, that the overall performance of
team ¢ in game g is the sum of the individual performances of the players on :
TeamPerformance(t,g) = ¥ic, Performance(i,g).
Even though the individual performances are not visible to the ratings engine, the players’ skill levels can still be estimated from the results of several games, as long as the team com-
positions vary across games. Microsoft’s TrueSkill™ ratings engine uses this model, along
with an efficient approximate inference algorithm, to serve hundreds of millions of users
every day. This model can be elaborated in numerous ways. For example, we might assume that weaker players have higher variance in their performance; we might include the player’s role on the team; and we might consider specific kinds of performance and skill—e.g., defending and attacking—in order to improve team composition and predictive accuracy.
15.1.3
Inference in relational probability models
The most straightforward approach to inference in RPMs s simply to construct the equivalent Bayesian network, given the known constant symbols belonging to each type. With B books
and C customers, the basic model given previously could be constructed with simple loops:*
forb=1toBdo add node Quality,, with no parents, prior ( 0.05,0.2,0.4,0.2,0.15) forc=1toCdo add node Honest, with no parents, prior ( 0.99,0.01 ) add node Kindness. with no parents, prior ( 0.1,0.1,0.2,0.3,03 ) forb=1toBdo add node Recommendation, with parents Honeste, Kindness.. Quality), and conditional distribution RecCPT (Honestc, Kindnessc, Quality,) Grounding Unrolling
This technique is called grounding or unrolling; it is the exact analog of propositionaliza-
tion for first-order logic (page 280). The obvious drawback is that the resulting Bayes net may be very large. Furthermore, if there are many candidate objects for an unknown relation or function—for example, the unknown author of By—then some variables in the network
may have many parents. Fortunately, it is often possible to avoid generating the entire implicit Bayes net. As we saw in the discussion of the variable elimination algorithm on page 433, every variable that is not an ancestor of a query variable or evidence variable is irrelevant to the query. Moreover, if the query is conditionally independent of some variable given the evidence, then that variable is also irrelevant. So, by chaining through the model starting from the query and evidence, we can identify just the set of variables that are relevant to the query. These are the only ones
that need to be instantiated to create a potentially tiny fragment of the implicit Bayes net.
Inference in this fragment gives the same answer as inference in the entire implicit Bayes net.
Another avenue for improving the efficiency of inference comes from the presence of repeated substructure in the unrolled Bayes net. This means that many of the factors constructed during variable elimination (and similar kinds of tables constructed by clustering algorithms) 4 Several statistical packages would view this code as defining the RPM, rather than just constructing a Bayes net to perform inference in the RPM. This view, however, misses an important role for RPM syntax: without a syntax with clear semantics, there is no way the model structure can be learned from data.
Section 152
Open-Universe Probability Models
507
will be identical; effective caching schemes have yielded speedups of three orders of magnitude for large networks.
Third, MCMC inference algorithms have some interesting properties when applied to RPMs with relational uncertainty. MCMC works by sampling complete possible worlds, 50 in each state the relational structure is completely known. In the example given earlier, each MCMC state would specify the value of Author(B,), and so the other potential authors are no longer parents of the recommendation nodes for B;.
For MCMC,
then, relational
uncertainty causes no increase in network complexity; instead, the MCMC process includes
transitions that change the relational structure, and hence the dependency structure, of the unrolled network.
Finally, it may be possible in some cases to avoid grounding the model altogether. Resolution theorem provers and logic programming systems avoid propositionalizing by instanti-
ating the logical variables only as needed to make the inference go through; that is, they /ift
the inference process above the level of ground propositional sentences and make each lifted step do the work of many ground steps.
The same idea can be applied in probabilistic inference. For example, in the variable
elimination algorithm, a lifted factor can represent an entire set of ground factors that assign
probabilities to random variables in the RPM, where those random variables differ only in the
constant symbols used to construct them. The details of this method are beyond the scope of
this book, but references are given at the end of the chapter. 15.2
Open-Universe
Probability Models
‘We argued earlier that database semantics was appropriate for situations in which we know
exactly the set of relevant objects that exist and can identify them unambiguously. (In partic-
ular, all observations about an object are correctly associated with the constant symbol that
names it.) In many real-world settings, however, these assumptions are simply untenable. For example, a book retailer might use an ISBN (International Standard Book Number)
constant symbol to name each book, even though a given “logical” book (e.g., “Gone With the Wind”) may have several ISBNs corresponding to hardcover, paperback, large print, reissues, and so on. It would make sense to aggregate recommendations across multiple ISBN, but the retailer may not know for sure which ISBNs are really the same book. (Note that we
are not reifying the individual copies of the book, which might be necessary for used-book
sales, car sales, and so on.) Worse still, each customer is identified by a login ID, but a dis-
honest customer may have thousands of IDs!
In the computer security field, these multiple
IDs are called sybils and their use to confound a reputation system is called a sybil attack.’ Thus, even a simple application in a relatively well-defined, online domain involves both existence uncertainty (what are the real books and customers underlying the observed data) and identity uncertainty (which logical terms really refer to the same object).
The phenomena of existence and identity uncertainty extend far beyond online book-
sellers. In fact they are pervasive:
+ A vision system doesn’t know what exists, if anything, around the next corner, and may not know if the object it sees now is the same one it saw a few minutes ago.
S The name “Sybi
comes from a famous case of multiple personality disorder.
Sybil
Sybil attack Existence uncertainty Identity uncertainty
508
Chapter 15 Probabilistic Programming +A
text-understanding system does not know in advance the entities that will be featured in a text, and must reason about whether phrases such as “Mary,” “Dr. Smith,” “she,” “his cardiologist,” “his mother,” and so on refer to the same object.
« An intelligence analyst hunting for spies never knows how many spies there really are and can only guess whether various pseudonyms, phone numbers, and sightings belong to the same individual.
Indeed, a major part of human cognition seems to require learning what objects exist and
being able to connect observations—which almost never come with unique IDs attached—to
Open universe probability model (OUPM)
hypothesized objects in the world.
Thus, we need to be able to define an open universe probability model (OUPM) based
on the standard semantics of first-order logic, as illustrated at the top of Figure
15.1.
A
language for OUPMs provides a way of easily writing such models while guaranteeing a unique, consistent probability distribution over the infinite space of possible worlds. 15.2.1
Syntax and semantics
The basic idea is to understand how ordinary Bayesian networks and RPMs manage to define a unique probability model and to transfer that insight to the first-order setting. In essence, a Bayes net generates each possible world, event by event, in the topological order defined
by the network structure, where each event is an assignment of a value to a variable. An RPM extends this to entire sets of events, defined by the possible instantiations of the logical
variables in a given predicate or function. OUPMs go further by allowing generative steps that
Number statement
add objects to the possible world under construction, where the number and type of objects may depend on the objects that are already in that world and their properties and relations. That is, the event being generated is not the assignment of a value to a variable, but the very existence of objects. One way o do this in OUPMs is to provide number statements that specify conditional distributions over the numbers of objects of various kinds. For example, in the book-
recommendation domain, we might want to distinguish between customers (real people) and their login IDs. (It’s actually login IDs that make recommendations, not customers!) Suppose
(to keep things simple) the number of customers is uniform between 1 and 3 and the number of books is uniform between 2 and 4:
#Customer ~ UniformInt(1,3) #Book ~ Uniformini(2,4).
(15.2)
‘We expect honest customers to have just one ID, whereas dishonest customers might have anywhere between 2 and 5 IDs:
#LoginID(Owner=c) ~ Origin function
if Honest(c) then Exactly(1) else Uniformint(2,5).
(15.3)
This number statement specifies the distribution over the number of login IDs for which customer c is the Owner.
The Owner function is called an origin function because it says
where each object generated by this number statement came from.
The example in the preceding paragraph uses a uniform distribution over the integers
between 2 and 5 to specify the number of logins for a dishonest customer.
This particular
distribution is bounded, but in general there may not be an a priori bound on the number of
Section 152
Open-Universe Probability Models
objects. The most commonly used distribution over the nonnegative integers is the Poisson distribution. The Poisson has one parameter, \, which is the expected number of objects, and a variable X sampled from Poisson(X) has the following distribution:
509
Poisson distribution
P(X=k)=Xe /K. The variance of the Poisson is also A, so the standard deviation is v/X.
This means that
for large values of ), the distribution is narrow relative to the mean—for example, if the number of ants in a nest is modeled by a Poisson with a mean of one million, the standard deviation is only a thousand, or 0.1%. For large numbers, it often makes more sense to use
the discrete log-normal distribution, which is appropriate when the log of the number of
Discrete log-normal distribution
magnitude distribution, uses logs to base 10: thus, a distribution OM(3, 1) has a mean of
Order-of-magnitude distribution
objects is normally distributed.
A particularly intuitive form, which we call the order-of-
10° and a standard deviation of one order of magnitude, i.c., the bulk of the probability mass falls between 102 and 10°.
The formal semantics of OUPMs begins with a definition of the objects that populate
possible worlds. In the standard semantics of typed first-order logic, objects are just numbered tokens with types. In OUPMs, each object is a generation history; for example, an object might be “the fourth login ID of the seventh customer.” (The reason for this slightly baroque construction will become clear shortly.) For types with no origin functions—e.g., the Customer and Book types in Equation (15.2)—the objects have an empty origin; for example, (Customer, ,2) refers to the second customer generated from that number statement.
For number statements with origin functions—e.g., Equation (15.3)—each object records its origin; for example, the object (LoginID, (Owner, (Customer, ,2)),3) is the third login belonging to the second customer. The number variables of an OUPM specify how many objects there are of each type with
each possible origin in each possible world; thus #L0ginlD guuer (Customer, 2)) () =4 means
that in world w, customer 2 owns 4 login IDs. As in relational probability models, the basic random variables determine the values of predicates and functions for all tuples of objects; thus, Honest(cusiomer, 2) (w) = true means that in world w, customer 2 is honest. A possible world is defined by the values of all the number variables and basic random variables. A
world may be generated from the model by sampling in topological order; Figure 15.4 shows an example. The probability of a world so constructed is the product of the probabilities for all the sampled values; in this case, 1.2672x 10~!!. Now it becomes clear why each object contains its origin: this property ensures that every world can be constructed by exactly
one generation sequence. If this were not the case, the probability of a world would be an unwieldy combinatorial sum over all possible generation sequences that create it.
Open-universe models may have infinitely many random variables, so the full theory in-
volves nontrivial measure-theoretic considerations.
For example, number statements with
Poisson or order-of-magnitude distributions allow for unbounded numbers of objects, lead-
ing to unbounded numbers of random variables for the properties and relations of those objects. Moreover, OUPMs can have recursive dependencies and infinite types (integers, strings, etc.). Finally, well-formedness disallows cyclic dependencies and infinitely receding ancestor chains; these conditions are undecidable in general, but certain syntactic sufficient conditions
can be checked easily.
Number variable
510
Chapter 15 Probabilistic Programming Variable #Customer #Book Honest cusiomer. 1)
Value 2 3 rue
Probability 03333 03333 099
1
0.1
Honest(cusiomer. 2) Kindness cusiomer..1)
Jalse 4
Quality gy, 1)
1 3
Kindness customer. 2) Quality gy, Quality gy,
2 3)
FLOgINID,e (Cusomer, 1)
#LoginID guner, (Customer, 2))
Recommendationtginip, (Owner,(Customer, 1))1),(Book 1) Recommendationtaginip, (Owner.(Customer, 1))1)(Book 2)
Recommendation(poginp,(Owner.(Customer, 1)1 (Book.3) RecommendationLoginip, (Owner. (Customer. 2)).1).(Book..1) RecommendationLoginip, (Owner. (Customer. 2)).1).(Book, 2) Recommendationt
ginip,(Owner.(Customer. 2)).1).(Boo
3)
i (0wner. (Custome 2)).2).(Book..1) Recommendationy.oginip,
Recommendation(g oginip, (Owner.(Customer, 2)).2).(Book. 2) Recommendation g oginip, (Owner.(Customer, 2)).2). (Book. 3
5 1 2 2 4 5 5 1 5 5 1
0.01 0.3
0.05 04
015
1.0 0.25
0.5 0.5
05
04 04
04
0.4 0.4 0.4
Figure 15.4 One particular world for the book recommendation OUPM. The number variables and basic random variables are shown in topological order, along with their chosen values and the probabilities for those values.
15.2.2
Inference in open-universe
probability models
Because of the potentially huge and sometimes unbounded size of the implicit Bayes net that corresponds to a typical OUPM, unrolling it fully and performing exact inference is quite impractical. Instead, we must consider approximate inference algorithms such as MCMC (see Section 13.4.2).
Roughly speaking, an MCMC algorithm for an OUPM is exploring the space of possible worlds defined by sets of objects and relations among them, as illustrated in Figure 15.1(top). A move between adjacent states in this space can not only alter relations and functions but also add or subtract objects and change the interpretations of constant symbols. Even though
each possible world may be huge, the probability computations required for each step— whether in Gibbs sampling or Metropolis-Hastings—are entirely local and in most cases take constant time. This is because the probability ratio between neighboring worlds depends on a subgraph of constant size around the variables whose values are changed. Moreover, a logical query can be evaluated incrementally in each world visited, usually in constant time. per world, rather than being recomputing from scratch.
Some special consideration needs to be given to the fact that a typical OUPM may have
possible worlds of infinite size. As an example, consider the multitarget tracking model in Figure 15.9: the function X (a.t), denoting the state of aircraft a at time 1, corresponds to an infinite sequence of variables for an unbounded number of aircraft at each step. For this
reason, MCMC for OUPMs samples not completely specified possible worlds but partial
Section 152
Open-Universe Probability Models
worlds, each corresponding to a disjoint set of complete worlds. A partial world is a minimal
self-supporting instantiation® of a subset of the relevant variables—that is, ancestors of the
evidence and query variables. For example, variables X (a,) for values of7 greater than the last observation time (or the query time, whichever is greater) are irrelevant, so the algorithm can consider just a finite prefix of the infinite sequence.
15.2.3
Examples
The standard “use case” for an OUPM
has three elements:
the model,
the evidence (the
Kknown facts in a given scenario), and the guery, which may be any expression, possibly with free logical variables. The answer is a posterior joint probability for each possible set of substitutions for the free variables, given the evidence, according to the model.” Every model includes type declarations, type signatures for the predicates and functions, one or
more number statements for each type, and one dependency statement for each predicate and
function. (In the examples below, declarations and signatures are omitted where the meaning
is clear.) As in RPMs, dependency statements use an if-then-else syntax to handle contextspecific dependencies. Citation matching
Millions of academic research papers and technical reports are to be found online in the form of pdf files. Such papers usually contain a section near the end called “References” or “Bibliography,” in which citations—strings of characters—are provided to inform the reader
of related work. These strings can be located and “scraped” from the pdf files with the aim of
creating a database-like representation that relates papers and researchers by authorship and
citation links. Systems such as CiteSeer and Google Scholar present such a representation to their users; behind the scenes, algorithms operate to find papers, scrape the citation strings, and identify the actual papers to which the citation strings refer. This is a difficult task because these strings contain no object identifiers and include errors of syntax, spelling, punctuation,
and content. To illustrate this, here are two relatively benign examples:
1. [Lashkari et al 94] Collaborative Interface Agents, Yezdi Lashkari, Max Metral, and Pattie Maes, Proceedings of the Twelfth National Conference on Articial Intelligence, MIT Press, Cambridge, MA, 1994. 2. Metral M. Lashkari, Y. and P. Maes. Collaborative interface agents. In Conference of the American Association for Artificial Intelligence, Seattle, WA, August 1994.
The key question is one of identity: are these citations of the same paper or different pa-
pers? Asked this question, even experts disagree or are unwilling to decide, indicating that
reasoning under uncertainty is going to be an important part of solving this problem.® Ad hoc approaches—such as methods based on a textual similarity metric—often fail miserably. For example, in 2002,
CiteSeer reported over 120 distinct books written by Russell and Norvig.
one in which the parents of every variable in the set are
7 As with Prolog, there may be infinitely many sets of substitutions of unbounded size; designing exploratory interfaces for such answers is an interesting visualization challenge. 8 The answer is yes, they are the same paper. The “National Conference on Articial Intelligence” (notice how the “fi” s missing, thanks to an error in seraping the ligature character) is another name for the AAAI conference; the conference took place in Seattle whereas the proceedings publisher is in Cambridge.
511
512
Chapter 15 Probabilistic Programming type Researcher, Paper, Citation random String Name(Researcher) random String Title(Paper) random Paper PubCited(Citation) random String Text(Citation) random Boolean Professor(Researcher) origin Researcher Author(Paper) #Researcher ~ OM(3,1) Name(r) ~ NamePrior() Professor(r) ~ Boolean(0.2) #Paper(Author=r) ~ if Professor(r) then OM(1.5,0.5) else OM(1,0.5) Title(p) ~ PaperTitlePrior() CitedPaper(c) ~ UniformChoice({Paper p}) Text(c) ~ HMMGrammar(Name(Author(CitedPaper(c))), Title(CitedPaper(c))) Figure 15.5 An OUPM for citation information extraction. For simplicity the model assumes one author per paper and omits details of the grammar and error models. In order to solve the problem using a probabilistic approach, we need a generative model
for the domain.
That is, we ask how these citation strings come to be in the world.
The
process begins with researchers, who have names. (We don’t need to worry about how the researchers came into existence; we just need to express our uncertainty about how many
there are.) These researchers write some papers, which have titles; people cite the papers,
combining the authors’ names and the paper’s title (with errors) into the text of the citation according to some grammar.
The basic elements of this model are shown in Figure 15.5,
covering the case where papers have just one author.”
Given just citation strings as evidence, probabilistic inference on this model to pick
out the most likely explanation for the data produces an error rate 2 to 3 times lower than CiteSeer’s (Pasula et al., 2003).
The inference process also exhibits a form of collective,
knowledge-driven disambiguation: the more citations for a given paper, the more accurately each of them is parsed, because the parses have to agree on the facts about the paper. Nuclear treaty monitoring
Verifying the Comprehensive Nuclear-Test-Ban Treaty requires finding all seismic events on Earth above a minimum
magnitude.
The UN CTBTO
maintains a network of sensors, the
International Monitoring System (IMS); its automated processing software, based on 100 years of seismology research, has a detection failure rate of about 30%. The NET-VISA system (Arora ef al., 2013), based on an OUPM, significantly reduces detection failures. The NET-VISA model (Figure 15.6) expresses the relevant geophysics directly. It describes distributions over the number of events in a given time interval (most of which are
9 The multi-author case has the same overall structure but is a bit more complicated. The parts of the model not shown—the NamePrior, ritlePrior, and HMMGrammar—are traditional probability models. For example, the NamePrior is a mixture of a categorical distribution over actual names and a letter trigram model (sec Section 23.1) to cover names not previously seen, both trained from data in the U.S. Census database.
Section 152
Open-Universe Probability Models
#SeismicEvents ~ Poisson(T * \e) Time(e) ~ UniformReal(0,T) EarthQuake(e) ~ Boolean(0.999) Location(e) ~ if Earthquake(e) then SpatialPrior() else UniformEarth() Depth(e) ~ if Earthquake(e) then UniformReal(0,700) else Exacly(0) Magnitude(e) ~ Exponential(log(10)) Detected(e,p.s) ~ Logistic(weights(s, p), Magnitude(e), Depth(e), Dist(e.s)) #Detections(site = 5) ~ Poisson(T * Af(s)) #Detections(event=e, phase=p, station=s) = if Detected(e, p.s) then 1 else0 OnserTime(a,s)if (event(a) = null) then ~ UniformReal(0,T) else = Time(event(a)) + GeoTT(Dist(event(a),s), Depth(event(a)).phase(a)) + Laplace(pu(s),o:(s)) Amplitude(a,s) if (event(a) = null) then ~ NoiseAmpModel(s) else = AmpModel(Magnitude(event(a)). Dist(event(a), 5), Depth(event(a)). phase(a) Azimuth(a,s) if (event(a) = null) then ~ UniformReal(0, 360) else = GeoAzimuth(Location(event(a)), Depth(event(a)), phase(a). Site(s)) + Laplace(0,7a(s)) Slowness(a,s) if (event(a) = null) then ~ UniformReal(0,20) else = GeoSlowness(Location(event(a)), Depth(event(a)), phase(a), Site(s)) + Laplace(0,7(s)) ObservedPhase(a,s) ~ CategoricalPhaseModel(phase(a)) Figure 15.6 A simplified version of the NET-VISA model (see text).
naturally occurring) as well as over their time, magnitude, depth, and location. The locations of natural events are distributed according to a spatial prior that is trained (like other parts
of the model) from historical data; man-made events are, by the treaty rules, assumed to oc-
cur uniformly over the surface of the Earth.
At every station s, each phase (seismic wave
type) p from an event ¢ produces either 0 or 1 detections (above-threshold signals); the detection probability depends on the event magnitude and depth and its distance from the station.
“False alarm” detections also occur according to a station-specific rate parameter. The measured arrival time, amplitude, and other properties of a detection d from a real event depend
on the properties of the originating event and its distance from the station.
Once trained, the model runs continuously. The evidence consists of detections (90% of
which are false alarms) extracted from raw IMS waveform data, and the query typically asks for the most likely event history, or bulletin, given the data. Results so far are encouraging;
for example, in 2009 the UN’s SEL3 automated bulletin missed 27.4% of the 27294 events in the magnitude range 3-4 while NET-VISA missed 11.1%. Moreover, comparisons with dense regional networks show that NET-VISA finds up to 50% more real events than the
final bulletins produced by the UN’s expert seismic analysts. NET-VISA also tends to associate more detections with a given event, leading to more accurate location estimates (see
Figure 15.7). As of January 1, 2018, NET-VISA has been deployed as part of the CTBTO ‘monitoring pipeline.
Despite superficial differences, the two examples are structurally similar: there are un-
Kknown objects (papers, earthquakes) that generate percepts according to some physical pro-
513
514
Chapter 15 Probabilistic Programming
i)
:
e (@
(b)
Figure 15.7 (a) Top: Example of seismic waveform recorded at Alice Springs, Australia. Bottom: the waveform after processing to detect the arrival times of seismic waves. Blue lines are the automatically detected arrivals; red lines are the true arrivals. (b) Location estimates for the DPRK nuclear test of February 12, 2013: UN CTBTO Late Event Bulletin (green triangle at top left): NET-VISA (blue square in center). The entrance to the underground test facility (small “x”) is 0.75km from NET-VISA's estimate. Contours show NET-VISA's posterior location distribution. Courtesy of CTBTO Preparatory Commission. cess (citation, seismic propagation). The percepts are ambiguous as to their origin, but when
multiple percepts are hypothesized to have originated with the same unknown object, that
object’s properties can be inferred more accurately. The same structure and reasoning patterns hold for areas such as database deduplication
and natural language understanding. In some cases, inferring an object’s existence involves
grouping percepts together—a process that resembles the clustering task in machine learning.
In other cases, an object may generate no percepts at all and still have its existence inferred— as happened, for example, when observations of Uranus led to the discovery of Neptune. The
existence of the unobserved object follows from its effects on the behavior and properties of
observed objects. 15.3
Keeping Track of a Complex World
Chapter 14 considered the problem of keeping track of the state of the world, but covered only the case of atomic representations (HMMs)
and factored representations (DBNs and
Kalman filters). This makes sense for worlds with a single object—perhaps a single patient
in the intensive care unit or a single bird flying through the forest. In this section, we see what
happens when two or more objects generate the observations. What makes this case different
from plain old state estimation is that there is now the possibility of uncertainty about which
object generated which observation. This is the identity uncertainty problem of Section 15.2 Data association
(page 507), now viewed in a temporal context. In the control theory literature, this is the data association problem—that is, the problem of associating observation data with the objects that generated them. Although we could view this as yet another example of open-universe
probabilistic modeling, it is important enough in practice to deserve its own section.
Section 153
Keeping Track of a Complex World
515
track termination detection failure
(©
@
false alarm —
®
'rack/
initiation
0; a positive affine transformation.? This fact was noted
in Chapter 5 (page 167) for two-player games of chance; here, we see that it applies to all kinds of decision scenarios.
Value function Ordinal utility function
As
in game-playing, in a deterministic environment an agent needs only a preference
ranking on states—the numbers don’t matter.
This is called a value function or ordinal
utility function.
Itis important to remember that the existence ofa utility function that describes an agent’s
preference behavior does not necessarily mean that the agent is explicirly maximizing that utility function in its own deliberations. As we showed in Chapter 2, rational behavior can be
generated in any number of ways. A rational agent might be implemented with a table lookup
(if the number of possible states is small enough).
By observing a rational agent’s behavior, an observer can learn about the utility function
that represents what the agent is actually trying to achieve (even if the agent doesn’t know it). ‘We return to this point in Section 16.7.
16.3
Utility Functions
Utility functions map from lotteries to real numbers. We know they must obey the axioms
of orderability, transitivity, continuity, substitutability, monotonicity, and decomposability. Is
that all we can say about utility functions? Strictly speaking, that is it: an agent can have any preferences it likes. For example, an agent might prefer to have a prime number of dollars in its bank account; in which case, if it had $16 it would give away $3. This might be unusual,
but we can’t call it irrational. An agent might prefer a dented 1973 Ford Pinto to a shiny new
Mercedes. The agent might prefer prime numbers of dollars only when it owns the Pinto, but when it owns the Mercedes, it might prefer more dollars to fewer. Fortunately, the preferences
of real agents are usually more systematic and thus easier to deal with.
3 Inthis nse, utilities resemble temperatures: a temperature in Fahrenheit 1.8 times the Cel plus3 . but converting from one to the other doesn’t make you hotter or colder.
s temperature
Section 163
Utility Functions
533
y assessment and utility scales If we want to build a decision-theoretic system that helps a human make decisions or acts on his or her behalf, we must first work out what the human’s utility function is. This process,
often called preference elicitation, involves presenting choices to the human and using the
observed preferences to pin down the underlying utility function.
Preference elicitation
Equation (16.2) says that there is no absolute scale for utilities, but it is helpful, nonethe-
less, to establish some scale on which utilities can be recorded and compared for any particu-
lar problem. A scale can be established by fixing the utilities of any two particular outcomes,
just as we fix a temperature scale by fixing the freezing point and boiling point of water. Typically, we fix the utility of a “best possible prize” at U(S) = ur and a “worst possible
catastrophe” at U (S) = u,. (Both of these should be finite.) Normalized utilities use a scale
with u; =0 and ur = 1. With such a scale, an England fan might assign a utility of 1 to England winning the World Cup and a utility of 0 to England failing to qualify.
Normalized utilties
Given a utility scale between ut and u , we can assess the utility of any particular prize
S by asking the agent to choose between S and a standard lottery [p,ur; (1 — p),u.]. The Standard lottery probability p is adjusted until the agent is indifferent between S and the standard lottery.
Assuming normalized utilities, the utility ofS is given by p. Once this is done for each prize, the utilities for all lotteries involving those prizes are determined.
Suppose, for example,
we want to know how much our England fan values the outcome of England reaching the
semi-final and then losing. We compare that outcome to a standard lottery with probability p
of winning the trophy and probability 1 — p of an ignominious failure to qualify. If there is indifference at p=0.3, then 0.3 is the value of reaching the semi-final and then losing.
In medical, transportation, environmental and other decision problems, people’s lives are
at stake. (Yes, there are things more important than England’s fortunes in the World Cup.) In such cases, u, is the value assigned to immediate death (or in the really worst cases, many deaths).
Although nobody feels
comfortable with putting a value on human life, it is a fact
that tradeoffs on matters of life and death are made all the time. Aircraft are given a complete
overhaul at intervals, rather than after every trip. Cars are manufactured in a way that trades off costs against accident survival rates.
We tolerate a level of air pollution that kills four
million people a year. Paradoxically, a refusal to put a monetary value on life can mean that life is undervalued. Ross Shachter describes a government agency that commissioned a study on removing asbestos from schools. The decision analysts performing the study assumed a particular dollar value for the life of a school-age child, and argued that the rational choice under that assump-
tion was to remove the asbestos. The agency, morally outraged at the idea of setting the value
of a life, rejected the report out of hand. It then decided against asbestos removal—implicitly asserting a lower value for the life of a child than that assigned by the analysts.
Currently several agencies of the U.S. government, including the Environmental Protec-
tion Agency, the Food and Drug Administration, and the Department of Transportation, use the value of a statistical life to determine the costs and benefits of regulations and interven-
Value of a statistical
One common “currency” used in medical and safety analysis is the micromort, a one in a
Micromort
tions. Typical values in 2019 are roughly $10 million. Some attempts have been made to find out the value that people place on their own lives. million chance of death. If you ask people how much they would pay to avoid a risk—for
534
Chapter 16 Making Simple Decisions example, to avoid playing Russian roulette with a million-barreled revolver—they will respond with very large numbers, perhaps tens of thousands of dollars, but their actual behavior reflects a much lower monetary value for a micromort. For example, in the UK, driving in a car for 230 miles incurs a risk of one micromort.
Over the life of your car—say, 92,000 miles—that’s 400 micromorts.
People appear to be
willing to pay about $12,000 more for a safer car that halves the risk of death.
Thus, their
car-buying action says they have a value of $60 per micromort. A number of studies have confirmed a figure in this range across many individuals and risk types. However, government
agencies such as the U.S. Department of Transportation typically set a lower figure; they will
QALY
spend only about $6 in road repairs per expected life saved. Of course, these calculations hold only for small risks. Most people won’t agree to kill themselves, even for $60 million. Another measure is the QALY or quality-adjusted life year. Patients are willing to accept a shorter life expectancy to avoid a disability.
For example, kidney patients on average are
indifferent between living two years on dialysis and one year at full health. 16.3.2
The uti
y of money
Utility theory has its roots in economics, and economics provides one obvious candidate for a utility measure:
money (or more specifically, an agent’s total net assets). The almost
universal exchangeability of money for all kinds of goods and services suggests that money plays a significant role in human utility functions.
Monotonic preference
Tt will usually be the case that an agent prefers more money to less, all other things being
equal.
We say that the agent exhibits a monotonic preference for more money.
This does
not mean that money behaves as a utility function, because it says nothing about preferences
between lotteries involving money.
Suppose you have triumphed over the other competitors in a television game show. The
host now offers you a choice: either you can take the $1,000,000 prize or you can gamble it
Expected monetary value
on the flip of a coin. If the coin comes up heads, you end up with nothing, but if it comes up tails, you get $2,500,000. If you're like most people, you would decline the gamble and pocket the million. Are you being irrational?
Assuming the coin is fair, the expected monetary value (EMV) of the gamble is %($0) +
%(52.500,000) = §$1,250,000, which is more than the original $1,000,000. But that does not necessarily mean that accepting the gamble is a better decision. Suppose we use S, to denote
the state of possessing total wealth $n, and that your current wealth is $k. Then the expected utilities of the two actions of accepting and declining the gamble are
EU(Accept) = U(Sk U (Sk12:5 )+ 00.000) EU(Decline) = U(Sk41,000000) To determine what to do, we need to assign utilities to the outcome states.
Utility is not
directly proportional to monetary value, because the utility for your first million is very high (or so they say), whereas the utility for an additional million is smaller. Suppose you assign a utility of 5 to your current financial status (S), a 9 to the state Sk+2.500.000 and an 8 to the. state Si+1.000.000- Then the rational action would be to decline, because the expected utility of accepting is only 7 (less than the 8 for declining). On the other hand, a billionaire would most likely have a utility function that is locally linear over the range of a few million more,
and thus would accept the gamble.
Section 16,3 Utility Functions U
150000
535
U
800000
-3
(a)
-5
(b)
Figure 16.2 The utility of money. (a) Empirical data for Mr. Beard over a limited range. (b) A typical curve for the full range. In a pioneering study of actual utility functions, Grayson (1960) found that the utility
of money was almost exactly proportional to the logarithm of the amount. (This idea was
first suggested by Bernoulli (1738); see Exercise 16.STPT.) One particular utility curve, for a certain Mr. Beard, is shown in Figure 16.2(a). The data obtained for Mr. Beard’s preferences are consistent with a utility function
U(Sksn) = —263.31 422.091og(n+ 150,000)
for the range between n = —$150,000 and n = $800, 000.
‘We should not assume that this is the definitive utility function for monetary value, but
itis likely that most people have a utility function that is concave for positive wealth. Going into debt is bad, but preferences between different levels of debt can display a reversal of the concavity associated with positive wealth. For example, someone already $10,000,000 in debt might well accept a gamble on a fair coin with a gain of $10,000,000 for heads and a Toss of $20,000,000 for tails.* This yields the S-shaped curve shown in Figure 16.2(b). If we restrict our attention to the positive part of the curves, where the slope is decreasing, then for any lottery L, the utility of being faced with that lottery is less than the utility of being handed the expected monetary value of the lottery as a sure thing:
U(L) 0, then life is
positively enjoyable and the agent avoids borh exits. As long as the actions in (4,1), (3,2), and (3,3) are as shown, every policy is optimal, and the agent obtains infinite total reward
because it never enters a terminal state. It turns out that there are nine optimal policies in all
for various ranges of r; Exercise 17.THRC asks you to find them.
The introduction of uncertainty brings MDPs closer to the real world than deterministic search problems. For this reason, MDPs have been studied in several fields, including Al,
operations research, economics, and control theory. Dozens of solution algorithms have been proposed, several of which we discuss in Section 17.2. First, however, we spell out in more detail the definitions of utilities, optimal policies, and models for MDPs.
17.1.1
Utilities over time
In the MDP example in Figure 17.1, the performance of the agent was measured by a sum of rewards for the transitions experienced. This choice of performance measure is not arbitrary,
but it is not the only possibility for the utility function> on environment histories, which we
write as Up([s0,a0,51,a1 .., ,)).
2 In this chapter we use U for the utility function (to be consistent with the rest of the book), but many works about MDPs use V (for value) instead.
Section 17.1
Sequential Decision Problems
565
&1
|
o
T
Geeh
—0.0274 0. It is easy to see, from the definition of utilities as discounted sums of rewards, that a similar transformation of rewards will leave the optimal policy unchanged in an MDP: R(s,a,5) = mR(s,a,5') +b. It turns out, however, that the additive reward decomposition of utilities leads to significantly
more freedom in defining rewards. Let &(s) be any function of the state s. Then, according to the shaping theorem, the following transformation leaves the optimal policy unchanged:
R(s,a,5') = R(s,a,8') +7D(s') — B(s) .
(17.9)
To show that this is true, we need to prove that two MDPs, M and M’, have identical optimal
policies as long as they differ only in their reward functions as specified in Equation (17.9). We start from the Bellman equation for Q, the Q-function for MDP M:
0(s.0) = L) P | s.0)[R(s.a.8) + 7 maxs Qs )]
Now let Q'(s,a)=Q(s,a) — ®(s) and plug it into this equation; we get
Q'(s.0) +®(s) = Y. P(s | 5.0)[R(s.a.5') +7 myX(Q'(:'st/) +@(s))]. which then simplifies to
Q50) = LPE15.0)[RG.0,5) +1() ~B(s) +7 max ()] LPE |5 @)R (s.a.8) +7 max @/ ()]
Shaping theorem
570
Chapter 17 Making Complex Decisions In other words, Q'(s,a) satisfies the Bellman equation for MDP M. Now we can extract the optimal policy for M’ using Equation (17.7):
i (s) = argmax Q' (s, a) = argmax O(s,@) — B(s) = argmax Q(s,a) = wj(s) a
a
a
The function ®(s) is often called a potential, by analogy to the electrical potential (volt-
age) that gives rise to electric fields. The term 7(s') — d(s) functions as a gradient of the
potential. Thus, if ®(s) has higher value in states that have higher utility, the addition of ~YP(s') — D(s) to the reward has the effect of leading the agent “uphill” in utility. At first sight, it may seem rather counterintuitive that we can modify the reward in this
way without changing the optimal policy. It helps if we remember that all policies are optimal with a reward function that is zero everywhere. This means, according to the shaping theorem,
that all policies are optimal for any potential-based reward of the form R(s,a,s') = y®(s') — @(s).
Intuitively, this is because with such a reward it doesn’t matter which way the agent
goes from A to B. (This is easiest to see when v=1: along any path the sum of rewards collapses to d(B) —D(A), so all paths are equally good.) So adding a potential-based reward to any other reward shouldn’t change the optimal policy.
The flexibility afforded by the shaping theorem means that we can actually help out the
agent by making the immediate reward more directly reflect what the agent should do. In fact, if we set @(s) =U(s), then the greedy policy 7 with respect to the modified reward R’
is also an optimal policy:
7G(s) = argmax ) P(s'|5,a)R (s,a,5)
= argmax Y P(s'| 5,a) [R(s,a,5') +7(s') — ®(s)] . & = argmax Y P(s 5,0)[R(s,a,5') + U (s') = U (s)) a
g
i
I
= argmax )"
P(s'|5,a)[R(s,a,s') + U (s)]
(by Equation (17.4)).
Of course, in order to set (s) =U (s), we would need to know U (s); so there is no free lunch, but there is still considerable value in defining a reward function that is helpful to the extent
possible.
This is precisely what animal trainers do when they provide a small treat to the
animal for each step in the target sequence.
17.1.4
Representing MDPs
The simplest way to represent P(s'|s,a) and R(s,a,s') is with big, three-dimensional tables of size [S|?|A|. This is fine for small problems such as the 4 x 3 world, for which the tables
have 112 x 4=484 entries each. In some cases, the tables are sparse—most entries are zero
because each state s can transition to only a bounded number of states s'—which means the
tables are of size O(|S]|A]). For larger problems, even sparse tables are far too big.
Dynamic decision
Just as in Chapter 16, where Bayesian networks were extended with action and utility nodes to create decision networks, we can represent MDPs by extending dynamic Bayesian networks (DBNS, see Chapter 14) with decision, reward, and utility nodes to create dynamic
decision networks, or DDNs. DDN are factored representations in the terminology of
Section 17.1
[ Prag/nplus,
Sequential Decision Problems
Plug/Unplug.,| LeftWheel,,,
Figure 17.4 A dynamic decision network for a mobile robot with state variables for battery level, charging status, location, and velocity, and action variables for the left and right wheel motors and for charging. Chapter 2; they typically have an exponential complexity advantage over atomic representations and can model quite substantial real-world problems.
Figure 17.4, which is based on the DBN in Figure 14.13(b) (page 486), shows some
elements of a slightly realistic model for a mobile robot that can charge itself. The state S; is
decomposed into four state variables: + X, consists of the two-dimensional location on a grid plus the orientation;
« X, is the rate of change of X;;
« Charging, is true when the robot is plugged in to a power source; « Battery, is the battery level, which we model as an integer in the range 0, The state space for the MDP is the Cartesian product of the ranges of these four variables. The
action is now a set A, of action variables, comprised of Plug/Unplug, which has three values (plug, unplug, and noop); LefitWheel for the power sent to the left wheel; and RightWheel for
the power sent to the right wheel. The set of actions for the MDP is the Cartesian product of
the ranges of these three variables. Notice that each action variable affects only a subset of the state variables. The overall transition model is the conditional distribution P(X,|X;,A,), which can be
computed as a product of conditional probabilities from the DDN. The reward here is a single variable that depends only on the location X (for, say, arriving at a destination) and Charging, as the robot has to pay for electricity used; in this particular model, the reward doesn’t depend
on the action or the outcome state.
The network in Figure 17.4 has been projected three steps into the future. Notice that the
network includes nodes for the rewards for t, 1 + 1, and 1 + 2, but the wtility for 1 + 3. This
571
572
Chapter 17 Making Complex Decisions Next
F
&
o
A,
A1
CurrentPiece;
CurrentPiece
Filled,
Filledyy
(a)
(b)
Figure 17.5 (a) The game of Tetris. The T-shaped piece at the top center can be dropped in any orientation and in any horizontal position. Ifa row is completed, that row disappears and the rows above it move down, and the agent receives one point. The next piece (here, the L-shaped piece at top right) becomes the current piece, and a new next piece appears, chosen at random from the seven piece types. The game ends if the board fills up to the top. (b) The DDN for the Tetris MDP. is because the agent must maximize the (discounted) sum of all future rewards, and U (X;3) represents the reward for all rewards from ¢ + 3 onwards. If a heuristic approximation to U is available, it can be included in the MDP representation in this way and used in lieu of further expansion. This approach is closely related to the use of bounded-depth search and heuristic evaluation functions for games in Chapter 5.
Another interesting and well-studied MDP is the game of Tetris (Figure 17.5(a)). The
state variables for the game are the CurrentPiece, the NextPiece, and a bit-vector-valued
variable Filled with one bit for each of the 10 x 20 board locations. Thus, the state space has
7 %7 % 22% 2 10
states. The DDN for Tetris is shown in Figure 17.5(b). Note that Filled,
is a deterministic function of Filled, and A,. It turns out that every policy for Tetris is proper (reaches a terminal state): eventually the board fills despite one’s best efforts to empty it.
17.2
Algorithms for MDPs
In this section, we present four different algorithms for solving MDPs. The first three, value
Monte Carlo planning
iteration, policy iteration, and linear programming, generate exact solutions offline. The fourth is a family of online approximate algorithms that includes Monte Carlo planning.
17.2.1 Value lteration
Value iteration
‘The Bellman equation (Equation (17.5)) s the basis of the value iteration algorithm for solving MDPs. If there are n possible states, then there are n Bellman equations, one for cach
Section 17.2
Algorithms for MDPs
573
function VALUE-ITERATION(mdp, €) returnsa utility function
inputs: mdp, an MDP with states S, actions A(s), transition model P(s’|s,a),
rewards R(s,a, "), discount ¢, the maximum error allowed in the utility of any state
local variables: U, U’, vectors of utilities for states in S, initially zero
4, the maximum relative change in the utility of any state
repeat
VU360 for each state s in S do U'ls] max,c o5y Q-VALUE(mdp, s,a. U) if |U[s] — Uls]| > & then 5 |U'[s] — U[s]| until§ < e(1-7)/7 return U Figure 17.6 The value iteration algorithm for calculating tilities of states. The termination condition is from Equation (17.12). state. The n equations contain n unknowns—the utilities of the
states. So we would like to
solve these simultaneous equations to find the utilities. There is one problem: the equations
are nonlinear, because the “max” operator is not a linear operator. Whereas systems of linear equations can be solved quickly using linear algebra techniques, systems of nonlinear equations are more problematic. One thing to try is an iterative approach. We start with arbitrary
initial values for the utilities, calculate the right-hand side of the equation, and plug it into the
left-hand side—thereby updating the utility of each state from the utilities of its neighbors.
We repeat this until we reach an equilibrium.
Let U;(s) be the utility value for state s at the ith iteration.
Bellman update, looks like this:
The iteration step, called a
Ui (s) = aeA(s) max Y7 P(s'|5,a)[R(s,a.s) + v Uls')],
(17.10)
where the update is assumed to be applied simultaneously to all the states at each iteration. If we apply the Bellman update infinitely often, we are guaranteed to reach an equilibrium
(see “convergence of value iteration” below), in which case the final utility values must be
solutions to the Bellman equations. In fact, they are also the unigue solutions, and the corre-
sponding policy (obtained using Equation (17.4)) is optimal. The detailed algorithm, including a termination condition when the utilities are “close enough,” is shown in Figure 17.6.
Notice that we make use of the Q-VALUE function defined on page 569. ‘We can apply value iteration to the 4x 3 world in Figure 17.1(a).
Starting with initial
values of zero, the utilities evolve as shown in Figure 17.7(a). Notice how the states at different distances from (4,3) accumulate negative reward until a path is found to (4,3), whereupon
the utilities start to increase.
We can think of the value iteration algorithm as propagating
information through the state space by means of local updates.
Bellman update
574
Chapter 17 Making Complex Decisions 1x107
558 g S
Iierations required
1x10°
0
5
10 15 20 25 30 Number of terations (a)
35
40
05
06
07 08 09 Discount factorY ®)
1
Figure 17.7 (a) Graph showing the evolution of the utilities of selected states using value iteration. (b) The number of value iterations required to guarantee an error of at most e =c-
Rumax, for different values of ¢, as a function of the discount factor 7. Convergence of value iteration
We said that value iteration eventually converges to a unique set of solutions of the Bellman
equations. In this section, we explain why this happens. We introduce some useful mathematical ideas along the way, and we obtain some methods for assessing the error in the utility function returned when the algorithm is terminated early; this is useful because it means that we don’t have to run forever. This section is quite technical.
Contraction
The basic concept used in showing that value iteration converges is the notion of a con-
traction. Roughly speaking, a contraction is a function of one argument that, when applied to two different inputs in turn, produces two output values that are “closer together.” by at least some constant factor, than the original inputs. For example, the function “divide by two™ is
a contraction, because, after we divide any two numbers by two, their difference is halved.
Notice that the “divide by two” function has a fixed point, namely zero, that is unchanged by
the application of the function. From this example, we can discern two important properties of contractions:
+ A contraction has only one fixed point; if there were two fixed points they would not
get closer together when the function was applied, so it would not be a contraction. « When the function is applied to any argument, the value must get closer to the fixed point (because the fixed point does not move), so repeated application of a contraction
always reaches the fixed point in the limit. Now, suppose we view the Bellman update (Equation (17.10)) as an operator B that is applied simultaneously to update the utility of every state. Then the Bellman equation becomes
U=BU and the Bellman update equation can be written as Uit1
Max norm
0.5 and Go otherwise. Once we have utilities c,(s) for all the conditional plans p of depth 1 in each physical state s, we can compute the utilities for conditional plans of depth 2 by considering each
possible first action, each possible subsequent percept, and then each way of choosing a
depth-1 plan to execute for each percept: [Stay; if Percept=A then Stay else Stay] [Stay; if Percept =A then Stay else Go|
[Go; if Percept=A then Stay else Stay]
Dominated plan
There are eight distinct depth-2 plans in all, and their utilities are shown in Figure 17.15(b). Notice that four of the plans, shown as dashed lines, are suboptimal across the entire belief space—we say these plans are dominated, and they need not be considered further. There are four undominated plans, cach of which is optimal in a specific region, as shown in Figure 17.15(c). The regions partition the belief-state space. We repeat the process for depth 3, and so on. In general, let p be a depth-d conditional plan whose initial action is @ and whose depth-(d — 1) subplan for percept e is p.e; then (17.18) ap(s) = );P(!\smlkts,a,:’) +7 L P(e|s)ape(s))This recursion naturally gives us a value iteration algorithm, which is given in Figure 17.16. The structure of the algorithm and its error analysis are similar to those of the basic value
iteration algorithm in Figure 17.6 on page 573; the main difference is that instead of computing one utility number for each state, POMDP-VALUE-ITERATION
maintains a collection of
undominated plans with their utility hyperplanes.
The algorithm’s complexity depends primarily on how many plans get generated. Given |A] actions and |E| possible observations, there are |A|?(£1"") distinct depth-d plans. Even for the lowly two-state world with d=8, that’s 2255 plans. The elimination of dominated plans is essential for reducing this doubly exponential growth: the number of undominated plans with d =8 is just 144. The utility function for these 144 plans is shown in Figure 17.15(d). Notice that the intermediate belief states have lower value than state A and state B, because in the intermediate states the agent lacks the information needed to choose a good action. This is why information has value in the sense defined in Section 16.6 and optimal policies in POMDPs often include information-gathering actions.
Section 17.5
Algorithms for Solving POMDPs
function POMDP-VALUE-ITERATION(pomdp, ¢) returns a utility function inputs: pomdp, a POMDP with states S, actions A(s). transition model P(s'|s.a). sensor model P(e|s). rewards R(s), discount y ¢, the maximum error allowed in the utility of any state local variables: U, U, sets of plans p with associated utility vectors o, U’ ¢ a set containing just the empty plan []. with a)(s) = R(s) repeat [
U’ the set of all plans consisting of an action and, for each possible next percept, aplanin U with utility vectors computed according to Equation (17.18) U’ REMOVE-DOMINATED-PLANS(U")
until MAX-DIFFERENCE(U,U’) return U
< e(1-7)/y
Figure 17.16 A high-level sketch of the value iteration algorithm for POMDPs. The REMOVE-DOMINATED-PLANS step and MAX-DIFFERENCE test are typically implemented as linear programs. Given such a utility function, an executable policy can be extracted by looking at which hyperplane is optimal at any given belief state b and executing the first action of the corresponding plan. In Figure 17.15(d), the corresponding optimal policy is still the same as for depth-1 plans: Stay when b(B) > 0.5 and Go otherwise. In practice, the value iteration algorithm in Figure 17.16 is hopelessly inefficient for larger problems—even the 4 x 3 POMDP is too hard. The main reason is that given n undominated
conditional plans at level d, the algorithm constructs |A|-nlE| conditional plans at level d + 1 before eliminating the dominated ones. With the four-bit sensor, |E| is 16, and n can be in the hundreds, so this is hopeless. Since this algorithm was developed in the 1970s, there have been several advances, in-
cluding more efficient forms of value iteration and various kinds of policy iteration algorithms.
Some
of these are discussed in the notes at the end of the chapter.
For general
POMDPs, however, finding optimal policies is very difficult (PSPACE-hard, in fact—that is, very hard indeed). The next section describes a different, approximate method for solving POMDPs, one based on look-ahead search.
17.5.2
Online algorithms for POMDPs
The basic design for an online POMDP agent is straightforward: it starts with some prior belief state; it chooses an action based on some deliberation process centered on its current
belief state; after acting, it receives an observation and updates its belief state using a filtering algorithm; and the process repeats.
One obvious choice for the deliberation process is the expectimax algorithm from Sec-
tion 17.2.4, except with belief states rather than physical states as the decision nodes in the tree. The chance nodes in the POMDP
tree have branches labeled by possible observations
and leading to the next belief state, with transition probabilities given by Equation (17.17). A fragment of the belief-state expectimax tree for the 4 x 3 POMDP is shown in Figure 17.17.
593
594
Chapter 17 Making Complex Decisions
Up.
1100
Righ,
o110/
1100
Down
Left
1010
0110
Figure 17.17 Part of an expectimax tree for the 4 x 3 POMDP with a uniform initial belief state. The belief states are depicted with shading proportional to the probability of being in each location. The time complexity of an exhaustive search to depth d is O(JA|" - [E[Y), where |A| is
the number of available actions and |E| is the number of possible percepts.
(Notice that
this is far less than the number of possible depth-d conditional plans generated by value iteration.)
As in the observable case, sampling at the chance nodes is a good way to cut
down the branching factor without losing too much accuracy in the final decision. Thus, the
complexity of approximate online decision making in POMDPs may not be drastically worse than that in MDPs.
For very large state spaces, exact filtering is infeasible, so the agent will need to run
an approximate filtering algorithm such as particle filtering (see page 492). Then the belief
POMCP
states in the expectimax tree become collections of particles rather than exact probability distributions. For problems with long horizons, we may also need to run the kind of long-range playouts used in the UCT algorithm (Figure 5.11). The combination of particle filtering and UCT applied to POMDPs goes under the name of partially observable Monte Carlo planning
or POMCP.
With a DDN
representation for the model, the POMCP
algorithm is, at least
in principle, applicable to very large and realistic POMDPs. Details of the algorithm are explored in Exercise 17.PoMC. POMCP is capable of generating competent behavior in the
4 x 3 POMDP. A short (and somewhat fortunate) example is shown in Figure 17.18. POMDP agents based on dynamic decision networks and online decision making have a
number of advantages compared with other, simpler agent designs presented in earlier chap-
ters. In particular, they handle partially observable, stochastic environments and can easily revise their “plans” to handle unexpected evidence. With appropriate sensor models, they can handle sensor failure and can plan to gather information. They exhibit “graceful degradation” under time pressure and in complex environments, using various approximation techniques.
So what is missing? The principal obstacle to real-world deployment of such agents is
the inability to generate successful behavior over long time-scales. Random or near-random
playouts have no hope of gaining any positive reward on, say, the task of laying the table
Summary
B 1000 111 0001
1010
1001
Figure 17.18 A sequence of percepts, belief states, and actions in the 4 x 3 POMDP with a wall-sensing error of ¢=0.2. Notice how the early Lefr moves are safe—they are very unlikely to fall into (4,2)—and coerce the agent’s location into a small number of possible locations. After moving Up, the agent thinks it is probably in (3,3). but possibly in (1.3). Fortunately, moving Right is a good idea in both cases, so it moves Right, finds out that it had been in (1,3) and is now in (2,3), and then continues moving Right and reaches the goal. for dinner, which might take tens of millions of motor-control actions.
It seems necessary
to borrow some of the hierarchical planning ideas described in Section 11.4. At the time of
writing, there are not yet satisfactory and efficient ways to apply these ideas in stochastic,
partially observable environments.
Summary
This chapter shows how to use knowledge about the world to make decisions even when the outcomes of an action are uncertain and the rewards for acting might not be reaped until many actions have passed. The main points are as follows: + Sequential decision problems in stochastic environments, also called Markov decision
processes, or MDPs, are defined by a transition model specifying the probabilistic
outcomes of actions and a reward function specifying the reward in each state. « The utility of a state sequence is the sum of all the rewards over the sequence, possibly discounted over time.
The solution of an MDP is a policy that associates a decision
with every state that the agent might reach. An optimal policy maximizes the utility of
the state sequences encountered when it is executed. + The utility of a state is the expected sum of rewards when an optimal policy is executed
from that state. The value iteration algorithm iteratively solves a set of equations
relating the utility of each state to those of its neighbors. + Policy iteration alternates between calculating the utilities of states under the current
policy and improving the current policy with respect to the current utilities.
* Partially observable MDPs, or POMDPs, are much more difficult to solve than are MDPs. They can be solved by conversion to an MDP in the continuous space of belief
states; both value iteration and policy iteration algorithms have been devised. Optimal behavior in POMDPs
includes information gathering to reduce uncertainty and there-
fore make better decisions in the future.
* A decision-theoretic agent can be constructed for POMDP environments.
The agent
uses a dynamic decision network to represent the transition and sensor models, to update its belief state, and to project forward possible action sequences.
We shall return MDPs and POMDPs in Chapter 22, which covers reinforcement learning methods that allow an agent to improve its behavior from experience.
595
596
Chapter 17 Making Complex Decisions Bibliographical and Historical Notes
Richard Bellman developed the ideas underlying the modern approach to sequential decision problems while working at the RAND Corporation beginning in 1949. According to his autobiography (Bellman, 1984), he coined the term “dynamic programming” to hide from a research-phobic Secretary of Defense, Charles Wilson, the fact that his group was doing mathematics. (This cannot be strictly true, because his first paper using the term (Bellman,
1952) appeared before Wilson became Secretary of Defense in 1953.) Bellman’s book, Dynamic Programming (1957), gave the new field a solid foundation and introduced the value iteration algorithm.
Shapley (1953b) actually described the value iteration algorithm independently of Bellman, but his results were not widely appreciated in the operations research community, perhaps because they were presented in the more general context of Markov games. Although the original formulations included discounting, its analysis in terms of stationary preferences was suggested by Koopmans (1972). The shaping theorem is due to Ng ef al. (1999). Ron Howard’s Ph.D. thesis (1960) introduced policy iteration and the idea of average reward for solving infinite-horizon problems. Several additional results were introduced
by Bellman and Dreyfus (1962). The use of contraction mappings in analyzing dynamic programming algorithms is due to Denardo (1967). Modified policy iteration is due to van Nunen (1976) and Puterman and Shin (1978). Asynchronous policy iteration was analyzed by Williams and Baird (1993), who also proved the policy loss bound in Equation (17.13). The general family of prioritized sweeping algorithms aims to speed up convergence to optimal policies by heuristically ordering the value and policy update calculations (Moore and Atkeson, 1993; Andre ez al., 1998; Wingate and Seppi, 2005).
The formulation of MDP-solving as a linear program is due to de Ghellinck (1960), Manne (1960), and D’Epenoux (1963). Although linear programming has traditionally been considered inferior to dynamic programming as an exact solution method for MDPs, de Farias and Roy (2003) show that it is possible to use linear programming and a linear representation of the utility function to obtain provably good approximate solutions to very large MDPs.
Papadimitriou and Tsitsiklis (1987) and Littman er al. (1995) provide general results on the
computational complexity of MDPs.
Yinyu Ye (2011) analyzes the relationship between
policy iteration and the simplex method for linear programming and proves that for fixed v,
the runtime of policy iteration is polynomial in the number of states and actions.
Seminal work by Sutton (1988) and Watkins (1989) on reinforcement learning methods
for solving MDPs played a significant role in introducing MDPs into the Al community. (Earlier work by Werbos (1977) contained many similar ideas, but was not taken up to the same extent.) Al researchers have pushed MDPs in the direction of more expressive representations that can accommodate much larger problems than the traditional atomic representations based on transition matrices.
The basic ideas for an agent architecture using dynamic decision networks were proposed by Dean and Kanazawa (1989a). Tatman and Shachter (1990) showed how to apply dynamic programming algorithms to DDN models. Several authors made the connection between MDPs and Al planning problems, developing probabilistic forms of the compact STRIPS representation for transition models (Wellman,
1990b; Koenig,
1991).
The book Planning
and Control by Dean and Wellman (1991) explores the connection in great depth.
Bibliographical and Historical Notes Later work on factored MDPs (Boutilier ez al., 2000; Koller and Parr, 2000; Guestrin et al., 2003b) uses structured representations of the value function as well as the transition model, with provable improvements in complexity. Relational MDPs (Boutilier et al., 2001;
Guestrin e al., 2003a) go one step further, using structured representations to handle domains
with many related objects. Open-universe MDPs and POMDPs (Srivastava et al., 2014b) also allow for uncertainty over the existence and identity of objects and actions.
Many authors have developed approximate online algorithms for decision making in MDPs, often borrowing explicitly from earlier AT approaches to real-time search and gameplaying (Werbos, 1992; Dean ez al., 1993; Tash and Russell, 1994). The work of Barto et al.
(1995) on RTDP (real-time dynamic programming) provided a general framework for understanding such algorithms and their connection to reinforcement learning and heuristic search.
The analysis of depth-bounded expectimax with sampling at chance nodes is due to Kearns
et al. (2002). The UCT algorithm described in the chapter is due to Kocsis and Szepesvari (2006) and borrows from earlier work on random playouts for estimating the values of states (Abramson, 1990; Briigmann, 1993; Chang er al., 2005).
Bandit problems were introduced by Thompson (1933) but came to prominence after
‘World War II through the work of Herbert Robbins (1952).
Bradt et al. (1956) proved the
first results concerning stopping rules for one-armed bandits, which led eventually to the
breakthrough results of John Gittins (Gittins and Jones, 1974; Gittins, 1989). Katehakis and
Veinott (1987) suggested the restart MDP as a method of computing Gittins indices. The text by Berry and Fristedt (1985) covers many variations on the basic problem, while the pellucid online text by Ferguson (2001) connects bandit problems with stopping problems. Lai and Robbins (1985) initiated the study of the asymptotic regret of optimal bandit
policies. The UCB heuristic was introduced and analyzed by Auer et al. (2002).
Bandit su-
perprocesses (BSPs) were first studied by Nash (1973) but have remained largely unknown
in AL Hadfield-Menell and Russell (2015) describe an efficient branch-and-bound algorithm
capable of solving relatively large BSPs. Selection problems were introduced by Bechhofer (1954). Hay et al. (2012) developed a formal framework for metareasoning problems, showing that simple instances mapped to selection rather than bandit problems. They also proved
the satisfying result that expected computation cost of the optimal computational strategy is
never higher than the expected gain in decision quality—although there are cases where the optimal policy may, with some probability, keep computing long past the point where any
possible gain has been used up. The observation that a partially observable MDP can be transformed into a regular MDP over belief states is due to Astrom (1965) and Aoki (1965).
The first complete algorithm
for the exact solution of POMDPs—essentially the value iteration algorithm presented in
this chapter—was proposed by Edward Sondik (1971) in his Ph.D. thesis. (A later journal paper by Smallwood and Sondik (1973) contains some errors, but is more accessible.)
Lovejoy (1991) surveyed the first twenty-five years of POMDP research, reaching somewhat pessimistic conclusions about the feasibility of solving large problems.
The first significant contribution within Al was the Witness algorithm (Cassandra et al., 1994; Kaelbling er al., 1998), an improved version of POMDP value iteration. Other algo-
rithms soon followed, including an approach due to Hansen (1998) that constructs a policy
incrementally in the form of a finite-state automaton whose states define the possible belief
states of the agent.
597
Factored MDP Relational MDP.
598
Chapter 17 Making Complex Decisions More recent work in Al has focused on point-based value iteration methods that, at each
iteration, generate conditional plans and a-vectors for a finite set of belief states rather than
for the entire belief space. Lovejoy (1991) proposed such an algorithm for a fixed grid of points, an approach taken also by Bonet (2002). An influential paper by Pineau ef al. (2003) suggested generating reachable points by simulating trajectories in a somewhat greedy fash-
ion; Spaan and Vlassis (2005) observe that one need generate plans for only a small, randomly
selected subset of points to improve on the plans from the previous iteration for all points in
the set. Shani ef al. (2013) survey these and other developments in point-based algorithms,
which have led to good solutions for problems with thousands of states. Because POMDPs are PSPACE-hard (Papadimitriou and Tsitsiklis, 1987), further progress on offline solution
methods may require taking advantage of various kinds of structure in value functions arising from a factored representation of the model.
The online approach for POMDPs—using look-ahead search to select an action for the
current belief state—was first examined by Satia and Lave (1973).
The use of sampling at
chance nodes was explored analytically by Keamns ef al. (2000) and Ng and Jordan (2000). ‘The POMCP algorithm is due to Silver and Veness (2011).
With the development of reasonably effective approximation algorithms for POMDPs,
their use as models for real-world problems has increased, particularly in education (Rafferty et al., 2016), dialog systems (Young et al., 2013), robotics (Hsiao et al., 2007; Huynh and
Roy, 2009), and self-driving cars (Forbes et al., 1995; Bai et al., 2015). An important largescale application is the Airborne Collision Avoidance System X (ACAS X), which keeps airplanes and drones from colliding midair. The system uses POMDPs with neural networks
to do function approximation. ACAS X significantly improves safety compared to the legacy
TCAS system, which was built in the 1970s using expert system technology (Kochenderfer, 2015; Julian er al., 2018).
Complex decision making has also been studied by economists and psychologists. They find that decision makers are not always rational, and may not be operating exactly as described by the models in this chapter. For example, when given a choice, a majority of people prefer $100 today over a guarantee of $200 in two years, but those same people prefer $200 Hyperbolic reward
in eight years over $100 in six years. One way to interpret this result is that people are not
using additive exponentially discounted rewards; perhaps they are using hyperbolic rewards (the hyperbolic function dips more steeply in the near term than does the exponential decay function). This and other possible interpretations are discussed by Rubinstein (2003).
The texts by Bertsekas (1987) and Puterman (1994) provide rigorous introductions to sequential decision problems and dynamic programming. Bertsekas and Tsitsiklis (1996) include coverage of reinforcement learning. Sutton and Barto (2018) cover similar ground but in a more accessible style. Sigaud and Buffet (2010), Mausam and Kolobov (2012) and
Kochenderfer (2015) cover sequential decision making from an Al perspective. Krishnamurthy (2016) provides thorough coverage of POMDPs.
TG
18
MULTIAGENT DECISION MAKING In which we examine what to do when more than one agent inhabits the environment.
18.1
Properties of Multiagent
Environments
So far, we have largely assumed that only one agent has been doing the sensing, planning, and
acting. But this represents a huge simplifying assumption, which fails to capture many realworld Al settings.
In this chapter, therefore, we will consider the issues that arise when an
agent must make decisions in environments that contain multiple actors. Such environments
are called multiagent systems, and agents in such a system face a multiagent planning Multiagent systems problem. However, as we will see, the precise nature of the multiagent planning problem— }13{25" Plarnine and the techniques that are appropriate for solving it—will depend on the relationships among the agents in the environment.
18.1.1
One decision maker
The first possibility is that while the environment contains multiple actors, it contains only one decision maker. In such a case, the decision maker develops plans for the other agents, and tells them what to do.
The assumption that agents will simply do what they are told
agent is called the benevolent agent assumption. However, even in this setting, plans involving Benevolent assumption
multiple actors will require actors to synchronize their actions. Actors A and B will have to
act at the same time for joint actions (such as singing a duet), at different times for mutually exclusive actions (such as recharging batteries when there is only one plug), and sequentially
when one establishes a precondition for the other (such as A washing the dishes and then B
drying them).
One special case is where we have a single decision maker with multiple effectors that
can operate concurrently—for example, a human who can walk and talk at the same time.
Such an agent needs to do multieffector planning to manage each effector while handling =45 e/iec*>" positive and negative interactions among the effectors.
When the effectors are physically
decoupled into detached units—as in a fleet of delivery robots in a factory—multieffector
planning becomes multibody planning.
A multibody problem is still a “standard” single-agent problem as long as the relevant sensor information collected by each body can be pooled—either centrally or within each body—to
form a common
Multibody planning
estimate of the world state that then informs the execution of
the overall plan; in this case, the multiple bodies can be thought of as acting as a single body. When communication constraints make this impossible, we have what is sometimes
called a decentralized planning problem; this is perhaps a misnomer, because the planning - Jeente!= —3. « Now suppose we change the rules to force O to reveal his strategy first, followed by E. Then the minimax value of this game is Up ., and because this game favors E we know that U is at most Up &
know U < 42.
With pure strategies, the value is +2 (see Figure 18.2(b)), so we
Combining these two arguments, we see that the true utility U of the solution to the original game must satisfy
Upo v(C)+Vv(D)
forall CCDCN
If a game is superadditive, then the grand coalition receives a value that is at least as high
as or higher than the total received by any other coalition structure. However, as we will see
shortly, superadditive games do not always end up with a grand coalition, for much the same reason that the players do not always arrive at a collectively desirable Pareto-optimal outcome in the prisoner’s dilemma.
18.3.2
Strategy in cooperative games
The basic assumption in cooperative game theory is that players will make strategic decisions about who they will cooperate with. Intuitively, players will not desire to work with unproductive players—they will naturally seek out players that collectively yield a high coalitional
value. But these sought-after players will be doing their own strategic reasoning. Before we can describe this reasoning, we need some further definitions.
An imputation for a cooperative game (N,) is a payoff vector that satisfies the follow-
Imputation
ing two conditions:
2= v(N)
x> v({i}) forallie N. The first condition says that an imputation must distribute the total value of the grand coali-
tion; the second condition, known as individual rationality, says that each player is at least
Individual rationality
as well off as if it had worked alone.
Given an imputation X = (x,...,x,) and a coalition C C N, we define x(C) to be the sum
¥jecXi—the total amount disbursed to C by the imputation x.
Next, we define the core of a game (N, v) as the set of all imputations X that satisfy the
condition x(C) > v(C) for every possible coalition C C N. Thus, if an imputation x is not in the core, then there exists some coalition C C N such that v(C) > x(C). The players in C
would refuse to join the grand coalition because they would be better off sticking with C.
The core of a game therefore consists of all the possible payoff vectors that no coalition
could object to on the grounds that they could do better by not joining the grand coalition.
Thus, if the core is empty, then the grand coalition cannot form, because no matter how the
grand coalition divided its payoff, some smaller coalition would refuse to join. The main
computational questions around the core relate to whether or not it is empty, and whether a
particular payoff distribution is in the core.
Core
628
Chapter 18 Multiagent Decision Making
The definition of the core naturally leads to a system of linear inequalities, as follows (the unknowns are variables xi....,.x,, and the values v(C) are constants): x>
v({i})
forallie N
v(C)
foralCCN
Lienxi = V(N) Yiecxi
=
Any solution to these inequalities will define an imputation in the core. We can formulate the
inequalities as a linear program by using a dummy objective function (for example, maximizing ¥y x;), which will allow us to compute imputations in time polynomial in the number
of inequalities.
The difficulty is that this gives an exponential number of inequalities (one
for each of the 2" possible coalitions). Thus, this approach yields an algorithm for checking
non-emptiness of the core that runs in exponential time. Whether we can do better than this
depends on the game being studied: for many classes of cooperative game, the problem of checking non-emptiness of the core is co-NP-complete. We give an example below. Before proceeding, let’s see an example of a superadditive game with an empty core. The game has three players N = {1,2,3}, and has a characteristic function defined as follows:
[
=2
o) = { 0 otherwise.
Now consider any imputation (x;,x2,x3) for this game. Since v(N) = 1, it must be the case that at least one player i has x; > 0, and the other two get a total payoff less than 1. Those two could benefit by forming a coalition without player i and sharing the value 1 among themselves. But since this holds for all imputations, the core must be empty.
The core formalizes the idea of the grand coalition being stable, in the sense that no
coalition can profitably defect from it. However, the core may contain imputations that are unreasonable, in the sense that one or more players might feel they were unfair.
N = {1,2}, and we have a characteristic function v defined as follows:
Suppose
v({1}) =v({2}) =5
v({1,2}) =20.
Here, cooperation yields a surplus of 10 over what players could obtain working in isolation,
and so intuitively, cooperation will make sense in this scenario. Now, it is easy to see that the
imputation (6, 14) is in the core of this game: neither player can deviate to obtain a higher
uiility. But from the point of view of player 1, this might appear unreasonable, because it gives 9/10 of the surplus to player 2. Thus, the notion of the core tells us when a grand Shapley value
coalition can form, but it does not tell us how to distribute the payoff.
The Shapley value is an elegant proposal for how to divide the v(N) value among the players, given that the grand coalition N formed. Formulated by Nobel laureate Lloyd Shapley in the early 1950s, the Shapley value is intended to be a fair distribution scheme. What does fair mean?
It would be unfair to distribute v(N) based on the eye color of
players, or their gender, or skin color. Students often suggest that the value v(N) should be
divided equally, which seems like it might be fair, until we consider that this would give the
same reward to players that contribute a lot and players that contribute nothing. Shapley’s insight was to suggest that the only fair way to divide the value v(N) was to do so according
Mond |
to how much each player contributed to creating the value v(N).
First we need to define the notion of a player’s marginal contribution. The marginal
Section 18.3 contribution that a player i makes to a coalition C
Cooperative Game Theory
629
is the value that i would add (or remove),
should i join the coalition C. Formally, the marginal contribution that player i makes to C is
denoted by mc;(C):
me;(C) = v(Cu{i}) - v(C). Now, a first attempt to define a payoff division scheme in line with Shapley’s suggestion
that players should be rewarded according to their contribution would be to pay each playeri the value that they would add to the coalition containing all other players:
mei(N —{i}). The problem is that this implicitly assumes that player i is the last player to enter the coalition.
So, Shapley suggested, we need to consider all possible ways that the grand coalition could form, that is, all possible orderings of the players N, and consider the value that i adds to the preceding players in the ordering. Then, a player should be rewarded by being paid the average marginal contribution that player i makes, over all possible orderings of the players, 10 the set of players preceding i in the ordering.
‘We let P denote all possible permutations (e.g., orderings) of the players N, and denote
members of P by p,p',... etc. Where p € P and i € N, we denote by p; the set of players
that precede i in the ordering p.
Then the Shapley value for a game G is the imputation
3(G) = (61(G),...,¢a(G)) defined as follows: 1
4(G) = fip();mc.(p,).
(18.1)
This should convince you that the Shapley value is a reasonable proposal. But the remark-
able fact is that it is the unique solution to a set of axioms that characterizes a “fair” payoff distribution scheme. We’ll need some more definitions before defining the axioms.
‘We define a dummy player as a player i that never adds any value to a coalition—that is,
mc;(C) =0 for all C C N — {i}. We will say that two players i and j are symmetric players
if they always make identical contributions to coalitions—that is, mc;(C) = me;(C) for all
C CN—{i,j}. Finally, where G = (N,v) and G’ = (N, V') are games with the same set of players, the game G + G’ is the game with the same player set, and a characteristic function
V" defined by v/(C) = v(C) + V/(C).
Given these definitions, we can define the fairness axioms satisfied by the Shapley value:
« Efficiency: ¥y 6i(G) = v(N). (All the value should be distributed.) « Dummy Player: 1f i is a dummy player in G then ¢;(G) = 0. (Players who never contribute anything should never receive anything.)
« Symmetry: 1f i and j are symmetric in G then ¢;(G) = ¢,(G). (Players who make identical contributions should receive identical payoffs.)
« Additivity: The value is additive over games: For all games G = (N, v) and G’ = (N, V), and for all players i € N, we have ¢;(G +G') = ¢i(G) + ¢i(G').
The additivity axiom is admittedly rather technical. If we accept it as a requirement, however,
we can establish the following key property: the Shapley value is the only way to distribute
coalitional value so as to satisfy these fairness axioms.
Dummy player
Symmetric players
630
Chapter 18 Multiagent Decision Making 18.3.3
Computation
in cooperative games
From a theoretical point of view, we now have a satisfactory solution. But from a computa-
tional point of view, we need to know how to compactly represent cooperative games, and
how to efficiently compute solution concepts such as the core and the Shapley value.
The obvious representation for a characteristic function would be a table listing the value
v(C) for all 2" coalitions. This is infeasible for large n. A number of approaches to compactly representing cooperative games have been developed, which can be distinguished by whether or not they are complete. A complete representation scheme is one that is capable of representing any cooperative game. The drawback with complete representation schemes is that there will always be some games that cannot be represented compactly. An alternative is to use a representation scheme that is guaranteed to be compact, but which is not complete.
Marginal contribution nets
Marginal conibation net
‘We now describe one representation scheme, called marginal contribution nets (MC-nets).
We will use a slightly simplified version to facilitate presentation, and the simplification makes it incomplete—the full version of MC-nets is a complete representation.
The idea behind marginal contribution nets is to represent the characteristic function of a
game (N,v) as a set of coalition-value rules, of the form: (C;,x;), where C; C N is a coalition
and x; is a number. To compute the value of a coalition C, we simply sum the values of all rules (C;,x;) such that C; C C. Thus, given a set of rules R = {(C1,x1),...,(C,x¢)}, the corresponding characteristic function is: v(C€) = Y {xi | (Cixi) eRand G; C C} Suppose we have a rule set R containing the following three rules:
{({1.2}5), ({212, ({344} Then, for example, we have: « v({1}) = 0 (because no rules apply), v({3}) = 4 v({1,3}) = v({2,3}) = v({1,2,3})
(third rule), 4 (third rule), 6 (second and third rules), and = 11 (first, second, and third rules).
‘With this representation we can compute the Shapley value in polynomial time.
The key
insight is that each rule can be understood as defining a game on its own, in which the players
are symmetric. By appealing to Shapley’s axioms of additivity and symmetry, therefore, the Shapley value ¢;(R) of player i in the game associated with the rule set R is then simply:
aR=
T
iec . 5er |{ 0fr otherwise The version of marginal contribution nets that we have presented here is not a complete repre-
sentation scheme: there are games whose characteristic function cannot be represented using rule sets of the form described above. A richer type of marginal contribution networks al-
Tows for rules of the form (¢,x), where ¢ is a propositional logic formula over the players
N: a coalition C satisfies the condition ¢ if it corresponds to a satisfying assignment for ¢.
Section 18.3
Cooperative Game Theory
e 11,121,841
(s
!l
A= (eaen) (Buas)
631
ool (nwes)
level3.
©2590)
a1
Figure 187 The coalition structure graph for N = {1,2,3,4}. Level 1 has coalition structures containing a single coalition; level 2 has coalition structures containing two coalitions, and 0 on. This scheme is a complete representation—in the worst case, we need a rule for every possible coalition. Moreover, the Shapley value can be computed in polynomial time with this scheme; the details are more involved than for the simple rules described above, although the
basic principle is the same; see the notes at the end of the chapter for references.
Coalition structures for maximum
social welfare
‘We obtain a different perspective on cooperative games if we assume that the agents share
a common purpose. For example, if we think of the agents as being workers in a company,
then the strategic considerations relating to coalition formation that are addressed by the core,
for example, are not relevant. Instead, we might want to organize the workforce (the agents)
into teams so as to maximize their overall productivity. More generally, the task is to find a coalition that maximizes the social welfare of the system, defined as the sum of the values of
the individual coalitions. We write the social welfare of a coalition structure CS as sw(CS), with the following definition:
sw(CSs) =Y v(C). cecs Then a socially optimal coalition structure CS* with respect to G maximizes this quantity.
Finding a socially optimal coalition structure is a very natural computational problem, which has been studied beyond the multiagent systems community:
it is sometimes called the set
partitioning partitioning problem. Unfortunately, the problem is NP-hard, because the number of possi- Set problem ble coalition structures grows exponentially in the number of players.
Finding the optimal coalition structure by naive exhaustive search is therefore infeasible
in general. An influential approach to optimal coalition structure formation is based on the
Coalition structure idea of searching a subspace of the coalition structure graph. The idea is best explained graph with reference to an example.
Suppose we have a game with four agents, N = {1,2,3,4}. There are fifteen possible
coalition structures for this set of agents.
We can organize these into a coalition structure
graph as shown in Figure 18.7, where the nodes at level £ of the graph correspond to all the coalition structures with exactly £ coalitions.
An upward edge in the graph represents
the division of a coalition in the lower node into two separate coalitions in the upper node.
632
Chapter 18 Multiagent Decision Making For example, there is an edge from {{1},{2,3,4}} to {{1},{2}.{3.4}} because this latter
coalition structure is obtained from the former by dividing the coalition {2,3,4} into the
coalitions {2} and {3,4}.
The optimal coalition structure CS* lies somewhere within the coalition structure graph,
and so to find this, it seems we would have to evaluate every node in the graph. But consider
the bottom two rows of the graph—Ilevels 1 and 2. Every possible coalition (excluding the empty coalition) appears in these two levels. (Of course, not every possible coalition structure
appears in these two levels.) Now, suppose we restrict our search for a possible coalition structure to just these two levels—we go no higher in the graph. Let CS’ be the best coalition
structure that we find in these two levels, and let CS* be the best coalition structure overall.
Let C* be a coalition with the highest value of all possible coalitions:
C* € argmax v(C). CeN
The value of the best coalition structure we find in the first two levels of the coalition structure
graph must be at least as much as the value of the best possible coalition: sw(CS') > v(C*).
This is because every possible coalition appears in at least one coalition structure in the first two levels of the graph. So assume the worst case, that is, sw(CS') = v(C*).
Compare the value of sw(CS') to sw(CS*). Since sw(CS') is the highest possible value
of any coalition structure, and there are n agents (n = 4 in the case of Figure 18.7), then the
highest possible value of sw(CS*) would be nv(C*) = n-sw(CS'). In other words, in the
worst possible case, the value of the best coalition structure we find in the first two levels of
the graph would be L the value of the best, where n is the number of agents. Thus, although
searching the first two levels of the graph does not guarantee to give us the optimal coalition
structure, it does guarantee to give s one that is no worse that & of the optimal. In practice it will often be much better than that.
18.4
Making Collective Decisions
‘We will now turn from agent design to mechanism design—the problem of designing the right game for a collection of agents to play. Formally, a mechanism consists of 1. A language for describing the set of allowable strategies that agents may adopt.
Center
2. A distinguished agent, called the center, that collects reports of strategy choices from the agents in the game. (For example, the auctioneer is the center in an auction.)
3. An outcome rule, known to all agents, that the center uses to determine the payoffs to
each agent, given their strategy choices. This section discusses some of the most important mechanisms.
Contract net protocol
18.4.1
Allocating tasks with the contract net
The contract net protocol is probably the oldest and most important multiagent problem-
solving technique studied in AL It is a high-level protocol for task sharing. As the name suggests, the contract net was inspired from the way that companies make use of contracts.
The overall contract net protocol has four main phases—see Figure 18.8. The process
starts with an agent identifying the need for cooperative action with respect to some task.
The need might arise because the agent does not have the capability to carry out the task
Section 18.4
problem recognition
27 awarding
633
lask\ o ouncement %
‘X’
T _X_
Making Collective Decisions
® «4 * bidding
'X'
Figure 18.8 The contract net task allocation protocol.
in isolation, or because a cooperative solution might in some way be better (faster, more efficient, more accurate).
The agent advertises the task to other agents in the net with a task announcement mes-
sage, and then acts as the manager of that task for its duration.
The task announcement
Task announcement Manager
message must include sufficient information for recipients to judge whether or not they are willing and able to bid for the task. The precise information included in a task announcement
will depend on the application area. It might be some code that needs to be executed; or it ‘might be a logical specification of a goal to be achieved. The task announcement might also include other information that might be required by recipients, such as deadlines, quality-ofservice requirements, and so on. ‘When an agent receives a task announcement, it must evaluate it with respect to its own
capabilities and preferences. In particular, each agent must determine, whether it has the
capability to carry out the task, and second, whether or not it desires to do so. On this basis, it may then submit a bid for the task. A bid will typically indicate the capabilities of the bidder that are relevant to the advertised task, and any terms and conditions under which the task will be carried out.
In general, a manager may receive multiple bids in response to a single task announcement. Based on the information in the bids, the manager selects the most appropriate agent
(or agents) to execute the task. Successful agents are notified through an award message, and
become contractors for the task, taking responsibility for the task until it is completed.
The main computational tasks required to implement the contract net protocol can be
summarized as follows:
&id
634
Chapter 18 Multiagent Decision Making « Task announcement processing. On receipt ofa task announcement, an agent decides if it wishes to bid for the advertised task.
« Bid processing. On receiving multiple bids, the manager must decide which agent to award the task to, and then award the task. * Award processing. Successful bidders (contractors) must attempt to carry out the task,
which may mean generating new subtasks, which are advertised via further task announcements.
Despite (or perhaps because of) its simplicity, the contract net is probably the most widely
implemented and best-studied framework for cooperative problem solving. It is naturally
applicable in many settings—a variation of it is enacted every time you request a car with
Uber, for example. 18.4.2
Allocating scarce resources with auctions
One of the most important problems in multiagent systems is that of allocating scarce re-
sources; but we may as well simply say “allocating resources,” since in practice most useful
Auction
resources are scarce in some sense. The auction is the most important mechanism for allo-
Bidder
there are multiple possible bidders. Each bidder i has a utility value v; for the item.
cating resources. The simplest setting for an auction is where there is a single resource and In some cases, each bidder has a private value for the item. For example, a tacky sweater
might be attractive to one bidder and valueless to another.
In other cases, such as auctioning drilling rights for an oil tract, the item has a com-
mon value—the tract will produce some amount of money, X, and all bidders value a dollar
equally—but there is uncertainty as to what the actual value of X is. Different bidders have different information, and hence different estimates of the item’s true value.
In either case,
bidders end up with their own v;. Given v;, each bidder gets a chance, at the appropriate time
or times in the auction, to make a bid b;. The highest bid, b,,q,, wins the item, but the price Ascending-bid auction English auction
paid need not be b,,; that’s part of the mechanism design. The best-known auction mechanism is the ascending-bid auction,’ or English auction,
in which the center starts by asking for a minimum (or reserve) bid by, If some bidder is willing to pay that amount, the center then asks for by, + d, for some increment d, and continues up from there. The auction ends when nobody is willing to bid anymore; then the last bidder wins the item, paying the price bid. How do we know if this is a good mechanism? One goal is to maximize expected revenue for the seller. Another goal is to maximize a notion of global utility. These goals overlap to
some extent, because one aspect of maximizing global utility is to ensure that the winner of
Efficient
the auction is the agent who values the item the most (and thus is willing to pay the most). We
say an auction is efficient if the goods go to the agent who values them most. The ascendingbid auction is usually both efficient and revenue maximizing, but if the reserve price is set too
high, the bidder who values it most may not bid, and if the reserve is set too low, the seller
Collusion
may get less revenue. Probably the most important things that an auction mechanism can do is encourage a sufficient number of bidders to enter the game and discourage them from engaging in collusion.
Collusion is an unfair or illegal agreement by two or more bidders to manipulate prices. It can 3 The word ion” comes from the L augeo, to increase.
Section 18.4
Making Collective Decisions
635
happen in secret backroom deals or tacitly, within the rules of the mechanism. For example, in 1999, Germany auctioned ten blocks of cellphone spectrum with a simultaneous auction
(bids were taken on all ten blocks at the same time), using the rule that any bid must be a minimum of a 10% raise over the previous bid on a block. There were only two credible bidders, and the first, Mannesman, entered the bid of 20 million deutschmark on blocks 1-5 and 18.18 million on blocks 6-10. Why 18.18M? One of T-Mobile’s managers said they “interpreted Mannesman’s first bid as an offer.” Both parties could compute that a 10% raise on 18.18M
is 19.99M; thus Mannesman’s bid was interpreted as saying “we can each get half the blocks for 20M:; let’s not spoil it by bidding the prices up higher”” And in fact T-Mobile bid 20M on blocks 6-10 and that was the end of the bidding.
The German government got less than they expected, because the two competitors were
able to use the bidding mechanism to come to a tacit agreement on how not to compete.
From the government’s point of view, a better result could have been obtained by any of these
changes to the mechanism: a higher reserve price; a sealed-bid first-price auction, so that the competitors could not communicate through their bids; or incentives to bring in a third bidder.
Perhaps the 10% rule was an error in mechanism design, because it facilitated the
precise signaling from Mannesman to T-Mobile.
In general, both the seller and the global utility function benefit if there are more bidders,
although global utility can suffer if you count the cost of wasted time of bidders that have no
chance of winning. One way to encourage more bidders is to make the mechanism easier for them. After all, if it requires too much research or computation on the part of the bidders, they may decide to take their money elsewhere. So it is desirable that the bidders have a dominant strategy. Recall that “dominant™ means that the strategy works against all other strategies, which in turn means that an agent
can adopt it without regard for the other strategies. An agent with a dominant strategy can just bid, without wasting time contemplating other agents’ possible strategies. A mechanism by
which agents have a dominant strategy is called a strategy-proof mechanism. If, as is usually Strategy-proof the case, that strategy involves the bidders revealing their true value, v;, then it is called a truth-revealing, or truthful, auction; the term incentive compatible is also used.
The
revelation principle states that any mechanism can be transformed into an equivalent truth-
revealing mechanism, so part of mechanism design is finding these equivalent mechanisms.
It turns out that the ascending-bid auction has most of the desirable properties. The bidder
with the highest value v; gets the goods at a price of b, +d, where b, is the highest bid among all the other agents and d is the auctioneer’s increment.*
Bidders have a simple dominant
strategy: keep bidding as long as the current cost is below your v;. The mechanism is not
quite truth-revealing, because the winning bidder reveals only that his v; > b, +d; we have a
lower bound on v; but not an exact amount.
A disadvantage (from the point of view of the seller) of the ascending-bid auction is that
it can discourage competition. Suppose that in a bid for cellphone spectrum there is one
advantaged company that everyone agrees would be able to leverage existing customers and
infrastructure, and thus can make a larger profit than anyone else. Potential competitors can
see that they have no chance in an ascending-bid auction, because the advantaged company
4 There is actually a small chance that the agent with highest v; fails to gt the goods, in the case in which by < v; < by +d. The chance of this can be made arbitrarily small by decreasing the increment d.
Truth-revealing
Revelation principle
636
Chapter 18 Multiagent Decision Making can always bid higher. Thus, the competitors may not enter at all, and the advantaged company ends up winning at the reserve price.
Another negative property of the English auction is its high communication costs. Either the auction takes place in one room or all bidders have to have high-speed, secure communi-
cation lines; in either case they have to have time to go through several rounds of bidding.
Sealed-bid auction
An alternative mechanism, which requires much less communication,
is the sealed-bid
auction. Each bidder makes a single bid and communicates it to the auctioneer, without the
other bidders seeing it. With this mechanism, there is no longer a simple dominant strategy.
If your value is v; and you believe that the maximum of all the other agents” bids will be b,, then you should bid b, + ¢, for some small ¢, if that is less than v;. Thus, your bid depends on
your estimation of the other agents’ bids, requiring you to do more work. Also, note that the
agent with the highest v; might not win the auction. This is offset by the fact that the auction is more competitive, reducing the bias toward an advantaged bidder.
Sealed-bid second-price auction
Vickrey auction
A small change in the mechanism for sealed-bid auctions leads to the sealed-bid second-
price auction, also known as a Vickrey auction.’
In such auctions, the winner pays the
price of the second-highest bid, b,, rather than paying his own bid. This simple modification completely eliminates the complex deliberations required for standard (or first-price) sealedbid auctions, because the dominant strategy is now simply to bid v;; the mechanism is truth-
revealing. Note that the utility of agent i in terms of his bid b;, his value v;, and the best bid among the other agents, b,, is
Ui
_ { (i=by) ifbi>b, 0
otherwise.
To see that b; = v; is a dominant strategy, note that when (v; — b,) is positive, any bid that wins the auction is optimal, and bidding v; in particular wins the auction. On the other hand, when
(vi—b,) is negative, any bid that loses the auction is optimal, and bidding v; in particular loses the auction. So bidding v; is optimal for all possible values of b,, and in fact, v; is the only bid that has this property. Because of its simplicity and the minimal computation requirements for both seller and bidders, the Vickrey auction is widely used in distributed Al systems.
Internet search engines conduct several trillion auctions each year to sell advertisements
along with their search results, and online auction sites handle $100 billion a year in goods, all using variants of the Vickrey auction. Note that the expected value to the seller is b,,
Revenue equivalence theorem
which is the same expected return as the limit of the English auction as the increment d goes to zero. This is actually a very general result: the revenue equivalence theorem states that, with a few minor caveats, any auction mechanism in which bidders have values v; known only
to themselves (but know the probability distribution from which those values are sampled),
will yield the same expected revenue. This principle means that the various mechanisms are not competing on the basis of revenue generation, but rather on other qualities.
Although the second-price auction is truth-revealing, it turns out that auctioning n goods
with an n+ 1 price auction is not truth-revealing. Many Internet search engines use a mech-
anism where they auction n slots for ads on a page. The highest bidder wins the top spot,
the second highest gets the second spot, and so on. Each winner pays the price bid by the
next-lower bidder, with the understanding that payment is made only if the searcher actually
5 Named after William Vickrey (1914-1996), who won the 1996 Nobel Prize in economics for this work and died of a heart attack three days later.
Section 18.4
Making Collective Decisions
637
clicks on the ad. The top slots are considered more valuable because they are more likely to be noticed and clicked on.
Imagine that three bidders, by, b, and b3, have valuations for a click of vy =200, v, = 180,
and v3 =100, and that n = 2 slots are available; and it is known that the top spot is clicked on
5% of the time and the bottom spot 2%. If all bidders bid truthfully, then b; wins the top slot
and pays 180, and has an expected return of (200— 180) x 0.05= 1. The second slot goes to by. But by can see that if she were to bid anything in the range 101-179, she would concede
the top slot to by, win the second slot, and yield an expected return of (200 — 100) x .02=2. Thus, b; can double her expected return by bidding less than her true value in this case. In general, bidders in this n+ 1 price auction must spend a lot of energy analyzing the bids of others to determine their best strategy; there is no simple dominant strategy.
Aggarwal et al. (2006) show that there is a unique truthful auction mechanism for this
multislot problem, in which the winner of slot j pays the price for slot j just for those addi-
tional clicks that are available at slot j and not at slot j+ 1. The winner pays the price for the lower slot for the remaining clicks. In our example, by would bid 200 truthfully, and would
pay 180 for the additional .05 —.02=.03 clicks in the top slot, but would pay only the cost of the bottom slot, 100, for the remaining .02 clicks. (200— 180) x .03+ (200— 100) x .02=2.6.
Thus, the total return to b; would be
Another example of where auctions can come into play within Al is when a collection of agents are deciding whether to cooperate on a joint plan. Hunsberger and Grosz (2000) show that this can be accomplished efficiently with an auction in which the agents bid for roles in
the joint plan. Common
goods
Now let’s consider another type of game, in which countries set their policy for controlling air pollution. Each country has a choice: they can reduce pollution at a cost of -10 points for
implementing the necessary changes, or they can continue to pollute, which gives them a net
utility of -5 (in added health costs, etc.) and also contributes -1 points to every other country
(because the air is shared across countries). Clearly, the dominant strategy for each country
is “continue to pollute,” but if there are 100 countries and each follows this policy, then each
country gets a total utility of -104, whereas if every country reduced pollution, they would
each have a utility of -10. This situation is called the tragedy of the commons: if nobody has
to pay for using a common resource, then it may be exploited in a way that leads to a lower total utility for all agents.
It is similar to the prisoner’s dilemma:
Tragedy of the commons
there is another solution
to the game that is better for all parties, but there appears to be no way for rational agents to
arrive at that solution under the current game. One approach for dealing with the tragedy of the commons s to change the mechanism to one that charges each agent for using the commons.
More generally, we need to ensure that
all externalities—effects on global utility that are not recognized in the individual agents’ transactions—are made explicit.
Setting the prices correctly is the difficult part. In the limit, this approach amounts to
creating a mechanism in which each agent is effectively required to maximize global utility,
but can do so by making a local decision. For this example, a carbon tax would be an example of a mechanism that charges for use of the commons
maximizes global utility.
in a way that, if implemented well,
Externalities
638
VCG
Chapter 18 Multiagent Decision Making It turns out there is a mechanism design, known as the Vickrey-Clarke-Groves or VCG
mechanism, which has two favorable properties.
First, it is utility maximizing—that is, it
maximizes the global utility, which is the sum of the utilities for all parties, };v;.
Second,
the mechanism is truth-revealing—the dominant strategy for all agents is to reveal their true value. There is no need for them to engage in complicated strategic bidding calculations. We will give an example using the problem of allocating some common goods. Suppose a city decides it wants to install some free wireless Internet transceivers. However, the number
of transceivers available is less than the number of neighborhoods that want them. The city
wants to maximize global utility, but if it says to each neighborhood council you value a free transceiver (and by the way we will give them to the parties the most)?” then each neighborhood will have an incentive to report a very VCG mechanism discourages this ploy and gives them an incentive to report It works as follows:
1. The center asks each agent to report its value for an item, v;. 2. The center allocates the goods to a set of winners, W, to maximize ;¢
“How much do that value them high value. The their true value. v;.
3. The center calculates for each winning agent how much of a loss their individual pres-
ence in the game has caused to the losers (who each got 0 utility, but could have got v;
if they were a winner).
4. Each winning agent then pays to the center a tax equal (o this loss. For example, suppose there are 3 transceivers available and 5 bidders, who bid 100, 50, 40, 20, and 10. Thus the set of 3 winners, W, are the ones who bid 100, 50, and 40 and the
global utility from allocating these goods is 190. For each winner, it is the case that had they
not been in the game, the bid of 20 would have been a winner. Thus, each winner pays a tax
of 20 to the center.
All winners should be happy because they pay a tax that is less than their value, and all losers are as happy as they can be, because they value the goods less than the required
tax. That’s why the mechanism is truth-revealing. In this example, the crucial value is 20; it would be irrational to bid above 20 if your true value was actually below 20, and vice versa.
Since the crucial value could be anything (depending on the other bidders), that means that is always irrational to bid anything other than your true value.
The VCG mechanism is very general, and can be applied to all sorts of games, not just auctions, with a slight generalization of the mechanism described above. For example, in a
ccombinatorial auction there are multiple different items available and each bidder can place
multiple bids, each on a subset of the items. For example, in bidding on plots of land, one bidder might want either plot X or plot Y but not both; another might want any three adjacent plots, and so on. The VCG mechanism can be used to find the optimal outcome, although
with 2V subsets of N goods to contend with, the computation of the optimal outcome is NPcomplete. With a few caveats the VCG mechanism is unique: every other optimal mechanism is essentially equivalent.
18.4.3
Social choice theory
Voting
The next class of mechanisms that we look at are voting procedures, of the type that are used for political decision making in democratic societies. The study of voting procedures derives from the domain of social choice theory.
Section 18.4 The basic setting is as follows.
Making Collective Decisions
639
As usual, we have a set N = {1,...,n} of agents, who
in this section will be the voters. These voters want to make decisions with respect to a set
Q= {wy,w,...} of possible outcomes. In a political election, each element of Q could stand for a different candidate winning the election.
Each voter will have preferences over Q. These are usually expressed not as quantitative
utilities but rather as qualitative comparisons:
we write w >~; w' to mean that outcome w is
ranked above outcome w' by agent i. In an election with three candidates, agent i might have
W
w3 = Wi The fundamental problem of social choice theory is to combine these preferences, using Social welfare a social welfare function, to come up with a social preference order: a ranking of the Function candidates, from most preferred down to least preferred. In some cases, we are only interested
in a social outcome—the most preferred outcome by the group as a whole. We will write Social outcome
w =" w' to mean that w is ranked above w’ in the social preference order.
A simpler setting is where we are not concerned with obtaining an entire ordering of
candidates, but simply want to choose a set of winners.
A social choice function takes as
input a preference order for each voter, and produces as output a set of winners.
Democratic societies want a social outcome that reflects the preferences of the voters.
Unfortunately, this is not always straightforward. Consider Condorcet’s Paradox, a famous
example posed by the Marquis de Condorcet (1743-1794). Suppose we have three outcomes, Q = {w,,wp,w,}, and three voters, N = {1,2,3}, with preferences as follows. Wa =1 Wp =1 We
We >2 Wa =2 Wh
(18.2)
Wp >3 We =3 Wa
Now, suppose we have to choose one of the three candidates on the basis of these preferences. The paradox is that: * 2/3 of the voters prefer w; over w;. + 2/3 of the voters prefer w; over ws.
* 2/3 of the voters prefer w) over ws. So, for each possible winner, we can point to another candidate who would be preferred by at least 2/3 of the electorate. It is obvious that in a democracy we cannot hope to make every
voter happy. This demonstrates that there are scenarios in which no matter which outcome we choose, a majority of voters will prefer a different outcome. A natural question is whether there is any “good” social choice procedure that really reflects the preferences of voters. To answer this, we need to be precise about what we mean when we say that a rule is “good.” ‘We will list some properties we would like a good social welfare function to satis
* The Pareto Condition:
above wj, then w; =* w;.
* The Condorcet
The Pareto condition simply says that if every voter ranks w;
Winner Condition:
An outcome is said to be a Condorcet winner if
a majority of candidates prefer it over all other outcomes. To put it another way, a
Condorcet winner is a candidate that would beat every other candidate in a pairwise
election. The Condorcet winner condition says that if w; is a Condorcet winner, then w; should be ranked first.
« Independence of Irrelevant Alternatives (IIA): Suppose there are a number of candidates, including w; and w;, and voter preferences are such that w; = w;. Now, suppose
Social choice function Condorcet's Paradox
Chapter 18 Multiagent Decision Making
Arrow's theorem
Simple majority vote
Plurality voting
one voter changed their preferences in some way, but nor about the relative ranking of wi and w;. The TTA condition says that, w; * w; should not change. « No Dictatorships: 1t should not be the case that the social welfare function simply outputs one voter’s preferences and ignores all other voters. These four conditions seem reasonable, but a fundamental theorem of social choice theory called Arrow’s theorem (due to Kenneth Arrow) tells s that it is impossible to satisfy all four conditions (for cases where there are at least three outcomes). That means that for any social choice mechanism we might care to pick, there will be some situations (perhaps unusual or pathological) that lead to controversial outcomes. However, it does not mean that democratic decision making is hopeless in most cases. We have not yet seen any actual voting procedures, so let’s now look at some. = With just two candidates, simple majority vote (the standard method in the US and UK) s the favored mechanism. We ask each voter which of the two candidates they prefer, and the one with the most votes is the winner. + With more than two outcomes, plurality voting is a common system.
We ask each
voter for their top choice, and select the candidate(s) (more than one in the case of ties) who get the most votes, even if nobody gets a majority. While it is common, plurality
voting has been criticized for delivering unpopular outcomes. A key problem is that it only takes into account the top-ranked candidate in each voter’s preferences.
Borda count
+ The Borda count (after Jean-Charles de Borda, a contemporary and rival of Condorcet)
is a voting procedure that takes into account all the information in a voter’s preference
ordering. Suppose we have k candidates. Then for each voter i, we take their preference
ordering -;, and give a score of k to the top ranked candidate, a score of k — 1 to the
second-ranked candidate, and so on down to the least-favored candidate in i’s ordering. The total score for each candidate is their Borda count, and to obtain the social outcome
~*, outcomes are ordered by their Borda count—highest to lowest. One practical prob-
lem with this system is that it asks voters to express preferences on all the candidates,
Approval voting
and some voters may only care about a subset of candidates.
+ In approval voting, voters submit a subset of the candidates that they approve of. The winner(s) are those who are approved by the most voters. This system is often used when the task is to choose multiple winners.
Instant runoff voting
« In instant runoff voting, voters rank all the candidates, and if a candidate has a major-
ity of first-place votes, they are declared the winner. If not, the candidate with the fewest
first-place votes is eliminated. That candidate is removed from all the preference rank-
ings (so those voters who had the eliminated candidate as their first choice now have
another candidate as their new first choice) and the process is repeated.
True majority rule voting
Eventually,
some candidate will have a majority of first-place votes (unless there is a tie).
+ In true majority rule voting, the winner is the candidate who beats every other can-
didate in pairwise comparisons. Voters are asked for a full preference ranking of all
candidates. We say that w beats ', if more voters have w > w’ than have w’ - w. This
system has the nice property that the majority always agrees on the winner, but it has the bad property that not every election will be decided:
example, no candidate wins a majority.
in the Condorcet paradox, for
Section 18.4
Making Collective Decisions
641
Strategic manipulation Besides Arrow’s Theorem,
another important negative results in the area of social choice
theory is the Gibbard-Satterthwaite Theorem.
This result relates to the circumstances
under which a voter can benefit from misrepresenting their preferenc
Gibbard— Satterthwaite Theorem
Recall that a social choice function takes as input a preference order for each voter, and
gives as output a set of winning candidates. Each voter has, of course, their own true prefer-
ences, but there is nothing in the definition of a social choice function that requires voters to
report their preferences truthfully; they can declare whatever preferences they like.
In some cases, it can make sense for a voter to misrepresent their preferences. For exam-
ple, in plurality voting, voters who think their preferred candidate has no chance of winning may vote for their second choice instead. That means plurality voting is a game in which voters have to think strategically (about the other voters) to maximize their expected utility.
This raises an interesting question: can we design a voting mechanism that is immune to
such manipulation—a mechanism that is truth-revealing? The Gibbard-Satterthwaite Theo-
rem tells us that we can not: Any social choice function that satisfies the Pareto condition for a domain with more than two outcomes is either manipulable or a dictatorship. That is, for any “reasonable” social choice procedure, there will be some circumstances under which a
voter can in principle benefit by misrepresenting their preferences. However, it does not tell
us how such manipulation might be done; and it does not tell us that such manipulation is
likely in practice. 18.4.4
Bargaining
Bargaining, or negotiation, is another mechanism that is used frequently in everyday life. It has been studied in game theory since the 1950s and more recently has become a task for automated agents. Bargaining is used when agents need to reach agreement on a matter of common interest. The agents make offers (also called proposals or deals) to each other under specific protocols, and either accept or reject cach offer. Bargaining with the alternating offers protocol
offers One influential bargaining protocol is the alternating offers bargaining model. For simplic- Alternating bargaining model ity we’ll again
assume just two agents. Bargaining takes place in a sequence of rounds. A;
begins, at round 0, by making an offer. If A, accepts the offer, then the offer is implemented. If Ay rejects the offer, then negotiation moves to the next round. This time A> makes an offer and A chooses to accept or reject it, and so on. If the negotiation never terminates (because
agents reject every offer) then we define the outcome to be the conflict deal. A convenient
simplifying assumption is that both agents prefer to reach an outcome—any outcome—in finite time rather than being stuck in the infinitely time-consuming conflict deal. ‘We will use the scenario of dividing a pie to illustrate alternating offers. The idea is that
Conflict deal
there is some resource (the “pie”) whose value is 1, which can be divided into two parts, one
part for each agent. Thus an offer in this scenario is a pair (x,1 —x), where x is the amount
of the pie that A; gets, and 1 — x is the amount that A, gets. The space of possible deals (the negotiation set) is thus:
{(1-2):0. Agent A can take this fact into account by offering (1 —72,72), an
Section 18.4
Making Collective Decisions
offer that A, may as well accept because A, can do no better than 7, at this point in time. (If you are worried about what happens with ties, just make the offer be (1 — (v, +¢€),72 +€) for
some small value of ¢.) So, the two strategies of A
offering (1 —~2,72), and A accepting that offer are in Nash
equilibrium. Patient players (those with a larger ~2) will be able to obtain larger pieces of the pie under this protocol: in this setting, patience truly is a virtue. Now consider the general case, where there are no bounds on the number of rounds. As
in the I-round case, A, can craft a proposal that A, should accept, because it gives A, the maximal
achievable amount, given the discount factors. It turns out that A; will get
I-m
-
and A will get the remainder. Negotiation
in task-oriented domains
In this section, we consider negotiation for task-oriented domains. In such a domain, a set of
tasks must be carried out, and each task is initially assigned to a set of agents. The agents may
Task-oriented domain
be able to benefit by negotiating on who will carry out which tasks. For example, suppose
some tasks are done on a lathe machine and others on a milling machine, and that any agent
using a machine must incur a significant setup cost. Then it would make sense for one agent 1o offer another “T have to set up on the milling machine anyway; how about if I do all your milling tasks, and you do all my lathe tasks?”
Unlike the bargaining scenario, we start with an initial allocation, so if the agents fail to
agree on any offers, they perform the tasks T that they were originally allocated.
To keep things simple, we will again assume just two agents. Let 7 be the set of all tasks
and let (7}, 7) denote the initial allocation of tasks to the two agents at time 0. Each task in T must be assigned to exactly one agent.
We assume we have a cost function ¢, which
for every set of tasks 7" gives a positive real number ¢(7”) indicating the cost to any agent
of carrying out the tasks 7”. (Assume the cost depends only on the tasks, not on the agent
carrying out the task.) The cost function is monotonic—adding more tasks never reduces the cost—and the cost of doing nothing is zero: ¢({}) =
0. As an example, suppose the cost of
setting up the milling machine is 10 and each milling task costs 1, then the cost of a set of two milling tasks would be 12, and the cost of a set of five would be 15.
An offer of the form (7;,7>) means that agent i is committed to performing the set of tasks 7;, at cost ¢(7;). The utility to agent i is the amount they have to gain from accepting the offer—the difference between the cost of doing this new set of tasks versus the originally
assigned set of tasks:
Ui((T,T2) = e(T3) = e(TY)
An offer (T, T5) is individually rational if U;((7},73)) > 0 for both agents. If a deal is not Individually rational individually rational, then at least one agent can do better by simply performing the tasks it
was originally allocated.
The negotiation set for task-oriented domains (assuming rational agents) is the set of
offers that are both individually rational and Pareto optimal. There is no sense making an
individually irrational offer that will be refused, nor in making an offer when there is a better offer that improves one agent’s utility without hurting anyone else.
Chapter 18 Multiagent Decision Making
The monotonic concession protocol Mornotonic concession protocol
The negotiation protocol we consider for task-oriented domains
is known as the monotonic
concession protocol. The rules of this protocol are as follows. + Negotiation proceeds in a series of rounds. + On the first round, both agents simultaneously propose a deal, D; = (T}, T3), from the negotiation set. (This is different from the alternating offers we saw before.)
+ An agreement is reached if the two agents propose deals D; and D;, respectively, such
that either (i) Uy (D) > Uy (Dy) or (i) Us(Dy) > Us(D3), that is, if one of the agents finds that the deal proposed by the other is at least as good or better than the proposal it made. If agreement is reached, then the rule for determining the agreement deal is as follows:
If each agent’s offer matches or exceeds that of the other agent, then one of
the proposals is selected at random. If only one proposal exceeds or matches the other’s proposal, then this is the agreement deal. « If no agreement is reached, then negotiation proceeds to another round of simultaneous
Concession
proposals. In round 7 + 1, each agent must either repeat the proposal from the previous
round or make a concession—a proposal that is more preferred by the other agent (i.e.,
has higher utility).
« If neither agent makes a concession, then negotiation terminates, and both agents im-
plement the conflict deal, carrying out the tasks they were originally assigned.
Since the set of possible deals is finite, the agents cannot negotiate indefinitely:
either the
agents will reach agreement, or a round will occur in which neither agent concedes. However, the protocol does not guarantee that agreement will be reached quickly: since the number of
possible deals is O(271), it is conceivable that negotiation will continue for a number of
rounds exponential in the number of tasks to be allocated. The Zeuthen strategy
Zeuthen strategy
So far, we have said nothing about how negotiation participants might or should behave when using the monotonic concession protocol for task-oriented domains. One possible strategy is the Zeuthen strategy.
The idea of the Zeuthen strategy is to measure an agent’s willingness to risk conflict.
Intuitively, an agent will be more willing to risk conflict if the difference in utility between its current proposal and the conflict deal is low.
In this case, the agent has little to lose if
negotiation fails and the conflict deal is implemented, and so is more willing to risk conflict, and less willing to concede. In contrast, if the difference between the agent’s current proposal and the conflict deal is high, then the agent has more to lose from conflict and is therefore less willing to risk conflict—and thus more willing to concede. Agent i’s willingness to risk conflict at round 7, denoted risk!, is measured as follows:
by conceding and accepting js offer
Until an agreement is reached, the value of risk{ will be a value between 0 and 1. Higher
values of risk! (nearer to 1) indicate that i has less to lose from conflict, and so is more willing to risk conflict.
Summary The Zeuthen strategy says that each agent’s first proposal should be a deal in the negoti-
ation set that maximizes its own utility (there may be more than one). After that, the agent
who should concede on round 7 of negotiation should be the one with the smaller value of risk—the one with the most to lose from conflict if neither concedes.
The next question to answer is how much should be conceded? The answer provided by the Zeuthen strategy is, “Just enough to change the balance of risk to the other agent.” That is, an agent should make the smallest concession that will make the other agent concede on
the next round. There is one final refinement to the Zeuthen strategy.
Suppose that at some point both
agents have equal risk. Then, according to the strategy, both should concede. But, knowing this, one agent could potentially “defect” by not conceding, and so benefit.
To avoid the
possibility of both conceding at this point, we extend the strategy by having the agents “flip a coin” to decide who should concede if ever an equal risk situation is reached.
With this strategy, agreement will be Pareto optimal and individually rational. However,
since the space of possible deals is exponential in the number of tasks, following this
strategy
may require O(2/"!) computations of the cost function at each negotiation step. Finally, the
Zeuthen strategy (with the coin flipping rule) is in Nash equilibrium.
Summary « Multiagent planning is necessary when there are other agents in the environment with which to cooperate or compete. Joint plans can be constructed, but must be augmented with some form of coordination if two agents are to agree on which joint plan to execute.
+ Game
theory describes rational behavior for agents in situations in which multiple
agents interact. Game theory is to multiagent decision making as decision theory is to
single-agent decision making. « Solution concepts in game theory are intended to characterize rational outcomes of a game—outcomes that might occur if every agent acted rationally.
+ Non-cooperative game theory assumes that agents must make their decisions indepen-
dently. Nash equilibrium is the most important solution concept in non-cooperative game theory. A Nash equilibrium is a strategy profile in which no agent has an incentive to deviate from its specified strategy. We have techniques for dealing with repeated games and sequential games.
+ Cooperative game theory considers settings in which agents can make binding agree-
ments to form coalitions in order to cooperate. Solution concepts in cooperative game attempt to formulate which coalitions are stable (the core) and how to fairly divide the
value that a coalition obtains (the Shapley value). « Specialized techniques are available for certain important classes of multiagent decision: the contract net for task sharing; auctions are used to efficiently allocate scarce
resources; bargaining for reaching agreements on matters of common interest; and vot-
ing procedures for aggregating preferences.
Chapter 18 Multiagent Decision Making Bibliographical and Historical Notes Itis a curiosity of the field that researchers in Al did not begin to seriously consider the issues
surrounding interacting agents until the 1980s—and the multiagent systems field did not really become established as a distinctive subdiscipline of Al until a decade later. Nevertheless,
ideas that hint at multiagent systems were present in the 1970s. For example, in his highly influential Society of Mind theory, Marvin Minsky (1986, 2007) proposed that human minds are constructed from an ensemble of agents. Doug Lenat had similar ideas in a framework he called BEINGS (Lenat, 1975). In the 1970s, building on his PhD work on the PLANNER
system, Carl Hewitt proposed a model of computation as interacting agents called the ac-
tor model, which has become established as one of the fundamental models in concurrent computation (Hewitt, 1977; Agha, 1986).
The prehistory of the multiagent systems field is thoroughly documented in a collection
of papers entitled Readings in Distributed Artificial Intelligence (Bond and Gasser, 1988).
The collection is prefaced with a detailed statement of the key research challenges in multi-
agent systems, which remains remarkably relevant today, more than thirty years after it was written.
Early research on multiagent systems tended to assume that all agents in a system
were acting with common interest, with a single designer. This
Cooperative distributed problem solving
is now recognized as a spe-
cial case of the more general multiagent setting—the special case is known as cooperative
distributed problem solving. A key system of this time was the Distributed Vehicle Moni-
toring Testbed (DVMT), developed under the supervision of Victor Lesser at the University
of Massachusetts (Lesser and Corkill, 1988). The DVMT modeled a scenario in which a col-
lection of geographically distributed acoustic sensor agents cooperate to track the movement
of vehicles.
The contemporary era of multiagent systems research began in the late 1980s, when it
was widely realized that agents with differing preferences are the norm in Al and society— from this point on, game theory began to be established as the main methodology for studying such agents. Multiagent planning has leaped in popularity in recent years, although it does have a long history. Konolige (1982) formalizes multiagent planning in first-order logic, while Pednault (1986) gives a STRIPS-style description.
The notion of joint intention, which is essential if
agents are to execute a joint plan, comes from work on communicative acts (Cohen and Perrault, 1979; Cohen and Levesque,
1990; Cohen et al., 1990). Boutilier and Brafman (2001)
show how to adapt partial-order planning to a multiactor setting.
Brafman and Domshlak
(2008) devise a multiactor planning algorithm whose complexity grows only linearly with
the number of actors, provided that the degree of coupling (measured partly by the tree width
of the graph of interactions among agents) is bounded.
Multiagent planning is hardest when there are adversarial agents.
As Jean-Paul Sartre
(1960) said, “In a football match, everything is complicated by the presence of the other team.” General Dwight D. Eisenhower said, “In preparing for battle I have always found that plans are useless, but planning is indispensable,” meaning that it is important to have a conditional plan or policy, and not to expect an unconditional plan to succeed.
The topic of distributed and multiagent reinforcement learning (RL) was not covered in
this chapter but is of great current interest. In distributed RL, the aim is to devise methods by
which multiple, coordinated agents learn to optimize a common utility function. For example,
Bibliographical and Historical Notes can we devise methods whereby separate subagents for robot navigation and robot obstacle avoidance could cooperatively achieve a combined control system that is globally optimal?
Some basic results in this direction have been obtained (Guestrin et al., 2002; Russell and Zimdars, 2003). The basic idea is that each subagent learns its own Q-function (a kind of
utility function; see Section 22.3.3) from its own stream of rewards. For example, a robot-
navigation component can receive rewards for making progress towards the goal, while the obstacle-avoidance component receives negative rewards for every collision. Each global decision maximizes the sum of Q-functions and the whole process converges to globally optimal solutions. The roots of game theory can be traced back to proposals made in the 17th century by
Christiaan Huygens and Gottfried Leibniz to study competitive and cooperative human in-
ientifically and mathematically. Throughout the 19th century, several leading reated simple mathematical examples to analyze particular examples of compet-
itive situations. The first formal results in game theory are due to Zermelo (1913) (who had, the year before, suggested a form of minimax search for games, albeit an incorrect one). Emile Borel (1921) introduced the notion of a mixed strategy. John von Neumann (1928) proved that
every two-person, zero-sum game has a maximin equilibrium in mixed strategies and a well-
defined value. Von Neumann’s collaboration with the economist Oskar Morgenstern led to the publication in 1944 of the Theory of Games and Economic Behavior, the defining book
for game theory. Publication of the book was delayed by the wartime paper shortage until a member of the Rockefeller family personally subsidized its publication. In 1950, at the age of 21, John Nash published his ideas concerning equilibria in general (non-zero-sum) games. His definition of an equilibrium solution, although anticipated in the work of Cournot (1838), became known as Nash equilibrium. After a long delay because of the schizophrenia he suffered from 1959 onward, Nash was awarded the Nobel Memorial Prize in Economics (along with Reinhart Selten and John Harsanyi) in 1994. The Bayes-Nash
equilibrium is described by Harsanyi (1967) and discussed by Kadane and Larkey (1982). Some issues in the use of game theory for agent control are covered by Binmore (1982). Aumann and Brandenburger (1995) show how different equilibria can be reached depending on the knowleedge each player has.
The prisoner’s dilemma was invented as a classroom exercise by Albert W. Tucker in 1950 (based on an example by Merrill Flood and Melvin Dresher) and is covered extensively
by Axelrod (1985) and Poundstone (1993). Repeated games were introduced by Luce and
Raiffa (1957), and Abreu and Rubinstein (1988) discuss the use of finite state machines for
repeated games—technically, Moore machines. The text by Mailath and Samuelson (2006) concentrates on repeated games. Games of partial information in extensive form were introduced by Kuhn (1953).
The
sequence form for partial-information games was invented by Romanovskii (1962) and independently by Koller ef al. (1996); the paper by Koller and Pfeffer (1997) provides a readable introduction to the field and describes a system for representing and solving sequential games. The use of abstraction to reduce a game tree to a size that can be solved with Koller’s
technique was introduced by Billings ef al. (2003). Subsequently, improved methods for equilibrium-finding enabled solution of abstractions with 102 states (Gilpin er al., 2008;
Zinkevich et al., 2008).
Bowling ef al. (2008) show how to use importance sampling to
647
Chapter 18 Multiagent Decision Making get a better estimate of the value ofa strategy. Waugh ef al. (2009) found that the abstraction approach is vulnerable to making systematic errors in approximating the equilibrium solution: it works for some games but not others.
Brown and Sandholm (2019) showed that, at least
in the case of multiplayer Texas hold "em poker, these vulnerabilities can be overcome by sufficient computing power.
They used a 64-core server running for 8 days to compute a
baseline strategy for their Pluribus program. human champion opponents.
With that strategy they were able to defeat
Game theory and MDPs are combined in the theory of Markov games, also called stochastic games (Littman, 1994; Hu and Wellman, 1998). Shapley (1953b) actually described the
value iteration algorithm independently of Bellman, but his results were not widely appre-
ciated, perhaps because they were presented in the context of Markov games. Evolutionary
game theory (Smith, 1982; Weibull, 1995) looks at strategy drift over time: if your opponent’s strategy is changing, how should you react?
Textbooks on game theory from an economics point of view include those by Myerson
(1991), Fudenberg and Tirole (1991), Osborne (2004), and Osborne and Rubinstein (1994).
From an Al perspective we have Nisan ef al. (2007) and Leyton-Brown and Shoham (2008). See (Sandholm, 1999) for a useful survey of multiagent decision making. Multiagent RL is distinguished from distributed RL by the presence of agents who cannot
coordinate their actions (except by explicit communicative acts) and who may not share the
same utility function. Thus, multiagent RL deals with sequential game-theoretic problems or
Markov games, as defined in Chapter 17. What causes problems is the fact that, while an
agent is learning to defeat its opponent’s policy, the opponent is changing its policy to defeat the agent. Thus, the environment is nonstationary (see page 444).
Littman (1994) noted this difficulty when introducing the first RL algorithms for zero-
sum Markov games. Hu and Wellman (2003) present a Q-learning algorithm for generalsum games that converges when the Nash equilibrium is unique; when there are multiple equilibria, the notion of convergence is not so easy to define (Shoham ef al., 2004). Assistance games were introduced under the heading of cooperative inverse reinforce-
ment learning by Hadfield-Menell ez al. (2017a). Malik ez al. (2018) introduced an efficient
Principal-agent game
POMDP solver designed specifically for assistance games.
They are related to principal-
agent games in economics, in which a principal (e.g., an employer) and an agent (e.g., an employee) need to find a mutually beneficial arrangement despite having widely different preferences. The primary differences are that (1) the robot has no preferences of its own, and (2) the robot is uncertain about the human preferences it needs to optimize.
Cooperative games were first studied by von Neumann and Morgenstern (1944). The notion of the core was introduced by Donald Gillies (1959), and the Shapley value by Lloyd Shapley (1953a). A good introduction to the mathematics of cooperative games is Peleg and Sudholter (2002). Simple games in general are discussed in detail by Taylor and Zwicker (1999). For an introduction to the computational aspects of cooperative game theory, see
Chalkiadakis er al. (2011).
Many compact representation schemes for cooperative games have been developed over
the past three decades, starting with the work of Deng and Papadimitriou (1994). The most
influential of these schemes is the marginal contribution networks model, which was intro-
duced by Teong and Shoham (2005). The approach to coalition formation that we describe was developed by Sandholm ef al. (1999); Rahwan ef al. (2015) survey the state of the art.
Bibliographical and Historical Notes The contract net protocol was introduced by Reid Smith for his PhD work at Stanford
University in the late 1970s (Smith, 1980). The protocol seems to be so natural that it is reg-
ularly reinvented to the present day. The economic foundations of the protocol were studied by Sandholm (1993).
Auctions and mechanism design have been mainstream topics in computer science and
AI for several decades:
see Nisan (2007) for a mainstream computer science perspective,
Krishna (2002) for an introduction to the theory of auctions, and Cramton er al. (2006) for a collection of articles on computational aspects of auctions. The 2007 Nobel Memorial Prize in Economics went to Hurwicz, Maskin, and Myerson
“for having laid the foundations of mechanism design theory” (Hurwicz, 1973). The tragedy of the commons, a motivating problem for the field, was analyzed by William Lloyd (1833) but named and brought to public attention by Garrett Hardin (1968). Ronald Coase presented
a theorem that if resources are subject to private ownership and if transaction costs are low
enough, then the resources will be managed efficiently (Coase, 1960). He points out that, in practice, transaction costs are high, so this theorem does not apply, and we should look to other solutions beyond privatization and the marketplace. Elinor Ostrom’s Governing the Commons (1990) described solutions for the problem based on placing management control over the resources into the hands of the local people who have the most knowledge of the situation. Both Coase and Ostrom won the Nobel Prize in economics for their work.
The revelation principle is due to Myerson (1986), and the revenue equivalence theorem was developed independently by Myerson (1981) and Riley and Samuelson (1981). Two
economists, Milgrom (1997) and Klemperer (2002), write about the multibillion-dollar spec-
trum auctions they were involved in. Mechanism design is used in multiagent planning (Hunsberger and Grosz, 2000; Stone et al., 2009) and scheduling (Rassenti er al., 1982). Varian (1995) gives a brief overview with connections to the computer science literature, and Rosenschein and Zlotkin (1994)
present a book-length treatment with applications to distributed AL Related work on distributed AT goes under several names, including collective intelligence (Tumer and Wolpert, 2000; Segaran, 2007) and market-based control (Clearwater,
1996).
Since 2001
there has
been an annual Trading Agents Competition (TAC), in which agents try to make the best profit on a series of auctions (Wellman ez al., 2001; Arunachalam and Sadeh, 2005).
The social choice literature is enormous, and spans erations on the nature of democracy through to highly procedures. Campbell and Kelly (2002) provide a good Handbook of Computational Social Choice provides a
the gulf from philosophical considtechnical analyses of specific voting starting point for this literature. The range of articles surveying research
topics and methods in this field (Brandt ez al., 2016). Arrow’s theorem lists desired properties
of a voting system and proves that is impossible to achieve all of them (Arrow, 1951). Dasgupta and Maskin (2008) show that majority rule (not plurality rule, and not ranked choice voting) is the most robust voting system. The computational complexity of manipulating elections was first studied by Bartholdi ez al. (1989).
We have barely skimmed the surface of work on negotiation in multiagent planning. Durfee and Lesser (1989) discuss how tasks can be shared out among agents by negotiation. Kraus et al. (1991) describe a system for playing Diplomacy, a board game requiring negoti-
ation, coalition formation, and dishonesty. Stone (2000) shows how agents can cooperate as
teammates in the competitive, dynamic, partially observable environment of robotic soccer. In
649
650
Chapter 18 Multiagent Decision Making a later article, Stone (2003) analyzes two competitive multiagent environments—RoboCup,
a robotic soccer competition, and TAC, the auction-based Trading Agents Competition—
and finds that the computational intractability of our current theoretically well-founded approaches has led to many multiagent systems being designed by ad hoc methods. Sarit Kraus has developed a number of agents that can negotiate with humans and other agents—
see Kraus (2001) for a survey. The monotonic concession protocol for automated negotiation was proposed by Jeffrey S. Rosenschein and his students (Rosenschein and Zlotkin, 1994). The alternating offers protocol was developed by Rubinstein (1982). Books on multiagent systems include those by Weiss (2000a), Young (2004), Vlassis (2008), Shoham and Leyton-Brown (2009), and Wooldridge (2009). The primary conference for multiagent systems is the International Conference on Autonomous Agents and Multi-
Agent Systems (AAMAS); there is also a journal by the same name. The ACM Conference on Electronic Commerce (EC) also publishes many relevant papers, particularly in the area of auction algorithms. The principal journal for game theory is Games and Economic Behavior.
TG
19
LEARNING FROM EXAMPLES In which we describe agents that can improve their behavior through diligent study of past experiences and predictions about the future.
An agent is learning if it improves its performance after making observations about the world.
Learning can range from the trivial, such as jotting down a shopping list, to the profound, as when Albert Einstein inferred a new theory of the universe. When the agent is a computer, we call it machine learning:
a computer observes some data, builds a model based on the
data, and uses the model as both a hypothesis about the world and a piece of software that
can solve problems.
‘Why would we want a machine to learn? Why not just program it the right way to begin
with? There are two main reasons.
First, the designers cannot anticipate all possible future
situations. For example, a robot designed to navigate mazes must learn the layout of each new
maze it encounters; a program for predicting stock market prices must learn to adapt when
conditions change from boom to bust. Second, sometimes the designers have no idea how
to program a solution themselves. Most people are good at recognizing the faces of family
members, but they do it subconsciously, so even the best programmers don’t know how to
program a computer to accomplish that task, except by using machine learning algorithms.
In this chapter, we interleave a discussion of various model classes—decision trees (Sec-
tion 19.3), linear models (Section 19.6), nonparametric models such as nearest neighbors (Section 19.7), ensemble models such as random forests (Section 19.8)—with practical advice on building machine learning systems (Section 19.9), and discussion of the theory of ‘machine learning (Sections 19.1 to 19.5).
19.1
Forms of Learning
Any component of an agent program can be improved by machine learning. The improve‘ments, and the techniques used to make them, depend on these factors:
« Which component is to be improved. * What prior knowledge the agent has, which influences the model it builds. * What data and feedback on that data is available.
Chapter 2 described several agent designs. The components of these agents include: 1. A direct mapping from conditions on the current state to actions. 2. A means to infer relevant properties of the world from the percept sequence. 3. Information about the way the world evolves and about the results of possible actions
the agent can take.
Machine learning
652
Chapter 19 Learning from Examples 4. Utility information indicating the desirability of world states. 5. Action-value information indicating the desirability of actions.
6. Goals that describe the most desirable states.
7. A problem generator, critic, and learning element that enable the system to improve. Each of these components can be learned. Consider a self-driving car agent that learns by observing a human driver. Every time the driver brakes, the agent might learn a condition— action rule for when to brake (component 1). By seeing many camera images that it is told contain buses, it can learn to recognize them (component 2).
By trying actions and ob-
serving the results—for example, braking hard on a wet road—it can learn the effects of its actions (component 3). Then, when it receives complaints from passengers who have been thoroughly shaken up during the trip, it can learn a useful component of its overall utility function (component 4).
The technology of machine learning has become a standard part of software engineering.
Any time you are building a software system, even if you don’t think of it as an AI agent,
components of the system can potentially be improved with machine learning. For example,
software to analyze images of galaxies under gravitational lensing was speeded up by a factor
of 10 million with a machine-learned model (Hezaveh et al., 2017), and energy use for cooling data centers was reduced by 40% with another machine-learned model (Gao, 2014). Turing Award winner David Patterson and Google Al head Jeff Dean declared the dawn of a “Golden
Age” for computer architecture due to machine learning (Dean et al., 2018).
We have seen several examples of models for agent components: atomic, factored, and
relational models based on logic or probability, and so on. Learning algorithms have been Prior knowledge
devised for all of these.
This chapter assumes little prior knowledge on the part of the agent: it starts from scratch
and learns from the data. In Section 21.7.2 we consider transfer learning, in which knowl-
edge from one domain is transferred to a new domain, so that learning can proceed faster with less data.
We do assume, however, that the designer of the system chooses a model
framework that can lead to effective learning.
Going from a specific set of observations to a general rule is called induction; from the
observations that the sun rose every day in the past, we induce that the sun will come up tomorrow.
This differs from the deduction we studied in Chapter 7 because the inductive
conclusions may be incorrect, whereas deductive conclusions are guaranteed to be correct if
the premises are correct. This chapter concentrates on problems where the input is a factored representation—a vector of attribute values.
It is also possible for the input to be any kind of data structure,
including atomic and relational.
When the output is one of a finite set of values (such as sunny/cloudy/rainy or true/false),
Classification
the learning problem is called classification. When it is a number (such as tomorrow’s tem-
Regression
mittedly obscure!) name regression. ! A better name would have been function approximation or mumeric prediction. But in 1886 Francis Galton ‘wrote an influential article on the concept of regression o the mean (e.g.. the children of tall parents are likely to be taller than average, but not as tall as the parents). Galton showed plots with what he called “regression lines,” and readers came to associate the word “regression” with the statistical technique of function approximation rather than with the topic of regression to the mean.
perature, measured either as an integer or a real number), the learning problem has the (ad-
Section 1.2 Supervised Learning
653
There are three types of feedback that can accompany the inputs, and that determine the Feedback
three main types of learning:
« In supervised learning the agent observes input-output pairs and learns a function that Supervised learning maps from input to output.
For example, the inputs could be camera images, each
one accompanied by an output saying “bus” or “pedestrian,” etc. An output like this is called a label. The agent learns a function that, when given a new image, predicts
the appropriate label. In the case of braking actions (component 1 above), an input is
Label
the current state (speed and direction of the car, road condition), and an output is the distance it took to stop. In this case a set of output values can be obtained by the agent
from its own percepts (after the fact); the environment is the teacher, and the agent learns a function that maps states to stopping distance.
« In unsupervised learning the agent learns patterns in the input without any explicit Unsupervised learning feedback. The most common unsupervised learning task is clustering: detecting poten-
tially useful clusters of input examples. For example, when shown millions of images
taken from the Internet, a computer vision system can identify a large cluster of similar images which an English speaker would call “cats.”
«+ In reinforcement learning the agent learns from a series of reinforcements:
rewards
and punishments. For example, at the end of a chess game the agent is told that it has
Reinforcement learning
won (a reward) or lost (a punishment). It is up to the agent to decide which of the actions prior to the reinforcement were most responsible for it, and to alter its actions
to aim towards more rewards in the future.
19.2
Super
ed Learning
More formally, the task of supervised learning is this: Given a training set of N example input-output pairs
Training set
(1, y1)s (32,32)s - (s yn) 5 where each pair was generated by an unknown function y = f(x),
discover a function / that approximates the true function f. The function / is called a hypothesis about the world. It is drawn from a hypothesis space H of possible functions. For example, the hypothesis space might be the set of polynomials
Hypothesis space
of degree 3; or the set of Javascript functions; or the set of 3-SAT Boolean logic formulas.
‘With alternative vocabulary, we can say that / is a model of the data, drawn from a model class 7, or we can say a function drawn from a function class. We call the output y; the
Model class Ground truth
ground truth—the true answer we are asking our model to predict. How do we choose a hypothesis space? We might have some prior knowledge about the data process that generated the data. If not, we can perform exploratory data analysis: examining Exploratory analysis the data with statistical tests and visualizations—histograms, scatter plots, box plots—to get
afeel for the data, and some insight into what hypothesis space might be appropriate. Or we can just try multiple hypothesis spaces and evaluate which one works best.
How do we choose a good hypothesis from within the hypothesis space? We could hope
for a consistent hypothesis: and / such that each x; in the training set has h(x;) = y;. With
continuous-valued outputs we can’t expect an exact match to the ground truth; instead we
Consistent hypothesis
654
Chapter 19 Learning from Examples Sinusoidal
Piecewise linear
Degree-12 polynomial
o
Data set 2
Data set 1
Linear
flf}
Figure 19.1 Finding hypotheses to fit data. Top row: four plots of best-fit functions from four different hypothesis spaces trained on data set 1. Bottom row: the same four functions, but trained on a slightly different data
set (sampled from the same f(x) function).
look for a best-fit function for which each h(x;) is close to y; (in a way that we will formalize in Section 19.4.2).
The true measure of a hypothesis is not how it does on the training set, but rather how
Test set Generalization
well it handles inputs it has not yet seen. We can evaluate that with a second sample of (x;,y;) pairs called a test set. We say that / generalizes well if it accurately predicts the outputs of the test set. Figure 19.1 shows that the function that a learning algorithm discovers depends on the hypothesis space H it considers and on the training set it is given. Each of the four plots in the top row have the same training set of 13 data points in the (x,y) plane. The four plots in the bottom row have a second set of 13 data points; both sets are representative of the
same unknown function f(x). Each column shows the best-fit hypothesis / from a different hypothesis space:
o Column 1: Straight lines; functions of the form i(x) = wjx- wy. There is no line that
would be a consistent hypothesis for the data points.
« Column 2: Sinusoidal functions of the form A(x) = wy.x -+sin(wox). This choice is not quite consistent, but fits both data sets very well.
e Column 3: Piecewise-linear functions where each line segment connects the dots from
one data point to the next. These functions are always consistent. o Column 4: Degree-12 polynomials, h(x) = ¥/2gwix’. These are consistent: we can always get a degree-12 polynomial to perfectly fit 13 distinct points. But just because
the hypothesis is consistent does not mean it is a good guess. One way to analyze hypothesis spaces is by the bias they impose (regardless of the train-
Bias
ing data set) and the variance they produce (from one training set to another). By bias we mean (loosely) the tendency of a predictive hypothesis to deviate from the
expected value when averaged over different training sets. Bias often results from restrictions
Section 1.2 Supervised Learning
655
imposed by the hypothesis space. For example, the hypothesis space of linear functions
induces a strong bias: it only allows functions consisting of straight lines. If there are any
patterns in the data other than the overall slope of a
line, a linear function will not be able
to represent those patterns. We say that a hypothesis is underfitting when it fails to find a Underfitting pattern in the data. On the other hand, the piecewise linear function has low bias; the shape of the function is driven by the data.
By variance we mean the amount of change in the hypothesis due to fluctuation in the
training data. The two rows of Figure 19.1 represent data sets that were each sampled from the same f(x) function. The data sets turned out to be slightly different.
Variance
For the first three
columns, the small difference in the data set translates into a small difference in the hypothe-
sis. We call that low variance. But the degree-12 polynomials in the fourth column have high
variance: look how different the two functions are at both ends of the x-axis. Clearly, at least
one of these polynomials must be a poor approximation to the true f(x). We say a function
is overfitting the data when it pays too much attention to the particular data set it is trained
on, causing it to perform poorly on unseen data.
Often there is a bias-variance tradeoff: a choice between more complex, low-bias hy-
potheses that fit the training data well and simpler, low-variance hypotheses that may generalize better.
Albert Einstein said in 1933, “the supreme goal of all theory is to make the
irreducible basic elements as simple and as few as possible without having to surrender the adequate representation of a single datum of experience.” In other words, Einstein recommends choosing the simplest hypothesis that matches the data. This principle can be traced further back to the 14th-century English philosopher William of Ockham.? His principle that “plurality [of entities] should not be posited without necessity” is called Ockham’s razor because it is used to “shave off”” dubious explanations.
Defining simplicity is not easy. It seems clear that a polynomial with only two parameters is simpler than one with thirteen parameters. We will make this intuition more precise in
Section
19.3.4.
However, in Chapter 21 we will see that deep neural network models can
often generalize quite well, even though they are very complex—some of them have billions of parameters. So the number of parameters by itself is not a good measure of a model’s fitness. Perhaps we should be aiming for “appropriateness,” not “simplicity” in a model class. We will consider this issue in Section 19.4.1. ‘Which hypothesis is best in Figure 19.1? We can’t be certain.
If we knew the data
represented, say, the number of hits to a Web site that grows from day to day, but also cycles depending on the time of day, then we might favor the sinusoidal function.
If we knew the
data was definitely not cyclic but had high noise, that would favor the linear function.
In some cases, an analyst is willing to say not just that a hypothesis is possible or im-
possible, but rather how probable it is. Supervised learning can be done by choosing the hypothesis 4 that is most probable given the data:
1" = argmax P(h|data) . heM
By Bayes’ rule this s equivalent to I* = argmax P(datal ) P(h) . heM
2 The name is often misspelled as “Occam.”
Overfitting Bias-variance tradeoff
Chapter 19 Learning from Examples Then we can say that the prior probability P() is high for a smooth degree-1 and lower for a degree-12 polynomial with large, sharp spikes. We allow functions when the data say we really need them, but we discourage them low prior probability. Why not let H be the class of all computer programs, or all Turing
or -2 polynomial unusual-looking by giving them a machines? The
problem is that there is a tradeoff between the expressiveness of a hypothesis space and the
computational complexity of finding a good hypothesis within that space. For example, fitting a straight line to data is an easy computation; fitting high-degree polynomials is somewhat harder; and fitting Turing machines is undecidable. A second reason to prefer simple hypothesis spaces is that presumably we will want to use h after we have learned it, and computing h(x) when h is a linear function is guaranteed to be fast, while computing an arbitrary Turing machine program is not even guaranteed to terminate.
For these reasons, most work on learning has focused on simple representations. In recent
years there has been great interest in deep learning (Chapter 21), where representations are
not simple but where the h(x) computation still takes only a bounded number of steps to
compute with appropriate hardware.
We will see that the expressiveness—complexity tradeoff is not simple: it is often the case,
as we saw with first-order logic in Chapter 8, that an expressive language makes it possible for a simple hypothesis to fit the data, whereas restricting the expressiveness of the language
means that any consistent hypothesis must be complex. 19.2.1
Example problem:
Restaurant waiting
We will describe a sample supervised learning problem in detail: the problem of deciding
whether to wait for a table at a restaurant. This problem will be used throughout the chapter
to demonstrate different model classes. For this problem the output, y, is a Boolean variable
that we will call WillWait; it is true for examples where we do wait for a table. The input, x,
is a vector of ten attribute values, each of which has discrete values:
1. Alternate: whether there is a suitable alternative restaurant nearby.
0PN AE W
656
Bar: whether the restaurant has a comfortable bar area to wait in. Fri/Sat: true on Fridays and Saturdays. Hungry: whether we are hungry right now.
Patrons: how many people are in the restaurant (values are None, Some, and Full). Price: the restaurant’s price range ($, $3, $$8). Raining: whether it is raining outside.
Reservation: whether we made a reservation.
Type: the kind of restaurant (French, Italian, Thai, or burger).
10. WaitEstimate: host’s wait estimate: 0-10, 10-30, 3060, or >60 minutes.
A set of 12 examples, taken from the experience of one of us (SR), is shown in Figure 19.2.
Note how skimpy these data are: there are 26 x 3% x 4% = 9,216 possible combinations of
values for the input attributes, but we are given the correct output for only 12 of them; each of
the other 9,204 could be either true or false; we don’t know. This is the essence of induction: we need to make our best guess at these missing 9,204 output values, given only the evidence of the 12 examples.
Section 19.3
Example
Learning Decision Trees
657
Input Attribut Alt Yes No Yes No
Yes
No
No Yes No
Some
$3$
Some
%$Full $ Some $ Full $ Full ~ $8$
None Some Full Full None Full
Figure 19.2 Examples for the restaurant domain. 19.3
Learning Decision Trees
A decision tree is a representation of a function that maps a vector of attribute values to
a single output value—a “decision.” A decision tree reaches its decision by performing a
Decision tree
sequence of tests, starting at the root and following the appropriate branch until a leaf is reached. Each internal node in the tree corresponds to a test of the value of one of the input
attributes, the branches from the node are labeled with the possible values of the attribute,
and the leaf nodes specify what value s to be returned by the function. In general, the input and output values can be discrete or continuous, but for now we will
Positive example) or false (a negative example). We call this Boolean classification. We will use j Negative consider only inputs consisting of discrete values and outputs that are either true (a positive
to index the examples (x; is the input vector for the jth example and y; is the output), and x;;
for the ith attribute of the jth example.
The tree representing the decision function that SR uses for the restaurant problem is
shown in Figure 19.3. Following the branches, we see that an example with Patrons = Full and WaitEstimate =0-10 will be classified as positive (i.e., yes, we will wait for a table).
19.3.1
Expressiveness of decision trees
A Boolean decision tree is equivalent to a logical statement of the form:
Output
A +A; is hard to represent
with a decision tree because the decision boundary is a diagonal line, and all decision tree
tests divide the space up into rectangular, axis-aligned boxes. We would have to stack a lot
of boxes to closely approximate the diagonal line. In other words, decision trees are good for some kinds of functions and bad for others. Is there any kind of representation that is efficient for all kinds of functions?
Unfortu-
nately, the answer is no—there are just too many functions to be able to represent them all with a small number of bits.
Even just considering Boolean functions with n Boolean at-
tributes, the truth table will have 2" rows, and each row can output true or false, so there are
22" different functions. With 20 attributes there are 248576 ~ 10300000 fynctions, so if we limit ourselves to a million-bit representation, we can’t represent all these functions.
19.3.2
Learning decision trees from examples
‘We want to find a tree that is consistent with the examples in Figure 19.2 and is as small as possible. Unfortunately, it is intractable to find a guaranteed smallest consistent tree. But with some simple heuristics, we can efficiently find one that is close to the smallest.
The
LEARN-DECISION-TREE algorithm adopts a greedy divide-and-conquer strategy: always test the most important attribute first, then recursively solve the smaller subproblems that are defined by the possible results of the test. By “most important attribute,” we mean the one
that makes the most difference to the classification of an example. That way, we hope to get to the correct classification with a small number of tests, meaning that all paths in the tree will be short and the tree as a whole will be shallow.
Figure 19.4(a) shows that Type is a poor attribute, because it leaves us with four possible
Pations? None
Some"
o N Ves|
Full
WaitEstimate? Altemate? No Yes
Reservation? No Yes
No
No /"\ Yes Yes
Bai No /"\ Yes
Figure 19.3 A decision tree for deciding whether to wait for a table.
Section 19.3
8
alian
659
mEan BE@EBD
Type? French
Learning Decision Trees
Patrons?
Thai
Burger
©lad ao
None
Some
Ful
.
No// (@)
(b)
\Yes
B
Figure 19.4 Splitting the examples by testing on attributes. At each node we show the positive (light boxes) and negative (dark boxes) examples remaining. (a) Splitting on Type brings us no nearer to distinguishing between positive and negative examples. (b) Splitting on Patrons does a good job of separating positive and negative examples. After splitting on Patrons, Hungry is a fairly good second test. outcomes, each of which has the same number of positive as negative examples. On the other
hand, in (b) we see that Patrons is a fairly important attribute, because if the value is None or Some, then we are left with example sets for which we can answer definitively (No and Yes,
respectively). If the value is Full, we are left with a mixed set of examples. There are four cases to consider for these recursive subproblems:
1. If the remaining examples are all positive (or all negative), then we are done: we can answer Yes or No. Figure 19.4(b) shows examples of this happening in the None and Some branches.
2. If there are some positive and some negative examples, then choose the best attribute to split them. Figure 19.4(b) shows Hungry being used to split the remaining examples. 3. If there are no examples left, it means that no example has been observed for this com-
bination of attribute values, and we return the most common output value from the set
of examples that were used in constructing the node’s parent. 4. If there are no attributes left, but both positive and negative examples, it means that these examples have exactly the same description, but different classifications. This can
happen because there is an error or noise in the data; because the domain is nondeterministic; or because we can’t observe an attribute that would distinguish the examples.
The best we can do is return the most common output value of the remaining examples. The LEARN-DECISION-TREE algorithm is shown in Figure 19.5. Note that the set of exam-
ples is an input to the algorithm, but nowhere do the examples appear in the tree returned by the algorithm. A tree consists of tests on attributes in the interior nodes, values of attributes on the branches, and output values on the leaf nodes. The details of the IMPORTANCE func-
tion are given in Section 19.3.3. The output of the learning algorithm on our sample training
set is shown in Figure 19.6. The tree is clearly different from the original tree shown in Fig-
Noise
660
Chapter 19 Learning from Examples function LEARN-DECISION-TREE(examples, attributes, parent_examples) returns a tree
if examples is empty then return PLURALITY-VALUE(parent_examples)
else if all examples have the same classification then return the classification
else if artributes is empty then return PLURALITY-VALUE(examples) else
A~ argmax, ¢ uipues IMPORTANCE (a, examples) tree < a new decision tree with root test A for each value v ofA do
exs 0, while carthquakes have —4.9+ 1.7x; —x, < x=
Section 19.6
Linear Regression and Classification
683
0. We can make the equation easier to deal with by changing it into the vector dot product form—with xo=1 we have —4.9%+ 17x -3 =0, and we can define the vector of weights, w=(-49,17,~1), and write the classification hypothesis Jw(x) = 1if w-x > 0 and 0 otherwise.
Alternatively, we can think of / as the result of passing the linear function w - x through a threshold function:
Iw(x) = Threshold(w-x) where Threshold(z) =1 if z > 0 and 0 otherwise.
Threshold function
The threshold function is shown in Figure 19.17(a).
Now that the hypothesis &y (x) has a well-defined mathematical form, we can think about
choosing the weights w to minimize the loss. In Sections 19.6.1 and 19.6.3, we did this both
in closed form (by setting the gradient to zero and solving for the weights) and by gradient
descent in weight space. Here we cannot do either of those things because the gradient is zero almost everywhere in weight space except at those points where w - x =0, and at those points the gradient is undefined.
There is, however, a simple weight update rule that converges to a solution—that is, to
a linear separator that classifies the data perfectly—provided the data are linearly separable. For a single example (x,y), we have wi — wita(y
(X)) X x;
(19.8)
which is essentially identical to Equation (19.6), the update rule for linear regression!
This
rule is called the perceptron learning rule, for reasons that will become clear in Chapter 21. Because we are considering a 0/1 classification problem, however, the behavior is somewhat
Perceptron learning rule
different. Both the true value y and the hypothesis output /(x) are either 0 or 1, so there are three possibilities:
« If the output is correct (i.e., y=hy(x)) then the weights are not changed.
« Ifyis 1 but hy(x) is 0, then w; is increased when the corresponding input x; is positive and decreased when x; is negative. This makes sense, because we want to make w - x
bigger so that /i (X) outputs a 1.
« If yis 0 but hy(x) is 1, then w; is decreased when the corresponding input x; is positive and increased when x; is negative.
This makes sense, because we want to make W - X
smaller so that A (x) outputs a 0.
Typically the learning rule is applied one example at a time, choosing examples at random (as in stochastic gradient descent). Figure 19.16(a) shows a training curve for this learning rule Training curve applied to the earthquake/explosion data shown in Figure 19.15(a). A training curve measures the classifier performance on a fixed training set as the learning process proceeds one update at a time on that training set. The curve shows the update rule converging to a zero-error
linear separator. The “convergence” process isn’t exactly preity, but it always works. This
particular run takes 657 steps to converge, for a data set with 63 examples, so each example
is presented roughly 10 times on average. Typically, the variation across runs is large.
684
Chapter 19 Learning from Examples 5 £
P Zoo
51 209
§
507
§07
8 0,
£ o, z &
Sos
0 100200300400500600 700 Number of weight updates (@)
06 20 = 04
Sos
025000 50000 75000 Number of weight updates
€06 08 h=gg
The solution for @ is the same as before.
S
The solution for 6}, the probability that a cherry
candy has a red wrapper, is the observed fraction of cherry candies with red wrappers, and
similarly for 6.
These results are very comforting, and it is easy to see that they can be extended to any
Bayesian network whose conditional probabilities are represented as tables. The most impor-
Section20.2
Learning with Complete Data
727
tant point is that with complete data, the maximum-likelihood parameter learning problem
for a Bayesian network decomposes into separate learning problems, one for each parameter. (See Exercise 20.NORX
for the nontabulated case, where each parameter affects several
conditional probabilities.) The second point is that the parameter values for a variable, given its parents, are just the observed frequencies of the variable values for each setting of the parent values. As before, we must be careful to avoid zeroes when the data set is small.
20.2.2
Naive Bayes models
Probably the most common Bayesian network model used in machine learning is the naive Bayes model first introduced on page 402. In this model, the “class™ variable C (which is to be predicted) is the root and the “attribute™ variables X; are the leaves. The model is “naive”
because it assumes that the attributes are conditionally independent of each other, given the class. (The model in Figure 20.2(b) is a naive Bayes model with class Flavor and just one
attribute, Wrapper.) In the case of Boolean variables, the parameters are
0=P(C=true),0y =P(X;=true|C=true),0,, = P(X; = true| C =false). The maximum-likelihood parameter values are found in exactly the same way as in Figure 20.2(b). Once the model has been trained in this way, it can be used to classify new examples for which the class variable C is unobserved. With observed attribute values ¥y, .., %, the probability of each class is given by P(C|x1,...,%,) = a P(C ]'[Px,\c
A deterministic prediction can be obtained by choosing the most likely class. Figure 20.3 shows the learning curve for this method when it is applied to the restaurant problem from
Chapter 19. The method learns fairly well but not as well as decision tree learning; this is presumably because the true hypothesis—which is a decision tree—is not representable exactly
using a naive Bayes model. Naive Bayes learning turns out to do surprisingly well in a wide
range of applications; the boosted version (Exercise 20.BNBX) is one of the most effective
general-purpose learning algorithms. Naive Bayes learning scales well to very large prob-
lems: with n Boolean attributes, there are just 2n+ 1 parameters, and no search is required
10 find hy, the maximum-likelihood naive Bayes hypothesis. Finally, naive Bayes learning systems deal well with noisy or missing data and can give probabilistic predictions when appropriate. Their primary drawback is the fact that the conditional independence assumption is seldom accurate; as noted on page 403, the assumption leads to overconfident probabilities that are often very close to 0 or 1, especially with large numbers of attributes.
20.2.3
Generative and discriminative models
We can distinguish two kinds of machine learning models used for classifiers: generative and discriminative.
A generative model models the probability distribution of each class.
For
example, the naive Bayes text classifier from Section 12.6.1 creates a separate model for each
possible category of text—one for sports, one for weather, and so on. Each model includes
the prior probability of the category—for example P(Category=weather)—as well as the conditional probability P(Inputs | Category =weather). From these we can compute the joint
probability P(Inputs, Category = weather)) and we can generate a random selection of words that is representative of texts in the weather category.
Generative model
Chapter 20 Learning Probabilistic Models
Proportion correct on fest set
728
Naive Bayes
0
20
40 60 Training set size
80
100
Figure 20.3 The learning curve for naive Bayes learning applied to the restaurant problem from Chapter 19; the learning curve for decision tree learning is shown for comparison. Discriminative model
A discriminative model directly learns the decision boundary between classes. That is,
it learns P(Category | Inputs). Given example inputs, a discriminative model will come up
with an output category, but you cannot use a discriminative model to, say, generate random
words that are representative of a category. Logistic regression, decision trees, and support vector machines are all discriminative model:
Since discriminative models put all their emphasis on defining the decision boundary— that is, actually doing the classification task they were asked to do—they tend to perform
better in the limit, with an arbitrary amount of training data. However, with limited data, in
some cases a generative model performs better. (Ng and Jordan, 2002) compare the generative naive Bayes classifier to the discriminative logistic regression classifier on 15 (small) data sets, and find that with the maximum amount of data, the discriminative model does better on
9 out of 15 data sets, but with only a small amount of data, the generative model does better
on 14 out of 15 data sets.
20.2.4
Maximum-likelihood parameter learning:
Continuous models
Continuous probability models such as the linear-Gaussian model were shown on page 422.
Because continuous variables are ubiquitous in real-world applications, it is important to know how to learn the parameters of continuous models from data. The principles for maximum-likelihood learning are identical in the continuous and discrete case:
Let us begin with a very simple case: learning the parameters of a Gaussian density function on a single variable. That is, we assume the data are generated as follows:
P = The parameters of this model are the mean 4 and the standard deviation o. (Notice that the
normalizing “constant” depends on o, so we cannot ignore it.) Let the observed values be
Section20.2
Learning with Complete Data
0 01020304 0506070809 x ®)
(a)
1
Figure 20.4 (a) A linear-Gaussian model described as y =6+ 6 plus Gaussian noise with fixed variance. (b) A set of 50 data points generated from this model and the best-fit line. xy. Then the log likelihood is
m?
L= Zlog e 5T= N(—logV27 — logo) — Setting the denvzuves to zero as usual, we obtain
LyN 5oL = —Flial-p=0 AL =_ S+ N, ELL(y-p)?=0 IyN 2 Se
@04)
That is, the maximum-likelihood value of the mean is the sample average and the maximum-
likelihood value of the standard deviation is the square root of the sample variance. Again, these are comforting results that confirm “commonsense” practice. Now consider a linear-Gaussian model with one continuous parent X and a continuous
child Y. As explained on page 422, ¥ has a Gaussian distribution whose mean depends linearly on the value ofX and whose standard deviation is fixed. distribution P(Y | X), we can maximize the conditional likelihood
POl =
1
_o 0o’
To learn the conditional
@5
Here, the parameters are 0y, 02, and o. The data are a collection of (x;, ;) pairs, as illustrated
in Figure 20.4. Using the usual methods (Exercise 20.LINR), we can find the maximumlikelihood values of the parameters. The point here is different. If we consider just the parameters #, and 0, that define the linear relationship between x and y, it becomes clear
that maximizing the log likelihood with respect to these parameters is the same as minimizing
the numerator (y— (6,x+6,))? in the exponent of Equation (20.5). This is the Ly loss, the
squared error between the actual value y and the prediction §;x + 5.
This is the quantity minimized by the standard linear regression procedure described in
Section 19.6. Now we can understand why: minimizing the sum of squared errors gives the maximum-likelihood straight-line model, provided that the data are generated with Gaussian
noise of fixed variance.
729
730
Chapter 20 Learning Probabilistic
0
02
04 06 Parameter § ()
Models
08
1
0
02
04 06 Parameter ®)
08
1
Figure 20.5 Examples of the Bera(a,b) distribution for different values of (a,b). 20.2.5
Bayesian parameter learning
Maximum-likelihood learning gives rise to simple procedures, but it has serious deficiencies with small data sets. For example, after seeing one cherry candy, the maximum-likelihood hypothesis is that the bag is 100% cherry (i.c., §=1.0). Unless one’s hypothesis prior is that bags must be either all cherry or all lime, this is not a reasonable conclusion. It is more likely
that the bag is a mixture of lime and cherry.
The Bayesian approach to parameter learning
starts with a hypothesis prior and updates the distribution as data arrive. The candy example in Figure 20.2(a) has one parameter, 0: the probability that a randomly selected piece of candy is cherry-flavored. In the Bayesian view, 0 is the (unknown)
value of a random variable © that defines the hypothesis space; the hypothesis prior is the
prior distribution over P(®). Thus, P(@=8) is the prior probability that the bag has a frac-
tion 0 of cherry candies.
Beta distribution
Hyperparameter
If the parameter 6 can be any value between 0 and 1, then P(®) is a continuous probability density function (see Section A.3). If we don’t know anything about the possible values of 6 we can use the uniform density function P(6) = Uniform(6;0, 1), which says all values are equally likely. A more flexible family of probability density functions is known as the beta distributions. Each beta distribution is defined by two hyperparameters® a and b such that Beta(0:a,b) = a 0!
(1-9)"",
(20.6)
for 0 in the range [0, 1]. The normalization constant a, which makes the distribution integrate to 1, depends on a and b. Figure 20.5 shows what the distribution looks like for various values of a and b. The mean value of the beta distribution is a/(a+ b), so larger values of a
suggest a belief that @ is closer to 1 than to 0. Larger values of a -+ b make the distribution
more peaked, suggesting greater certainty about the value of ©. It turns out that the uniform
density function is the same as Beta(1,1): the mean is 1/2, and the distribution is flat.
3 They are called hyperparameters because they parameterize a distribution over 0, which is itselfa parameter.
Section20.2
Learning with Complete Data
731
Figure 20.6 A Bayesian network that corresponds to a Bayesian learning process. Posterior distributions for the parameter variables ©, ©;, and @, can be inferred from their prior distributions and the evidence in Flavor; and Wrapper;. Besides its flexibility, the beta family has another wonderful property: if Beta(a, b), then, after a data point is observed, the posterior distribution for © distribution. In other words, Beta is closed under update. The beta family conjugate prior for the family of distributions for a Boolean variable.* Let’s works. Suppose we observe a cherry candy: then we have
© has a prior is also a beta is called the see how this Conjugate prior
P(0|Dy=cherry) = a P(Dy=cherry|0)P(6)
= o 0-Beta(f;a,b) = o’ 6-0°'(1-9)"" = ' 0°(1-0)""" = o/ Beta(:a+1,b).
Thus, after seeing a cherry candy, we simply increment the a parameter to get the posterior;
similarly, after seeing a lime candy, we increment the b parameter. Thus, we can view the @ and b hyperparameters as virtual counts, in the sense that a prior Beta(a, b) behaves exactly
as if we had started out with a uniform prior Beta(1,1) and seen a— 1 actual cherry candies and b — 1 actual lime candies. By examining a sequence of beta distributions for increasing values ofa and b, keeping the proportions fixed, we can see vividly how the posterior distribution over the parameter
© changes as data arrive. For example, suppose the actual bag of candy is 75% cherry. Figure 20.5(b) shows the sequence Beta(3,1), Beta(6,2), Beta(30,10). Clearly, the distribution is converging to a narrow peak around the true value of ©. For large data sets, then, Bayesian learning (at least in this case) converges to the same answer as maximum-likelihood learning. Now let us consider a more complicated case. The network in Figure 20.2(b) has three parameters, 6, 01, and 6, where 6 is the probability ofa red wrapper on a cherry candy and 4 Other conjugate priors include the Dirichlet family for the parameters of a discrete multivalued distribution and the Normal-Wishart family for the parameters of a Gaussian ion. See Bernardo and Smith (1994).
Virtual count
732
Chapter 20 Learning Probabilistic
Models
0, is the probability of a red wrapper on a lime candy. The Bayesian hypothesis prior must
cover all three parameters—that is, we need to specify P(®,0,,0;).
Parameter independence
parameter independence:
Usually, we assume
P(0,0,.0,) = P(O)P(©,)P(©,).
‘With this assumption, each parameter can have its own beta distribution that is updated sepa-
rately as data arrive. Figure 20.6 shows how we can incorporate the hypothesis prior and any data into a Bayesian network, in which we have a node for each parameter variable. The nodes ©,0;,0, have no parents. For the ith observation of a wrapper and corresponding flavor of a piece of candy, we add nodes Wrapper; and Flavor;. Flavor; is dependent on the flavor parameter ©:
P(Flavor;=cherry|©=0) = 0. Wrapper; is dependent on ©; and ©,: P(Wrapper; = red| Flavor; =cherry,©,=6,) = 6 P(Wrapper; = red| Flavor; =lime,® = 0,) = 0, .
Now, the entire Bayesian learning process for the original Bayes net in Figure 20.2(b) can be formulated as an inference problem in the derived Bayes net shown in Figure 20.6, where the.
P>
data and parameters become nodes. Once we have added all the new evidence nodes, we can
then query the parameter variables (in this case, ©,©;,0,). Under this formulation there is just one learning algorithm—the inference algorithm for Bayesian networks. Of course, the nature of these networks is somewhat different from those of Chapter 13
because of the potentially huge number of evidence variables representing the training set and the prevalence of continuous-valued parameter variables. Exact inference may be impos-
sible except in very simple cases such as the naive Bayes model. Practitioners typically use approximate inference methods such as MCMC
(Section 13.4.2); many statistical software
packages incorporate efficient implementations of MCMC for this purpose. 20.2.6
Bayesian linear regression
Here we illustrate how to apply a Bayesian approach to a standard statistical task:
linear
regression. The conventional approach was described in Section 19.6 as minimizing the sum of squared errors and reinterpreted in Section 20.2.4 as maximizing likelihood assuming a Gaussian error model. These produce a single best hypothesis: a straight line with specific values for the slope and intercept and a fixed variance for the prediction error at any given
point. There is no measure of how confident one should be in the slope and intercept values.
Furthermore, if one is predicting a value for an unseen data point far from the observed
data points, it seems to make no sense to assume a prediction error that is the same as the
prediction error for a data point right next to an observed data point.
It would seem more
sensible for the prediction error to be larger, the farther the data point is from the observed
data, because a small change in the slope will cause a large change in the predicted value for a distant point. The Bayesian approach fixes both of these problems. The general idea, as in the preceding section, is to place a prior on the model parameters—here, the coefficients of the linear model and the noise variance—and then to compute the parameter posterior given the data. For
multivariate data and unknown noise model, this leads to rather a lot of linear algebra, so we
Section20.2
Learning with Complete Data
733
focus on a simple case: univariable data, a model that is constrained to go through the origin, and known noise:
and the model is
a normal distribution with variance o2. Then we have just one parameter ¢ 5 (ML) .
P(y[x,0) = N (y:6x,07) =
(20.7)
As the log likelihood is quadratic in 0, the appropriate form for a conjugate prior on is also a Gaussian. This ensures that the posterior for ¢ will also be Gaussian. We’ll assume a mean 0 and variance o7 for the prior, so that
P(0)=N(9:€u. We don’t necessarily expect the top halves of photos to look like bottom halves, so there is a scale beyond which spatial invariance no longer holds.
Local spatial invariance can be achieved by constraining the / weights connecting a local
region to a unit in the hidden layer to be the same for each hidden unit. (That is, for hidden
units i and j, the weights wy ..., w;,; are the same as wi j,...,wy,;.)
This makes the hidden
units into feature detectors that detect the same feature wherever it appear in the image.
Typically, we want the first hidden layer to detect many kinds of features, not just one; so
for each local image region we might have d hidden units with d distinct sets of weights.
This means that there are d/ weights in all—a number that is not only far smaller than n?,
Convolutional neural network (CNN) Kernel Convolution
but is actually independent of n, the image size. Thus, by injecting some prior knowledge— namely, knowledge of adjacency and spatial invariance—we can develop models that have far fewer parameters and can learn much more quickly. A convolutional neural network (CNN) is one that contains spatially local connections,
at least in the early layers, and has patterns of weights that are replicated across the units
in each layer. A pattern of weights that is replicated across multiple local regions is called a kernel and the process of applying the kernel to the pixels of the image (or to spatially
organized units in a subsequent layer) is called convolution.*
Kernels and convolutions are easiest to illustrate in one dimension rather than two or
more, so we will assume an input vector x of size n, corresponding to n pixels in a one-
3 Similar ideas can be applied to process time-series data sources such as audio waveforms. These typically exhibit temporal invariance—a word sounds the same no matter what time of day it is uttered. Recurrent neural neworks (Setion 216 auomatiall exhibittemmpors invaranee. # In the terminology of signa correlation, not a convolution. But “convolution” is used within the field of neural networks.
Section 21.3
Convolutional Networks
Figure 21.4 An example ofa one-dimensional convolution operation with a kernel of size 1=3 and a stride s=2. The peak response is centered on the darker (lower intensity) input pixel. The results would usually be fed through a nonlinear activation function (not shown) before going to the next hidden layer. dimensional image, and a vector kernel k of size 1. (For simplicity we will assume that / is an odd number.) All the ideas carry over straightforwardly to higher-dimensional case: ‘We write the convolution operation using the * symbol, for example:
operation is defined as follows:
=
!
£
|k
i (141)/2 -
z = x
k.
The
(21.8)
In other words, for each output position i, we take the dot product between the kernel k and a
snippet of x centered on x; with width /.
The process is illustrated in Figure 21.4 for a kernel vector [+1,—1,+1], which detects a darker point in the 1D image. (The 2D version might detect a darker line.) Notice that in
this example the pixels on which the kernels are centered are separated by a distance of 2 pixels; we say the kernel is applied with a stride s=2. Notice that the output layer has fewer Stride pixels: because of the stride, the number of pixels is reduced from 1 to roughly n/s. (In two dimensions, the number of pixels would be roughly n/s,s,, where s, and s, are the strides in the x and y directions in the image.) We say “roughly” because of what happens at the edge of the image: in Figure 21.4 the convolution stops at the edges of the image, but one can also
pad the input with extra pixels (either zeroes or copies of the outer pixels) so that the kernel can be applied exactly |n/s| times. For small kernels, we typically use s=1, so the output
has the same dimensions as the image (see Figure 21.5). The operation of applying a kernel across an image can be implemented in the obvious way by a program with suitable nested loops; but it can also be formulated as a single matrix
operation, just like the application of the weight matrix in Equation (21.1). For example, the convolution illustrated in Figure 21.4 can be viewed as the following matrix multiplication: 0 -1 0
0 +1 +1
0 0 -1
0 0 +1
5
oo
-1+ 0 +1 0 0
§ =[9]. 4
21.9)
hom
4+l 0 0
In this weight matrix, the kernel appears in each row, shifted according to the stride relative
to the previous row, One wouldn’t necessarily construct the weight matrix explicitly—it is
761
762
Chapter 21
Deep Learning
Figure 21.5 The first two layers of a CNN for a 1D image with a kernel size /=3 and a stride s= 1. Padding is added at the left and right ends in order to keep the hidden layers the same size as the input. Shown in red is the receptive field ofa unit in the second hidden layer. Generally speaking, the deeper the unit, the larger the receptive field. mostly zeroes, after all—but the fact that convolution is a linear matrix operation serves as a
reminder that gradient descent can be applied easily and effectively to CNNG, just as it can to plain vanilla neural networks. As mentioned earlier, there will be d kernels, not just one; so, with a stride of 1, the
output will be d times larger.
This means that a two-dimensional input array becomes a
three-dimensional array of hidden units, where the third dimension is of size d.
It is im-
portant to organize the hidden layer this way, so that all the kernel outputs from a particular image location stay associated with that location. Unlike the spatial dimensions of the image,
Receptive field
however, this additional “kernel dimension” does not have any adjacency properties, so it does not make sense to run convolutions along it. CNNs were inspired originally by models of the visual cortex proposed in neuroscience. In those models, the receptive
field of a neuron is the portion of the sensory input that can
affect that neuron’s activation. In a CNN, the receptive field of a unit in the first hidden layer
is small—just the size of the kernel, i.e., / pixels.
In the deeper layers of the network, it
can be much larger. Figure 21.5 illustrates this for a unit in the second hidden layer, whose
receptive field contains five pixels. When the stride is 1, as in the figure, a node in the mth
hidden layer will have a receptive field of size (I — 1)m+ 1; so the growth is linear in m. (In a 2D image, each dimension of the receptive field grows linearly with m, so the area grows quadratically.) When the stride is larger than 1, each pixel in layer m represents s pixels in layer m — 1; therefore, the receptive field grows as O(ls™)—that is, exponentially with depth.
The same effect occurs with pooling layers, which we discuss next. 21.3.1
Pooling
Pooling and downsampling
A pooling layer in a neural network summarizes a set of adjacent units from the preceding layer with a single value. Pooling works just like a convolution layer, with a kernel size / and stride s, but the operation that is applied is fixed rather than learned. Typically, no activation function is associated with the pooling layer. There are two common forms of pooling: * Average-pooling computes the average value of its / inputs.
Downsampling
This is identical to con-
volution with a uniform kernel vector k= (1/1,..., 1/I]. If we set [ =s, the effect is to coarsen the resolution of the image—to downsample it—by a factor of s. An object
that occupied, say, 10s pixels would now occupy only 10 pixels after pooling. The same
Section 21.3
Convolutional Networks
learned classifier that would be able to recognize the object at original image would now be able to recognize that object in if it was too big to recognize in the original image. In other facilitates multiscale recognition. It also reduces the number
a size of 10 pixels in the the pooled image, even words, average-pooling of weights required in
subsequent layers, leading to lower computational cost and possibly faster learning.
* Max-pooling computes the maximum value of its / inputs.
It can also be used purely
for downsampling, but it has a somewhat different semantics. Suppose we applied maxpooling to the hidden layer [5,9,4] in Figure 21.4: the result would be a 9, indicating that somewhere in the input image there is a darker dot that is detected by the kernel.
In other words, max-pooling acts as a kind of logical disjunction, saying that a feature
exists somewhere in the unit’s receptive field.
If the goal is to classify the image into one of ¢ categories, then the final layer of the network will be a softmax with ¢ output units.
The early layers of the CNN
are image-sized, so
somewhere in between there must be significant reductions in layer size. Convolution layers and pooling layers with stride larger than 1 all serve to reduce the layer size. It’s also possible to reduce the layer size simply by having a fully connected layer with fewer units than the preceding layer. CNNs often have one or two such layers preceding the final softmax layer. 21.3.2
Tensor operations in CNNs
We saw in Equations (21.1) and (21.3) that the use of vector and matrix notation can be helpful in keeping mathematical derivations simple and elegant and providing concise descriptions of computation graphs. Vectors and matrices are one-dimensional and two-dimensional special cases of tensors, which (in deep learning terminology) are simply multidimensional arrays Tensor of any dimension.’ For CNNs, tensors are a way of keeping track of the “shape™ of the data as it progresses
through the layers of the network. This is important because the whole notion of convolution
depends on the idea of adjacency:
adjacent data elements are assumed to be semantically
related, so it makes sense to apply operators to local regions of the data. Moreover, with suitable language primitives for constructing tensors and applying operators, the layers them-
selves can be described concisely as maps from tensor inputs to tensor outputs.
A final reason for describing CNNs in terms of tensor operations is computational effi-
ciency: given a description of a network as a sequence of tensor operations, a deep learning software package can generate compiled code that is highly optimized for the underlying
computational substrate. Deep learning workloads are often run on GPUs (graphics processing units) or TPUs (tensor processing units), which make available a high degree of parallelism. For example, one of Google’s third-generation TPU pods has throughput equivalent
to about ten million laptops. Taking advantage of these capabilities is essential if one is train-
ing a large CNN on a large database of images. Thus, it is common to process not one image at a time but many images in parallel; as we will see in Section 21.4, this also aligns nicely with the way that the stochastic gradient descent algorithm calculates gradients with respect
to a minibatch of training examples.
Let us put all this together in the form of an example.
256 x 256 RGB
images with a minibatch size of 64.
atical definition of tensors requires that cert
Suppose we are training on
The input in this case will be a four-
ariances hold under a change of b:
763
764
Chapter 21
Deep Learning
dimensional tensor of size 256 x 256 x 3 x 64.
Feature map Channel
Then we apply 96 kernels of size 5x5x 3
with a stride of 2 in both x and y directions in the image. This gives an output tensor of size 128 x 128 x 96 x 64. Such a
tensor is often called a feature map, since it shows how each
feature extracted by a kernel appears across the entire image; in this case it is composed of
96 channels, where each channel carries information from one feature. Notice that unlike the
input tensor, this feature map no longer has dedicated color channels; nonetheless, the color information may still be present in the various feature channels if the learning algorithm finds
color to be useful for the final predictions of the network.
21.3.3 Residual network
Residual networks
Residual networks are a popular and successful approach to building very deep networks that avoid the problem of vanishing gradients.
Typical deep models use layers that learn a new representation at layer i by completely re-
placing the representation at layer i — 1. Using the matrix—vector notation that we introduced
in Equation (21.3), with z(!) being the values of the units in layer i, we have
20 = f(20) = g (WD)
Because each layer completely replaces the representation from the preceding layer, all of the layers must learn to do something useful. Each layer must, at the very least, preserve the
task-relevant information contained in the preceding layer. If we set W = 0 for any layer i, the entire network ceases to function. If we also set W(~!) = 0, the network would not even
be able to learn: layer i would not learn because it would observe no variation in its input from layer i — 1, and layer i — 1 would not learn because the back-propagated gradient from
layer i would always be zero. Of course, these are extreme examples, but they illustrate the
need for layers to serve as conduits for the signals passing through the network.
The key idea of residual networks is that a layer should perturb the representation from the previous layer rather than replace it entirely. If the learned perturbation is small, the next
layer is close to being a copy of the previous layer. This is achieved by the following equation for layer i in terms of layer i — 1:
20 = g (@D 4 @),
Residual
1.10)
where g, denotes the activation functions for the residual layer. Here we think of f as the
residual, perturbing the default behavior of passing layer i — 1 through to layer i. The function used to compute the residual is typically a neural network with one nonlinear layer combined
with one linear layer:
f(z) =Vg(Wz), where W and V are learned weight matrices with the usual bias weights added.
Residual networks make it possible to learn significantly deeper networks reliably. Consider what happens if we set V=0 for a particular layer in order to disable that layer. Then the residual f disappears and Equation (21.10) simplifies to 20 = g,z ). Now suppose that g, con:
function to its inputs: z
s of ReLU activation functions and that z¢~!) also applies a ReLU =ReLU(in"""). In that case we have
21 = g,(z0V) = ReLU(2"")) = ReLU(ReLU(in"""))) = ReLU(in(""1) = 2(-1) |
Section 21.4 Learning Algorithms where the penultimate step follows because ReLU(ReLU(x))=ReLU(x).
In other words,
in residual nets with ReLU activations, a layer with zero weights simply passes its inputs
through with no change. The rest of the network functions just as if the layer had never existed. Whereas traditional networks must learn to propagate information and are subject to catastrophic failure of information propagation for bad choices of the parameters, residual
networks propagate information by default. Residual networks
are often used with convolutional layers in vision applications, but
they are in fact a general-purpose tool that makes deep networks more robust and allows
researchers to experiment more freely with complex and heterogeneous network designs. At the time of writing, it is not uncommon
to see residual networks with hundreds of layers.
The design of such networks is evolving rapidly, so any additional specifics we might provide would probably be outdated before reaching printed form. Readers desiring to know the best architectures for specific applications should consult recent research publications.
21.4
Learning Algorithms
Training a neural network consists of modifying the network’s parameters so as to minimize
the loss function on the training set. In principle, any kind of optimization algorithm could
be used. In practice, modern neural networks are almost always trained with some variant of
stochastic gradient descent (SGD).
‘We covered standard gradient descent and its stochastic version in Section 19.6.2. Here,
the goal i to minimize the loss L(w), where w represents all of the parameters of the network. Each update step in the gradient descent process looks like this: W
w—aVyL(w),
where « is the learning rate. For standard gradient descent, the loss L is defined with respect
to the entire training set. For SGD, it is defined with respect to a minibatch of m examples chosen randomly at each step.
As noted in Section 4.2, the literature on optimization methods for high-dimensional
continuous spaces includes innumerable enhancements to basic gradient descent.
We will
not cover all of them here, but it is worth mentioning a few important considerations that are
particularly relevant to training neural networks:
+ For most networks that solve real-world problems, both the dimensionality of w and the
size of the training set are very large. These considerations militate strongly in favor
of using SGD with a relatively small minibatch size m: stochasticity helps the algo-
rithm escape small local minima in the high-dimensional weight space (as in simulated annealing—see page 114); and the small minibatch size ensures that the computational
cost of each weight update step is a small constant, independent of the training set size. * Because the gradient contribution of each training example in the SGD minibatch can
be computed independently, the minibatch size is often chosen so as to take maximum advantage of hardware parallelism in GPUs or TPUs.
« To improve convergence, it is usually a good idea to use a learning rate that decreases over time. Choosing the right schedule is usually a matter of trial and error. + Near a local or global minimum of the loss function with respect to the entire training set, the gradients estimated from small minibatches will often have high variance and
765
766
Chapter 21
Deep Learning
Figure 21.6 Tllustration of the back-propagation of gradient information in an arbitrary computation graph. The forward computation of the output of the network proceeds from left to right, while the back-propagation of gradients proceeds from right to left. may point in entirely the wrong direction, making convergence difficult. One solution
Momentum
is to increase the minibatch size as training proceeds; another is to incorporate the idea
of momentum, which keeps a running average of the gradients of past minibatches in order to compensate for small minibatch sizes.
+ Care must be taken to mitigate numerical instabilities that may arise due to overflow,
underflow, and rounding error. These are particularly problematic with the use of exponentials in softmax,
sigmoid,
and tanh activation functions, and with the iterated
computations in very deep networks and recurrent networks (Section 21.6) that lead to
vanishing and exploding activations and gradients. Overall, the process of learning the weights of the network is usually one that exhibits diminishing returns.
We run until it is no longer practical to decrease the test error by running
longer. Usually this does not mean we have reached a global or even a local minimum of the loss function. Instead, it means we would have to make an impractically large number of very small steps to continue reducing the cost, or that additional steps would only cause overfitting, or that estimates of the gradient are too inaccurate to make further progress.
21.4.1
Computing
gradients in computation graphs
On page 755, we derived the gradient of the loss function with respect to the weights in a specific (and very simple) network. We observed that the gradient could be computed by back-propagating error information from the output layer of the network to the hidden layers. ‘We also said that this result holds in general for any feedforward computation graph. Here,
we explain how this works. Figure 21.6 shows a generic node in a computation graph. (The node & has in-degree and out-degree 2, but nothing in the analysis depends on this.) During the forward pass, the node
computes some arbitrary function / from its inputs, which come from nodes f and g. In turn, h feeds its value to nodes j and k.
The back-propagation process passes messages back along each link in the network. At each node, the incoming messages are collected and new messages are calculated to pass
Section 21.4 Learning Algorithms back to the next layer. As the figure shows, the messages are all partial derivatives of the loss
L. For example, the backward message dL/dh; is the partial derivative of L with respect to
Jj’s first input, which is the forward message from & to j. Now, / affects L through both j and k, so we have
AL/3h=OL/3h;+ AL/ .
@111y
‘With this equation, the node / can compute the derivative of L with respect to / by summing
the incoming messages from j and k. Now, to compute the outgoing messages dL/d fy, and 9L/dgp, we use the following equations:
OL
i
9L ah
ISy
and
AL
g,
JL Jh
9h dgi
(21.12)
In Equation (21.12), JL/dh was already computed by Equation (21.11), and 9h/d f;, and dh/dgy are just the derivatives of i with respect to its first and second arguments, respec-
tively. For example, if 4 is a multiplication node—that is, h(f,g)= f - g—then dh/d fi=g and 9h/dg, = f. Software packages for deep learning typically come with a library of node types (addition, multiplication, sigmoid, and so on), each of which knows how to compute its own derivatives as needed for Equation (21.12). The back-propagation process begins with the output nodes, where each initial message 9L/35; is calculated directly from the expression for L in terms of the predicted value § and the true value y from the training data.
At each internal node, the incoming backward
messages are summed according to Equation (21.11) and the outgoing messages are generated
from Equation (21.12). The process terminates at each node in the computation graph that represents a weight w (e.g., the light mauve ovals in Figure 21.3(b)). At that point, the sum of the incoming messages to w is JL/Jw—precisely the gradient we need to update w. Exercise 21.BPRE asks you to apply this process to the simple network in Figure 21.3 in order
to rederive the gradient expressions in Equations (21.4) and (21.5).
‘Weight-sharing, as used in convolutional networks (Section 21.3) and recurrent networks
(Section 21.6), is handled simply by treating each shared weight as a single node with multiple outgoing arcs in the computation graph. During back-propagation, this results in multiple incoming gradient messages. By Equation (21.11), this means that the gradient for the shared
weight is the sum of the gradient contributions from each place it is used in the network.
It is clear from this description of the back-propagation process that its computational
cost is linear in the number of nodes in the computation graph, just like the cost of the forward computation. Furthermore, because the node types are typically fixed when the network
is designed, all of the gradient computations can be prepared in symbolic form in advance
and compiled into very efficient code for each node in the graph.
Note also that the mes-
sages in Figure 21.6 need not be scalars: they could equally be vectors, matrices, or higher-
dimensional tensors, so that the gradient computations can be mapped onto GPUs or TPUs to benefit from parallelism.
One drawback of back-propagation is that it requires storing most of the intermediate
values that were computed during forward propagation in order to calculate gradients in the backward pass. This means that the total memory cost of training the network is proportional to the number of units in the entire network.
Thus, even if the network itself is represented
only implicitly by propagation code with lots of loops, rather than explicitly by a data struc-
ture, all of the intermediate results of that propagation code have to be stored explicitly.
767
768
Chapter 21 21.4.2
Batch normalization
Deep Learning
Batch normalization
Batch normalization is a commonly used technique that improves the rate of convergence of SGD by rescaling the values generated at the internal layers of the network from the examples within each minibatch.
Although the reasons for its effectiveness are not well understood at
the time of writing, we include it because it confers significant benefits in practice. To some
extent, batch normalization seems to have effects similar to those of the residual network. Consider a node z somewhere in the network: the values of z for the m examples in a
minibatch are zj., ... ,z,. Batch normalization replaces each z; with a new quantity Z;:
where 1 is the mean value of z across the minibatch, o is the standard deviation of zy, ...z,
€ is a small constant added to prevent division by zero, and and 3 are learned parameters.
Batch normalization standardizes the mean and variance of the values, as determined by the values of 3 and . This makes it much simpler to train a deep network. Without batch
normalization, information can get lost if a layer’s weights are too small, and the standard
deviation at that layer decays to near zero. Batch normalization prevents this from happening.
It also reduces the need for careful initialization of all the weights in the network to make sure
that the nodes in each layer are in the right operating region to allow information to propagate. With batch normalization, we usually include 3 and , which may be node-specific or layer-specific, among the parameters of the network, so that they are included in the learning process.
After training, 3 and ~ are fixed at their learned values.
21.5
Generalization
So far we have described how to fit a neural network to its training set, but in machine learn-
ing the goal is to generalize to new data that has not been seen previously, as measured by performance on a test set. In this section, we focus on three approaches to improving gener-
alization performance: choosing the right network architecture, penalizing large weights, and
randomly perturbing the values passing through the network during training. 21.5.1
Choosing a network architecture
A great deal of effort in deep learning research has gone into finding network architectures that generalize well. Indeed, for each particular kind of data—images, speech, text, video, and so on—a good deal of the progress in performance has come from exploring different kinds of network architectures and varying the number of layers, their connectivity, and the types of node in each layer.®
Some neural network architectures are explicitly designed to generalize well on particular
types of data: convolutional networks encode the idea that the same feature extractor is useful at all locations across a spatial grid, and recurrent networks encode the idea that the same
update rule is useful at all points in a stream of sequential data.
To the extent that these
assumptions are valid, we expect convolutional architectures to generalize well on images and recurrent networks to generalize well on text and audio signals.
& Noting that much of this incremental, exploratory work is arried out by graduate students, some have called the process graduate student descent (GSD).
Section 21.5
01
3.layer 1llayer
008 Test-set error
Generalization
006 004 002
o
1
2
3
4
5
Number of weights (x 10")
6
7
Figure 21.7 Test-set error as a function of layer width (as measured by total number of weights) for three-layer and eleven-layer convolutional networks. The data come from early versions of Google’s system for transcribing addresses in photos taken by Street View cars (Goodfellow et al., 2014). One of the most important empirical findings in the field of deep learning is that when
comparing two networks with similar numbers of weights, the deeper network usually gives better generalization performance.
Figure 21.7 shows this effect for at least one real-world
application—recognizing house numbers. The results show that for any fixed number of parameters, an eleven-layer network gives much lower test-set error than a three-layer network. Deep learning systems perform well on some but not all tasks. For tasks with high-
dimensional inputs—images, video, speech signals, etc.—they perform better than any other
pure machine learning approaches. Most of the algorithms described in Chapter 19 can handle high-dimensional input only if it is preprocessed using manually designed features to reduce the dimensionality. This preprocessing approach, which prevailed prior to 2010, has not yielded performance comparable to that achieved by deep learning systems. Clearly, deep learning models are capturing some important aspects of these tasks. In particular, their success implies that the tasks can be solved by parallel programs with a relatively
small number of steps (10 to 10° rather than, say, 107). This is perhaps not surprising, because these tasks are typically solved by the brain in less than a second, which is time enough for only a few tens of sequential neuron firings. Moreover, by examining the internal-layer representations learned by deep convolutional networks for vision tasks, we find evidence
that the processing steps seem to involve extracting a sequence of increasingly abstract representations of the scene, beginning with tiny edges, dots, and corner features and ending with
entire objects and arrangements of multiple objects. On the other hand, because they are simple circuits, deep learning models lack the compositional and quantificational expressive power that we see in first-order logic (Chapter 8) and context-free grammars (Chapter 23).
Although deep learning models generalize well in many cases, they may also produce
unintuitive errors. They tend to produce input-output mappings that are discontinuous, so
that a small change to an input can cause a large change in the output. For example, it may
769
770
Adversarial example
Chapter 21
Deep Learning
be possible to alter just a few pixels in an image of a dog and cause the network to classify the dog as an ostrich or a school bus—even though the altered image still looks exactly like a
dog. An altered image of this kind is called an adversarial example. In low-dimensional spaces it is hard to find adversarial examples. But for an image with
a million pixel values, it is often the case that even though most of the pixels contribute to
the image being classified in the middle of the “dog” region of the space, there are a few dimensions where the pixel value is near the boundary to another category. An adversary
with the ability to reverse engineer the network can find the smallest vector difference that
would move the image over the border.
When adversarial examples were first discovered, they set off two worldwide scrambles:
one to find learning algorithms and network architectures that would not be susceptible to adversarial attack, and another to create ever-more-effective adversarial attacks against all
kinds of learning systems.
So far the attackers seem to be ahead.
In fact, whereas it was
assumed initially that one would need access to the internals of the trained network in order
to construct an adversarial example specifically for that network, it has turned out that one can construct robust adversarial examples that fool multiple networks with different architec-
tures, hyperparameters, and training sets. These findings suggest that deep learning models
recognize objects in ways that are quite different from the human visual system. 21.5.2
Neural architecture search
Unfortunately, we don’t yet have a clear set of guidelines to help you choose the best network
architecture for a particular problem. Success in deploying a deep learning solution requires experience and good judgment. From the earliest days of neural network research, attempts have been made to automate
the process of architecture selection. We can think of this as a case of hyperparameter tuning (Section 19.4.4), where the hyperparameters determine the depth, width, connectivity, and
Neural architecture search
other attributes of the network.
However, there are so many choices to be made that simple
possible network architectures.
Many of the search techniques and learning techniques we
approaches like grid search can’t cover all possibilities in a reasonable amount of time. Therefore, it is common to use neural architecture search to explore the state space of covered earlier in the book have been applied to neural architecture search.
Evolutionary algorithms have been popular because it is sensible to do both recombination (joining parts of two networks together) and mutation (adding or removing a layer or changing a parameter value). Hill climbing can also be used with these same mutation operations.
Some researchers have framed the problem as reinforcement learning, and some
as Bayesian optimization.
Another possibility is to treat the architectural possibilities as a
continuous differentiable space and use gradient descent to find a locally optimal solution.
For all these search techniques, a major challenge is estimating the value of a candidate
network.
The straightforward way to evaluate an architecture is to train it on a test set for
multiple batches and then evaluate its accuracy on a validation set. But with large networks that could take many GPU-days.
Therefore, there have been many attempts to speed up this estimation process by eliminating or at least reducing the expensive training process. We can train on a smaller data set. We can train for a small number of batches and predict how the network would improve with more batches.
We can use a reduced version of the network architecture that we hope
Section 21.5
Generalization
771
retains the properties of the full version. We can train one big network and then search for subgraphs of the network that perform better; this search can be fast because the subgraphs
share parameters and don’t have to be retrained. Another approach is to learn a heuristic evaluation function (as was done for A* search).
That is, start by choosing a few hundred network architectures and train and evaluate them.
That gives us a data set of (network, score) pairs. Then learn a mapping from the features of a network to a predicted
score. From that point on we can generate a large number of candidate
networks and quickly estimate their value. After a search through the space of networks, the best one(s) can be fully evaluated with a complete training procedure.
21.5.3 Weight decay In Section 19.4.3 we saw that regularization—limiting the complexity of a model—can aid
generalization. This is true for deep learning models as well. In the context of neural networks
we usually call this approach weight decay.
Weight decay consists of adding a penalty AX; ;W
to the loss function used to train the
neural network, where \ is a hyperparameter controlling the strength of the penalty and the
sum is usually taken over all of the weights in the network. Using A=0 is equivalent to not using weight decay, while using larger values of A encourages the weights to become small. It is common to use weight decay with A near 104
Choosing a specific network architecture can be seen as an absolute constraint on the
hypothesis space: a function is either representable within that architecture or it is not. Loss
function penalty terms such as weight decay offer a softer constraint: functions represented
with large weights are in the function family, but the training set must provide more evidence in favor of these functions than is required to choose a function with small weights. It is not straightforward to interpret the effect of weight decay in a neural network. In
networks with sigmoid activation functions, it is hypothesized that weight decay helps to keep the activations near the linear part of the sigmoid, avoiding the flat operating region
that leads to vanishing gradients. With ReLU activation functions, weight decay seems to be beneficial, but the explanation that makes sense for sigmoids no longer applies because the ReLU’s output is either linear or zero.
Moreover, with residual connections, weight decay
encourages the network to have small differences between consecutive layers rather than
small absolute weight values.
Despite these differences in the behavior of weight decay
across many architectures, weight decay is still widely useful.
One explanation for the beneficial effect of weight decay is that it implements a form of maximum a posteriori (MAP) learning (see page 723). Letting X and y stand for the inputs
and outputs across the entire training set, the maximum a posteriori hypothesis /ap satisfies
Inap = argmax P(y| X, W)P(W) w = argmin[ log P(y|X, W) — log P(W)] The first term is the usual cross-entropy loss; the second term prefers weights that are likely
under a prior distribution. This aligns exactly with a regularized loss function if we set
logP(W) = -A Y W3, 7
which means that P(W) is a zero-mean Gaussian prior.
Weight decay
772
Chapter 21 21.5.4
Dropout
Deep Learning
Dropout
Another way that we can intervene to reduce the test-set error of a network—at the cost of making it harder to fit the training set—is to use dropout. At each step of training, dropout
applies one step of back-propagation learning to a new version of the network that is created
by deactivating a randomly chosen subset of the units. This is a rough and very low-cost approximation to training a large ensemble of different networks (see Section 19.8).
More specifically, let us suppose we are using stochastic gradient descent with minibatch
size m.
For each minibatch, the dropout algorithm applies the following process to every
node in the network: with probability p, the unit output is multiplied by a factor of 1/p;
otherwise, the unit output is fixed at zero. Dropout is typically applied to units in the hidden
layers with p=0.5; for input units, a value of p=0.8 turns out to be most effective. This process
produces a thinned network with about half as many units as the original, to which
back-propagation is applied with the minibatch of m training examples. The process repeats in the usual way until training is complete. At test time, the model is run with no dropout.
We can think of dropout from several perspectives:
+ By introducing noise at training time, the model is forced to become robust to noise. + As noted above, dropout approximates the creation of a large ensemble of thinned net-
works. This claim can be verified analytically for linear models, and appears to hold experimentally for deep learning models.
+ Hidden units trained with dropout must learn not only to be useful hidden units; they
must also learn to be compatible with many other possible sets of other hidden units
that may or may not be included in the full model.
This is similar to the selection
processes that guide the evolution of genes: each gene must not only be effective in its own function, but must work well with other genes, whose identity in future organisms
may vary considerably.
+ Dropout applied to later layers in a deep network forces the final decision to be made
robustly by paying attention to all of the abstract features of the example rather than focusing on just one and ignoring the others. For example, a classifier for animal images might be able to achieve high performance on the training set just by looking at the
animal’s nose, but would presumably fail on a test case where the nose was obscured or damaged. With dropout, there will be training cases where the internal “nose unit” is
zeroed out, causing the learning process to find additional identifying features. Notice
that trying to achieve the same degree of robustness by adding noise to the input data
would be difficult: there is no easy way to know in advance that the network is going to
focus on noses, and no easy way to delete noses automatically from each image.
Altogether, dropout forces the model to learn multiple, robust explanations for each input.
This causes the model to generalize well, but also makes it more difficult to fit the training
set—it is usually necessary to use a larger model and to train it for more iterations.
21.6
Recurrent Neural Networks
Recurrent neural networks (RNNs) are distinct from feedforward networks in that they allow
cycles in the computation graph. In all the cases we will consider, each cycle has a delay,
50 that units may take as input a value computed from their own output at an earlier step in
Section 21.6
(@)
Recurrent Neural Networks
773
(b)
Figure 21.8 (a) Schematic diagram of a basic RNN where the hidden layer z has recurrent connections; the A symbol indicates a delay. (b) The same network unrolled over three time steps to create a feedforward network. Note that the weights are shared across all time steps. the computation. (Without the delay, a cyclic circuit may reach an inconsistent state.) This
allows the RNN to have internal state, or memory: inputs received at earlier time steps affect
the RNN’s response to the current input. RNNs can also be used to perform more general computations—after all, ordinary com-
puters are just Boolean circuits with memory—and to model real neural systems, many of
which contain cyclic connections. Here we focus on the use of RNNs to analyze sequential data, where we assume that a new input vector X, arrives at each time step.
As tools for analyzing sequential data, RNNs can be compared to the hidden Markov
models, dynamic Bayesian networks, and Kalman filters described in Chapter 14. (The reader
may find it helpful to refer back to that chapter before proceeding.) Like those models, RNNs. make a Markov assumption (see page 463):
the hidden state z, of the network suffices
to capture the information from all previous inputs. Furthermore, suppose we describe the RNN'’s update process for the hidden state by the equation z, =f,,(z,_;,x;) for some param-
eterized function f,. Once trained, this function represents a time-homogeneous process
(page 463)—effectively a universally quantified assertion that the dynamics represented by
fw hold for all time steps. Thus, RNNs add expressive power compared to feedforward networks, just as convolutional networks do, and just as dynamic Bayes nets add expressive power compared to regular Bayes nets. Indeed, if you tried to use a feedforward network to analyze sequential data, the fixed size of the input layer would force the network to examine only a finite-length window of data, in which case the network would fail to detect
long-distance dependencies. 21.6.1
Training a basic RNN
The basic model we will consider has an input layer x, a hidden layer z with recurrent con-
nections, and an output layer y, as shown in Figure 21.8(a). We assume that both x and y are
observed in the training data at each time step. The equations defining the model refer to the values of the variables indexed by time step 7:
% = fo(z-1,%)=g (W21 + Wiexi) = g (inz,) 9 = g(Weyz) =gy(iny,),
(21.13)
Memory
774
Chapter 21
Deep Learning
where g, and g, denote the activation functions for the hidden and output layers, respectively.
As usual, we assume an extra dummy input fixed at +1 for each unit as well as bias weights
associated with those inputs. Given a sequence of input vectors Xi,...,x and observed outputs y;,...,yr, we can turn this model into a feedforward network by “unrolling” it for T steps, as shown in Figure 21.8(b). Notice that the weight matrices W, W__, and W are shared across all time steps.
In the unrolled network, it is easy to see that we can calculate gradients to train the
weights in the usual way; the only difference is that the sharing of weights across layers makes the gradient computation a little more complicated.
To keep the equations simple, we will show the gradient calculation for an RNN with
just one input unit, one hidden unit, and one output unit.
For this case, making the bias
weights explicit, we have z,=g:(w.z—1 + WioX; + woz) and § =gy (w.,z + woy). As in Equations (21.4) and (21.5), we will assume a squared-error loss L—in this case, summed
over the time steps. The derivations for the input-layer and output-layer weights w,. and w.., are essentially identical to Equation (21.4), so we leave them as an exercise. For the hidden-layer weight w. ., the first few steps also follow the same pattern as Equation (21.4):
%
= %
7
T
5
Z’Z(,Vl’fl);%
r
I
™=
M
I
y(iny) = );. 2y
$1)8(inys)
=200, = 30, i) g(weys + ) a
) wzy =20y = )&y (iny
(21.14)
I
Now the gradient for the hidden unit z; can be obtained from the previous time step as follows:
9z
T
W,
(Wezio1 + Wy
+wo2)
(21.15)
where the last line uses the rule for derivatives of products: (uv)/dx=vdu/dx+udv/dx.
Looking at Equation (21.15), we notice two things. First, the gradient expression is re-
cursive: the contribution to the gradient from time step 7 is calculated using the contribution
Back-propagation through time
from time step 7 — 1. If we order the calculations in the right way, the total run time for computing the gradient will be linear in the size of the network. This algorithm is called backpropagation through time, and is usually handled automatically by deep learning software
Exploding gradient
terms proportional o w.. TT_ g.(in.,). For sigmoids, tanhs, and ReLUs, g’ < 1, 5o our simple RNN will certainly suffer from the vanishing gradient problem (see page 756) if w.. < 1. On the other hand, if w.. > 1, we may experience the exploding gradient problem. (For the
systems. Second, if we iterate the recursive calculation, we see that gradients at 7 will include
general case, these outcomes depend on the first eigenvalue of the weight matrix W...) The next section describes a more elaborate RNN design intended to mitigate this issue.
Section 217 21.6.2
Long short-term memory
Unsupervised Learning and Transfer Learning
775
RNNs
Several specialized RNN architectures have been designed with the goal of enabling informa-
tion to be preserved over many time steps. One of the most popular is the long short-term
short-term ‘memory or LSTM. The long-term memory component of an LSTM, called the memory cell Long memory and denoted by ¢, is essentially copied from time step to time step. (In contrast, the basic RNN
multiplies its memory by a weight matrix at every time step, as shown in Equation (21.13).)
Memory cell
New information enters the memory by adding updates; in this way, the gradient expressions
do not accumulate multiplicatively over time. LSTMs also include gating units, which are Gating unit vectors that control the flow of information in the LSTM via elementwise multiplication of
the corresponding information vector: « The forget gate f determines if each element of the memory cell is remembered (copied Forget gate to the next time step) or forgotten (reset to zero).
« The input gate i determines if each element of the memory cell is updated additively Input gate by new information from the input vector at the current time step.
« The output gate o determines if each element of the memory cell is transferred to the Output gate short-term memory z, which plays a similar role to the hidden state in basic RNNs. Whereas the word “gate” in circuit design usually connotes a Boolean function, gates in
LSTMs are soft—for example, elements of the memory cell vector will be partially forgotten
if the corresponding elements of the forget-gate vector are small but not zero. The values for the gating units are always in the range [0, 1] and are obtained as the outputs of a sigmoid function applied to the current input and the previous hidden state. In detail, the update equations for the LSTM are as follows: £, =
o(Weyx +W_yz,1)
o(Weix+ Weizi1) (WeoX, +Wooz,1)
€ = ¢ 1 Of +i O tanh(Wy X, + Wez,1)
7, =
tanh(¢,) © o,
where the subscripts on the various weight matrices W indicate the origin and destination of
the corresponding links. The ® symbol denotes elementwise multiplication. LSTMs were among the first practically usable forms of RNN. They have demonstrated excellent performance on a wide range of tasks including speech recognition and handwriting recognition. Their use in natural language processing is discussed in Chapter 24. 21.7
Unsupervised
Learning and Transfer Learning
The deep learning systems we have discussed so far are based on supervised learning, which requires each training example to be labeled with a value for the target function.
Although
such systems can reach a high level of test-set accuracy—as shown by the ImageNet com-
petition results, for example—they often require far more labeled data than a human would for the same task. For example, a child needs to see only one picture of a giraffe, rather than thousands, in order to be able to recognize giraffes reliably in a wide range of settings
and views. Clearly, something is missing in our deep learning story; indeed, it may be the
776
Chapter 21
Deep Learning
case that our current approach to supervised deep learning renders some tasks completely
unattainable because the requirements for labeled data would exceed what the human race
(or the universe) can supply. Moreover, even in cases where the task is feasible, labeling large data sets usually requires scarce and expensive human labor.
For these reasons, there is intense interest in several learning paradigms that reduce the
dependence on labeled data. As we saw in Chapter 19, these paradigms include unsuper-
vised learning, transfer learning, and semisupervised learning. Unsupervised learning
algorithms learn solely from unlabeled inputs x, which are often more abundantly available than labeled examples. Unsupervised learning algorithms typically produce generative models, which can produce realistic text, images, audio, and video, rather than simply predicting labels for such data. Transfer learning algorithms require some labeled examples but are able to improve their performance further by studying labeled examples for different tasks, thus making it possible to draw on more existing sources of data. Semisupervised learning algo-
rithms require some labeled examples but are able to improve their performance further by also studying unlabeled examples. This section covers deep learning approaches to unsupervised and transfer learning; while semisupervised learning is also an active area of research in the deep learning community, the techniques developed so far have not proven broadly
effective in practice, so we do not cover them. 21.7.1
Unsupervised learning
Supervised learning algorithms all have essentially the same goal: given a training set of inputs x and corresponding outputs y=
f(x), learn a function A that approximates
f well.
Unsupervised learning algorithms, on the other hand, take a training set of unlabeled exam-
ples x. Here we describe two things that such an algorithm might try to do. The first is to
learn new representations—for example, new features of images that make it easier to iden-
tify the objects in an image. The second is to learn a generative model—typically in the form
of a probability distribution from which new samples can be generated. (The algorithms for learning Bayes nets in Chapter 20 fall in this category.) Many algorithms are capable of both representation learning and generative modeling. Suppose we learn a joint model Ay (x,z), where z is a set of latent, unobserved variables that represent the content of the data x in some way. In keeping with the spirit of the chapter, we do not predefine the meanings of the z variables; the model is free to learn to associate
z with x however it chooses. For example, a model trained on images of handwritten digits
might choose to use one direction in z space to represent the thickness of pen strokes, another to represent ink color, another to represent background color, and so on. With images of faces, the learning algorithm might choose one direction to represent gender and another to
capture the presence or absence of glasses, as illustrated in Figure 21.9. A learned probability model Py (x,z) achieves both representation learning (it has constructed meaningful z vectors from the raw x vectors) and generative modeling: grate z out of Py (x,z) we obtain Py (x).
if we inte-
Probabilistic PCA: A simple generative model PPCA
There have been many proposals for the form that Py (x,z) might take. One of the simplest
is the probabilistic principal components analysis (PPCA) model.” In a PPCA model, z
Section 217
Unsupervised Learning and Transfer Learning
Figure 21.9 A demonstration of how a generative model has learned to use different directions in z space to represent different aspects of faces. We can actually perform arithmetic in 2 space. The images here are all generated from the learned model and show what happens when we decode different points in z space. We start with the coordinates for the concept of “man with glasses.” subtract off the coordinates for “man,” add the coordinates for “woman.” and obtain the coordinates for “woman with glasses.” Images reproduced with permission from (Radford et al., 2015).
is chosen from a zero-mean, spherical Gaussian, then x is generated from z by applying a weight matrix W and adding spherical Gauss ian noise: P(z) = N(z0,I)
By(x|z) = N(x;Wz,0°1).
The weights W (and optionally the noise parameter o) can be learned by maximizing the likelihood of the data, given by
Pu(x) :/Ry(x.z)dz=N(x:0.WWT+rIZI).
(21.16)
The maximization with respect to W can be done by gradient methods or by an efficient iterative EM
algorithm
(see Section 20.3).
Once W
has been learned, new data samples
can be generated directly from Py (x) using Equation (21.16). Moreover, new observations. x that have very low probability according to Equation (21.16) can be flagged as potential
anomalies. With PPCA, we usually assume that the dimensionality ofz is much less than the dimensionality of x, so that the model learns to explain the data as well as possible in terms of a
small number of features. These features can be extracted for use in standard classifiers by computing Z, the expectation of Py (z|x). Generating data from a probabilistic PCA model is straightforward: first sample z from its fixed Gaussian prior, then sample x from a Gaussian with mean Wz.
As we will see
shortly, many other generative models resemble this process, but use complicated mappings
defined by deep models rather than linear mappings from z-space to x-space.
7 Standard PCA involves fitting a multivariate Gaussian to the raw input data and then selecting out the longest axes—the principal components—of that ellipsoidal distribution.
777
778
Chapter 21
Deep Learning
Autoencoders
Autoencoder
Many unsupervised deep learning algorithms are based on the idea of an autoencoder. An autoencoder is a model containing two parts: an encoder that maps from x to a representation
2 and a decoder that maps from a representation Z to observed data x. In general, the encoder
is just a parameterized function f and the decoder is just a parameterized function g. The
model is trained so that x ~ g(f(x)), so that the encoding decoding process. The functions f and g can be simple single matrix or they can be represented by a deep neural A very simple autoencoder is the linear autoencoder, a shared weight matrix W:
process is roughly inverted by the linear models parameterized by a network. where both f and g are linear with
One way to train this model is to minimize the squared error ¥; [[x; — g(f(x;))|[* so that
x ~ g(f(x)). The idea is to train W so that a low-dimensional 2 will retain as much information as possible to reconstruct the high-dimensional data x.
This linear autoencoder
turns out to be closely connected to classical principal components analysis (PCA). When
z is m-dimensional, the matrix W should learn to span the m principal components of the data—in other words, the set of m orthogonal directions in which the data has highest vari-
ance, or equivalently the m eigenvectors of the data covariance matrix that have the largest
eigenvalues—exactly as in PCA. The PCA model is a simple generative model that corresponds to a simple linear autoencoder. The correspondence suggests that there may be a way to capture more complex kinds Variational autoencoder Variational posterior
of generative models using more complex kinds of autoencoders.
coder (VAE) provides one way to do this.
The variational autoen-
Variational methods were introduced briefly on page 458 as a way to approximate the
posterior distribution in complex probability models, where summing or integrating out a large number of hidden variables is intractable.
The idea is to use a variational posterior
Q(z), drawn from a computationally tractable family of distributions, as an approximation to
the true posterior. For example, we might choose Q from the family of Gaussian distributions
with a diagonal covariance matrix. Within the chosen family of tractable distributions, Q is
optimized to be as close as possible to the true posterior distribution P(z|x).
For our purposes, the notion of “as close as possible” is defined by the KL divergence, which we mentioned on page 758. This is given by
Da(Q@IP(alx) = [ 0(a)ioggt
which is an average (with respect to Q) of the log ratio between Q and P. It is easy to see
Variational lower bound ELBO
that Dk (Q(z)||P(z|x)) > 0, with equality when Q and P coincide. We can then define the
variational lower bound £ (sometimes called the evidence lower bound, or ELBO) on the log likelihood of the data:
L(x,0) =logP(x) — Dg1(Q(2)||P(2]x)) QL17) We can see that £ is a lower bound for logP because the KL divergence is nonnegative. Variational learning maximizes £ with respect to parameters w rather than maximizing log P(x),
in the hope that the solution found, w", is close to maximizing log P(x) as well.
Section 217
Unsupervised Learning and Transfer Learning
779
As written, £ does not yet seem to be any easier to maximize than log P. Fortunately, we
can rewrite Equation (21.17) to reveal improved computational tractability:
£ = logP(x)— / Q(z)logp?z(T))()dz = 7/Q(z)logQ(z)dz+/Q(z)logP(x)P(z\x)dz = H(Q)+Fz-glogP(z,x)
where H(Q) is the entropy of the Q distribution.
For some variational families @ (such
as Gaussian distributions), H(Q) can be evaluated analytically.
Moreover, the expectation,
E,olog P(z,x), admits an efficient unbiased estimate via samples of z from Q. For each sample, P(z,x) can usually be evaluated efficiently—for example, if P is a Bayes net, P(z,X) is just a product of conditional probabilities because z and x comprise all the variables. Variational autoencoders provide a means of performing variational learning in the deep learning setting. Variational learning involves maximizing £ with respect to the parameters
of both P and Q. For a variational autoencoder, the decoder g(z) is interpreted as defining
log P(x|z). For example, the output of the decoder might define the mean of a conditional
Gaussian. Similarly, the output of the encoder f(x) is interpreted as defining the parameters of
Q—for example, Q might be a Gaussian with mean f(x). Training the variational autoencoder
then consists of maximizing £ with respect to the parameters of both the encoder f and the
decoder g, which can themselves be arbitrarily complicated deep networks.
Deep autoregressive models An autoregressive model (or AR model) is one in which each element x; of the data vector x
is predicted based on other elements of the vector. Such a model has no latent variables. If x
Autoregressive model
is of fixed size, an AR model can be thought of as a fully observable and possibly fully connected Bayes net. This means that calculating the likelihood ofa given data vector according to an AR model is trivial; the same holds for predicting the value of a single missing variable given all the others, and for sampling a data vector from the model.
The most common application of autoregressive models is in the analysis of time series
data, where an AR model of order k predicts x; given x,_j,...,x,_. In the terminology of Chapter 14, an AR model is a non-hidden Markov model. In the terminology of Chapter 23, an n-gram model of letter or word sequences is an AR model of order n — 1. In classical AR models, where the variables are real-valued, the conditional distribution
P(% | t»....X_1) is a linear-Gaussian model with fixed variance whose mean is a weighted linear combination of X, ..., X j—in other words, a standard linear regression model. The maximum likelihood solution is given by the Yule-Walker equations, which are closely Yule-Walker equations related to the normal equations on page 680.
A deep autoregressive model is one in which the linear-Gaussian
model is replaced
by an arbitrary deep network with a suitable output layer depending on whether x; is dis-
crete or continuous. Recent applications of this autoregressive approach include DeepMind’s ‘WaveNet model
for speech generation (van den Oord er al., 2016a).
WaveNet is trained
on raw acoustic signals, sampled 16,000 times per second, and implements a nonlinear AR model of order 4800 with a multilayer convolutional structure.
In tests it proves to be sub-
stantially more realistic than previous state-of-the-art speech generation systems.
Deep autoregressive model
780
Generative adversarial network GAN) enerator Discriminator Implicit model
Chapter 21
Deep Learning
Generative adversarial networks A generative adversarial network (GAN)
is actually a pair of networks that combine to
form a generative system. One of the networks, the generator, maps values from z to X in
order to produce samples from the distribution Py (x). A typical scheme samples z from a unit
Gaussian of moderate dimension and then passes it through a deep network /,, to obtain x. The other network, the discriminator, is a classifier trained to classify inputs x as real (drawn
from the training set) or fake (created by the generator). GANs are a kind of implicit model
in the sense that samples can be generated but their probabilities are not readily available; in a Bayes net, on the other hand, the probability of a sample is just the product of the conditional
probabilities along the sample generation path.
The generator is closely related to the decoder from the variational autoencoder frame-
work. The challenge in implicit modeling is to design a loss function that makes it possible to train the model using samples from the distribution, rather than maximizing the likelihood assigned to training examples from the data set.
Both the generator and the discriminator are trained simultaneously, with the generator
learning to fool the discriminator and the discriminator learning to accurately separate real from fake data.
The competition between generator and discriminator can be described in
the language of game theory (see Chapter 18). The idea is that in the equilibrium state of the
game, the generator should reproduce the training distribution perfectly, such that the discrim-
inator cannot perform better than random guessing. GANs have worked particularly well for
image generation tasks. For example, GAN can create photorealistic, high-resolution images of people who have never existed (Karras ef al., 2017). Unsupervised
translation
Translation tasks, broadly construed, consist of transforming an input x that has rich structure into an output y that also has rich structure.
In this context, “rich structure” means that the
data are multidimensional and have interesting statistical dependencies among the various
dimensions. Images and natural language sentences have a rich structure, but a single number, such as a clas ID, does not. Transforming a sentence from English to French or converting a photo of a night scene into an equivalent photo taken during the daytime are both examples
of translation tasks.
Supervised translation consists of gathering many (x,y) pairs and training the model to
map each X to the corresponding y.
For example, machine translation systems are often
trained on pairs of sentences that have been translated by professional human translators. For
other kinds of translation, supervised training data may not be available. For example, con-
sider a photo ofa night scene containing many moving cars and pedestrians. It is presumably
not feasible to find all of the cars and pedestrians and return them to their original positions in
Unsupervised translation
the night-time photo in order to retake the same photo in the daytime. To overcome this difficulty, it is possible to use unsupervised translation techniques that are capable of training
on many examples of x and many separate examples of y but no corresponding (x,y) pairs. These approaches are generally based on GANS; for example, one can train a GAN gen-
erator to produce a realistic example ofy when conditioned on x, and another GAN generator to perform the reverse mapping. The GAN training framework makes it possible to train a
generator to generate any one of many possible samples that the discriminator accepts as a
Section 217
Unsupervised Learning and Transfer Learning
781
realistic example of y given x, without any need for a specific paired y as is traditionally needed in supervised learning. More detail on unsupervised translation for images is given in Section 25.7.5. 21.7.2
Transfer learning and multitask learning
In transfer learning, experience with one learning task helps an agent learn better on another Transfer learning task. For example, a person who has already learned to play tennis will typically find it casier to learn related sports such as racquetball and squash; a pilot who has learned to fly one type of commercial passenger airplane will very quickly learn to fly another type: a student who has already learned algebra finds it easier to learn calculus.
‘We do not yet know the mechanisms of human transfer learning.
For neural networks,
learning consists of adjusting weights, so the most plausible approach for transfer learning is
to copy over the weights learned for task A to a network that will be trained for task B. The
weights are then updated by gradient descent in the usual way using data for task B. It may be a good idea to use a smaller learning rate in task B, depending on how similar the tasks are and how much data was used in task A.
Notice that this approach requires human expertise in selecting the tasks: for example,
weights learned during algebra training may not be very useful in a network intended for racquetball.
Also, the notion of copying weights requires a simple mapping between the
input spaces for the two tasks and essentially identical network architectures.
One reason for the popularity of transfer learning is the availability of high-quality pre-
trained models.
For example, you could download a pretrained visual object recognition
model such as the ResNet-50 model trained on the COCO data set, thereby saving yourself
weeks of work. From there you can modify the model parameters by supplying additional images and object labels for your specific task.
Suppose you want to classify types of unicycles. You have only a few hundred pictures
of different unicycles, but the COCO data set has over 3,000 images in each of the categories
of bicycles, motorcycles, and skateboards. This means that a model pretrained on COCO
already has experience with wheels and roads and other relevant features that will be helpful
in interpreting the unicycle images.
Often you will want to freeze the first few layers of the pretrained model—these layers
serve as feature detectors that will be useful for your new model. Your new data set will be
allowed to modify the parameters of the higher levels only; these are the layers that identify problem-specific features and do classification. However, sometimes the difference between sensors means that even the lowest-level layers need to be retrained.
As another example, for those building a natural language system, it is now common
to start with a pretrained model
such as the ROBERTA
model (see Section 24.6), which
already “knows” a great deal about the vocabulary and syntax of everyday language. The next step is to fine-tune the model in two ways. First, by giving it examples of the specialized vocabulary used in the desired domain; perhaps a medical domain (where it will learn about
“mycardial infarction”) or perhaps a financial domain (where it will learn about “fiduciary
responsibility”). Second, by training the model on the task it is to perform. If it is to do question answering, train it on question/answer pairs. One very important kind of transfer learning involves transfer between simulations and the real world.
For example, the controller for a self-driving car can be trained on billions
782
Chapter 21
Deep Learning
of miles of simulated driving, which would be impossible in the real world. Then, when the
Multitask learning
controller is transitioned to the real vehicle, it adapts quickly to the new environment. Multitask learning is a form of transfer learning in which we simultaneously
train a
model on multiple objectives. For example, rather than training a natural language system on part-of-speech tagging and then transferring the learned weights to a new task such as document classification, we train one system simultaneously on part-of-speech tagging, document
classification, language detection, word prediction, sentence difficulty modeling, plagiarism
detection, sentence entailment, and question answering. The idea is that to solve any one of these tasks, a model might be able to take advantage of superficial features of the data. But to
solve all eight at once with a common representation layer, the model is more likely to create a common representation that reflects real natural language usage and content.
21.8
Applications
Deep learning has been applied successfully to many important problem areas in Al For indepth explanations, we refer the reader to the relevant chapters: Chapter 22 for the use of deep learning in reinforcement learning systems, Chapter 24 for natural language processing, Chapter 25 (particularly Section 25.4) for computer vision, and Chapter 26 for robotics. 21.8.1
Vision
We begin with computer vision, which is the application area that has arguably had the biggest impact on deep learning, and vice versa. Although deep convolutional networks had been in use since the 1990s for tasks such as handwriting recognition, and neural networks had begun to surpass generative probability models for speech recognition by around 2010, it was the success of the AlexNet deep learning system in the 2012 ImageNet competition that propelled
deep learning into the limelight. The ImageNet competition was a supervised learning task with 1,200,000 images in 1,000 different categories, and systems were evaluated on the “top-5” score—how often the correct category appears in the top five predictions. AlexNet achieved an error rate of 15.3%, whereas the next best system had an error rate of more than 25%.
AlexNet had five convolutional
layers interspersed with max-pooling layers, followed by three fully connected layers. It used
ReLU activation functions and took advantage of GPUs to speed up the process of training
60 million weight
Since 2012, with improvements
in network design, training methods, and computing
resources, the top-5 error rate has been reduced to less than 2%—well below the error rate of
a trained human (around 5%). CNNs have been applied to a wide range of vision tasks, from self-driving cars to grading cucumbers.® Driving, which is covered in Section 25.7.6 and in several sections of Chapter 26, is among the most demanding of vision tasks: not only must
the algorithm detect, localize, track, and recognize pigeons, paper bags, and pedestrians, but it has to do it in real time with near-perfect accuracy.
8 The widely known tale of the Japanese cucumber farmer who built his own cucumber-s ing robot using TensorFlow is, it tums out, mostly mythical. The algorithm was developed by the farmer’s son, who worked previously as a software engineer at Toyota, and its low accuracy—about 70%—meant that the cucumbers il had to be sorted by hand (Zeeberg, 2017).
Section 218 Applications 21.8.2
Natural language processing
Deep learning has also had a huge impact on natural language processing (NLP) applications. such as machine translation and speech recognition. Some advantages of deep learning for these applications include the possibility of end-to-end learning, the automatic generation
of internal representations for the meanings of words, and the interchangeability of learned
encoders and decoders. End-to-end learning refers to the construction of entire systems as a single, learned func-
tion f. For example, an f for machine translation might take
as input an English sentence
S and produce an equivalent Japanese sentence S; = £(Sg). Such an f can be learned from training data in the form of human-translated pairs of sentences (or even pairs of texts, where
the alignment of corresponding sentences or phrases is part of the problem to be solved). A more classical pipeline approach might first parse Sg, then extract its meaning, then re-express
the meaning in Japanese as S, then post-edit S, using a language model for Japanese. This pipeline approach has two major drawbacks:
first, errors are compounded at each stage; and
second, humans have to determine what constitutes a “parse tree” and a “meaning representation,” but there is no easily accessible ground truth for these notions, and our theoretical
ideas about them are almost certainly incomplete.
At our present stage of understanding, then, the classical pipeline approach—which, at
least naively, seems to correspond to how a human translator works—is outperformed by the end-to-end method made possible by deep learning. For example, Wu ef al. (2016b) showed that end-to-end translation using deep learning reduced translation errors by 60% relative to a previous pipeline-based system. As of 2020, machine translation systems are approaching human performance for language pairs such as French and English for which very large paired data sets are available, and they are usable for other language pairs covering the majority of Earth’s population. There is even some evidence that networks trained on multiple languages do in fact learn an internal meaning representation: for example, after learning to translate Portuguese to English and English to Spanish, it is possible to translate Portuguese directly
into Spanish without any Portuguese/Spanish sentence pairs in the training set.
One of the most significant findings to emerge from the application of deep learning
to language tasks is that a great deal deal of mileage comes from re-representing individual words as vectors in a high-dimensional space—so-called word embeddings (see Section 24.1).
The vectors are usually extracted from the weights of the first hidden layer of
a network trained on large quantities of text, and they capture the statistics of the lexical
contexts in which words are used. Because words with similar meanings are used in similar
contexts, they end up close to each other in the vector space. This allows the network to generalize effectively across categories of words, without the need for humans to predefine those categories. For example, a sentence beginning “John bought a watermelon and two pounds of ... s likely to continue with “apples” or “bananas” but not with “thorium” or “geography.” Such a prediction is much easier to make if “apples” and “bananas” have similar representations in the internal layer. 21.8.3
Reinforcement learning
In reinforcement learning (RL), a decision-making agent learns from a sequence of reward
signals that provide some indication of the quality of its behavior. The goal is to optimize the
sum of future rewards. This can be done in several ways: in the terminology of Chapter 17,
783
784
Chapter 21
Deep Learning
the agent can learn a value function, a Q-function, a policy, and so on. From the point of view of deep learning, all these are functions that can be represented by computation graphs.
For example, a value function in Go takes a board position as input and returns an estimate of how advantageous the position is for the agent. While the methods of training in RL differ from those of supervised learning, the ability of multilayer computation graphs to represent
Deep reinforcement learning
complex functions over large input spaces has proved to be very useful. The resulting field of research is called deep reinforcement learning.
In the 1950s, Arthur Samuel experimented with multilayer representations of value func-
tions in his work on reinforcement learning for checkers, but he found that in practice a linear
function approximator worked best. (This may have been a consequence of working with a computer roughly 100 billion times less powerful than a modern tensor processing unit.) The first major successful demonstration of deep RL was DeepMind’s Atari-playing agent, DQN (Mnih et al., 2013). Different copies of this agent were trained to play each of several different Atari video games, and demonstrated skills such as shooting alien spaceships, bouncing balls with paddles, and driving simulated racing cars. In each case, the agent learned a Qfunction from raw image data with the reward signal being the game score. Subsequent work has produced deep RL systems that play at a superhuman level on the majority of the 57 different Atari games. DeepMind’s ALPHAGO system also used deep RL to defeat the best human players at the game of Go (see Chapter 5).
Despite its impressive successes, deep RL still faces significant obstacles: it is often difficult to get good performance, and the trained system may behave very unpredictably if the environment differs even a little from the training data (Irpan, 2018).
Compared to
other applications of deep learning, deep RL is rarely applied in commercial settings. It is, nonetheless, a very active area of research.
Summary
This chapter described methods for learning functions represented by deep computational graphs. The main points were: + Neural networks represent complex nonlinear functions with a network of parameterized linear-threshold units.
« The back-propagation algorithm implements a gradient descent in parameter space to minimize the loss function. « Deep learning works well for visual object recognition, speech recognition, natural language processing, and reinforcement learning in complex environments.
+ Convolutional networks are particularly well suited for image processing and other tasks
where the data have a grid topology. « Recurrent networks are effective for sequence-processing tasks including language modeling and machine translation.
Bibliographical and Historical Notes Bibliographical and
Historical Notes
The literature on neural networks is vast. Cowan and Sharp (1988b,
1988a) survey the early
history, beginning with the work of McCulloch and Pitts (1943). (As mentioned in Chap-
ter 1, John McCarthy has pointed to the work of Nicolas Rashevsky (1936, 1938) as the earliest mathematical model of neural learning.) Norbert Wiener, a pioneer of cybernetics and control theory (Wiener, 1948), worked with McCulloch and Pitts and influenced a num-
ber of young researchers, including Marvin Minsky, who may have been the first to develop a working neural network in hardware, in 1951 (see Minsky and Papert, 1988, pp. ix-x). Alan Turing (1948) wrote a research report titled Intelligent Machinery that begins with the
sentence “I propose to investigate the question as to whether it is possible for machinery to show intelligent behaviour” and goes on to describe a recurrent neural network architecture
he called “B-type unorganized machines” and an approach to training them. Unfortunately, the report went unpublished until 1969, and was all but ignored until recently.
The perceptron, a one-layer neural network with a hard-threshold activation function, was
popularized by Frank Rosenblatt (1957). After a demonstration in July 1958, the New York
Times described it as “the embryo of an electronic computer that [the Navy] expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence.” Rosenblatt
(1960) later proved the perceptron convergence theorem, although it had been foreshadowed by purely mathematical work outside the context of neural networks (Agmon, 1954; Motzkin and Schoenberg, 1954). Some early work was also done on multilayer networks, including Gamba perceptrons (Gamba ef al., 1961) and madalines (Widrow, 1962). Learning Machines (Nilsson,
1965) covers much of this early work and more.
The subsequent demise
of early perceptron research efforts was hastened (or, the authors later claimed, merely explained) by the book Perceptrons (Minsky and Papert, 1969), which lamented the field’s lack of mathematical rigor. The book pointed out that single-layer perceptrons could represent only linearly separable concepts and noted the lack of effective learning algorithms for multilayer networks. These limitations were already well known (Hawkins, 1961) and had been acknowledged by Rosenblatt himself (Rosenblatt, 1962). The papers collected by Hinton and Anderson (1981), based on a conference in San Diego in 1979, can be regarded as marking a renaissance of connectionism. The two-volume “PDP” (Parallel Distributed Processing) anthology (Rumelhart and McClelland, 1986) helped
to spread the gospel, so to speak, particularly in the psychology and cognitive science com-
munities. The most important development of this period was the back-propagation algorithm
for training multilayer networks.
The back-propagation algorithm was discovered independently several times in different
contexts (Kelley, 1960; Bryson,
1962; Dreyfus, 1962; Bryson and Ho, 1969; Werbos,
1974;
Parker, 1985) and Stuart Dreyfus (1990) calls it the “Kelley-Bryson gradient procedure.” Although Werbos had applied it to neural networks, this idea did not become widely known
until a paper by David Rumelhart, Geoff Hinton, and Ron Williams (1986) appeared in Narure giving a nonmathematical presentation of the algorithm. Mathematical respectability was enhanced by papers showing that multilayer feedforward networks are (subject to technical conditions) universal function approximators (Cybenko,
1988,
1989).
The late 1980s and
early 1990s saw a huge growth in neural network research: the number of papers mushroomed by a factor of 200 between 1980-84 and 1990-94.
785
786
Chapter 21
Deep Learning
In the late 1990s and early 2000s, interest in neural networks waned as other techniques such as Bayes nets, ensemble methods, and kernel machines came to the fore. Interest in deep
models was sparked when Geoff Hinton’s research on deep Bayesian networks—generative models with category variables at the root and evidence variables at the leaves—began to bear fruit, outperforming kernel machines on small benchmark
data sets (Hinton er al., 2006).
Interest in deep learning exploded when Krizhevsky ef al. (2013) used deep convolutional networks to win the ImageNet competition (Russakovsky et al., 2015).
Commentators often cite the availability of “big data” and the processing power of GPUs
as the main contributing factors in the emergence of deep learning.
Architectural improve-
ments were also important, including the adoption of the ReLU activation function instead of the logistic sigmoid (Jarrett et al., 2009; Nair and Hinton, 2010; Glorot et al., 2011) and later the development of residual networks (He et al., 2016).
On the algorithmic side, the use of stochastic gradient descent (SGD) with small batches
was essential in allowing neural networks to scale to large data sets (Bottou and Bousquet, 2008). Batch normalization (loffe and Szegedy, 2015) also helped in making the training pro-
cess faster and more reliable and has spawned several additional normalization techniques (Ba etal.,2016; Wu and He, 2018; Miyato et al., 2018). Several papers have studied the empirical behavior of SGD on large networks and large data sets (Dauphin et al., 2015; Choromanska et al., 2014; Goodfellow et al., 2015b). On the theoretical side, some progress has been made
on explaining the observation that SGD applied to overparameterized networks often reaches
a global minimum with a training error of zero, although so far the theorems to this effect assume a network with layers far wider than would ever occur in practice (Allen-Zhu et al.,
2018; Du ef al., 2018). Such networks have more than enough capacity to function as lookup tables for the training data.
The last piece of the puzzle, at least for vision applications, was the use of convolutional
networks. These had their origins in the descriptions of the mammalian visual system by neurophysiologists David Hubel and Torsten Wiesel (Hubel and Wiesel, 1959, 1962, 1968).
They described “simple cells” in the visual system of a cat that resemble edge detectors,
as well as “complex cells” that are invariant to some transformations such as small spatial
translations. In modern convolutional networks, the output of a convolution is analogous to a
simple cell while the output of a pooling layer is analogous to a complex cell. The work of Hubel and Wiesel inspired many of the early connectionist models of vision (Marr and Poggio, 1976). The neocognitron (Fukushima, 1980; Fukushima and Miyake,
1982), designed as a model of the visual cortex, was essentially a convolutional network in terms of model architecture, although an effective training algorithm for such networks had
to wait until Yann LeCun and collaborators showed how to apply back-propagation (LeCun
et al., 1995). One of the early commercial successes of neural networks was handwritten digit recognition using convolutional networks (LeCun ez al., 1995).
Recurrent neural networks (RNNs) were commonly proposed as models of brain function
in the 1970s, but no effective learning algorithms were associated with these proposals. The
method of back-propagation through time appears in the PhD thesis of Paul Werbos (1974),
and his later review paper (Werbos, 1990) gives several additional references to rediscoveries of the method in the 1980s. One of the most influential early works on RNNs was due to Jeff Elman (1990), building on an RNN architecture suggested by Michael Jordan (1986).
Williams and Zipser (1989) present an algorithm for online learning in RNNs. Bengio et al.
Bibliographical and Historical Notes (1994) analyzed the problem of vanishing gradients in recurrent networks. The long shortterm memory (LSTM) architecture (Hochreiter,
1991; Hochreiter and Schmidhuber,
1997;
Gers et al., 2000) was proposed as a way of avoiding this problem. More recently, effective RNN designs have been derived automatically (Jozefowicz er al., 2015; Zoph and Le, 2016).
Many methods have been tried for improving generalization in neural networks. Weight decay was suggested by Hinton (1987) and analyzed mathematically by Krogh and Hertz
(1992). The dropout method is due to Srivastava et al. (2014a). Szegedy et al. (2013) intro-
duced the idea of adversarial examples, spawning a huge literature.
Poole et al. (2017) showed that deep networks (but not shallow ones) can disentangle complex functions into flat manifolds in the space of hidden units. Rolnick and Tegmark
(2018) showed that the number of units required to approximate a certain class of polynomials
of n variables grows exponentially for shallow networks but only linearly for deep networks. White et al. (2019) showed that their BANANAS
system could do neural architecture
search (NAS) by predicting the accuracy of a network to within 1% after training on just 200 random sample architectures. Zoph and Le (2016) use reinforcement learning to search the space of neural network architectures. Real er al. (2018) use an evolutionary algorithm to do model selection, Liu er al. (2017) use evolutionary algorithms on hierarchical representations, and Jaderberg ef al. (2017) describe population-based training. Liu er al. (2019)
relax the space of architectures to a continuous differentiable space and use gradient descent to find a locally optimal solution.
Pham et al. (2018) describe the ENAS
(Efficient Neural
Architecture Search) system, which searches for optimal subgraphs of a larger graph. It is fast because it does not need to retrain parameters. The idea of searching for a subgraph goes back to the “optimal brain damage™ algorithm of LeCun et al. (1990).
Despite this impressive array of approaches, there are critics who feel the field has not yet
matured. Yu et al. (2019) show that in some cases these NAS algorithms are no more efficient
than random architecture selection. For a survey of recent results in neural architecture search,
see Elsken ef al. (2018).
Unsupervised learning constitutes a large subfield within statistics, mostly under the
heading of density estimation. Silverman (1986) and Murphy (2012) are good sources for classical and modem techniques in this area. Principal components analysis (PCA) dates back to Pearson (1901); the name comes from independent work by Hotelling (1933). The probabilistic PCA model (Tipping and Bishop, 1999) adds a generative model for the principal components themselves. The variational autoencoder is due to Kingma and Welling (2013) and Rezende ef al. (2014); Jordan et al. (1999) provide an introduction to variational methods for inference in graphical models. For autoregressive models, the classic text is by Box er al. (2016). The Yule-Walker equations for fitting AR models were developed independently by Yule (1927) and Walker (1931).
Autoregressive models with nonlinear dependencies were developed by several authors (Frey, 1998; Bengio and Bengio, 2001; Larochelle and Murray, 2011). The autoregressive WaveNet
model (van den Oord et al., 2016a) was based on earlier work on autoregressive image gen-
eration (van den Oord et al., 2016b).
Generative adversarial networks, or GANs, were first
proposed by Goodfellow et al. (2015a), and have found many applications in AL Some theoretical understanding of their properties is emerging, leading to improved GAN models and algorithms (Li and Malik, 2018b, 2018a; Zhu et al., 2019). Part of that understanding involves
protecting against adversarial attacks (Carlini ef al., 2019).
787
788
Hopfield network
Boltzmann machine
Chapter 21
Deep Learning
Several branches of research into neural networks have been popular in the past but are not actively explored today. Hopfield networks (Hopfield, 1982) have symmetric connections between each pair of nodes and can learn to store patterns in an associative memory, 5o that an entire pattern can be retrieved by indexing into the memory using a fragment of the pattern.
Hopfield networks are deterministic; they were later generalized to stochastic
Boltzmann machines (Hinton and Sejnowski, 1983, 1986). Boltzmann machines are possi-
bly the earliest example of a deep generative model. The difficulty of inference in Boltzmann
machines led to advances in both Monte Carlo techniques and variational techniques (see
Section 13.4).
Research on neural networks for Al has also been intertwined to some extent with research into biological neural networks. The two topics coincided in the 1940s, and ideas for convolutional networks and reinforcement learning can be traced to studies of biological sys-
Computational
tems; but at present, new ideas in deep learning tend to be based on purely computational or statistical concerns.
The field of computational neuroscience aims to build computational
models that capture important and specific properties of actual biological systems. Overviews
are given by Dayan and Abbott (2001) and Trappenberg (2010). For modern neural nets and deep learning, the leading textbooks are those by Goodfellow et al. (2016) and Charniak (2018). There are also many hands-on guides associated with the various open-source software packages for deep learning.
Three of the leaders of the
field—Yann LeCun, Yoshua Bengio, and Geoff Hinton—introduced the key ideas to non-Al researchers in an influential Nature article (2015). The three were recipients of the 2018 Turing Award. Schmidhuber (2015) provides a general overview, and Deng et al. (2014)
focus on signal processing tasks.
The primary publication venues for deep learning research are the conference on Neural
Information Processing Systems (NeurIPS), the International Conference on Machine Learning (ICML), and the International Conference on Learning Representations (ICLR). The main
journals are Machine Learning, the Journal of Machine Learning Research, and Neural Com-
putation. Increasingly, because of the fast pace of research, papers appear first on arXiv.org and are often described in the research blogs of the major research centers.
TS
D2
REINFORCEMENT LEARNING In which we see how experiencing rewards and punishments can teach an agent how to maximize rewards in the future. ‘With supervised learning, an agent learns by passively observing example input/output
pairs provided by a “teacher” In this chapter, we will see how agents can actively learn from
their own experience, without a teacher, by considering their own ultimate success or failure.
22.1
Learning from Rewards
Consider the problem of learning to play chess.
Let’s imagine treating this as a supervised
learning problem using the methods of Chapters 19-21. The chess-playing agent function
takes as input a board position and returns a move, so we train this function by supplying
examples of chess positions, each labeled with the correct move. Now, it so happens that we have available databases of several million grandmaster games, each a sequence of positions. and moves. The moves made by the winner are, with few exceptions, assumed to be good,
if not always perfect. Thus, we have a promising training set. The problem is that there are
relatively few examples (about 108) compared to the space of all possible chess positions (about 10°). In a new game, one soon encounters positions that are significantly different
from those in the database, and the trained agent function is likely to fail miserably—not least because it has no idea of what its moves are supposed to achieve (checkmate) or even what
effect the moves have on the positions of the pieces. And of course chess is a tiny part of the real world. For more realistic problems, we would need much vaster grandmaster databases,
and they simply don’t exist.'
Analternative is reinforcement learning (RL), in which an agent interacts with the world
and periodically receives rewards (or, in the terminology of psychology, reinforcements) that reflect how well it is doing.
For example, in chess the reward is 1 for winning, 0 for
losing, and § for a draw. We have already seen the concept of rewards in Chapter 17 for Markov decision processes (MDPs). Indeed, the goal is the same in reinforcement learning:
maximize the expected sum of rewards.
Reinforcement learning differs from “just solving
an MDP” because the agent is not given the MDP as a problem to solve; the agent is in the MDP. It may not know the transition model or the reward function, and it has to act in order
to learn more. Imagine playing a new game whose rules you don’t know; after a hundred or 50 moves, the referee tells you “You lose.” That is reinforcement learning in a nutshell.
From our point of view as designers of Al systems, providing a reward signal to the agent is usually much easier than providing labeled examples of how to behave. First, the reward ! As Yann LeCun and Alyosha Efros have pointed out, “the Al revolution will not be supervised.”
{3570
790
Chapter 22 Reinforcement Learning function is often (as we saw for chess) very concise and easy to specify: it requires only a few lines of code to tell the chess agent if it has won or lost the game or to tell the car-racing agent that it has won or lost the race or has crashed.
Second, we don’t have to be experts,
capable of supplying the correct action in any situation, as would be the case if we tried to apply supervised learning. It turns out, however, that a
Sparse
little bit of expertise can go a long way in reinforcement
learning. The two examples in the preceding paragraph—the win/loss rewards for chess and racing—are what we call sparse rewards, because in the vast majority of states the agent is given no informative reward signal at all. In games such as tennis and cricket, we can easily supply additional rewards for each point won or for each run scored. In car racing, we could reward the agent for making progress around the track in the right direction. When learning to crawl, any forward motion is an achievement. much easier.
These intermediate rewards make learning
As long as we can provide the correct reward signal to the agent, reinforcement learning provides a very general way to build Al systems. This is particularly true for simulated environments, where there is no shortage of opportunities to gain experience. The addition of deep learning as a tool within RL systems has also made new applications possible, including learning to play Atari video games from raw visual input (Mnih ef al., 2013), controlling robots (Levine ef al., 2016), and playing poker (Brown and Sandholm, 2017). Literally hundreds of different reinforcement learning algorithms have been devised, and
Model-based reinforcement Iearning
many of them can employ as tools a wide range of learning methods from Chapters 19-21. In this chapter, we cover the basic ideas and give some sense of the variety of approaches through a few examples. We categorize the approaches as follows: o Model-based reinforcement learni
In these approaches the agent uses a transition
model of the environment to help interpret the reward signals and to make decisions about how to act. The model may be initially unknown, in which case the agent learns the model from observing the effects of its actions, or it may already be known—for example, a chess program may know the rules of chess even if it does not know how to choose good moves. In partially observable environments, the transition model is also useful for state estimation (see Chapter 14). Model-based reinforcement learning
Model-free reinforcement learning
Action-utiity learning Qlearning Q-function
Policy search
systems often learn a utility function U (s), defined (as in Chapter 17) in terms of the sum of rewards from state s onward.” o Model-free reinforcement learning: In these approaches the agent neither knows nor learns a transition model for the environment. Instead, it learns a more direct represen-
tation of how to behave. This comes in one of two varieties:
o Action-utility learning: We introduced action-utility functions in Chapter 17. The
most common form of action-utility learning is Q-learning, where the agent learns
a Q-function, or quality-function, Q(s,a), denoting the sum of rewards from state
s onward if action a is taken. Given a Q-function, the agent can choose what to do in s by finding the action with the highest Q-value.
o Policy search:
The agent learns a policy 7(s) that maps directly from states to
actions. In the terminology of Chapter 2, this a reflex agent. 2 Inthe RL literature, which draws more on operations research than economics, utility functions are often called value functions and denoted V (s).
Section 22.2
Passive Reinforcement Learning
791
08516 | 0.9078 | 0.9578 0.8016
(@)
(®)
Figure 22.1 (a) The optimal policies for the stochastic environment with R(s.a,s') = — 0.04 for transitions between nonterminal states. There are two policies because in state (3,1) both Left and Up are optimal. We saw this before in Figure 17.2. (b) The utilities of the states in the4 x 3 world, given policy 7. Passive
‘We begin in Section 22.2 with passive reinforcement learning, where the agent’s policy = reinforcement
is fixed and the task is to learn the utilities of states (or of state-action pairs); this could
="
reinforcement learning, where the agent must also figure out what to do. The principal
(e reinforcement
also involve learning a model of the environment. (An understanding of Markov decision processes, as described in Chapter 17, is essential for this section.) Section 22.3 covers active issue is exploration: an agent must experience as much as possible of its environment in
order to learn how to behave in it. Section 22.4 discusses how an agent can use inductive
learning (including deep learning methods) to learn much faster from its experiences. We also discuss other approaches that can help scale up RL to solve real problems, including providing intermediate pseudorewards to guide the learner and organizing behavior into a hierarchy of actions. Section 22.5 covers methods for policy search. In Section 22.6, we explore apprenticeship learning: training a learning agent using demonstrations rather than
reward signals. Finally, Section 22.7 reports on applications of reinforcement learning. 22.2
Passive Reinforcement Learning
We start with the simple case of a fully observable environment with a small number of
actions and states, in which an agent already has a fixed policy 7 (s) that determines its actions.
The agent is trying to learn the utility function U™ (s)—the expected total discounted reward if policy m is executed beginning in state s. We call this a passive learning agent.
The passive learning task is similar to the policy evaluation task, part of the policy iteration algorithm described in Section 17.2.2. The difference is that the passive learning agent does not know the transition model P(s'|s,a), which specifies the probability of reaching state s’ from state s after doing action a@; nor does it know the reward function R(s,a,s’), which specifies the reward for each transition.
‘We will use as our example the 4 x 3 world introduced in Chapter 17. Figure 22.1 shows
the optimal policies for that world and the corresponding utilities. The agent executes a set
Fassive leaming
792
Chapter 22 Reinforcement Learning
Trial
of trials in the environment using its policy 7. In each trial, the agent starts in state (1,1) and
experiences a sequence of state transitions until it reaches one of the terminal states, (4,2) or
(4,3). Tts percepts supply both the current state and the reward received for the transition that just occurred to reach that state. Typical trials might look like this:
P20
0203
B e
6
a 1>‘““(12 (4,2)
Note that each transition is annotated with both the action ldkcn and the reward received at the next state. The object is to use the information about rewards to learn the expected utility U™ (s) associated with each nonterminal state s. The utility is defined to be the expected sum
of (discounted) rewards obtained if policy 7 is followed. As in Equation (17.2) on page 567, we write
U™(s) =E | Y 1=0
+'R(S:,7(S:),S41) |»
2.1)
where R(S;,7(S;),5.1) is the reward received when action 7(S,) is taken in state S, and reaches state S..1. Note that S, is a random variable denoting the state reached at time ¢ when executing policy 7, starting from state So=s. We will include a discount factor
in all of
our equations, but for the 4 x 3 world we will set = 1, which means no discounting. Direct utility estimation
Reward-to-go
22.2.1
Direct utility estimation
The idea of direct utility estimation is that the utility of a
state is defined as the expected
total reward from that state onward (called the expected reward-to-go), and that each trial
provides a sample of this quantity for each state visited. For example, the first of the three
trials shown earlier provides a sample total reward of 0.76 for state (1,1), two samples of 0.80 and 0.88 for (1,2), two samples of 0.84 and 0.92 for (1,3), and so on. Thus, at the end of each
sequence, the algorithm calculates the observed reward-to-go for each state and updates the
estimated utility for that state accordingly, just by keeping a running average for each state
in a table. In the limit of infinitely many trials, the sample average will converge to the true
expectation in Equation (22.1). This means that we have reduced reinforcement learning to a standard supervised learn-
ing problem in which each example is a (state, reward-to-go) pair. We have a lot of powerful algorithms for supervised learning, so this approach seems promising, but it ignores an im-
portant constraint: The utility of a state is determined by the reward and the expected utility
of the successor states. More specifically, the utility values obey the Bellman equations for a
fixed policy (see also Equation (17.14)):
Uils) = X P(s' | 5.m1(8)) [R(s. mi(5), ) + 7 Ui(s))] s
(222)
By ignoring the connections between states, direct utility estimation misses opportunities for
learning.
For example, the second of the three trials given earlier reaches the state (3,2),
which has not previously been visited.
from the first trial to have a high utility.
The next transition reaches (3,3), which is known
The Bellman equation suggests immediately that
(3,2) is also likely to have a high utility, because it leads to (3,3), but direct utility estimation
Section 22.2
Passive Reinforcement Learning
793
learns nothing until the end of the trial. More broadly, we can view direct utility estimation
as searching for U in a hypothesis space that is much larger than it needs to be, in that it includes many functions that violate the Bellman equations.
often converges very slowly. 22.2.2
Adaptive dynamic
For this reason, the algorithm
programming
An adaptive dynamic programming
(or ADP) agent takes advantage of the constraints
among the utilities of states by learning the transition model that connects them and solving the corresponding Markov decision process using dynamic programming. For a passive learning agent, this means plugging the learned transition model P(s'[5,7(s)) and the observed rewards R(s,7(s),s')
into Equation (22.2) to calculate the utilities of the states.
As
we remarked in our discussion of policy iteration in Chapter 17, these Bellman equations are linear when the policy 7 is fixed, so they can be solved using any linear algebra package. Alternatively, we can adopt the approach of modified policy iteration (see page 578),
using a simplified value iteration process to update the utility estimates after each change to the learned model. Because the model usually changes only slightly with each observation, the value iteration process can use the previous utility estimates as initial values and typically
converge very quickly.
Learning the transition model is easy, because the environment is fully observable. This
means that we have a supervised learning task where the input for each training example is a
state-action pair, (s,a), and the output is the resulting state, s'. The transition model P(s'| s,a)
is represented as a table and it is estimated directly from the counts that are accumulated in
Ny - The counts record how often state is reached when executing in s. For example, in the three trials given on page 792, Right is executed four times in (3,3) and the resulting state
is (3,2) twice and (4,3) twice, so P((3,2)|(3,3),Right) and P((4,3)|(3,3),Right) are both estimated to be §. The full agent program for a passive ADP agent is shown in Figure 22.2. Its performance on the 4 x 3 world is shown in Figure 22.3. In terms of how quickly its value estimates
improve, the ADP agent is limited only by its ability to learn the transition model. In this
sense, it provides a standard against which to measure any other reinforcement learning al-
gorithms. It is, however, intractable for large state spaces. In backgammon, for example, it would involve solving roughly 10*° equations in 10*° unknowns. 22.2.3
Temporal-difference learning
Solving the underlying MDP as in the preceding section is not the only way to bring the
Bellman equations to bear on the learning problem.
Another way is to use the observed
transitions to adjust the utilities of the observed states so that they agree with the constraint
equations. Consider, for example, the transition from (1,3) to (2,3) in the second trial on page 792. Suppose that as a result of the first trial, the utility estimates are U™(1,3)=0.88
and U™(2,3)=0.96. Now, if this transition from (1,3) to (2,3) occurred all the time, we would expect the utilities to obey the equation
U™(1,3) = —0.04+ U™(2,3), 50 U7(1,3) would be 0.92. Thus, its current estimate of 0.84 might be a lttle low and should be increased. More generally, when a transition occurs from state s to state s’ via action 7(s),
Adaptive dynamic programming
794
Chapter 22 Reinforcement Learning function PASSIVE-ADP-LEARNER(percepr) returns an action
inputs: percept, a percept indicating the current state s’ and reward signal r
persistent: 7, a fixed policy
mdp, an MDP with model P, rewards R, actions A, discount y U, atable of utilities for states, initially empty
Nyjs.a» 2 table of outcome count vectors indexed by state and action, initially zero 5. a, the previous state and action, initially null if " is new then U[s'] 0 s is not null then
increment Ny s, alls’]
Rls,a,s'1r
add a to A[s] P(- | 5,a) =+ > [envada |06 | | [Caenviada |->-[de 03] ([ |»a et Score00 ] [Una [-2.1] | [Scow=03] [puera [-09] | [Seorer=11 |puema [-19] | [Seorer =17 Beam 2
Beam 2
Beam 2
Fypomests | [ Viord [Score] | [Fipoesis | [Word [Score] | [Fypomesis [-05 | |[Lapueraal] > |~>[envada]-03 | | [Lapvera |>foe s Scors: =21] [puena [-2.1] | [Scorer=12] (o ]-07] | [Scorer =15
++
Figure 24.8 Beam search with beam size of b=2. The score of each word is the logprobability generated by the target RNN softmax, and the score of each hypothesis is the sum of the word scores. At timestep 3, the highest scoring hypothesis La entrada can only generate low-probability continuations, so it “falls off the beam.” using a greedy decoder to translate into Spanish the English sentence we saw before, The
front door is red. The correct translation is “La puerta de entrada es roja”—literally “The door of entry is red.” Suppose the target RNN correctly generates the first word La for The. Next, a greedy decoder might propose entrada for front. But this is an error—Spanish word order should put the noun puerta before the modifier. Greedy decoding is fast—it only considers one choice at each timestep and can do so quickly—but the model has no mechanism to correct mistakes. We could try to improve the attention mechanism so that it always attends to the right word and guesses correctly every time.
But for many sentences it is infeasible to guess
correctly all the words at the start of the sentence until you have seen what’s at the end.
A better approach is to search for an optimal decoding (or at least a good one) using
one of the search algorithms from Chapter 3. A common choice is a beam search (see Sec-
tion 4.1.3). In the context of MT decoding, beam search typically keeps the top k hypotheses at each stage, extending each by one word using the top k choices of words, then chooses the best k of the resulting > new hypotheses. When all hypotheses in the beam generate the special token, the algorithm outputs the highest scoring hypothesis.
A visualization of beam search is given in Figure 24.8. As deep learning models become
more accurate, we can usually afford to use a smaller beam size.
Current state-of-the-art
neural MT models use a beam size of 4 to 8, whereas the older generation of statistical MT models would use a beam size of 100 or more.
24.4 Transformer Self-attention
The influential article “Attention is all you need” (Vaswani et al., 2018) introduced the transformer architecture, which uses a self-attention mechanism that can model long-distance
context without a sequential dependency.
24.4.1 Self-attention
The Transformer Architecture
Self-attention
Previously, in sequence-to-sequence models, attention was applied from the target RNN to the source RNN. Self-attention extends this mechanism so that each sequence of hidden states also attends to itself—the source to the source, and the target to the target. This allows
the model to additionally capture long-distance (and nearby) context within each sequence.
Section 24.4
The Transformer Architecture
869
The most straightforward way of applying self-attention is where the attention matrix is
directly formed by the dot product of the input vectors. However, this is problematic. The dot product between a vector and itself will always be high, so each hidden state will be biased
towards attending to itself. The transformer solves this by first projecting the input into three different representations using three different weight matrices:
« The query vector g;=W,x; is the one being attended from, like the target in the standard attention mechanism. + The key vector k;=W,x; attention mechanism.
is the one being attended to, like the source in the basic
« The value vector v;=W,x; is the context that is being generated.
Query vector
Key vector Value vector
In the standard attention mechanism, the key and value networks are identical, but intuitively
it makes sense for these to be separate representations. The encoding results of the ith word,
¢;, can be calculated by applying an attention mechanism to the projected vectors:
rj = (ak)/Va ay = & f(L)
a-X i
k
lij Vs
where d is the dimension of k and q. Note that i and j are indexes in the same sentence, since we are encoding the context using self-attention. In each transformer layer, self-attention uses
the hidden vectors from the previous layer, which initially is the embedding layer. There are several details worth mentioning here.
First of all, the self-attention mecha-
nism is asymmetric, as r;; is different from rj;. Second, the scale factor v/ was added to improve numerical stability. Third, the encoding for all words in a sentence can be calculated
simultaneously, as the above equations can be expressed using matrix operations that can be
computed efficiently in parallel on modern specialized hardware. The choice of which context to use is completely learned from training examples, not
prespecified. The context-based summarization, ¢;, is a sum over all previous positions in the
sentence. In theory, and information from the sentence could appear in ¢;, but in practice, sometimes important information gets lost, because it is essentially averaged out over the whole sentence.
One way to address that is called multiheaded attention.
We divide the
sentence up into m equal pieces and apply the attention model to each of the m pieces. Each piece has its own set of weights.
Then the results are concatenated together to form ¢;. By
concatenating rather than summing, we make it easier for an important subpiece to stand out.
24.4.2
From self-attention to transformer
Self-attention is only one component of the transformer model. Each transformer layer con-
sists of several sub-layers. At each transformer layer, self-attention is applied first. The output of the attention module is fed through feedforward layers, where the same feedforward
weight matrices are applied independently at each position. A nonlinear activation function, typically ReLU, is applied after the first feedforward layer. In order to address the potential
vanishing gradient problem, two residual connections (are added into the transformer layer. A single-layer transformer in shown in Figure 24.9. In practice, transformer models usually
Multiheaded attention
870
Chapter 24 Deep Learning for Natural Language Processing
{
Transformer
Output Vectors,
{
Layer
Residual
Connection
]
Rssiual)
Self-Attention-
Residual
P11
L input Vectors Figure 24.9 A single-layer transformer consists of self-attention, a feedforward network, and residual connections.
Clas: Adverb
Class Pronoun
Cla PastTenseVerb
t t 1 Feedforward | | Feedforward | | Feedforward | | f f i “Transformer Layer f f i ‘Transformer Layer f f i “Transformer Layer t f f Positional Positional Positional Embedding 1| |Embedding2| |Embedding3| + + + Embedding | [Embedding | [Embedding | lookup Tookup Tookup. f f Yesterday
Figure 24.10
they
cut
Class. Determiner
t Feedforward | | Feedforward f f f
f
f
f
f f Positional Positional |Embedding4| |Embedding 5 + + [Embedding | [ Embedding Tookup Tookup f f the
rope
Using the transformer architecture for POS tagging.
have six or more layers.
As with the other models that we’ve learned about, the output of
layer i is used as the input to layer i+ 1.
oo
The transformer architecture does not explicitly capture the order of words in the sequence, since context is modeled only through self-attention, which is agnostic to word order. To capture the ordering of the words, the transformer uses a technique called positional em-
bedding. If our input sequence has a maximum length of n, then we learn n new embedding
Section 245 Pretraining and Transfer Learning
871
vectors—one for each word position. The input to the first transformer layer is the sum of the word embedding at position 7 plus the positional embedding corresponding to position 7.
Figure 24.10 illustrates the transformer architecture for POS tagging, applied to the same
sentence used in Figure 24.3. At the bottom, the word embedding and the positional embed-
dings are summed to form the input for a three-layer transformer. The transformer produces
one vector per word, as in RNN-based POS tagging. Each vector is fed into a final output layer and softmax layer to produce a probability distribution over the tags.
In this section, we have actually only told half the transformer story: the model we de-
scribed here is called the transformer encoder. It is useful for text classification tasks. The
full transformer architecture was originally designed as a sequence-to-sequence model for machine translation. decoder.
Therefore, in addition to the encoder, it also includes a transformer
The encoder and decoder are nearly identical, except that the decoder uses a ver-
sion of self-attention where each word can only attend to the words before it, since text is
Transformer encoder Transformer decoder
generated left-to-right. The decoder also has a second attention module in each transformer
layer that attends to the output of the transformer encoder.
2
Pretraining and Transfer Learning
Getting enough data to build a robust model can be a challenge.
In computer vision (see
Chapter 25), that challenge was addressed by assembling large collections of images (such as ImageNet) and hand-labeling them. For natural language, it is more common
to work with text that is unlabeled.
The dif-
ference is in part due to the difficulty of labeling: an unskilled worker can easily label an image as “cat” or “sunset,” but it requires extensive training to annotate a sentence with partof-speech tags or parse trees. The difference is also due to the abundance of text: the Internet adds over 100 billion words of text each day, including digitized books, curated resources
such as Wikipedia, and uncurated social media posts. Projects such as Common Crawl provide easy access to this data. Any running text can
be used to build n-gram or word embedding models, and some text comes with structure that
can be helpful for a variety of tasks—for example, there are many FAQ sites with questionanswer pairs that can be used to train a question-answering system.
Similarly, many Web
sites publish side-by-side translations of texts, which can be used to train machine translation
systems. Some text even comes with labels of a sort, such as review sites where users annotate their text reviews with a 5-star rating system.
‘We would prefer not to have to go to the trouble of creating a new data set every time we want a new NLP model. In this section, we introduce the idea of pretraining: a form
of transfer learning (see Section 21.7.2) in which we use a large amount of shared general-
domain language data to train an initial version of an NLP model. From there, we can use a
smaller amount of domain-specific data (perhaps including some labeled data) to refine the model.
The refined model can learn the vocabulary, idioms, syntactic structures, and other
linguistic phenomena that are specific to the new domain. 24.5.1
Pretrained word embeddings
In Section 24.1, we briefly introduced word embeddings. We saw that how similar words like banana and apple end up with similar vectors, and we saw that we can solve analogy
Pretraining
872
Chapter 24 Deep Learning for Natural Language Processing problems with vector subtraction. This indicates that the word embeddings are capturing substantial information about the words.
In this section we will dive into the details of how word embeddings are created using an
entirely unsupervised process over a large corpus of text. That is in contrast to the embeddings
from Section 24.1, which were built during the process of supervised part of speech tagging,
and thus required POS tags that come from expensive hand annotation. We will concentrate
on one specific model for word embeddings,
the GloVe (Global
Vectors) model. The model starts by gathering counts of how many times each word appears
within a window of another word, similar to the skip-gram model. First choose window size
(perhaps 5 words) and let X;; be the number of times that words i and j co-occur within
a window,
and let X; be the number of times word i co-occurs with any other word.
Let
P,;=X;;/X; be the probability that word j appears in the context of word i. As before, let E; be the word embedding for word i. Part of the intuition of the GloVe model is that the relationship between two words can
best be captured by comparing them both to other words. Consider the words ice and steam. Now consider the ratio of their probabilities of co-occurrence with another word, w, that is:
Puice/Pustean
When w is the word solid the ratio will be high (meaning solid applies more to ice) and when w is the word gas it will be low (meaning gas applies more to steam). And when w is a non-content word like rhe, a word like warer that is equally relevant to both, or an equally irrelevant word like fashion, the ratio will be close to 1.
The GloVe model starts with this intuition and goes through some mathematical reason-
ing (Pennington ef al., 2014) that converts ratios of probabilities into vector differences and
dot products, eventually arriving at the constraint
E;-E}=log(P;). In other words, the dot product of two word vectors is equal to the log probability of their
co-occurrence. That makes intuitive sense: two nearly-orthogonal vectors have a dot product
close to 0, and two nearly-identical normalized vectors have a dot product close to 1. There
is a technical complication wherein the GloVe model creates two word embedding vectors
for each word, E; and EJ; computing the two and then adding them together at the end helps limit overfitting.
Training a model like GloVe is typically much less expensive than training a standard
neural network:
a new model can be trained from billions of words of text in a few hours
using a standard desktop CPU. It is possible to train word embeddings on a specific domain, and recover knowledge in
that domain. For example, Tshitoyan e al. (2019) used 3.3 million scientific abstracts on the
subject of material science to train a word embedding model. They found that, just as we saw
that a generic word embedding model can answer “Athens is to Greece as Oslo is to what?” with “Norway,” their material science model can answer “NiFe is to ferromagnetic as IrMn is to what?” with “antiferromagnetic.”
Their model does not rely solely on co-occurrence of words; it seems to be capturing more complex scientific knowledge. When asked what chemical compounds can be classified
as “thermoelectric™ or “topological insulator,” their model is able to answer correctly.
For
example, CsAgGapSey never appears near “thermoelectric” in the corpus, but it does appear
Section 245 Pretraining and Transfer Learning
873
near “chalcogenide.” “band gap,” and “optoelectric,” which are all clues enabling it to be classified as similar to “thermoelectric.”
Furthermore, when trained only on abstracts up
to the year 2008 and asked to pick compounds that are “thermoelectric” but have not yet appeared in abstracts, three of the model’s top five picks were discovered to be thermoelectric in papers published between 2009 and 2019. 24.5.2
Pretrained contextual representations
‘Word embeddings are better representations than atomic word tokens, but there is an impor-
tant issue with polysemous words. For example, the word rose can refer to a flower or the past tense of rise. Thus, we expect to find at least two entirely distinct clusters of word contexts for rose: one similar to flower names such as dahlia, and one similar to upsurge.
No
single embedding vector can capture both of these simultaneously. Rose is a clear example of a word with (at least) two distinct meanings, but other words have subtle shades of meaning that depend on context, such as the word need in you need to see this movie versus humans
need oxygen to survive. And some idiomatic phrases like break the bank are better analyzed as a whole rather than as component words.
Therefore, instead of just learning a word-to-embedding table, we want to train a model to
generate contextual representations of each word in a sentence. A contextual representation
maps both a word and the surrounding context of words into a word embedding vector. In
other words, if we feed this model the word rose and the context the gardener planted a rose bush, it should produce a contextual embedding that is similar (but not necessarily identical)
to the representation we get with the context the cabbage rose had an unusual fragrance, and very different from the representation of rose in the context the river rose five feet.
Figure 24.11 shows a recurrent network that creates contextual word embeddings—the boxes that are unlabeled in the figure. We assume we have already built a collection of
noncontextual word embeddings. We feed in one word at a time, and ask the model to predict the next word. So for example in the figure at the point where we have reached the word “car,” the the RNN node at that time step will receive two inputs: the noncontextual word embedding for “car” and the context, which encodes information from the previous words “The red.” The RNN node will then output a contextual representation for “car.” The network
as a whole then outputs a prediction for the next word, “is.” We then update the network’s weights to minimize the error between the prediction and the actual next word.
This model is similar to the one for POS tagging in Figure 24.5, with two important
differences. First, this model is unidirectional (left-to-right), whereas the POS model is bidi-
rectional. Second, instead of predicting the POS tags for the current word, this model predicts
the next word using the prior context. Once the model is built, we can use it to retrieve representations for words and pass them on to some other task: we need not continue to predict
the next word. Note that computing a contextual representations always requires two inputs, the current word and the context.
24.5.3
Masked
language models
A weakness of standard language models such as n-gram models is that the contextualization
of each word is based only on the previous words of the sentence. Predictions are made from
left to right. But sometimes context from later in a sentence—for example, feet in the phrase
rose five feet—helps to clarify earlier words.
Contextual representations
874
Chapter 24 Deep Learning for Natural Language Processing
t
is
Feedforward |
[Feedforward
Feedforward
t
big
car
red
t
{
Feedforward |
t
[Feedforward
Contextual
representations (RNN output)
RNN}——[RNN|——{RNN|——{RNN}-——[RNN
]\t“’,"‘“f::‘"\,:;“
(word embeddings)
1
f
‘°°; X
> X
Figure 26.11 A simple triangular robot that can translate, and needs to avoid a rectangular
obstacle. On the left is the workspace, on the right is the configuration space.
Once we have a path, the task of executing a sequence of actions to follow the path is
called trajectory tracking control. A trajectory is a path that has a time associated with
[rajecio tracking
each point on the path. A path just says “go from A to B to C, etc.” and a trajectory says Trajectory “start at A, take 1 second to get to B, and another 1.5 seconds to get to C, ete.” 26.5.1
Configuration space
Imagine a simple robot, R, in the shape of a right triangle as shown by the lavender triangle in the lower left corner of Figure 26.11. The robot needs to plan a path that avoids a rectangular obstacle, O. The physical space that a robot moves about in is called the workspace. This Workspace particular robot can move in any direction in the x — y plane, but cannot rotate. The figure shows five other possible positions of the robot with dashed outlines; these are each as close
to the obstacle as the robot can get. The body of the robot could be represented as a set of (x,y) points (or (x,y,z) points
for a three-dimensional robot), as could the obstacle. With this representation, avoiding the obstacle means that no point on the robot overlaps any point on the obstacle. Motion planning
would require calculations on sets of points, which can be complicated and time-consuming.
We can simplify the calculations by using a representation scheme in which all the points
that comprise the robot are represented as a single point in an abstract multidimensional
space, which we call the configuration space, or C-space. The idea is that the set of points Configuration space that comprise the robot can be computed if we know (1) the basic measurements of the robot
(for our triangle robot, the length of the three sides will do) and (2) the current pose of the robot—its position and orientation.
For our simple triangular robot, two dimensions suffice for the C-space: if we know the
(x,y) coordinates of a specific point on the robot—we’ll use the right-angle vertex—then we
can calculate where every other point of the triangle is (because we know the size and shape of the triangle and because the triangle cannot rotate). In the lower-left corner of Figure 26.11,
the lavender triangle can be represented by the configuration (0,0).
If we change the rules so that the robot can rotate, then we will need three dimensions,
(x,y,0), to be able to calculate where every point is. Here
is the robot’s angle of rotation
in the plane. If the robot also had the ability to stretch itself, growing uniformly by a scaling
factor s, then the C-space would have four dimensions, (x,y,6,5).
~C-space
Chapter 26 Robotics For now we’ll stick with the simple two-dimensional C-space of the non-rotating triangle
robot. The next task is to figure out where the points in the obstacle are in C-space. Consider
the five dashed-line triangles on the left of Figure 26.11 and notice where the right-angle vertex is on each of these. Then imagine all the ways that the triangle could slide about.
Obviously, the right-angle vertex can’t go inside the obstacle, and neither can it get any
C-space obstacle
closer than it is on any of the five dashed-line triangles. So you can see that the area where
the right-angle vertex can’t go—the C-space obstacle—is the five-sided polygon on the right
of Figure 26.11 labeled Cyp
In everyday language we speak of there being multiple obstacles for the robot—a table, a
chair, some walls. But the math notation is a bit easier if we think of all of these as combit
into one “obstacle” that happens to have disconnected components.
In general, the C:
obstacle is the set of all points ¢ in C such that, if the robot were placed in that configuration, its workspace geometry would intersect the workspace obstacle.
Let the obstacles in the workspace be the set of points O, and let the set of all points on
the robot in configuration ¢ be .A(g). Then the C-space obstacle is defined as
Free space Degrees of freedom (DOF)
Forward kinematics
Cons={q:q€CandA(q)NO# {}}
and the free space is Cjree = C — Cops.
The C-space becomes more interesting for robots with moving parts. Consider the two-
link arm from Figure 26.12(a). It is bolted to a table so the base does not move, but the arm
has two joints that move independently—we call these degrees of freedom (DOF). Moving
the joints alters the (x,y) coordinates of the elbow, the gripper, and every point on the arm. The arm’s configuration space is two-dimensional: (Bguous ferp), Where Oy, is the angle of the shoulder joint, and 6, is the angle of the elbow joint. Knowing the configuration for our two-link arm means we can determine where each
point on the arm is through simple trigonometry.
ping is a function
In general, the forward kinematics map-
¢p:C—=W
that takes in a configuration and outputs the location of a particular point b on the robot when
the robot is in that configuration. A particularly useful forward kinematics mapping is that for
the robot’s end effector, ¢z The set of all points on the robot in a particular configuration ¢ is denoted by A(q) C W:
Alg) =U{(@)}-
;
Inverse kinematics
The inverse problem, of mapping a desired location for a point on the robot to the config-
uration(s) the robot needs to be in for that to happen, is known as inverse kinematics:
IKy:x €W = {g €C st d(q) =x}.
Sometimes the inverse kinematics mapping might take not just a position, but also a desired
orientation as input. When we want a manipulator to grasp an object, for instance, we can
compute a desired position and orientation for its gripper, and use inverse kinematics to de-
termine a goal configuration for the robot. Then a planner needs to find a way to get the robot from its current configuration to the goal configuration without intersecting obstacles.
Workspace obstacles are often depicted as simple geometric forms—especially in robotics textbooks, which tend to focus on polygonal obstacles. But how do the obstacles look in configuration space?
Section 26.5 Planning and Control
Felby Fshou,
(b)
(a)
Figure 26.12 (a) Workspace representation of a robot arm with two degrees of freedom. The workspace is a box with a flat obstacle hanging from the ceiling. (b) Configuration space of the same robot. Only white regions in the space are configurations that are free of collisions. The dot in this diagram corresponds to the configuration of the robot shown on the left.
For the two-link arm, simple obstacles in the workspace, like a vertical line, have very
complex C-space counterparts, as shown in Figure 26.12(b). The different shadings of the occupied space correspond to the different objects in the robot’s workspace: the dark region
surrounding the entire free space corresponds to configurations in which the robot collides
with itself. It is easy to see that extreme values of the shoulder or elbow angles cause such a violation. The two oval-shaped regions on both sides of the robot correspond to the table on which the robot is mounted. The third oval region corresponds to the left wall.
Finally, the most interesting object in configuration space is the vertical obstacle that
hangs from the ceiling and impedes the robot’s motions.
This object has a funny shape in
configuration space: it is highly nonlinear and at places even concave. With a little bit of imagination the reader will recognize the shape of the gripper at the upper left end. ‘We encourage the reader to pause for a moment and study this diagram. The shape of this
obstacle in C-space is not at all obvious!
The dot inside Figure 26.12(b) marks the configu-
ration of the robot in Figure 26.12(a). Figure 26.13 depicts three additional configurations,
both in workspace and in configuration space. In configuration conf-1, the gripper is grasping the vertical obstacle.
We see that even if the robot’s workspace is represented by flat polygons, the shape of
the free space can be very complicated. In practice, therefore, one usually probes a configu-
ration space instead of constructing it explicitly. A planner may generate a configuration and
then test to see if it is in free space by applying the robot kinematics and then checking for collisions in workspace coordinates.
941
942
Chapter 26 Robotics
conf-3
conf-1
conf-2
() (a) Figure 26.13 Three robot configurations, shown in workspace and configuration space. 26.5.2 Motion planning
Motion planning
The motion planning problem is that of finding a plan that takes a robot from one configuration to another without colliding with an obstacle. It is a basic building block for movement and manipulation.
In Section 26.5.4 we will discuss how to do this under complicated dy-
namics, like steering a car that may drift off the path if you take a curve too fast. For now, we
will focus on the simple motion planning problem of finding a geometric path that is collision
free. Motion planning is a quintessentially continuous-state search problem, but it is often
Piano mover's problem
possible to discretize the space and apply the search algorithms from Chapter 3.
The motion planning problem is sometimes referred to as the piano mover’s problem. It
gets its name from a mover’s struggles with getting a large, irregular-shaped piano from one room to another without hitting anything. We are given:
« a workspace world W in either R? for the plane or R? for three dimensions,
an obstacle region O C W,
+ arobot with a configuration space C and set of points A(g) for g € C, * astarting configuration ¢ € C, and
« a goal configuration g € C. The obstacle region induces a C-space obstacle C,p, and its corresponding free space Cyee defined as in the previous section. We need to find a continuous path through free space. We
will use a parameterized curve, 7(r), to represent the path, where 7(0) = g, and 7(1) = g, and 7(r) for every
between 0 and 1 is some point in Cyy.. That is, ¢ parameterizes how
far we are along the path, from start to goal. Note that ¢ acts somewhat like time in that as
1 increases the distance along the path increases, but7 is always a point on the interval [0, 1] and is not measured in seconds.
Section 265
Planning and Control
943
4
o
Figure 26.14 A visibility graph. Lines connect every pair of vertices that can “see” each other—lines that don’t go through an obstacle. The shortest path must lie upon these lines. The motion planning problem can be made more complex in various ways: defining the
goal as
a set of possible configurations rather than a single configuration; defining the goal
in the workspace rather than the C-space; defining a cost function (e.g.,
path length) to be
minimized; satisfying constraints (e.g., if the path involves carrying a cup of coffee, making
sure that the cup is always oriented upright so the coffee does not spill).
The spaces of motion planning: Let’s take a step back and make sure we understand the
spaces involved in motion planning. First, there is the workspace or world W. Points in W are points in the everyday three-dimensional world. Next, we have the space of configurations,
C. Points g in C are d-dimensional, with d the robot’s number of degrees of freedom, and
map to sets of points .A(g) in W. Finally, there is the space of paths. The space of paths is a
space of functions. Each point in this space maps to an entire curve through C-space. This space is co-dimensional! Intuitively, we need d dimensions for each configuration along the path, and there are as many configurations on a path as there are points in the number line interval [0, 1]. Now let’s consider some ways of solving the motion planning problem. Visi
ility graphs
For the simplified case of two-dimensional configuration spaces and polygonal C-space ob-
stacles, vi ity graphs are a convenient way to solve the motion planning problem with a guaranteed shortest-path solution. Let Vops C C be the set of vertices of the polygons making
up Cops, and let V = VoU {gs.q5}-
We construct a graph G = (V,E) on the vertex set V with edges e;; € E connecting a
vertex v; to another vertex v; if the line connecting the two vertices
is collision-free—that is,
if {Avi+ (1= A)v;: A€ [0,1]}NCops = { }. When this happens, we say the two vertices “can
see each other,” which is where “visibility” graphs got their name. To solve the motion planning problem, all we need to do is run a discrete graph search (e.g., best-first search) on the graph G with starting state g, and goal g,. In Figure 26.14
we see a visibility graph and an optimal three-step solution. An optimal search on visibility
graphs will always give us the optimal path (if one exists), or report failure if no path exists.
Voronoi
diagrams
Visibility graphs encourage paths that run immediately adjacent to an obstacle—if you had to walk around a table to get to the door, the shortest path would be to stick as close to the table
as possible. However, if motion or sensing is nondeterministic, that would put you at risk of
bumping into the table. One way to address this is to pretend that the robot’s body is a bit
Visibility graph
Chapter 26 Robotics
Figure 26.15 A Voronoi diagram showing the set of points (black lines) equidistant to two
or more obstacles in configuration space.
larger than it actually is, providing a buffer zone. Another way is to accept that path length is not the only metric we want to optimize. Section 26.8.2 shows how to learn a good metric from human examples of behavior.
Voronoi diagram Region
Voronoi graph
A third way is to use a different technique, one that puts paths as far away from obstacles
as possible rather than hugging close to them. A Voronoi diagram is a representation that
allows us to do just that. To get an idea for what a Voronoi diagram does, consider a space
where the obstacles are, say, a dozen small points scattered about a plane. Now surround each
of the obstacle points with a region consisting of all the points in the plane that are closer to that obstacle point than to any other obstacle point. Thus, the regions partition the plane. The
Voronoi diagram consists
and vertices of the regions.
of the set of regions, and the Voronoi graph consists of the edges
‘When obstacles are areas, not points, everything stays pretty much the same. Each region
still contains all the points that are closer to one obstacle than to any other, where distance is
measured to the closest point on an obstacle. The boundaries between regions still correspond
to points that are equidistant between two obstacles, but now the boundary may be a curve
rather than a straight line. Computing these boundaries can be prohibitively expensive in high-dimensional spaces.
To solve the motion planning problem, we connect the start point g, to the closest point
on the Voronoi graph via a straight line, and the same for the goal point g,. We then use discrete graph search to find the shortest path on the graph. For problems like navigating
through corridors indoors, this gives a nice path that goes down the middle of the corridor.
However, in outdoor settings it can come up with inefficient paths, for example suggesting an
unnecessary 100 meter detour to stick to the middle of a wide-open 200-meter space.
Section 26.5
Planning and Control
945
joal
(a)
(b)
Figure 26.16 (a) Value function and path found for a discrete grid cell approximation of the configuration space. (b) The same path visualized in workspace coordinates. Notice how the robot bends its elbow to avoid a collision with the vertical obstacle. Cell decomposition
An alternative approach to motion planning is to discretize the C-space. Cell decomposition ~Cell decomposition methods decompose the free space into a finite number of contiguous regions, called cells. These cells are designed so that the path-planning problem within a single cell can be solved by simple means (e.g., moving along a straight line). The path-planning problem then becomes a discrete graph search problem (as with visibility graphs and Voronoi graphs) to find a path through a sequence of cel The simplest cell decomposition consists of a regularly spaced grid. Figure 26.16(a) shows a square grid decomposition of the space and a solution path that is optimal for this grid size. Grayscale shading indicates the value of each free-space grid cell—the cost of the shortest path from that cell to the goal. (These values can be computed by a deterministic form of the VALUE-ITERATION algorithm given in Figure 17.6 on page 573.) Figure 26.16(b)
shows the corresponding workspace trajectory for the arm. Of course, we could also use the A" algorithm to find a shortest path.
This grid decomposition has the advantage that it
from three limitations.
simple to implement, but it suffers
First, it is workable only for low-dimensional configuration spaces,
because the number of grid cells increases exponentially with d, the number of dimensions. (Sounds familiar? This is the curse of dimensionality.) Second, paths through discretized
state space will not always be smooth. We see in Figure 26.16(a) that the diagonal parts of the path are jagged and hence very difficult for the robot to follow accurately. The robot can attempt to smooth out the solution path, but this is far from straightforward. Third, there is the problem of what to do with cells that are “mixed”—that
is, neither
entirely within free space nor entirely within occupied space. A solution path that includes
Chapter 26 Robotics such a cell may not be a real solution, because there may be no way to safely cross the
cell. This would make the path planner unsound.
On the other hand, if we insist that only
completely free cells may be used, the planner will be incomplete, because it might be the case that the only paths to the goal go through mixed cells—it might be that a corridor is actually wide enough for the robot to pass, but the corridor is covered only by mixed cells. The first approach to this problem is further subdivision of the mixed cells—perhaps using cells of half the original size. This can be continued recursively until a path is found
that lies entirely within free cells. This method works well and is complete if there is a way to
decide if a given cell is a mixed cell, which is easy only if the configuration space boundaries have relatively simple mathematical descriptions.
It is important to note that cell decomposition does not necessarily require explicitly rep-
Collision checker
resenting the obstacle space C,p,. We can decide to include a cell or not by using a collision checker. This is a crucial notion to motion planning. A collision checker is a function 7(g)
that maps to 1 if the configuration collides with an obstacle, and 0 otherwise. It is much easier
to check whether a specific configuration is in collision than to explicitly construct the entire
obstacle space Cyps. Examining the solution path shown in Figure 26.16(a), we can see an additional difficulty that will have to be resolved. The path contains arbitrarily sharp corners, but a physical robot has momentum and cannot change direction instantaneously. This problem can be solved by storing, for each grid cell, the exact continuous state (position and velocity) that was attained
when the cell was reached in the search. Assume further that when propagating information to nearby grid cells, we use this continuous state as a basis, and apply the continuous robot
motion model for jumping to nearby cells. So we don’t make an instantaneous 90° turn; we
make a rounded turn governed by the laws of motion. We can now guarantee that the resulting
Hybrid A
trajectory is smooth and can indeed be executed by the robot. One algorithm that implements
this is hybrid A*.
Randomized motion planning Randomized motion planning does graph search on a random decomposition of the configuration space, rather than a regular cell decomposition. The key idea is to sample a random set
of points and to create edges between them if there is a very simple way to get from one to
Probabilistic roadmap (PRM) Simple planner
the other (e.g., via a straight line) without colliding; then we can search on this graph.
A probabilistic roadmap (PRM) algorithm is one way to leverage this idea. We assume
access to a collision checker 7 (defined on page 946), and to a simple planner B(q;,q>) that
returns a path from g, to ¢, (or failure) but does so quickly. This simple planner is not going
to be complete—it might return failure even if a solution actually exists. Its job is to quickly
try to connect gy and g5 and let the main algorithm know if it succeeds. We will use it to
Milestone
define whether an edge exists between two vertices.
The algorithm starts by sampling M milestones—points in Cy—in addition to the points g, and g,. It uses rejection sampling, where configurations are sampled randomly and collision-checked using + until a total of M milestones are found.
uses the simple planner to try to connect pairs of milestones.
Next, the algorithm
If the simple planner returns
success, then an edge between the pair is added to the graph; otherwise, the graph remains as
is. We try to connect each milestone either to its k nearest neighbors (we call this &-PRM), or
to all milestones in a sphere of a radius r. Finally, the algorithm searches for a path on this
Section 26.5
@
Planning and Control
947
7
Figure 26.17 The probabilistic roadmap (PRM) algorithm. Top left: the start and goal configurations. Top right: sample M collision-free milestones (here M = 5). Bottom left: connect each milestone to its k nearest neighbors (here k = 3). Bottom right: find the shortest path from the start to the goal on the resulting graph. graph from g to g. If no path is found, then M more milestones are sampled, added to the graph, and the process is repeated. Figure 26.17 shows a roadmap with the path found between two configurations. PRMs
Probabilistically are not complete, but they are what is called probabilistically complete—they will eventu- comp lete ally find a path, if one exists. Intuitively, this is because they keep sampling more milestones.
PRMs work well even in high-dimensional configuration spaces. PRM:s are also popular for multi-query planning, in which we have multiple motion Multi-query planning planning problems within the same C-space. Often, once the robot reaches a goal, it is called upon to reach another goal in the same workspace. PRMs are really useful, because the robot can dedicate time up front to constructing a roadmap, and amortize the use of that roadmap over multiple queries.
Rapidly-exploring random trees Rapidly exploring : ! of PRMs called rapidly exploring random trees (RRTS) is} popular for singleAn extension random trées query planning. We incrementally build two trees, one with g, as the oot and one with gg " * as the root.
Random
milestones are chosen, and an attempt is made to connect each new
milestone to the existing trees. If a milestone connects both trees, that means a solution has been found, as in Figure 26.18. If not, the algorithm finds the closest point in each tree and
adds to the tree a new edge that extends from the point by a distance § towards the milestone.
This tends to grow the tree towards previously unexplored sections of the space. Roboticists love RRTs for their ease of use. However, RRT solutions are typically nonoptimal and lack smoothness. Therefore, RRTs are often followed by a post-processing step.
The most common one is “short-cutting,” in which we randomly select one of the vertices on
the solution path and try to remove it by connecting its neighbors to each other (via the simple
Chapter 26 Robotics
Gsample
qs Figure 26.18 The bidirectional RRT algorithm constructs two trees (one from the start, the other from the goal) by incrementally connecting each sample to the closest node in each tree, if the connection is possible. When a sample connects to both trees, that means we have found a solution path.
Figure 26.19 Snapshots of a trajectory produced by an RRT and post-processed with short-
cutting. Courtesy of Anca Dragan.
planner). We do this repeatedly for as many steps as we have compute time for. Even then, RRT"
the trajectories might look a little unnatural due to the random positions of the milestone that were selected, as shown in Figure 26.19.
RRT is a modification to RRT that makes the algorithm asymptotically optimal: the
solution converges to the optimal solution as more and more milestones are sampled. The key idea is to pick the nearest neighbor based on a notion of cost to come rather than distance
from the milestone only, and to rewire the tree, swapping parents of older vertices if it is
cheaper to reach them via the new milestone.
Trajectory optimization for kinematic planning Randomized sampling algorithms tend to first construct a complex but feasible path and then optimize it. Trajectory optimization does the opposite: it starts with a simple but infeasible
path, and then works to push it out of collision. The goal is to find a path that optimizes a cost
Section 265
Planning and Control
949
function! over paths. That is, we want to minimize the cost function J(7), where 7(0) = g, and 7(1) = g, J is called a functional because it is a function over functions.
The argument to J is
7, which is itself a function: 7(r) takes as input a point in the [0, 1] interval and maps it to
a configuration. A standard cost functional trades off between two important aspects of the robot’s motion: collision avoidance and efficiency,
J = Jobs + Mo
where the efficiency J;- measures the length of the path and may also measure smoothness. A convenient way to define efficiency is with a quadratic: it integrates the squared first derivative of 7 (we will see in a bit why this does in fact incentivize short paths):
Ja = [ 30 Pas. e
For the obstacle term, assume we can compute the distance d(x) from any point x € W to the nearest obstacle edge. This distance is positive outside of obstacles, 0 at the edge, and
negative inside. This is called a signed distance field. We can now define a cost field in the
workspace, call it ¢, that has high cost inside of obstacles, and a small cost right outside. With
Signed distance field
this cost, we can make points in the workspace really hate being inside obstacles, and dislike being right next to them (avoiding the visibility graph problem of their always hanging out by the edges of obstacles). Of course, our robot is not a point in the workspace, so we have some more work to do—we need to consider all points b on the robot’s body:
(fn(T \—ob () [l db ds ,»m—// ew
This
is called a path mtegral~n does not just integrate ¢ along the way for each body point,
but it multiplies by the derivative to make the cost invariant to retiming of the path. Imagine a robot sweeping through the cost field, accumulating cost as is moves. Regardless of how fast
Path integral
or slow the arm moves through the field, it must accumulate the exact same cost.
The simplest way to solve the optimization problem above and find a path is gradient
descent. If you are wondering how to take gradients of functionals with respect to functions,
something called the calculus of variations is here to help. It is especially easy for functionals of the form
'
Il
/0 F(s,7(5),#(5))ds
which are integrals of functions that depend just on the parameter s, the value of the function
Euler-Lagrange at s, and the derivative of the function at s. In such a case, the Euler-Lagrange equation equation says that the gradient is
Vo J(s)
)=
JF
= o
d
OF
5O " warm
If we look closely at Jo; and Jyp, they both follow this pattern. In particular for J,7, we have F(s,7(s),7(s)) = [|#(s)[|. To get a bit more comfortable with this, let’s compute the gradient
1 Roboticists like to minimize a cost function J, whereas in other parts of Al we try to maximize a utility function
U orareward R.
950
Chapter 26 Robotics
Figure 26.20 Trajectory optimization for motion planning. Two point-obstacles with circular bands of decreasing cost around them. The optimizer starts with the straight line trajectory, and lets the obstacles bend the line away from collisions, finding the minimum path through the cost field.
for Jo5 only. We see that F does not have a direct dependence on 7(s), so the first term in the formula is 0. We are left with d
Vel(s)=0-2(s9)
since the partial of F with respect to 7(s) is 7(s).
Notice how we made things easier for ourselves when defining Jo;—it’s a nice quadratic
of the derivative (and we even put a % in front so that the 2 nicely cancels out). In practice, you will see this trick happen a lot for optimization—the art is not just in choosing how to
optimize the cost function, but also in choosing a cost function that will play nicely with how you will optimize it. Simplifying our gradient, we get
V. J(s) = —#(s). Now, since J is a quadratic, setting this gradient to 0 gives us the solution for 7 if we
didn’t have to deal with obstacles. Integrating once, we get that the first derivative needs to be constant; integrating again we get that 7(s) = a- s+ b, with a and b determined by the endpoint constraints for 7(0) and 7(1). The optimal path with respect to J.g is thus the
straight line from start to goal! It is indeed the most efficient way to go from one to the other if there are no obstacles to worry about.
Of course, the addition of Jp, is what makes things difficult—and we will spare you
deriving its gradient here. The robot would typically initialize its path to be a straight line,
which would plow right through some obstacles. It would then calculate the gradient of the cost about the current path, and the gradient would serve to push the path away from the obstacles (Figure 26.20). Keep in mind that gradient descent will only find a locally optimal solution—just like hill climbing. Methods such as simulated annealing (Section 4.1.2) can be
used for exploration, to make it more likely that the local optimum is a good one. 26.5.3
Control theory
Trajectory tracking control
We have covered how to plan motions, but not how to actually move—to apply current to motors, to produce torque, to move the robot.
This is the realm of control theory, a field
of increasing importance in Al There are two main questions to deal with: how do we turn
Section 26.5 Planning and Control
951
Figure 26.21 The task of reaching to grasp a bottle solved with a trajectory optimizer. Left: the initial trajectory, plotted for the end effector. Middle: the final trajectory after optimiza-
tion. Right: the goal configuration. Courtesy of Anca Dragan. See Ratliff ef al. (2009).
a mathematical description of a path into a sequence of actions in the real world (open-loop
control), and how do we make sure that we are staying on track (closed-loop control)?
From configurations to torques for open-loop tracking: Our path 7(1) gives us configurations. The robot starts at rest at g, = 7(0). From there the robot’s motors will turn currents into torques, leading to motion. But what torques should the robot aim for, such that it ends
up at g = 7(1)?
This is where the idea of a dynamics model (or transition model) comes in. We can give
the robot a function f that computes the effects torques have on the configuration. Remem-
Dynamics model
ber F = ma from physics? Well, there is something like that for torques too, in the form
u= f"(.4.G), with u a torque, ¢ a velocity, and g an acceleration.? If the robot is at config-
uration ¢ and velocity ¢, and applied torque u, that would lead to acceleration G = f(q,q,u). The tuple (g.¢) is a dynamic state, because it includes velocity, whereas g is the kinematic
state and is not sufficient for computing exactly what torque to apply. f is a deterministic
dynamics model in the MDP over dynamic states with torques as actions. f~! is the inverse
Dynamic state
Kinematic state
dynamics, telling us what torque to apply if we want a particular acceleration, which leads Inverse dynamics to a change in velocity and thus a change in dynamic state.
Now, naively, we could think of 7 € [0, 1] as “time” on a scale from 0 to 1 and select our
torque using inverse dynamics:
) = £ (r(0,7(0),5(1))
(262)
assuming that the robot starts at (7(0),7(0)). In reality though, things are not that easy.
The path 7 was created as a sequence of points, without taking velocities and accelera-
tions into account. As such, the path may not satisfy 7(0) = 0 (the robot starts at 0 velocity), or even that 7 is differentiable (let alone twice differentiable).
Further, the meaning of the
endpoint “1” is unclear: how many seconds does that map to? In practice, before we even think of tracking a reference path, we usually retime it, that Retiming is, transform it into a trajectory £(r) that maps the interval [0, 7] for some time duration 7'
into points in the configuration space C. (The symbol is the Greek letter Xi.) Retiming is trickier than you might think, but there are approximate ways to do it, for instance by picking
a maximum velocity and acceleration, and using a profile that accelerates to that maximum 2 We omit the details of /=" here, but they involve mass, inertia, gravity, and Coriolis and centrifugal forces.
952
Chapter 26 Robotics
e
(@)
s
(®)
(c)
Figure 26.22 Robot arm control using (a) proportional control with gain factor 1.0, (b) proportional control with gain factor 0.1, and (c) PD (proportional derivative) control with gain factors 0.3 for the proportional component and 0.8 for the differential component. In all cases the robot arm tries to follow the smooth line path, but in (a) and (b) deviates substantially from
the path.
velocity, stays there as long as it can, and then decelerates back to 0. Assuming we can do this, Equation (26.2) above can be rewritten as
ur) = 1760 E0,€0)-
(263)
Even with the change from 7 to £, an actual trajectory, the equation of applying torques from
Control law
above (called a control law) has a problem in practice. Thinking back to the reinforcement
Stiction
and inertias exactly, and f might not properly account for physical phenomena like stiction in the motors (the friction that tends to prevent stationary surfaces from being set in motion—to
learning section, you might guess what it is. The equation works great in the situation where [ is exact, but pesky reality gets in the way as usual: in real systems, we can’t measure masses
make them stick). So, when the robot arm starts applying those torques but f is wrong, the
errors accumulate and you deviate further and further from the reference path.
Rather than just letting those errors accumulate, a robot can use a control process that
looks at where it thinks it is, compares that to where it wanted to be, and applies a torque to minimize the error.
P controller Gain factor
A controller that provides force in negative proportion to the observed error is known as
a proportional controller or P controller for short. The equation for the force
u(t) = Kp(&(1) — q:)
where g, is the current configuration, and Kp is a constant representing the gain factor of the
controller. Kp regulates how strongly the controller corrects for deviations between the actual state g; and the desired state &(t).
Figure 26.22(a) illustrates what can go wrong with proportional control. Whenever a deviation occurs—whether due to noise or to constraints on the forces the robot can apply—the
robot provides an opposing force whose magnitude is proportional to this deviation. Intuitively, this might appear plausible, since deviations should be compensated by a counterforce to keep the robot on track. However, as Figure 26.22(a) illustrates, a proportional controller
can cause the robot to apply too much force, overshooting the desired path and zig-zagging
Section 265 back and forth. This is the result of the natural inertia of the robot:
Planning and Control
953
once driven back to its
reference position the robot has a velocity that can’t instantaneously be stopped.
In Figure 26.22(a), the parameter Kp = 1. At first glance, one might think that choosing a smaller value for Kp would remedy the problem, giving the robot a gentler approach to the desired path.
Kp
Unfortunately, this is not the case. Figure 26.22(b) shows a trajectory for
=1, still exhibiting oscillatory behavior. The lower value of the gain parameter helps, but
does not solve the problem. In fact, in the absence of friction, the P controller is essentially a
spring law; so it will oscillate indefinitely around a fixed target location.
There are a number of controllers that are superior to the simple proportional control law.
A controller is said to be stable if small perturbations lead to a bounded error between the
robot and the reference signal. It is said to be strictly stable if it is able to return to and then
stay on its reference path upon such perturbations. Our P controller appears to be stable but not strictly stable, since it fails to stay anywhere near its reference trajectory.
The simplest controller that achieves strict stability in our domain is a PD controller. The letter ‘P’ stands again for proportional, and ‘D’ stands for derivative. PD controllers are
Stable Strictly stable PD controller
described by the following equation:
(26.4) + Kp(€(r) = dr) = a1) £() ult) = Kp( As this equation suggests, PD controllers extend P controllers by a differential component,
which adds to the value of u(r) a term that is proportional to the first derivative of the error
&(t) — g, over time. What is the effect of such a term? In general, a derivative term damp-
ens the system that is being controlled.
To see this, consider a situation where the error is
changing rapidly over time, as is the case for our P controller above. The derivative of this error will then counteract the proportional term, which will reduce the overall response to
the perturbation. However, if the same error persists and does not change, the derivative will
vanish and the proportional term dominates the choice of control.
Figure 26.22(c) shows the result of applying this PD controller to our robot arm, using as gain parameters Kp = .3 and K = .8. Clearly, the resulting path is much smoother, and does
not exhibit any obvious oscillations.
PD controllers do have failure modes, however. In particular, PD controllers may fail to
regulate an error down to zero, even in the absence of external perturbations.
Often such a
situation is the result ofa systematic external force that is not part of the model. For example,
an autonomous car driving on a banked surface may find itself systematically pulled to one
side.
Wear and tear in robot arms causes similar systematic errors.
In such situations, an
over-proportional feedback is required to drive the error closer to zero. The solution to this
problem lies in adding a third term to the control law, based on the integrated error over time:
u(t) = KP({(f)*q:)+K:A’({(-Y)*q\)d»“rKv(f(')*d:)-
(26.5)
Here K; is a third gain parameter. The term [}(€(s) calculates the integral of the error over time. The effect of this term is that long-lasting deviations between the reference signal and the actual state are corrected.
Integral terms, then, ensure that a controller does not exhibit
systematic long-term error, although they do pose a danger of oscillatory behavior. A controller with all three terms is called a PID controller (for proportional integral
derivative). PID controllers are widely used in industry, for a variety of control problems. Think of the three terms as follows—proportional: try harder the farther away you are from
PID controller
954
Chapter 26 Robotics the path; derivative: try even harder if the error is increasing; integral: try harder if you haven’t made progress for a long time.
Computed torque control
A middle ground between open-loop control based on inverse dynamics and closed-loop PID control is called computed torque control. We compute the torque our model thinks we. will need, but compensate for model inaccuracy with proportional error terms:
u(t) = §7 (60, £0).60)) +m(E)) (Kp(€(1) — 1) +Kp(£(1) = 1)) Feedforward component Feedback component
feedforward
(26.6)
feedback
The first term is called the feedforward component because it looks forward to where the
robot needs to go and computes what torque might be required. The second is the feedback
component because it feeds the current error in the dynamic state back into the control law.
m(q) is the inertia matrix at configuration g—unlike normal PD control, the gains change with the configuration of the system. Plans versus policies
Let’s take a step back and make sure we understand the analogy between what happened so far in this chapter and what we learned in the search, MDP, and reinforcement learning chapters.
‘With motion in robotics, we are really considering an underlying MDP where the states are
dynamic states (configuration and velocity), and the actions are control inputs, usually in the form of torques. If you take another look at our control laws above, they are policies, not plans—they tell the robot what action to take from any state it might reach. However,
they are usually far from optimal policies. Because the dynamic state is continuous and high dimensional (as is the action space), optimal policies are computationally difficult to extract. Instead, what we did here is to break up the problem. We come up with a plan first, in a simplified state and action space: we use only the kinematic state, and assume that states are reachable from one another without paying attention to the underlying dynamics. This is motion planning, and it gives us the reference path. If we knew the dynamics perfectly, we
could turn this into a plan for the original state and action space with Equation (26.3).
But because our dynamics model is typically erroneous, we turn it instead into a policy
that tries to follow the plan—getting back to it when it drifts away. When doing this, we introduce suboptimality in two ways:
first by planning without considering dynamics, and
second by assuming that if we deviate from the plan, the optimal thing to do is to return to the original plan. In what follows, we describe techniques that compute policies directly over the dynamic state, avoiding the separation altogether.
26.5.4
Optimal
control
Rather than using a planner to create a kinematic path, and only worrying about the dynamics
of the system after the fact, here we discuss how we might be able to do it all at once. We'll
take the trajectory optimization problem for kinematic paths, and turn it into true trajectory
optimization with dynamics: we will optimize directly over the actions, taking the dynamics (or transitions) into account. This brings us much closer to what we’ve seen in the search and MDP chapters.
If we
know the system’s dynamics, then we can find a sequence of actions to execute, as we did in
Chapter 3. If we’re not sure, then we might want a policy, as in Chapter 17.
Section 265
Planning and Control
955
In this section, we are looking more directly at the underlying MDP the robot works
in. We're switching from the familiar discrete MDPs to continuous ones.
We will denote
our dynamic state of the world by x, as is common practice—the equivalent of s in discrete MDPs. Let x, and x; be the starting and goal states.
‘We want to find a sequence of actions that, when executed by the robot, result in state-
action pairs with low cumulative cost. The actions are torques which we denote with u(r)
for ¢ starting at 0 and ending at 7. Formally, we want to find the sequence of torques u that minimize a cumulative cost J:
min Ji J(x(t),u(r))dr
(26.7)
u
subject to the constraints
Ve, x(1) = f(x(0),u(t))
X(0) = xy, X(T) =%,
How is this connected to motion planning and trajectory tracking control? Well, imagine
we take the notion of efficiency and clearance away from the obstacles and put it into the cost function J, just as we did before in trajectory optimization over kinematic state. The dynamic
state is the configuration and velocity, and torques u change it via the dynamics
f from open-
loop trajectory tracking. The difference is that now we're thinking about the configurations
and the torques at the same time. Sometimes, we might want to treat collision avoidance as a
hard constraint as well, something we’ve also mentioned before when we looked at trajectory optimization for the kinematic state only.
To solve this optimization problem, we can take gradients of J—not with respect to the
sequence 7 of configurations anymore, but directly with respect to the controls . It is sometimes helpful to include the state sequence x as a decision variable too, and use the dynamics
constraints to ensure that x and u are consistent. There are various trajectory optimization
techniques using this approach; two of them go by the names multiple shooting and direct
collocation. None of these techniques will find the global optimal solution, but in practice they can effectively make humanoid robots walk and make autonomous cars drive. Magic happens when in the problem above, J is quadratic and f is linear in x and 1. We want to minimize
min / XTQx+ulRudt
o
subjectto
Vi, i(r) = Ax(r) + Bu(t).
‘We can optimize over an infinite horizon rather than a finite one, and we obtain a policy
from any state rather than just a sequence of controls. Q and R need to be positive definite matrices for this to work. This gives us the linear quadratic regulator (LQR). With LQR,
Linear quadratic regulator (LQR)
the optimal value function (called cost to go) is quadratic, and the optimal policy is linear. The policy looks like u = —Kx, where finding the matrix K requires solving an algebraic Riceati equation—no local optimization, no value iteration, no policy iteration are needed! Riccati equation Because of the ease of finding the optimal policy, LQR finds many uses in practice despite the fact that real problems seldom actually have quadratic costs and linear dynamics. A LQR really useful method is called iterative LQR (ILQR), which works by starting with a solu- Iterative (ILQR) tion and then iteratively computing a linear approximation of the dynamics and a quadratic approximation of the cost around it, then solving the resulting LQR system to arrive at a new
solution. Variants of LQR are also often used for trajectory tracking.
956
Chapter 26 Robotics 26.6_Planning Uncertain
Movements
In robotics, uncertainty arises from partial observability of the environment and from the stochastic (or unmodeled) effects of the robot’s actions. Errors can also arise from the use
of approximation algorithms
such as particle filtering, which does not give the robot an exact
belief state even if the environment is modeled perfectly.
The majority of today’s robots use deterministic algorithms for decision making, such as the path-planning algorithms of the previous section, or the search algorithms that were introduced in Chapter 3. These deterministic algorithms are adapted in two ways: first, they
deal with the continuous state space by turning it into a discrete space (for example with Most likely state
visibility graphs or cell decomposition).
Second, they deal with uncertainty in the current
state by choosing the most likely state from the probability distribution produced by the state estimation algorithm. That approach makes the computation faster and makes a better fit for the deterministic search algorithms. In this section we discuss methods for dealing with
uncertainty that are analogous to the more complex search algorithms covered in Chapter 4.
First, instead of deterministic plans, uncertainty calls for policies. We already discussed
how trajectory tracking control turns a plan into a policy to compensate for errors in dynamics.
Online replanning Model predictive
control (MPC)
Sometimes though, if the most likely hypothesis changes enough, tracking the plan designed for a different hypothesis is too suboptimal. This is where online replanning comes in: we can recompute a new plan based on the new belief. Many robots today use a technique called model predictive control (MPC), where they plan for a shorter time horizon, but replan
at every time step. (MPC is therefore closely related to real-time search and game-playing
algorithms.) This effectively results in a policy: at every step, we run a planner and take the
first action in the plan; if new information comes along, or we end up not where we expected, that’s OK, because we are going to replan anyway and that will tell us what to do next. Second, uncertainty calls for information gathering actions. When we consider only the information we have and make a plan based on it (this is called separating estimation from
control), we are effectively solving (approximately) a new MDP at every step, corresponding
to our current belief about where we are or how the world works. But in reality, uncertainty is
better captured by the POMDP framework: there is something we don’t directly observe, be it the robot’s location or configuration, the location of objects in the world, or the parameters
of the dynamics model itself—for example, where exactly is the center of mass of link two on this arm?
What we lose when we don’t solve the POMDP
is the ability to reason about future
information the robot will get: in MDPs we only plan with what we know, not with what we might eventually know. Remember the value of information? Well, robots that plan using
their current belief as if they will never find out anything more fail to account for the value of
information. They will never take actions that seem suboptimal right now according to what
they know, but that will actually result in a lot of information and enable the robot to do well.
What does such an action look like for a navigation robot?
The robot could get close
to a landmark to get a better estimate of where it is, even if that landmark is out of the way
according to what it currently knows. This action is optimal only if the robot considers the
new observations it will get, as opposed to looking only at the information it already has.
Guarded movement
To get around this, robotics techniques sometimes define information gathering actions
explicitly—such as moving a hand until it touches a surface (called guarded movements)—
Section 26.6 initial configuration
Planning Uncertain Movements
-
motion
envelope —
957
cy \2
Figure 26.23 A two-dimensional environment, velocity uncertainty cone, and envelope of possible robot motions. The intended velocity is v, but with uncertainty the actual velocity could be anywhere in C,. resulting in a final configuration somewhere in the motion envelope, which means we wouldn’t know if we hit the hole or not. and make sure the robot does that before coming up with a plan for reaching its actual goal. Each guarded motion consists of (1) a motion command
and (2) a termination condition,
which is a predicate on the robot’s sensor values saying when to stop. Sometimes, the goal itself could be reached via a sequence of guarded moves guaranteed to succeed regardless of uncertainty. As an example, Figure 26.23 shows a two-dimensional
configuration space with a narrow vertical hole. It could be the configuration space for inser-
tion ofa rectangular peg into a hole or a car key into the ignition. The motion commands are constant velocities. The termination conditions are contact with a surface. To model uncer-
tainty in control, we assume that instead of moving in the commanded direction, the robot’s actual motion lies in the cone C, about it.
The figure shows what would happen if the robot attempted to move straight down from
the initial configuration. Because of the uncertainty in velocity, the robot could move anywhere in the conical envelope, possibly going into the hole, but more likely landing to one side of it. Because the robot would not then know which side of the hole it was on, it would
not know which way to move.
A more sensible strategy is shown in Figures 26.24 and 26.25. In Figure 26.24, the robot deliberately moves to one side of the hole. The motion command is shown in the figure, and the termination test is contact with any surface. In Figure 26.25, a motion command is
given that causes the robot to slide along the surface and into the hole. Because all possible
velocities in the motion envelope are to the right, the robot will slide to the right whenever it
is in contact with a horizontal surface.
It will slide down the right-hand vertical edge of the hole when it touches it, because
all possible velocities are down relative to a vertical surface.
It will keep moving until it
reaches the bottom of the hole, because that is its termination condition.
In spite of the
control uncertainty, all possible trajectories of the robot terminate in contact with the bottom of the hole—that is, unless surface irregularities cause the robot to stick in one place.
Other techniques beyond guarded movements change the cost function to incentivize ac-
tions we know will lead to information—Tlike the coastal navigation heuristic which requires
the robot to stay near known landmarks. More generally, techniques can incorporate the ex-
pected information gain (reduction of entropy of the belief) as a term in the cost function,
Coastal navigation
958
Chapter 26
Robotics
initial configuration
7Cv
~
v motion
envelope
Figure 26.24 The first motion command and the resulting envelope of possible robot motions. No matter what actual motion ensues, we know the final configuration will be to the left of the hole.
DN
motion
v
/ envelope
Figure 26.25 The second motion command and the envelope of possible motions. Even with error, we will eventually get into the hole. leading to the robot explicitly reasoning about how much information each action might bring
when deciding what to do. While more difficult computationally, such approaches have the
advantage that the robot invents its own information gathering actions rather than relying on human-provided heuristics and scripted strategies that often lack flexibility. 26.7
Reinforcement Learning in Robotics
Thus far we have considered tasks in which the robot has access to the dynamics model of
the world. In many tasks, it is very difficult to write down such a model, which puts us in the
domain of reinforcement learning (RL).
One challenge of RL in robotics is the continuous nature of the state and action spaces,
which we handle either through discretization, or, more commonly, through function approxi-
mation. Policies or value functions are represented as combinations of known useful features,
or as deep neural networks. Neural nets can map from raw inputs directly to outputs, and thus largely avoid the need for feature engineering, but they do require more data. A bigger challenge is that robots operate in the real world. We have seen how reinforcement learning can be used to learn to play chess or Go by playing simulated games. But when
a real robot moves in the real world, we have to make sure that its actions are safe (things
Section 26.7
Reinforcement Learning in Robotics
(@)
Figure 26.26 Training a robust policy. (a) Multiple simulations are run of a robot hand ma-
nipulating objects, with different randomized parameters for physics and lighting. Courtesy of Wojciech Zaremba. (b) The real-world environment, with a single robot hand in the center
of a cage, surrounded by cameras and range finders. (c) Simulation and real-world training yields multiple different policies for grasping objects; here a pinch grasp and a quadpod grasp. Courtesy of OpenAL See Andrychowicz et al. (2018a).
break!), and we have to accept that progress will be slower than in a simulation because the
world refuses to move faster than one second per second. Much of what is interesting about using reinforcement learning in robotics boils down to how we might reduce the real world
sample complexity—the number of interactions with the physical world that the robot needs before it has learned how to do the task.
26.7.1
Exploiting models
A natural way to avoid the need for many real-world samples is to use as much knowledge of the world’s dynamics as possible. For instance, we might not know exactly what the coefficient of friction or the mass of an object is, but we might have equations that describe
the dynamics as a function of these parameters.
In such a case, model-based reinforcement learning (Chapter 22) is appealing, where
the robot can alternate between fitting the dynamics parameters and computing a better policy. Even if the equations are incorrect because they fail to model every detail of physics,
researchers have experimented with learning an error term, in addition to the parameters, that can compensate for the inaccuracy of the physical model. Or, we can abandon the equations
and instead fit locally linear models of the world that each approximate the dynamics in a region of the state space, an approach that has been successful in getting robots to master
complex dynamic tasks like juggling. A model of the world can also be useful in reducing the sample complexity of model-free
reinforcement learning methods by doing sim-to-real transfer: transferring policies that work ~sim-to-rea!
959
960
Chapter 26 Robotics in simulation to the real world. The idea is to use the model as a simulator for a policy search (Section 22.5). To learn a policy that transfers well, we can add noise to the model during
Domain randomization
training, thereby making the policy more robust. Or, we can train policies that will work with a variety of models by sampling different parameters in the simulations—sometimes referred to as domain randomization. An example is in Figure 26.26, where a dexterous manipulation task is trained in simulation by varying visual attributes, as well as physical attributes like friction or damping.
Finally, hybrid approaches that borrow ideas from both model-based and model-free al-
gorithms are meant to give us the best of both. The hybrid approach originated with the Dyna
architecture, where the idea was to iterate between acting and improving the policy, but the
policy improvement would come in two complementary ways:
1) the standard model-free
way of using the experience to directly update the policy, and 2) the model-based way of using the experience to fit a model, then plan with it to generate a policy.
More recent techniques have experimented with fitting local models, planning with them
to generate actions, and using these actions as supervision to fit a policy, then iterating to get better and better models around the areas that the policy needs.
This has been successfully
applied in end-to-end learning, where the policy takes pixels as input and directly outputs
torques as actions—it enabled the first demonstration of deep RL on physical robots. Models can also be exploited for the purpose of ensuring safe exploration.
Learning
slowly but safely may be better than learning quickly but crashing and burning half way through. So arguably, more important than reducing real-world samples is reducing realworld samples in dangerous states—we
don’t want robots falling off cliffs, and we don’t
want them breaking our favorite mugs or, even worse, colliding with objects and people. An approximate model, with uncertainty associated to it (for example by considering a range of
values for its parameters), can guide exploration and impose constraints on the actions that the robot is allowed to take in order to avoid these dangerous states. This is an active area of research in robotics and control.
26.7.2
Motion primitive
Exploiting other information
Models are useful, but there is more we can do to further reduce sample complexity. When setting up a reinforcement learning problem, we have to select the state and action spaces, the representation of the policy or value function, and the reward function we’re using. These decisions have a large impact on how easy or how hard we are making the problem. One approach is to use higher-level motion primitives instead of low-level actions like torque commands. A motion primitive is a parameterized skill that the robot has. For exam-
ple, a robotic soccer player might have the skill of “pass the ball to the player at (x,y).” All the policy needs to do is to figure out how to combine them and set their parameters, instead
of reinventing them. This approach often learns much faster than low-level approaches, but does restrict the space of possible behaviors that the robot can learn.
Another way to reduce the number of real-world samples required for learning is to reuse information from previous learning episodes on other tasks, rather than starting from scratch.
This falls under the umbrella of metalearning or transfer learning.
Finally, people are a great source of information. In the next section, we talk about how
to interact with people, and part of it is how to use their actions to guide the robot’s learning.
Section 26.8
26.8
Humans and Robots
961
Humans and Robots
Thus far, we’ve focused on a robot planning and learning how to act in isolation. This is useful for some robots, like the rovers we send out to explore distant planets on our behalf.
But, for the most part, we do not build robots to work in isolation. We build them to help us, and to work in human environments, around and with us.
This raises two complementary challenges. First is optimizing reward when there are
people acting in the same environment as the robot. We call this the coordination problem
(see Section 18.1). When the robot’s reward depends on not just its own actions, but also the
actions that people take, the robot has to choose its actions in a way that meshes well with
theirs. When the human and the robot are on the same team, this turns into collaboration.
Second is the challenge of optimizing for what people actually want. If a robot is to
help people, its reward function needs to incentivize the actions that people want the robot to
execute. Figuring out the right reward function (or policy) for the robot is itself an interaction problem. We will explore these two challenges in turn. 26.8.1
Coordination
Let’s assume for now, as we have been, that the robot has access to a clearly defined reward
function. But, instead of needing to optimize it in isolation, now the robot needs to optimize
it around a human who is also acting. For example, as an autonomous car merges on the
highway, it needs to negotiate the maneuver with the human driver coming in the target lane—
should it accelerate and merge in front, or slow down and merge behind? Later, as it pulls to
astop sign, preparing to take a right, it has to watch out for the cyclist in the bicycle lane, and for the pedestrian about to step onto the crosswalk.
Or, consider a mobile robot in a hallway.
Someone heading straight toward the robot
steps slightly to the right, indicating which side of the robot they want to pass on. The robot has to respond, clarifying its intentions. Humans
as approximately
rational agents
One way to formulate coordination with a human is to model it as a game between the robot
and the human (Section 18.2). With this approach, we explicitly make the assumption that
people are agents incentivized by objectives. This does not automatically mean that they are
perfectly rational agents (i.e., find optimal solutions in the game), but it does mean that the
robot can structure the way it reasons about the human via the notion of possible objectives that the human might have. In this game: « the state of the environment captures the configurations of both the robot and human agents; call itx = (xg,xg); + each agent can take actions, ug and uy respectively;
« cach agent has an objective that can be represented as a cost, Jg and Jy: each agent
wants to get to its goal safely and efficiently; + and, as in any game, each objective depends on the state and on the actions of both
agents: Jg(x, g, yr) and Jy (x, 1y, ug). Think of the car-pedestrian interaction—the car should stop if the pedestrian crosses, and should go forward if the pedestrian waits.
Three important aspects complicate this game. First is that the human and the robot don’t
necessarily know each other’s objectives. This makes it an incomplete information game.
Incomplete information game
962
Chapter 26 Robotics Second is that the state and action spaces are continuous, as they’ve been throughout this
chapter. We learned in Chapter 5 how to do tree search to tackle discrete games, but how do we tackle continuous spaces? Third, even though at the high level the game model makes sense—humans do move, and they do have objectives—a human’s behavior might not always be well-characterized as a solution to the game. The game comes with a computational challenge not only for the robot, but for us humans too. It requires thinking about what the robot will do in response to what the person does, which depends on what the robot thinks the person will do, and pretty soon we get to “what do you think I think you think I think”— it’s turtles all the way down! Humans can’t deal with all of that, and exhibit certain suboptimalities.
This means that the
robot should account for these suboptimalities.
So, then, what is an autonomous car to do when the coordination problem is this hard?
We will do something similar to what we’ve done before in this chapter. For motion planning and control, we took an MDP and broke it up into planning a trajectory and then tracking it with a controller. Here too, we will take the game, and break it up into making predictions about human actions, and deciding what the robot should do given these predictions.
Predicting human action Predicting human actions is hard because they depend on the robot’s actions and vice versa. One trick that robots use is to pretend the person is ignoring the robot. The robot assumes people are noisily optimal with respect to their objective, which is unknown to the robot and
is modeled as no longer dependent on the robot’s actions: Jj (x,
). In particular, the higher
the value of an action for the objective (the lower the cost to go), the more likely the human
is to take it. The robot can create a model for P(uy | x,Jy), for instance using the softmax function from page 811:
Plug | %, Jig) o< &= Q)
(268)
with Q(x, ups3Jy) the Q-value function corresponding to Ji; (the negative sign is there because in robotics we like to minimize cost, not maximize reward). Note that the robot does not
assume perfectly optimal actions, nor does it assume that the actions are chosen based on
reasoning about the robot at all.
Armed with this model, the robot uses the human’s ongoing actions as evidence about Jp;.
If we have an observation model for how human actions depend on the human’s objective,
each human action can be incorporated to update the robot’s belief over what objective the person has
b (J) o< b(Ju)Purg | %) -
(26.9)
An example is in Figure 26.27: the robot is tracking a human’s location and as the human moves, the robot updates its belief over human goals. As the human heads toward the
windows, the robot increases the probability that the goal is to look out the window, and
decreases the probability that the goal is going to the kitchen, which is in the other direction.
This is how the human’s past actions end up informing the robot about what the human will do in the future. Having a belief about the human’s goal helps the robot anticipate what next actions the human will take. The heatmap in the figure shows the robot’s future
predictions: red is most probable; blue least probable.
Section 26.8
Humans and Robots
(b)
(©)
Figure 26.27 Making predictions by assuming that people are noisily rational given their goal: the robot uses the past actions to update a belief over what goal the person is heading 10, and then uses the belief to make predictions about future actions. (a) The map ofa room. (b) Predictions after seeing a small part of the person’s trajectory (white path):; (¢) Predictions after seeing more human actions: the robot now knows that the person is not heading to the hallway on the left, because the path taken so far would be a poor path if that were the person’s goal. Images courtesy of Brian D. Ziebart. See Ziebart et al. (2009). The same can happen in driving. We might not know how much another driver values
efficiency, but if we see them accelerate as someone is trying to merge in front of them, we
now know a bit more about them. And once we know that, we can better anticipate what they will do in the future—the same driver is likely to come closer behind us, or weave through traffic to get ahead. Once the robot can make predictions about human future actions, it has reduced its prob-
lem to solving an MDP. The human actions complicate the transition function, but as long as
the robot can anticipate what action the person will take from any future state, the robot can
calculate P(¥' | x,ug): it can compute P(up | x) from P(ug | x,Ju) by marginalizing over Ji,
and combine it with P(x’ | x,ug,up), the transition (dynamics) function for how the world updates based on both the robot’s and the human’s actions. In Section 26.5 we focused on how to solve this in continuous state and action spaces for deterministic dynamics, and in
Section 26.6 we discussed doing it with stochastic dynamics and uncertainty.
Splitting prediction from action makes it easier for the robot to handle interaction, but
sacrifices performance much from control.
plitting estimation from motion did, or splitting planning
A robot with this split no longer understands that its actions can influence what people end up doing. In contrast, the robot in Figure 26.27 anticipates where people will go and then optimizes for reaching its own goal and avoiding collisions with them. In Figure 26.28, we have an autonomous car merging on the highway. If it just planned in reaction to other cars, it might have to wait a long time while other cars occupy its target lane. In contrast, a car that reasons about prediction and action jointly knows that different actions it could take will
result in different reactions from the human. If it starts to assert itself, the other cars are likely to slow down a bit and make room. Roboticists are working towards coordinated interactions like this so robots can work better with humans.
963
Chapter 26 Robotics
(b)
Figure 26.28 (a) Left: An autonomous car (middle lane) predicts that the human driver (left lane) wants to keep going forward, and plans a trajectory that slows down and merges behind. Right: The car accounts for the influence its actions can have on human actions, and realizes
it can merge in front and rely on the human driver to slow down. (b) That same algorithm produces an unusual strategy at an intersection: the car realizes that it can make it more
likely for the person (bottom) to proceed faster through the intersection by starting to inch
backwards. Images courtesy of Anca Dragan. See Sadigh et al. (2016). Human
predi
ions about the robot
Incomplete information is often two-sided:
the robot does not know the human’s objective
and the human, in turn, does not know the robot’s objective—people need to be making predictions about robots. As robot designers, we are not in charge of how the human makes
predictions; we can only control what the robot does. However, the robot can act in a way to make it easier for the human to make correct predictions.
The robot can assume that
the human is using something roughly analogous to Equation (26.8) to estimate the robot’s objective Jg, and thus the robot will act so that its true objective can be easily inferred. A special case of the game is when the human
and the robot are on the same team,
working toward the same goal or objective: Jy = Jg. Imagine getting a personal home robot
Joint agent
that is helping you make dinner or clean up—these are examples of collaboration.
We can now define a joint agent whose actions are tuples of human-robot actions,
(upr, ug) and who optimizes for Jy (x,up,ug) = Jg(x,ug,u), and we're solving a regular planning problem. We compute the optimal plan or policy for the joint agent, and voila, we now know what the robot and human should do.
This would work really well if people were perfectly optimal. The robot would do its part
of the joint plan, the human theirs. Unfortunately, in practice, people don’t seem to follow the perfectly laid out joint-agent plan; they have a mind of their own! We've already learned one way to handle this though, back in Section 26.6. We called it model predictive control
(MPC): the idea was to come up with a plan, execute the first action, and then replan. That
way, the robot always adapts its plan to what the human is actually doing.
Let’s work through an example. Suppose you and the robot are in your kitchen, and have
decided to make waffles. You are slightly closer to the fridge, so the optimal joint plan would
Section 26.8
Humans and Robots
have you grab the eggs and milk from the fridge, while the robot fetches the flour from the cabinet. The robot knows this because it can measure quite precisely where everyone is. But
suppose you start heading for the flour cabinet. You are going against the optimal joint plan. Rather than sticking to it and stubbornly also going for the flour, the MPC robot recalculates
the optimal plan, and now that you are close enough to the flour it is best for the robot to grab the waffle iron instead.
If we know that people might deviate from optimality, we can account for it ahead of time.
In our example, the robot can try to anticipate that you are going for the flour the moment you take your first step (say, using the prediction technique above). Even if it is still technically
optimal for you to turn around and head for the fridge, the robot should not assume that’s
what is going to happen. Instead, the robot can compute a plan in which you keep doing what you seem to want.
Humans as black box agents ‘We don’t have to treat people as objective-driven, intentional agents to get robots to coordinate with us. An alternative model is that the human is merely some agent whose policy 7y “messes” with the environment dynamics.
The robot does not know 7y, but can model the
problem as needing to act in an MDP with unknown dynamics. We have seen this before: for general agents in Chapter 22, and for robots in particular in Section 26.7. The robot can fit a policy model 7y to human data, and use it to compute an optimal policy for itself. Due to scarcity of data, this has been mostly used so far at the task level. For
instance, robots have learned through interaction what actions people tend to take (in response
to its own actions) for the task of placing and drilling screws in an industrial assembly task. Then there is also the model-free reinforcement learning alternative:
the robot can start
with some initial policy or value function, and keep improving it over time via trial and error.
26.8.2
Learning to do what humans want
Another way interaction with humans comes into robotics is in Jg itself—the robot’s cost or reward function. The framework of rational agents and the associated algorithms reduce the
problem of generating good behavior to specifying a good reward function.
as for many other Al agents, getting the cost right is still difficult.
But for robots,
Take autonomous cars: we want them to reach the destination, to be safe, to drive com-
fortably for their passengers, to obey traffic laws, etc. A designer of such a system needs to trade off these different components of the cost function. The designer’s task is hard because robots are built to help end users, and not every end user is the same. We all have different
preferences for how aggressively we want our car to drive, etc.
Below, we explore two alternatives for trying to get robot behavior to match what we
actually want the robot to do.
The first is to learn a cost function from human input.
The
second is to bypass the cost function and imitate human demonstrations of the task.
Preference learning: Learning cost functions Imagine that an end user is showing a robot how to do a task. For instance, they are driving
the car in the way they would like it to be driven by the robot. Can you think of a way for the robot to use these actions—we call them “demonstrations”—to figure out what cost function it should optimize?
965
966
Chapter 26 Robotics
Figure 26.29 Left: A mobile robot is shown a demonstration that stays on the dirt road. Middle: The robot infers the desired cost function, and uses it in a new scene, knowing to
put lower cost on the road there. Right: The robot plans a path for the new scene that also
stays on the road, reproducing the preferences behind the demonstration. Images courtesy of Nathan Ratliff and James A. Bagnell. See Ratliff er al. (2006).
We have actually already seen the answer to this back in Section 26.8.1. There, the setup
was a little different: we had another person taking actions in the same space as the robot, and the robot needed to predict what the person would do. But one technique we went over for making these predictions was to assume that people act to noisily optimize some cost function
Jy, and we can use their ongoing actions as evidence about what cost function that is. We
can do the same here, except not for the purpose of predicting human behavior in the future, but rather acquiring the cost function the robot itself should optimize.
If the person drives
defensively, the cost function that will explain their actions will put a lot of weight on safety
and less so on efficiency. The robot can adopt this cost function as its own and optimize it when driving the car itself.
Roboticists have experimented with different algorithms for making this cost inference
computationally tractable. In Figure 26.29, we see an example of teaching a robot to prefer staying on the road to going over the grassy
terrain. Traditionally in such methods, the cost
function has been represented as a combination of hand-crafted features, but recent work has also studied how to represent it using a deep neural network, without feature engineering.
There are other ways for a person to provide input. A person could use language rather
than demonstration to instruct the robot.
A person could act as a critic, watching the robot
perform a task one way (or two ways) and then saying how well the task was done (or which way was better), or giving advice on how to improve.
Learning policies directly via imitation An alternative is to bypass cost functions and learn the desired robot policy directly. In our car example, the human’s demonstrations make for a convenient data set of states labeled by
the action the robot should take at each state: D = {(x;,u;)}. The robot can run supervised Behavioral cloning Generalization
learning to fit a policy
behavioral cloning.
: x — u, and execute that policy. This is called imitation learning or
A challenge with this approach is in generalization to new states. The robot does not
know why the actions in its database have been marked as optimal. It has no causal rule; all
it can do is run a supervised learning algorithm to try to learn a policy that will generalize to unknown states. However, there is no guarantee that the generalization will be correct.
Section 26.8
Humans and Robots
Figure 26.30 A human teacher pushes the robot down to teach it to stay closer to the table. The robot appropriately updates its understanding of the desired cost function and starts optimizing it. Courtesy of Anca Dragan. See Bajcsy f al. (2017).
Figure 26.31 A programming interface that involves placing specially designed blocks in the robot’s workspace to select objects and specify high-level actions. Images courtesy of Maya Cakmak. See Sefidgar ef al. (2017). The ALVINN autonomous car project used this approach, and found that even when
starting from a
state in D, 7 will make small errors, which will take the car off the demon-
strated trajectory. There, 7 will make a larger error, which will take the car even further off the desired course.
‘We can address this at training time if we interleave collecting labels and learning: start
with a demonstration, learn a policy, then roll out that policy and ask the human for what action to take at every state along the way, then repeat. The robot then learns how to correct its mistakes as it deviates from the human’s desired actions.
Alternatively, we can address it by leveraging reinforcement learning. The robot can fit a
dynamics model based on the demonstrations, and then use optimal control (Section 26.5.4)
to generate a policy that optimizes for staying close to the demonstration.
A version of this
has been used to perform very challenging maneuvers at an expert level in a small radiocontrolled helicopter (see Figure 22.9(b)). The DAGGER (Data Aggregation) system starts with a human expert demonstration. From that it learns a policy, 7 and uses the policy to generate a data set D. Then from
D it generates a new policy 75 that best imitates the original human data. This repeats, and
967
968
Chapter 26 Robotics on the nth iteration it uses 7, to generate more data, to be added to D, which is then used to create 7, . In other words, at each iteration the system gathers new data under the current
policy and trains the next policy using all the data gathered so far. Related recent techniques use adversarial training:
they alternate between training a
classifier to distinguish between the robot’s learned policy and the human’s demonstrations,
and training a new robot policy via reinforcement learning to fool the classifier. These ad-
vances enable the robot to handle states that are near demonstrations, but generalization to
far-off states or to new dynamics is a work in progress.
Teaching interfaces and the correspondence problem.
So far, we have imagined the
case of an autonomous car or an autonomous helicopter, for which human demonstrations use
the same actions that the robot can take itself: accelerating, braking, and steering. But what
happens if we do this for tasks like cleaning up the Kitchen table? We have two choices here: either the person demonstrates using their own body while the robot watches, or the person physically guides the robot’s effectors. Correspondence problem
The first approach is appealing because it comes naturally to end users. Unfortunately,
it suffers from the correspondence problem: how to map human actions onto robot actions.
People have different kinematics and dynamics than robots. Not only does that make it difficult to translate or retarget human motion onto robot motion (e.g., retargeting a five-finger
human grasp to a two-finger robot grasp), but often the high-level strategy a person might use is not appropriate for the robot.
Kinesthetic teaching Keyframe
Visual programming
The second approach, where the human teacher moves the robot’s effectors into the right positions, is called kinesthetic teaching. It is not easy for humans to teach this way, espe-
cially to teach robots with multiple joints. The teacher needs to coordinate all the degrees of
freedom as it is guiding the arm through the task. Researchers have thus investigated alter-
natives, like demonstrating keyframes as opposed to continuous trajectories, as well as the use of visual programming to enable end users to program primitives for a task rather than
demonstrate from scratch (Figure 26.31). Sometimes both approaches are combined.
26.9 Deliberative Reactive
Alternative Robo
Frameworks
Thus far, we have taken a view of robotics based on the notion of defining or learning a reward function, and having the robot optimize that reward function (be it via planning or learning), sometimes in coordination or collaboration with humans.
view of robotics, to be contrasted with a reactive view.
26.9.1
This is a deliberative
Reactive controllers
In some cases, it is easier to set up a good policy for a robot than to model the world and plan.
Then, instead of a rational agent, we have a reflex agent.
For example, picture a legged robot that attempts to lift a leg over an obstacle. We could
give this robot a rule that says lift the leg a small height 7 and move it forward, and if the leg
encounters an obstacle, move it back and start again at a higher height. You could say that is modeling an aspect of the world, but we can also think of& as an auxiliary variable of the robot controller, devoid of direct physical meaning. One such example is the six-legged (hexapod) robot, shown in Figure 26.32(a), designed for walking through rough terrain. The robot’s sensors are inadequate to obtain accurate
Section 26.9
969
Alternative Robotic Frameworks
retract, lift higher
liftup
setdown push backward (®)
Figure 26.32 (a) Genghis, a hexapod robot. (Image courtesy of Rodney A. Brooks.) (b) An
augmented finite state machine (AFSM) that controls one leg. The AFSM
reacts to sensor
feedback: if a leg is stuck during the forward swinging phase, it will be lifted increasingly higher.
‘models of the terrain for path planning. Moreover, even if we added high-precision cameras and rangefinders, the 12 degrees of freedom (two for each leg) would render the resulting path planning problem computationally difficult. It is possible, nonetheless, to specify a controller directly without an explicit environ-
mental model.
(We have already seen this with the PD controller, which was able to keep a
complex robot arm on target without an explicit model of the robot dynamics.)
For the hexapod robot we first choose a gait, or pattern of movement of the limbs. One
statically stable gait is to first move the right front, right rear, and left center legs forward (keeping the other three fixed), and then move the other three.
flat terrain.
Gait
This gait works well on
On rugged terrain, obstacles may prevent a leg from swinging forward.
This
problem can be overcome by a remarkably simple control rule: when a leg’s forward motion is blocked, simply retract it, lift it higher; and try again. The resulting controller is shown in Figure 26.32(b) as a simple finite state machine; it constitutes a reflex agent with state, where the internal state is represented by the index of the current machine state (s; through s4).
26.9.2
Subsumption
architectures
The subsumption architecture (Brooks, 1986) is a framework for assembling reactive controllers out of finite state machines.
Nodes in these machines may contain tests for certain
Subsumption architecture
sensor variables, in which case the execution trace of a finite state machine is conditioned
on the outcome of such a test. Arcs can be tagged with messages that will be generated when traversing them, and that are sent to the robot’s motors or to other finite state machines.
Additionally, finite state machines possess internal timers (clocks) that control the time it
takes to traverse an arc. The resulting machines are called augmented finite state machines
(AFSMs), where the augmentation refers to the use of clocks.
An example of a simple AFSM is the four-state machine we just talked about, shown in
Figure 26.32(b).
This AFSM
implements a cyclic controller, whose execution mostly does
not rely on environmental feedback. The forward swing phase, however, does rely on sensor feedback.
If the leg is stuck, meaning that it has failed to execute the forward swing, the
Augmented finite
state machine {AFSM)
970
Chapter 26 Robotics robot retracts the leg, lifts it up a little higher, and attempts to execute the forward swing once
again. Thus, the controller is able to react to contingencies arising from the interplay of the robot and its environment.
The subsumption architecture offers additional primitives for synchronizing AFSMs, and for combining output values of multiple, possibly conflicting AFSMs. In this way, it enables the programmer to compose increasingly complex controllers in a bottom-up fashion. In our example, we might begin with AFSMs for individual legs, followed by an AFSM for coordinating multiple legs. On top of this, we might implement higher-level behaviors such as collision avoidance, which might involve backing up and turning. The idea of composing robot controllers from AFSMs is quite intriguing. Imagine how difficult it would be to generate the same behavior with any of the configuration-space path-
planning algorithms described in the previous section. First, we would need an accurate model of the terrain. The configuration space of a robot with six legs, each of which is driven
by two independent motors, totals 18 dimensions (12 dimensions for the configuration of the
legs, and six for the location and orientation of the robot relative to its environment).
Even
if our computers were fast enough to find paths in such high-dimensional spaces, we would
have to worry about nasty effects such as the robot sliding down a slope.
Because of such stochastic effects, a single path through configuration space would al-
most certainly be too brittle, and even a PID controller might not be able to cope with such
contingencies. In other words, generating motion behavior deliberately is simply too complex a problem in some cases for present-day robot motion planning algorithms. Unfortunately, the subsumption architecture has its own problems. First, the AFSMs
are driven by raw sensor input, an arrangement that works if the sensor data is reliable and contains all necessary
information for decision making,
but fails if sensor data has to be
integrated in nontrivial ways over time. Subsumption-style controllers have therefore mostly
been applied to simple tasks, such as following a wall or moving toward visible light sources. Second, the lack of deliberation makes it difficult to change the robot’s goals.
A robot
with a subsumption architecture usually does just one task, and it has no notion of how to
modify its controls to accommodate different goals (just like the dung beetle on page 41). Third, in many real-world problems, the policy we want is often too complex to encode explicitly. Think about the example from Figure 26.28, of an autonomous car needing to negotiate a lane change with a human driver.
goes into the target lane.
We might start off with a simple policy that
But when we test the car, we find out that not every driver in the
target lane will slow down to let the car in. We might then add a bit more complexity: make
the car nudge towards the target lane, wait for a response form the driver in that lane, and
then either proceed or retreat back. But then we test the car, and realize that the nudging needs to happen at a different speed depending on the speed of the vehicle in the target lane, on whether there is another vehicle in front in the target lane, on whether there is a vehicle
behind the car in the initial, and so on. The number of conditions that we need to consider
to determine the right course of action can be very large, even for such a deceptively simple
maneuver. This in turn presents scalability challenges for subsumption-style architectures. All that said, robotics is a complex problem with many approaches: deliberative, reactive, or a mixture thereof; based on physics, cognitive models, data, or a mixture thereof. The right approach s still a subject for debate, scientific inquiry, and engincering prowess.
Section 26.10
Application Domains
971
Figure 2633 (a) A patient with a brain-machine interface controlling a robot arm to grab a drink. Tmage courtesy of Brown University. (b) Roomba, the robot vacuum cleaner. Photo by HANDOUT/KRT/Newscom. 26.10
Application Domains
Robotic technology is already permeating our world, and has the potential to improve our independence, health, and productivity. Here are some example applications. Home care:
Robots have started to enter the home to care for older adults and people
with motor impairments, assisting them with activities of daily living and enabling them to live more independently.
These include wheelchairs and wheelchair-mounted arms like the
Kinova arm from Figure 26.1(b). Even though they start off as being operated by a human di-
rectly, these robots are gaining more and more autonomy. On the horizon are robots operated
by brain-machine interfaces, which have been shown to enable people with quadriplegia to use a robot arm to grasp objects and even feed themselves (Figure 26.33(a)). Related to these are prosthetic limbs that intelligently respond to our actions, and exoskeletons that give us superhuman strength or enable people who can’t control their muscles from the waist down
to walk again.
Personal robots are meant to assist us with daily tasks
like cleaning and organizing, free-
ing up our time. Although manipulation still has a way to go before it can operate seamlessly
in messy, unstructured human environments, navigation has made some headway. In particu-
lar, many homes already enjoy a mobile robot vacuum cleaner like the one in Figure 26.33(b). Health care:
Robots assist and augment surgeons, enabling more precise, minimally
invasive, safer procedures with better patient outcomes. The Da Vinci surgical robot from Figure 26.34(a) is now widely deployed at hospitals in the U.S. Services: Mobile robots help out in office buildings, hotels, and hospitals.
Savioke has
put robots in hotels delivering products like towels or toothpaste to your room. The Helpmate and TUG
robots carry food and medicine in hospitals (Figure 26.34(b)), while Dili-
gent Robotics’ Moxi robot helps out nurses with back-end logistical responsibilities. Co-Bot
roams the halls of Carnegie Mellon University, ready to guide you to someone’s office. We
can also use telepresence robots like the Beam to attend meetings and conferences remotely,
or check in on our grandparents.
Telepresence robots
972
Chapter 26 Robotics
(b)
Figure 26.34 (a) Surgical robot in the operating room. Photo by Patrick Landmann/Science Source. (b) Hospital delivery robot. Photo by Wired.
Figure 26.35 (a) Autonomous car BOsS which won the DARPA Urban Challenge. Photo by Tangi Quemener/AFP/Getty Images/Newscom. Courtesy of Sebastian Thrun. (b) Aerial view showing the perception and predictions of the Waymo autonomous car (white vehicle with green track). Other vehicles (blue boxes) and pedestrians (orange boxes) are shown with anticipated trajectories. Road/sidewalk boundaries are in yellow. Photo courtesy of Waymo. Autonomous cars: Some of us are occasionally distracted while driving, by cell phone
calls, texts, or other distractions. The sad result: more than a million people die every year in traffic accidents. Further, many of us spend a lot of time driving and would like to recapture
some of that time. All this has led to a massive ongoing effort to deploy autonomous cars. Prototypes have existed since the 1980s, but progress was
stimulated by the 2005 DARPA
Grand Challenge, an autonomous vehicle race over 200 challenging kilometers of unrehearsed desert terrain.
Stanford’s Stanley vehicle completed the course in less than seven
hours, winning a $2 million prize and a place in the National Museum of American History.
Section 26.10
Application Domains
973
(b)
Figure 26.36 (a) A robot mapping an abandoned coal mine. (b) A 3D map of the mine acquired by the robot. Courtesy of Sebastian Thrun.
Figure 26.35(a) depicts BOss, which in 2007 won the DARPA
Urban Challenge, a compli-
cated road race on city streets where robots faced other robots and had to obey traffic rules.
In 2009, Google started an autonomous driving project (featuring many of the researchers who had worked on Stanley and BOSs), which has now spun off as Waymo. In 2018 Waymo started driverless testing (with nobody in the driver seat) in the suburbs of Pheonix,
zona.
Ari-
In the meantime, other autonomous driving companies and ride-sharing companies
are working on developing their own technology, while car manufacturers have been selling cars with more and more assistive intelligence, such as Tesla’s driver assist, which is meant
for highway driving. Other companies are targeting non-highway driving applications including college campuses and retirement communities. Still other companies are focused on non-passenger applications such as trucking, grocery delivery, and valet parking. Entertainment:
Disney has been using robots (under the name animatronics) in their
Driver assist
Animatronics
parks since 1963. Originally, these robots were restricted to hand-designed, open-loop, unvarying motion (and speech), but since 2009 a version called autonomatronics can generate Autonomatronics autonomous actions. Robots
also take the form of intelligent toys for children; for example,
Anki’s Cozmo plays games with children and may pound the table with frustration when it loses. Finally, quadrotors like Skydio’s R1 from Figure 26.2(b) act as personal photographers and videographers, following us around to take action shots as we ski or bike. Exploration and hazardous environments: Robots have gone where no human has gone before, including the surface of Mars. Robotic arms assist astronauts in deploying and retrieving satellites and in building the International Space Station. Robots also help explore
under the sea. They are routinely used to acquire maps of sunken ships. Figure 26.36 shows a robot mapping an abandoned coal mine, along with a 3D model of the mine acquired using range sensors. In 1996, a team of researches released a legged robot into the crater of an active volcano to acquire data for climate research. Robots are becoming very effective tools
for gathering information in domains that are difficult (or dangerous) for people to access. Robots have assisted people in cleaning up nuclear waste, most notably in Three Mile Island, Chernobyl, and Fukushima. Robots were present after the collapse of the World Trade
974
Chapter 26 Robotics Center, where they entered structures deemed too dangerous for human
search and rescue
crews. Here too, these robots are initially deployed via teleoperation, and as technology advances they are becoming more and more autonomous, with a human operator in charge but not having to specify every single command.
Industry: The majority of robots today are deployed in factories, automating tasks that
are difficult, dangerous, or dull for humans. (The majority of factory robots are in automobile
factories.) Automating these tasks is a positive in terms of efficiently producing what society needs. At the same time, it also means displacing some human workers from their jobs. This
has important policy and economics implications—the need for retraining and education, the need for a fair division of resources, etc. These topics are discussed further in Section 27.3.5.
Summary
Robotics is about physically embodied agents, which can change the state of the physical world. In this chapter, we have leaned the following: « The most common types of robots arc manipulators (robot arms) and mobile robots. They have sensors for perceiving the world and actuators that produce motion, which then affects the world via effectors. « The general robotics problem involves stochasticity (which can be handled by MDPs), partial observability (which can be handled by POMDPs), and acting with and around other agents (which can be handled with game theory). The problem is made even
harder by the fact that most robots work in continuous and high-dimensional state and
action spaces. They also operate in the real world, which refuses to run faster than real time and in which failures lead to real things being damaged, with no “undo” capability. Ideally, the robot would solve the entire problem in one go: observations in the form
of raw sensor feeds go in, and actions in the form of torques or currents to the motors
come out. In practice though, this is too daunting, and roboticists typically decouple different aspects of the problem and treat them independently.
* We typically separate perception (estimation) from action (motion generation). Percep-
tion in robotics involves computer vision to recognize the surroundings through cam-
eras, but also localization and mapping.
« Robotic perception concerns itself with estimating decision-relevant quantities from sensor data. To do so, we need an internal representation and a method for updating this internal representation over time.
+ Probabilistic filtering algorithms such as particle filters and Kalman filters are useful
for robot perception. These techniques maintain the belief state, a posterior distribution over state variables.
« For generating motion, we use configuration spaces, where a point specifies everything
we need to know to locate every body point on the robot. For instance, for a robot arm with two joints, a configuration consists of the two joint angles.
« We typically decouple the motion generation problem into motion planning, concerned
with producing a plan, and trajectory tracking control, concerned with producing a
policy for control inputs (actuator commands) that results in executing the plan.
Bibliographical and Historical Notes
« Motion planning can be solved via graph scarch using cell decomposition; using randomized motion planning algorithms, which sample milestones in the continuous configuration space; or using trajectory optimization, which can iteratively push a straight-line path out of collision by leveraging a signed distance field. + A path found by a search algorithm can be executed using the path as the reference
trajectory for a PID controller, which constantly corrects for errors between where the robot is and where it is supposed o be, or via computed torque control, which adds a
feedforward term that makes use of inverse dynamics to compute roughly what torque to send to make progress along the trajectory.
« Optimal control unites motion planning and trajectory tracking by computing an optimal trajectory directly over control inputs.
This
is especially easy when we have
quadratic costs and linear dynamics, resulting in a linear quadratic regulator (LQR). Popular methods make use of this by linearizing the dynamics and computing secondorder approximations of the cost (ILQR). « Planning under uncertainty unites perception and action by online replanning (such as model predictive control) and information gathering actions that aid perception.
+ Reinforcement learning is applied in robotics, with techniques striving to reduce the
required number of interactions with the real world. Such techniques tend to exploit models, be it estimating models and using them to plan, or training policies that are robust with respect to different possible model parameters.
« Interaction with humans
requires the ability to coordinate the robot’s actions with
theirs, which can be formulated as a game. We usually decompose the solution into
prediction, in which we use the person’s ongoing actions to estimate what they will do in the future, and action, in which we use the predictions to compute the optimal motion for the robot. + Helping humans also requires the ability to learn or infer what they want. Robots can
approach this by learning the desired cost function they should optimize from human
input, such as demonstrations, corrections, or instruction in natural language. Alterna-
tively, robots can imitate human behavior, and use reinforcement learning to help tackle
the challenge of generalization to new states. Bibliographical and
Historical Notes
The word robot was popularized by Czech playwright Karel Capek in his 1920 play R.U.R. (Rossum’s Universal Robots).
The robots, which were grown chemically rather than con-
structed mechanically, end up resenting their masters and decide to take over. It appears that
it was Capek’s brother, Josef, who first combined the Czech words “robota” (obligatory work)
and “robotnik™ (serf) to yield “robot™ in his 1917 short story Opilec (Glanc, 1978). The term
robotics was invented for a science fiction story (Asimov, 1950). The idea of an autonomous machine predates the word “robot” by thousands of years. In 7th century BCE Greek mythology, a robot named Talos was built by Hephaistos, the Greek god of metallurgy, to protect the island of Crete.
The legend is that the sorceress Medea
defeated Talos by promising him immortality but then draining his life fluid. Thus, this is the
975
976
Chapter 26 Robotics first example of a robot making a mistake in the process of changing its objective function. In 322 BCE, Aristotle anticipated technological unemployment, speculating “If every tool, when ordered, or even of its own accord, could do the work that befits it.. . then there would be no need either of apprentices for the master workers or of slaves for the lords.” In the 3rd century BCE an actual humanoid robot called the Servant of Philon could pour wine or water into a cup; a series of valves cut off the flow at the right time. Wonderful automata were built in the 18th century—Jacques Vaucanson’s mechanical duck from 1738
being one early example—but the complex behaviors they exhibited were entirely fixed in advance. Possibly the earliest example of a programmable robot-like device was the Jacquard loom (1805), described on page 15. Grey Walter’s “turtle,” built in 1948, could be considered the first autonomous mobile
robot, although its control system was not programmable. The “Hopkins Beast” built in 1960 at Johns Hopkins University, was much more sophisticated; it had sonar and photocell sensors, pattern-recognition hardware, and could recognize the cover plate of a standard AC power outlet. It was capable of searching for outlets, plugging itself in, and then recharging its batteries! Still, the Beast had a limited repertoire of skills.
The first general-purpose mobile robot was “Shakey,” developed at what was then the
Stanford Research Institute (now SRI) in the late 1960s (Fikes and Nilsson, 1971; Nilsson,
1984). Shakey was the first robot to integrate perception, planning, and execution, and much
subsequent research in Al was influenced by this remarkable achievement. Shakey appears
on the cover of this book with project leader Charlie Rosen (1917-2002). Other influential projects include the Stanford Cart and the CMU Rover (Moravec, 1983). Cox and Wilfong (1990) describe classic work on autonomous vehicles. The first commercial robot was an arm called UNIMATE,
for universal automation, de-
veloped by Joseph Engelberger and George Devol in their compnay, Unimation. In 1961, the first UNIMATE robot was sold to General Motors for use in manufacturing TV picture tubes. 1961 was also the year when Devol obtained the first U.S. patent on a robot.
In 1973, Toyota and Nissan started using an updated version of UNIMATE for auto body
spot welding. This initiated a major revolution in automobile manufacturing that took place mostly in Japan and the U.S., and that is still ongoing. Unimation followed up in 1978 with the development of the Puma robot (Programmable Universal Machine for Assembly), which
was the de facto standard for robotic manipulation for the two decades that followed. About
500,000 robots are sold each year, with half of those going to the automotive industry. In manipulation,
the first major effort at creating a hand-eye machine was Heinrich
Emst’s MH-1, described in his MIT Ph.D. thesis (Ernst, 1961).
The Machine Intelligence
project at Edinburgh also demonstrated an impressive early system for vision-based assembly called FREDDY (Michie, 1972).
Research on mobile robotics has been stimulated by several important competitions.
AAAT’s annual mobile robot competition began in 1992.
The first competition winner was
CARMEL (Congdon et al., 1992). Progress has been steady and impressive: in recent com-
petitions robots entered the conference complex, found their way to the registration desk, registered for the conference, and even gave a short talk. The RoboCup
competition,
launched in 1995 by Kitano and colleagues (1997), aims
to “develop a team of fully autonomous humanoid robots that can win against the human world champion team in soccer” by 2050.
Some competitions use wheeled robots, some
Bibliographical and Historical Notes
977
humanoid robots, and some software simulations. Stone (2016) describes recent innovations
in RoboCup.
The DARPA Grand Challenge, organized by DARPA in 2004 and 2005, required autonomous vehicles to travel more than 200 kilometers through the desert in less than ten hours (Buehler ez al., 2006). In the original event in 2004, no robot traveled more than eight miles, leading many to believe the prize would never be claimed. In 2005, Stanford’s robot
Stanley won the competition in just under seven hours (Thrun, 2006).
DARPA then orga-
nized the Urban Challenge, a competition in which robots had to navigate 60 miles in an
urban environment with other traffic.
Carnegie Mellon University’s robot BOSS took first
place and claimed the $2 million prize (Urmson and Whittaker, 2008). Early pioneers in the development of robotic cars included Dickmanns and Zapp (1987) and Pomerleau (1993).
The field of robotic mapping has evolved from two distinct origins. The first thread began with work by Smith and Cheeseman (1986), who applied Kalman filters to the simultaneous localization and mapping (SLAM) problem. This algorithm was first implemented by Moutarlier and Chatila (1989) and later extended by Leonard and Durrant-Whyte (1992); see Dissanayake et al. (2001) for an overview of early Kalman filter variations. The second thread
began with the development of the occupancy grid representation for probabilistic mapping,
which specifies the probability that each (x,y) location is occupied by an obstacle (Moravec
Occupancy grid
and Elfes, 1985).
Kuipers and Levitt (1988) were among the first to propose topological rather than metric mapping, motivated by models of human spatial cognition. A seminal paper by Lu and Milios (1997) recognized the sparseness of the simultaneous localization and mapping problem, which gave rise to the development of nonlinear optimization techniques by Konolige (2004)
and Montemerlo and Thrun (2004), as well as hierarchical methods by Bosse et al. (2004). Shatkay and Kaelbling (1997) and Thrun er al. (1998) introduced the EM algorithm into the
field of robotic mapping for data association. An overview of probabilistic mapping methods can be found in (Thrun et al., 2005). Early mobile robot localization techniques are surveyed by Borenstein ef al. (1996). Although Kalman
filtering was well known
as a localization method in control theory for
decades, the general probabilistic formulation of the localization problem did not appear in the Al literature until much later, through the work of Tom Dean and colleagues (Dean er al., 1990) and of Simmons and Koenig (1995). The latter work introduced the term Markov
localization. The first real-world application of this technique was by Burgard ef al. (1999), Markov localization through a series of robots that were deployed in museums.
Monte Carlo localization based
on particle filters was developed by Fox ef al. (1999) and is now widely used. The Rao-
Blackwellized particle filter combines particle filtering for robot localization with exact
filtering for map building (Murphy and Russell, 2001; Montemerlo et al., 2002).
Rao-Blackwellized particle filter
A great deal of early work on motion planning focused on geometric algorithms for de-
terministic and fully observable motion planning problems. The PSPACE-hardness of robot
motion planning was shown in a seminal paper by Reif (1979). The configuration space rep-
resentation is due to Lozano-Perez (1983). A series of papers by Schwartz and Sharir on what
they called piano movers problems (Schwartz et al., 1987) was highly influential.
Recursive cell decomposition for configuration space planning was originated in the work
of Brooks and Lozano-Perez (1985) and improved significantly by Zhu and Latombe (1991). The earliest skeletonization algorithms were based on Voronoi diagrams (Rowat, 1979) and
Piano movers
978
Chapter 26 Robotics
Visibility graph
visibility graphs (Wesley and Lozano-Perez, 1979). Guibas ez al. (1992) developed efficient
techniques for calculating Voronoi diagrams incrementally, and Choset (1996) generalized Voronoi diagrams to broader motion planning problems. John Canny (1988) established the first singly exponential algorithm for motion planning.
The seminal text by Latombe (1991) covers a variety of approaches to motion planning, as
do the texts by Choset et al. (2005) and LaValle (2006). Kavraki et al. (1996) developed the theory of probabilistic roadmaps. Kuffner and LaValle (2000) developed rapidly exploring random trees (RRTS). Involving optimization in geometric motion planning began with elastic bands (Quinlan
and Khatib, 1993), which refine paths when the configuration-space obstacles change. Ratliff et al. (2009) formulated the idea as the solution to an optimal control problem,
allowing
the initial trajectory to start in collision, and deforming it by mapping workspace obstacle gradients via the Jacobian into the configuration space. Schulman et al. (2013) proposed a practical second-order alternative. The control of robots as dynamical systems—whether for manipulation or navigation—
has generated a vast literature. While this chapter explained the basics of trajectory tracking
control and optimal control, it left out entire subfields, including adaptive control, robust
control, and Lyapunov analysis. Rather than assuming everything about the system is known
a priori, adaptive control aims to adapt the dynamics parameters and/or the control law online.
Robust control, on the other hand, aims to design controllers that perform well in spite of
uncertainty and external disturbances. Lyapunov analysis was originally developed in the 1890s for the stability analysis of
general nonlinear systems, but it was not until the early 1930s that control theorists realized its true potential.
With the development of optimization methods, Lyapunov
analysis was
extended to control barrier functions, which lend themselves nicely to modern optimization tools. These methods are widely used in modern robotics for real-time controller design and
safety analysis. Crucial works in robotic control include a trilogy on impedance control by Hogan (1985) and a general study of robot dynamics by Featherstone (1987). Dean and Wellman (1991)
were among the first to try to tie together control theory and Al planning systems. Three clas-
Haptic feedback
sic textbooks on the mathematics of robot manipulation are due to Paul (1981), Craig (1989), and Yoshikawa (1990). Control for manipulation is covered by Murray (2017). The area of grasping is also important in robotics—the problem of determining a stable grasp is quite difficult (Mason and Salisbury, 1985). Competent grasping requires touch sensing, or haptic feedback, to determine contact forces and detect slip (Fearing and Hollerbach, 1985). Understanding how to grasp the the wide variety of objects in the world is a daunting
task.
(Bousmalis et al., 2017) describe a system that combines real-world experimentation
with simulations guided by sim-to-real transfer to produce robust grasping.
Potential-field control, which attempts to solve the motion planning and control problems
Vector field histogram
simultaneously, was developed for robotics by Khatib (1986). In mobile robotics, this idea was viewed as a practical solution to the collision avoidance problem, and was later extended
into an algorithm called vector field histograms by Borenstein (1991). ILQR is currently widely used at the intersection of motion planning and control and is
due to Li and Todorov (2004). It is a variant of the much older differential dynamic program-
ming technique (Jacobson and Mayne, 1970).
Bibliographical and Historical Notes Fine-motion planning with limited sensing was investigated by Lozano-Perez et al. (1984)
and Canny and Reif (1987). Landmark-based navigation (Lazanas and Latombe, 1992) uses many of the same ideas in the mobile robot arena. Navigation functions, the robotics version of a control policy for deterministic MDPs, were introduced by Koditschek (1987). Key work
applying POMDP methods (Section 17.4) to motion planning under uncertainty in robotics is
due to Pineau et al. (2003) and Roy et al. (2005). Reinforcement learning in robotics took off with the seminal work by Bagnell and Schneider (2001) and Ng ef al. (2003), who developed the paradigm in the context of autonomous helicopter control.
Kober er al. (2013) offers an overview of how reinforcement
learning changes when applied to the robotics problem. Many of the techniques implemented on physical systems build approximate dynamics models, dating back to locally weighted linear models due to Atkeson ef al. (1997). But policy gradients played their role as well,
enabling (simplified) humanoid robots to walk (Tedrake er al., 2004), or a robot arm to hit a baseball (Peters and Schaal, 2008).
Levine ez al. (2016) demonstrated the first deep reinforcement learning application on a
real robot. At the same time, model-free RL in simulation was being extended to continuous domains (Schulman et al., 2015a; Heess et al., 2016; Lillicrap et al., 2015). Other work
scaled up physical data collection massively to showcase the learning of grasps and dynamics
models (Pinto and Gupta, 2016; Agrawal ez al., 2017; Levine et al., 2018). Transfer from simulation to reality or sim-to-real (Sadeghi and Levine, 2016; Andrychowicz et al., 2018a), metalearning (Finn er al., 2017), and sample-efficient model-free reinforcement learning
(Andrychowicz ef al., 2018b) are active areas of research.
Early methods for predicting human actions made use of filtering approaches (Madha-
van and Schlenoff, 2003), but seminal work by Ziebart ef al. (2009) proposed prediction by modeling people as approximately rational agents. Sadigh ef al. (2016) captured how these predictions should actually depend on what the robot decides to do, building toward a gametheoretic setting. For collaborative settings, Sisbot et al. (2007) pioneered the idea of account-
ing for what people want in the robot’s cost function. Nikolaidis and Shah (2013) decomposed
collaboration into learning how the human will act, but also learning how the human wants
the robot to act, both achievable from demonstrations. For learning from demonstration see
Argall et al. (2009). Akgun et al. (2012) and Sefidgar et al. (2017) studied teaching by end users rather than by experts. Tellex et al. (2011) showed how robots can infer what people want from natural language
instructions. Finally, not only do robots need to infer what people want and plan on doing, but
people too need to make the same inferences about robots. Dragan et al. (2013) incorporated amodel of the human’s inferences into robot motion planning.
The field of human-robot interaction is much broader than what we covered in this
chapter, which focused primarily on the planning and learning aspects. Thomaz ef al. (2016) provides a survey of interaction more broadly from a computational perspective. Ross et al. (2011) describe the DAGGER system. The topic of software architectures for robots engenders much religious debate. The
good old-fashioned Al candidate—the three-layer architecture—dates back to the design of Shakey and is reviewed by Gat (1998). The subsumption architecture is due to Brooks (1986),
although similar ideas were developed independently by Braitenberg, whose book, Vehicles (1984), describes a
series of simple robots based on the behavioral approach.
979
980
Chapter 26 Robotics The success of Brooks’s
six-legged walking robot was followed by many other projects.
Connell, in his Ph.D. thesis (1989), developed an entirely reactive mobile robot that was ca-
pable of retrieving objects. Extensions of the paradigm to multirobot systems can be found in work by Parker (1996) and Mataric (1997). GRL (Horswill, 2000) and COLBERT (Kono-
lige, 1997) abstract the ideas of concurrent behavior-based robotics into general robot control
languages. Arkin (1998) surveys some of the most popular approaches in this field. Two early textbooks, by Dudek and Jenkin (2000) and by Murphy (2000), cover robotics
generally. More recent overviews are due to Bekey (2008) and Lynch and Park (2017). Anex-
cellent book on robot manipulation addresses advanced topics such as compliant motion (Ma-
son, 2001). Robot motion planning is covered in Choset ef al. (2005) and LaValle (2006).
Thrun ez al. (2005) introduces probabilistic robotics. The Handbook of Robotics (Siciliano and Khatib, 2016) is a massive, comprehensive overview of all of robotics.
The premiere conference for robotics is Robotics: Science and Systems Conference, fol-
lowed by the IEEE International Conference on Robotics and Automation.
Human-Robot
Interaction is the premiere venue for interaction. Leading robotics journals include IEEE
Robotics and Automation, the International Journal of Robotics Research, and Robotics and
Autonomous Systems.
TR 97
PHILOSOPHY, ETHICS, AND SAFETY OF Al In which we consider the big questions around the meaning of Al how we can ethically develop and apply it, and how we can keep it safe.
Philosophers have been asking big questions for a long time: How do minds work? s it possible for machines to act intelligently in the way that people do? Would such machines have real, conscious minds? To these, we add new ones: What are the ethical implications of intelligent machines in day-to-day use? Should machines be allowed to decide to kill humans? Can algorithms be fair and unbiased? What will humans do if machines can do all kinds of work? And how do we control machines that may become more intelligent than us? 27.1
The Limits of Al
In 1980, philosopher John Searle introduced a distinction between weak Al—the idea that
machines could act as if they were intelligent—and strong AI—the assertion that machines
that do so are actually consciously thinking (not just simulating thinking). Over time the definition of strong Al shifted to refer to what is
also called “human-level AI” or “general
AI’—programs that can solve an arbitrarily wide variety of tasks, including novel ones, and do so as well as a human.
Critics of weak AT who objected to the very possibility of intelligent behavior in machines now appear as shortsighted as Simon Newcomb, who in October 1903 wrote “aerial flight is one of the great class of problems with which man can never cope”—just two months before the Wright brothers’ flight at Kitty Hawk.
The rapid progress of recent years does
not, however, prove that there can be no limits to what Al can achieve. Alan Turing (1950),
the first person to define Al was also the first to raise possible objections to Al foreseeing almost all the ones subsequently raised by others. 27.1.1
The argument from informality
Turing’s “argument from informality of behavior” says that human behavior is far too com-
plex to be captured by any formal set of rules—humans must be using some informal guide-
lines that (the argument claims) could never be captured in a formal set of rules and thus
could never be codified in a computer program. A key proponent of this view was Hubert Dreyfus, who produced a series of influential critiques of artificial intelligence:
What Computers
Can’t Do (1972), the sequel What
weak Al
Strong Al
982
Chapter 27 Philosophy, Ethics, and Safety of AT Computers Still Can’t Do (1992), and, with his brother Stuart, Mind Over Machine (1986).
Good Old-Fashioned AT'(GOFAI)
Similarly, philosopher Kenneth Sayre (1993) said “Artificial intelligence pursued within the cult of computationalism stands not even a ghost of a chance of producing durable results.” The technology they criticize came to be called Good Old-Fashioned AI (GOFAI).
GOFAI corresponds to the simplest logical agent design described in Chapter 7, and we saw there that it is indeed difficult to capture every contingency of appropriate behavior in a set of necessary and sufficient logical rules; we called that the qualification problem.
But
as we saw in Chapter 12, probabilistic reasoning systems are more appropriate for openended domains, and as we saw in Chapter 21, deep learning systems do well on a variety of “informal” tasks. Thus, the critique is not addressed against computers per se, but rather against one particular style of programming them with logical rules—a style that was popular
in the 1980s but has been eclipsed by new approaches.
One of Dreyfus’s strongest arguments is for situated agents rather than disembodied log-
ical inference engines. An agent whose understanding of “dog” comes only from a limited set of logical sentences such as “Dog(x) = Mammal(x)” is at a disadvantage compared to an agent that has watched dogs run, has played fetch with them, and has been licked by one. As philosopher Andy Clark (1998) says, “Biological brains are first and foremost the control systems for biological bodies. Biological bodies move and act in rich real-world surroundings.”
Embodied cognition
According to Clark, we are “good at frisbee, bad at logic.”
The embodied cognition approach claims that it makes no sense to consider the brain
separately: cognition takes place within a body, which is embedded in an environment. We need to study the system as a whole; the brain’s functioning exploits regularities in its envi-
ronment, including the rest of its body.
Under the embodied cognition approach, robotics,
vision, and other sensors become central, not peripheral. Overall, Dreyfus
saw areas where Al did not have complete answers and said that Al is
therefore impossible; we now see many of these same areas undergoing continued research and development leading to increased capability, not impossibility. 27.1.2
The argument from disability
The “argument from disability” makes the claim that “a machine can never do X.” As examples of X, Turing lists the following: Be kind, resourceful, beautiful, friendly, have initiative, have a sense of humor, tell right from wrong, make mistakes, fall in love, enjoy strawberries and cream, make someone fall in love with it, learn from experience, use words properly, be the subject of its own thought, have as much diversity of behavior as man, do something really new. In retrospect, some of these are rather easy—we’re all familiar with computers that “make
mistakes.” Computers with metareasoning capabilities (Chapter 5) can examine heir own computations, thus being the subject of their own reasoning. A century-old technology has the proven ability to “make someone fall in love with it"—the teddy bear. Computer chess
expert David Levy predicts that by 2050 people will routinely fall in love with humanoid robots.
As for a robot falling in love, that is a common theme in fiction,' but there has
been only limited academic speculation on the subject (Kim ef al., 2007). Computers have
! For example, the opera Coppélia (1870), the novel Do Androids Dream of Electric Sheep? (1968), the movies AIQ001), Wall-E (2008), and Her (2013).
Section 27.1
The Limits of AT
done things that are “really new;” making significant discoveries in astronomy, mathematics, chemistry, mineralogy, biology, computer science, and other fields, and creating new forms of art through style transfer (Gatys e al., 2016). Overall, programs exceed human performance in some tasks and lag behind on others. The one thing that it is clear they can’t do is be exactly human. 27.1.3
The
mathematical
objection
Turing (1936) and Godel (1931) proved that certain mathematical questions are in princi-
ple unanswerable by particular formal systems. Godel’s incompleteness theorem (see Sec-
tion 9.5) is the most famous example of this. Briefly, for any formal axiomatic framework F
powerful enough to do arithmetic, it is possible to construct a so-called Godel sentence G(F)
with the following properties: « G(F) is a sentence of F, but cannot be proved within F. « If F is consistent, then G(F) is true.
Philosophers such as J. R. Lucas (1961) have claimed that this theorem shows that machines
are mentally inferior to humans, because machines are formal systems that are limited by the incompleteness theorem—they cannot establish the truth of their own Godel sentence—
while humans have no such limitation. This has caused a lot of controversy, spawning a vast literature, including two books by the mathematician/physicist Sir Roger Penrose (1989, 1994).
Penrose repeats Lucas’s claim with some fresh twists, such as the hypothesis that
humans are different because their brains operate by quantum gravity—a theory that makes
multiple false predictions about brain physiology.
‘We will examine three of the problems with Lucas’s claim. First, an agent should not be
ashamed that it cannot establish the truth of some sentence while other agents can. Consider the following sentence:
Lucas cannot consistently assert that this sentence is true. If Lucas asserted this sentence, then he would be contradicting himself, so therefore Lucas
cannot consistently assert it, and hence it is true. We have thus demonstrated that there is a
true sentence that Lucas cannot consistently assert while other people (and machines) can. But that does not make us think any less of Lucas.
Second, Gadel’s incompleteness theorem and related results apply to mathematics, not
to computers. No entity—human or machine—can prove things that are impossible to prove. Lucas and Penrose falsely assume that humans can somehow get around these limits, as when
Lucas (1976) says “we must assume our own consistency, if thought is to be possible at all.” But this is an unwarranted assumption: humans are notoriously inconsistent. This is certainly
true for everyday reasoning, but it is also true for careful mathematical thought. A famous example is the four-color map problem. Alfred Kempe (1879) published a proof that was widely accepted for 11 years until Percy Heawood (1890) pointed out a flaw. Third, Gédel’s incompleteness theorem technically applies only to formal systems that are powerful enough to do arithmetic. This
includes Turing machines, and Lucas’s claim is
in part based on the assertion that computers are equivalent to Turing machines. This is not
quite true. Turing machines are infinite, whereas computers (and brains) are finite, and any computer can therefore be described as a (very large) system in propositional logic, which is not subject to Gédel’s incompleteness theorem. Lucas assumes that humans can “change their
983
984
Chapter 27 Philosophy, Ethics, and Safety of AT minds” while computers cannot, but that is also false—a computer can retract a conclusion
after new evidence or further deliberation; it can upgrade its hardware; and it can change its
decision-making processes with machine learning or software rewriting. 27.1.4
Measuring Al
Alan Turing, in his famous paper “Computing Machinery and Intelligence” (1950), suggested that instead of
asking whether machines can think, we should ask whether machines can pass
a behavioral test, which has come to be called the Turing test. The test requires a program
to have a conversation (via typed messages) with an interrogator for five minutes.
The in-
terrogator then has to guess if the conversation is with a program or a person; the program
passes the test if it fools the interrogator 30% of the time. To Turing, the key point was not
the exact details of the test, but instead the idea of measuring intelligence by performance on some kind of open-ended behavioral task, rather than by philosophical speculation.
Nevertheless, Turing conjectured that by the year 2000 a computer with a storage of a
billion units could pass the test, but here we are on the other side of 2000, and we still can’t
agree whether any program has passed. Many people have been fooled when they didn’t know they might be chatting with a computer.
The ELIZA program and Internet chatbots such as
MGONZ (Humphrys, 2008) and NATACHATA (Jonathan et al., 2009) fool their correspondents
repeatedly, and the chatbot CYBERLOVER
has attracted the attention of law enforcement be-
cause of its penchant for tricking fellow chatters into divulging enough personal information that their identity can be stolen. In 2014, a chatbot called Eugene Goostman fooled 33% of the untrained amateur judges
in a Turing test. The program claimed to be a boy from Ukraine with limited command of
English; this helped explain its grammatical errors. Perhaps the Turing test is really a test of
human gullibility. So far no well-trained judge has been fooled (Aaronson, 2014).
Turing test competitions have led to better chatbots, but have not been a focus of research
within the AT community. Instead, Al researchers who crave competition are more likely
to concentrate on playing chess or Go or StarCraft II, or taking an 8th grade science exam,
or identifying objects in images.
In many of these competitions, programs have reached
or surpassed human-level performance, but that doesn’t mean the programs are human-like
outside the specific task. The point is to improve basic science and technology and to provide useful tools, not to fool judges.
27.2
Can Machines Really Think?
Some philosophers claim that a machine that acts intelligently would not be actually thinking,
but would be only a simulation of thinking. But most Al researchers are not concerned with
the distinction, and the computer scientist Edsger Dijkstra (1984) said that “The question of whether Machines Can Think ... is about as relevant as the question of whether Submarines Can Swim.” The American Heritage Dictionary’s first definition of swim is “To move through
water by means of the limbs, fins, or tail,” and most people agree that submarines, being limbless, cannot swim. The dictionary also defines fly as “To move through the air by means of wings or winglike parts,” and most people agree that airplanes, having winglike parts,
can fly. However, neither the questions nor the answers have any relevance to the design or
capabilities of airplanes and submarines; rather they are about word usage in English. (The
Section 27.2
985
Can Machines Really Think?
fact that ships do swim (“priver”) in Russian amplifies this point.) English speakers have
not yet settled on a precise definition for the word “think”—does it require “a brain” or just “brain-like parts?”
Again, the issue was addressed by Turing. He notes that we never have any direct ev-
idence about the internal mental states of other humans—a kind of mental solipsism. Nev-
ertheless, Turing says, “Instead of arguing continually over this point, it is usual to have the polite convention that everyone thinks.” Turing argues that we would also extend the polite
convention to machines, if only we had experience with ones that act intelligently. How-
Polite convention
ever, now that we do have some experience, it seems that our willingness to ascribe sentience
depends at least as much on humanoid appearance and voice as on pure intelligence. 27.2.1
The Chinese room
The philosopher John Searle rejects the polite convention. His famous Chinese room argu- Chinese room ment (Searle, 1990) goes as follows: Imagine a human, who understands only English, inside a room that contains a rule book, written in English, and various stacks of paper. Pieces of paper containing indecipherable symbols are slipped under the door to the room. The human follows the instructions in the rule book, finding symbols in the stacks, writing symbols on new pieces of paper, rearranging the stacks, and so on. Eventually, the instructions will cause
one or more symbols to be transcribed onto a piece of paper that is passed back to the outside world. From the outside, we see a system that is taking input in the form of Chinese sentences and generating fluent, intelligent Chinese responses. Searle then argues: it is given that the human does not understand Chinese. The rule book
and the stacks of paper, being just pieces of paper, do not understand Chinese. Therefore, there is no understanding of Chinese. And Searle says that the Chinese room is doing the same thing that a computer would do, so therefore computers generate no understanding. Searle (1980) is a proponent of biological naturalism, according to which mental states Biological naturalism are high-level emergent features that are caused by low-level physical processes in the neurons, and it is the (unspecified) properties of the neurons that matter: according to Searle’s biases, neurons have “it” and transistors do not. There have been many refutations of Searle’s
argument, but no consensus. His argument could equally well be used (perhaps by robots) to argue that a human cannot have true understanding; after all, a human is made out of cells,
the cells do not understand, therefore there is no understanding. In fact, that is the plot of Terry Bisson’s (1990) science fiction story They're Made Out of Meat, in which alien robots
explore Earth and can’t believe that hunks of meat could possibly be sentient. How they can be remains a mystery.
27.2.2
Consciousness and qualia
Running through all the debates about strong Al is the issue of consciousness:
awareness
of the outside world, and of the self, and the subjective experience of living. The technical
term for the intrinsic nature of experiences is qualia (from the Latin word meaning, roughly,
“of what kind”). The big question is whether machines can have qualia. In the movie 2001, when astronaut David Bowman is disconnecting the “cognitive circuits™ of the HAL 9000
computer, it says “I'm afraid, Dave. Dave, my mind is going. I can feel it.” Does HAL actually have feelings (and deserve sympathy)? Or is the reply just an algorithmic response, no different from “Error 404: not found”?
Consciousness Qualia
986
Chapter 27 Philosophy, Ethics, and Safety of AT There is a similar question for animals: pet owners are certain that their dog or cat has
consciousness, but not all scientists agree. Crickets change their behavior based on tempera-
ture, but few people would say that crickets experience the feeling of being warm or cold.
One reason that the problem of consciousness is hard is that it remains ill-defined, even
after centuries of debate.
But help may be on the way. Recently philosophers have teamed
with neuroscientists under the auspices of the Templeton Foundation to start a series of ex-
periments that could resolve some of the issues. Advocates of two leading theories of con-
sciousness (global workspace theory and integrated information theory) have agreed that the
experiments could confirm one theory over the other—a rarity in philosophy.
Alan Turing (1950) concedes that the question of consciousness is a difficult one, but
denies that it has much relevance to the practice of Al: “I do not wish to give the impression
that I think there is no mystery about consciousness ... But I do not think these mysteries
necessarily need to be solved before we can answer the question with which we are concerned in this paper” We agree with Turing—we are interested in creating programs that behave intelligently.
Individual aspects of consciousness—awareness,
self-awareness, attention—
can be programmed and can be part of an intelligent machine. The additional project of making a machine conscious in exactly the way humans are is not one that we are equipped
to take on. We do agree that behaving intelligently will require some degree of awareness,
which will differ from task to task, and that tasks involving interaction with humans will
require a model of human subjective experience. In the matter of modeling experience, humans have a clear advantage over machines, because they can use their own subjective apparatus to appreciate the subjective experience of others.
For example, if you want to know what it’s like when someone hits their thumb
with a hammer, you can hit your thumb with a hammer. Machines have no such capability— although unlike humans, they can run each other’s code. 27.3
The Ethics of Al
Given that Al is a powerful technology, we have a moral obligation to use it well, to promote the positive aspects and avoid or mitigate the negative ones.
The positive aspects are many. For example, Al can save lives through improved med-
ical diagnosis, new medical discoveries, better prediction of extreme weather events, and
safer driving with driver assistance and (eventually) self-driving technologies. There are also many opportunities to improve lives. Microsoft’s Al for Humanitarian Action program ap-
plies Al to recovering from natural disasters, addressing the needs of children, protecting refugees, and promoting human rights. Google’s Al for Social Good program supports work on rainforest protection, human rights jurisprudence, pollution monitoring, measurement of fossil fuel emissions, cris counseling, news fact checking, suicide prevention, recycling, and other issues. The University of Chicago’s Center for Data Science for Social Good applies machine learning to problems in criminal justice, economic development, education, public health, energy, and environment.
Al applications in crop management and food production help feed the world. Optimization of business processes using machine learning will make businesses more productive,
increasing wealth and providing more employment. Automation can replace the tedious and
dangerous tasks that many workers face, and free them to concentrate on more interesting
Section 27.3
The Ethics of Al
987
aspects. People with disabilities will benefit from Al-based assistance in seeing, hearing, and
mobility. Machine translation already allows people from different cultures to communicate.
Software-based Al solutions have near zero marginal cost of production, and so have the potential to democratize access to advanced technology (even as other aspects of software have the potential to centralize power).
Despite these many positive aspects, we shouldn’t ignore the negatives. Many new tech-
nologies have had unintended negative side effects:
nuclear fission brought Chernobyl and
the threat of global destruction; the internal combustion engine brought air pollution, global warming, and the paving of paradise. Other technologies can have negative effects even when
used as intended, such as sarin gas, AR-15 rifles, and telephone solicitation. Automation will
create wealth, but under current economic conditions much of that wealth will flow to the
owners of the automated systems, leading to increased income inequality. This can be disruptive to a well-functioning society. In developing countries, the traditional path to growth
through low-cost manufacturing for export may be cut off, as wealthy countries adopt fully
automated manufacturing facilities on-shore. Our ethical and governance decisions will dic-
tate the level of inequality that AT will engender.
All scientists and engineers face ethical considerations of what projects they should or
should not take on, and how they can make sure the execution of the project is safe and beneficial. In 2010, the UK’s Engineering and Physical Sciences Research Council held a meeting to develop a set of Principles of Robotics. In subsequent years other government agencies, nonprofit organizations, and companies created similar sets of principles. The gist is that every organization that creates Al technology, and everyone in the organization, has a responsibility to make sure the technology contributes to good, not harm. The most commonly-cited
principles are:
Ensure safety
Establish accountability
Promote collaboration
Avoid concentration of power
Ensure fairness Respect privacy
Provide transparency
Limit harmful uses of Al
Uphold human rights and values Reflect diversity/inclusion Acknowledge legal/policy implications
Contemplate implications for employment
Note that many of the principles, such as “ensure safety,” have applicability to all software or
hardware systems, not just Al systems. Several principles are worded in a vague way, making them difficult to measure or enforce.
That is in part because Al is a big field with many
subfields, each of which has a different set of historical norms and different relationships between the Al developers and the stakeholders. Mittelstadt (2019) suggests that the subfields
should each develop more specific actionable guidelines and case precedents. 27.3.1
Lethal autonomous weapons
The UN defines a lethal autonomous weapon as one that locates, selects, and engages (i.c., Kills) human targets without human supervision. Various weapons fulfill some of these criteria. For example, land mines have been used since the 17th century: they can select and engage targets in a limited sense according to the degree of pressure exerted or the quan-
tity of metal present, but they cannot go out and locate targets by themselves. (Land mines are banned under the Ottawa Treaty.) Guided missiles, in use since the 1940s, can chase
targets, but they have to be pointed in the right general direction by a human. Auto-firing
Negative side effects
988
Chapter 27 Philosophy, Ethics, and Safety of AT radar-controlled guns have been used to defend naval ships since the 1970s; they are mainly intended to destroy incoming missiles, but they could also attack manned aircraft. Although
the word “autonomous” is often used to describe unmanned air vehicles or drones, most such
weapons are both remotely piloted and require human actuation of the lethal payload.
At the time of writing, several weapons systems seem to have crossed the line into full au-
tonomy. For example Israel’s Harop missile is a “loitering munition” with a ten-foot wingspan and a fifty-pound warhead. It searches for up to six hours in a given geographical region for any target that meets a given criterion and then destroys it. The criterion could be “emits a radar signal resembling antiaircraft radar” or “looks like a tank.”
The Turkish manufac-
turer STM advertises its Kargu quadcopter—which carries up to 1.5kg of explosives—as
capable of “Autonomous hit ... . targets selected on images . . . tracking moving targets . ... anti-
personnel ... face recognition.” Autonomous weapons have been called the “third revolution in warfare” after gunpowder and nuclear weapons. Their military potential is obvious. For example, few experts doubt
that autonomous fighter aircraft would defeat any human pilot. Autonomous aircraft, tanks,
and submarines can be cheaper, faster, more maneuverable, and have longer range than their manned counterparts.
Since 2014, the United Nations in Geneva has conducted regular discussions under the
auspices of the Convention on Certain Conventional Weapons (CCW) on the question of whether to ban lethal autonomous
weapons.
At the time of writing, 30 nations, ranging
in size from China to the Holy See, have declared their support for an international treaty, while other key countries—including Israel, Russia, South Korea, and the United States—are
opposed to a ban. The debate over autonomous weapons includes legal, ethical and practical aspects. The legal issues are governed primarily by the CCW, which requires the possibility of discrim-
inating between combatants and non-combatants, the judgment of military necessity for an attack,
and the assessment of proportionality
possibility of collateral damage.
The feas
between the military value of a target and the
ity of meeting these criteria
is an engineering
question—one whose answer will undoubtedly change over time. At present, discrimination seems feasible in some circumstances and will undoubtedly improve rapidly, but necessity and proportionality are not presently feasible: they require that machines make subjective
and situational judgments that are considerably more difficult than the relatively simple tasks
of searching for and engaging potential targets. For these reasons, it would be legal to use autonomous weapons only in circumstances where a human operator can reasonably predict that the execution of the mission will not result in civilians being targeted or the weapons
conducting unnecessary or disproportionate attacks. This means that, for the time being, only
very restricted missions could be undertaken by autonomous weapons.
On the ethical side, some find it simply morally unacceptable to delegate the decision to
kill humans to a machine.
For example, Germany’s ambassador in Geneva has stated that
it “will not accept that the decision over life and death is taken solely by an autonomous
system” while Japan “has no plan to develop robots with humans out of the loop, which may
be capable of committing murder.” Gen. Paul Selva, at the time the second-ranking military
officer in the United States, said in 2017, “I don’t think it’s reasonable for us to put robots in charge of whether or not we take a human life.” Finally, Anténio Guterres, the head of the United Nations, stated in 2019 that “machines with the power and discretion to take lives
Section 27.3
The Ethics of Al
without human involvement are politically unacceptable, morally repugnant and should be prohibited by international law.” More than 140 NGOs in over 60 countries are part of the Campaign to Stop Killer Robots, and an open letter organized in 2015 by the Future of Life Institute organized an open letter was signed by over 4,000 Al researchers? and 22,000 others. Against this, it can be argued that as technology improves it ought to be possible to develop weapons that are less likely than human soldiers or pilots to cause civilian casualties. (There is also the important benefit that autonomous weapons reduce the need for human sol-
diers and pilots to risk death.) Autonomous systems will not succumb to fatigue, frustration,
hysteria, fear, anger, or revenge, and need not “shoot first, ask questions later” (Arkin, 2015). Just as guided munitions have reduced collateral damage compared to unguided bombs, one may expect intelligent weapons to further improve the precision of attacks. (Against this, see Benjamin (2013) for an analysis of drone warfare casualties.) This, apparently, is the position of the United States in the latest round of negotiations in Geneva.
Perhaps counterintuitively, the United States is also one of the few nations whose own
policies currently preclude the use of autonomous weapons. The 2011 U.S. Department of Defense (DOD) roadmap says: “For the foreseeable future, decisions over the use of force
[by autonomous systems] and the choice of which individual targets to engage with lethal force will be retained under human control.” The primary reason for this policy is practical: autonomous systems are not reliable enough to be trusted with military decisions. The issue of reliability came to the fore on September 26,
1983, when Soviet missile
officer Stanislav Petrov’s computer display flashed an alert of an incoming missile attack. According to protocol, Petrov should have initiated a nuclear counterattack, but he suspected the alert was a bug and treated it as such. He was correct, and World War III was (narrowly)
averted. We don’t know what would have happened if there had been no human in the loop. Reliability is a very serious concern for military commanders, who know well the complexity of battlefield situations. Machine learning systems that operate flawlessly in training
may perform poorly when deployed. Cyberattacks against autonomous weapons could result
in friendly-fire casualties; disconnecting the weapon from all communication may prevent
that (assuming it has not already been compromised), but then the weapon cannot be recalled if it is malfunctioning.
The overriding practical issue with autonomous weapons is that they they are scalable
weapons of mass destruction, in the sense that the scale of an attack that can be launched is
proportional to the amount of hardware one can afford to deploy. A quadcopter two inches
in diameter can carry a lethal explosive charge, and one million can fit in a regular shipping
container. Precisely because they are autonomous, these weapons would not need one million
human supervisors to do their work.
As weapons of mass destruction, scalable autonomous weapons have advantages for the
attacker compared to nuclear weapons and carpet bombing: they leave property intact and can be applied selectively to eliminate only those who might threaten an occupying force. They
could certainly be used to wipe out an entire ethnic group or all the adherents ofa particular religion. In many situations, they would also be untraceable. These characteristics make them particularly attractive to non-state actors.
the two authors of this book.
989
990
Chapter 27 Philosophy, Ethics, and Safety of AT These considerations—particularly
those characteristics that advantage the attacker—
suggest that autonomous weapons will reduce global and national security for all parties.
The rational response for governments seems to be to engage in arms control discussions Dual use
rather than an arms race. The process of designing a treaty is not without its difficulties, however.
Al is a dual
use technology: Al technologies that have peaceful applications such as flight control, visual tracking, mapping, navigation, and multiagent planning, can easily be applied to military
purposes.
It is easy to turn an autonomous quadcopter into a weapon simply by attaching
an explosive and commanding it to seek out a target.
Dealing with this will require care-
ful implementation of compliance regimes with industry cooperation, as has already been
demonstrated with some success by the Chemical Weapons Convention.
27.3.2
Surveillance, security, and privacy
In 1976, Joseph Weizenbaum warned that automated speech recognition technology could
lead to widespread wiretapping, and hence to a loss of civil liberties. Today, that threat has
been realized, with most electronic communication going through central servers that can
be monitored, and cities packed with microphones and cameras that can identify and track
Surveillance camera
individuals based on their voice, face, and gait. Surveillance that used to require expensive and scarce human resources can now be done at a mass scale by machines. As of 2018, there were as many as 350 million surveillance cameras in China and 70 million in the United States. China and other countries have begun exporting surveillance technology to low-tech countries, some with reputations for mistreating their citizens and
disproportionately targeting marginalized communities. Al engineers should be clear on what
uses of surveillance are compatible with human rights, and decline to work on applications that are incompatible. As more of our institutions operate online, we become more vulnerable to cybercrime
Cybersecurity
(phishing, credit card fraud, botnets, ransomware) and cyberterrorism (including potentially deadly attacks such as shutting down hospitals and power plants or commandeering selfdriving cars). Machine learning can be a powerful tool for both sides in the cybersecurity battle. Attackers can use automation to probe for insecurities and they can apply reinforce-
ment learning for phishing attempts and automated blackmail. Defenders can use unsuper-
vised learning to detect anomalous incoming traffic patterns (Chandola ef al., 2009; Malhotra
et al., 2015) and various machine learning techniques to detect fraud (Fawcett and Provost,
1997; Bolton and Hand, 2002). As attacks get more sophisticated, there is a greater responsi-
bility for all engineers, not just the security experts, to design secure systems from the start.
One forecast (Kanal, 2017) puts the market for machine learning in cybersecurity at about
$100 billion by 2021.
As we interact with computers for increasing amounts of our daily lives, more data on us
is being collected by governments and corporations. Data collectors have a moral and legal
responsibility to be good stewards of the data they hold.
In the U.S., the Health Insurance
Portability and Accountability Act (HIPAA) and the Family Educational Rights and Privacy Act (FERPA) protect the privacy of medical and student records. The European Union’s General Data Protection Regulation (GDPR) mandates that companies design their systems
with protection of data in mind and requires that they obtain user consent for any collection
or processing of data.
Section 27.3
991
The Ethics of Al
Balanced against the individual’s right to privacy is the value that society gains from sharing data. We want to be able to stop terrorists without oppressing peaceful dissent, and we want to cure diseases without compromising any individual’s right to keep their health history private. One key practice is de-identification: eliminating personally identifying in- De-identification formation (such as name and social security number) so that medical researchers can use the data to advance the common good. The problem is that the shared de-identified data may
be subject to re-identification.
For example, if the data strips out the name, social security
number, and street address, but includes date of birth, gender, and zip code, then, as shown by
Latanya Sweeney (2000), 87% of the U.S. population can be uniquely re-identified. Sweeney emphasized this point by re-identifying the health record for the governor of her state when
he was admitted to the hospital. In the Netflix Prize competition, de-identified records of in-
dividual movie ratings were released, and competitors were asked to come up with a machine learning algorithm that could accurately predict which movies an individual would like. But researchers were able to re-identify individual users by matching the date of a rating in the
Netflix Prize
Netflix database with the date of a similar ranking in the Internet Movie Database (IMDB), where users sometimes use their actual names (Narayanan and Shmatikov, 2006).
This risk can be mitigated somewhat by generalizing fields: for example, replacing the
exact birth date with just the year of birth, or a broader range like “20-30 years old.” Deleting a field altogether can be seen as a form of generalizing to “any.” But generalization alone does not guarantee that records are safe from re-identification; it may be that there is only one person in zip code 94720 who is 90-100 years old. A useful property is k-anonymity: K-anonymity a database is k-anonymized if every record in the database is indistinguishable from at least
k— 1 other records. If there are records that are more unique than this, they would have to be
further generalized.
An alternative to sharing de-identified records is to keep all records private, but allow
aggregate querying. An API for queries against the database is provided, and valid queries Aggregate querying receive a response that summarizes the data with a count or average. But no response is given if it would violate certain guarantees of privacy. For example, we could allow an epidemiologist to ask, for each zip code, the percentage of people with cancer. For zip codes with at least n people a percentage would be given (with a small amount of random noise),
but no response would be given for zip codes with fewer than n people..
Care must be taken to protect against de-identification using multiple queries. For exam-
ple, if the query “average salary and number of employees of XYZ company age 30-40” gives the response [$81,234, 12] and the query “average salary and number of employees of XYZ company age 30-41 gives the response [$81,199, 131, and if we use LinkedIn to find the one 41-year-old at XYZ company, then we have successfully identified them, and can compute their exact salary, even though all the responses involved 12 or more people. The system must
be carefully designed to protect against this, with a combination of limits on the queries that
can be asked (perhaps only a predefined set of non-overlapping age ranges can be queried) and the precision of the results (perhaps both queries give the answer “about $81,000). A stronger guarantee is differential privacy, which assures that an attacker cannot use Differential privacy queries to re-identify any individual in the database, even if the attacker can make multiple
queries and has access to separate linking databases. The query response employs a random-
ized algorithm that adds a small amount of noise to the result. Given a database D, any record
in the database r, any query Q, and a possible response y to the query, we say that the database
992
Chapter 27 Philosophy, Ethics, and Safety of AT
D has edifferential privacy if the log probability of the response y varies by less than € when we add the record r: |log P(Q(D)=y) —logP(Q(D+r)=y)| < €. In other words, whether any one person decides to participate in the data base or not makes no appreciable difference to the answers anyone can get, and therefore there is no privacy
disincentive to participate. Many databases are designed to guarantee differential privacy.
Federated learning
So far we have considered the issue of sharing de-identified data from a central database.
An approach called federated learning (Konecny et al., 2016) has no central database; in-
stead, users maintain their own local databases that keep their data private. However, they
can share parameters of a machine learning model that is enhanced with their data, without
the risk of revealing any of the private data. Imagine a speech understanding application that users can run locally on their phone. The application contains a baseline neural network, which is then improved by local training on the words that are heard on the user’s phone. Periodically, the owners of the application poll a subset of the users and ask them for the parameter values of their improved local network, but not for any of their raw data. The
parameter values are combined together to form a new improved model which is then made
available to all users, so that they all get the benefit of the training that is done by other users.
For this scheme to preserve privacy, we have to be able to guarantee that the model
parameters shared by each user cannot be reverse-engineered. If we sent the raw parameters,
Secure aggregation
there is a chance that an adversary inspecting them could deduce whether, say, a certain word had been heard by the user’s phone. One way to eliminate this risk is with secure aggregation (Bonawitz et al., 2017).
The idea is that the central server doesn’t need to know the exact
parameter value from each distributed user; it only needs to know the average value for each
parameter, over all polled users. So each user can disguise their parameter values by adding
a unique mask to each value; as long as the sum of the masks is zero, the central server will be able to compute the correct average. Details of the protocol make sure that it is efficient in terms of communication (less than half the bits transmitted correspond to masking), is robust to individual users failing to respond, and is secure in the face of adversarial users,
cavesdroppers, or even an adversarial central server. 27.3.3
Fairness and bias
Machine learning is augmenting and sometimes replacing human decision-making in im-
portant situations: whose loan gets approved, to what neighborhoods police officers are de-
Societal bias
ployed, who gets pretrial release or parole. But machine learning models can perpetuate
societal bias. Consider the example of an algorithm to predict whether criminal defendants
are likely to re-offend, and thus whether they should be released before trial. It could well be
that such a system picks up the racial or gender prejudices of human judges from the examples in the training set. Designers of machine learning systems have a moral responsibility to
ensure that their systems are in fact fair. In regulated domains such as credit, education, em-
ployment, and housing, they have a legal responsibility as well. But what is faimess? There are multiple criteria; here are six of the most commonly-used concepts:
o Individual fairness: A requirement that individuals are treated similarly to other similar individuals, regardless of what class they are in.
Section 27.3
993
The Ethics of Al
* Group fairness: A requirement that two classes are treated similarly, as measured by
some summary statistic. o Fairness through unawareness:
If we delete the race and gender attributes from the
data set, then it might seem that the system cannot discriminate on those attributes.
Unfortunately, we know that machine learning models can predict latent variables (such as race and gender) given other correlated variables (such as zip code and occupation). Furthermore, deleting those attributes makes it impossible to verify equal opportunity
or equal outcomes. Still, some countries (e.g., Germany) have chosen this approach for
their demographic statistics (whether or not machine learning models are involved). o Equal outcome: The idea that each demographic class gets the same results; they have demographic parity. For example, suppose we have to decide whether we should Demographic parity approve loan applications; the goal is to approve those applicants who will pay back the loan and not those who will default on the loan. Demographic parity says that
both males and females should have the same percentage of loans approved. Note that this is a group faimess criterion that does nothing to ensure individual fairness; a wellqualified applicant might be denied and a poorly-qualified applicant might be approved, as long as the overall percentages are equal. Also, this approach favors redress of past biases over accuracy of prediction. Ifa man and a woman are equal in every way, except the woman receives a lower salary for the same job, should she be approved because she would be equal if not for historical biases, or should she be denied because the lower salary does in fact make her more likely to default?
Equal opportunity: The idea that the people who truly have the ability to pay back the loan should have an equal chance of being correctly classified as such, regardless of their sex. This approach is also called “balance.” It can lead to unequal outcomes and
ignores the effect of bias in the societal processes that produced the training data.
Equal impact: People with similar likelihood to pay back the loan should have the same expected utility, regardless of the class they belong to. This goes beyond equal
opportunity in that it considers both the benefits of a true prediction and the costs of a
false prediction. Let us examine how these issues play out in a particular context. COMPAS is a commercial system for recidivism (re-offense) scoring. It assigns to a defendant in a criminal case a risk score, which is then used by a judge to help make decisions: Is it safe to release
the defendant before trial, or should they be held in jail? If convicted, how long should the
sentence be? Should parole be granted? Given the significance of these decisions, the system has been the subject of intense scrutiny (Dressel and Farid, 2018). COMPAS is designed to be well calibrated: all the individuals who are given the same
score by the algorithm should have approximately the same probability of re-offending, regardless of race. For example, among all people that the model assigns a risk score of 7 out of 10, 60% of whites and 61% of blacks re-offend. The designers thus claim that it meets the
desired fairness goal.
On the other hand, COMPAS
does not achieve equal opportunity:
the proportion of
those who did not re-offend but were falsely rated as high-risk was 45% for blacks and 23%
for whites. In the case State v. Loomis, where a judge relied on COMPAS
to determine the
sentence of the defendant, Loomis argued that the secretive inner workings of the algorithm
Well calibrated
994
Chapter 27 Philosophy, Ethics, and Safety of AT violated his due process rights. Though the Wisconsin Supreme Court found that the sentence
given would be no different without COMPAS
in this case, it did issue warnings about the
algorithm’s accuracy and risks to minority defendants. Other researchers have questioned whether it is appropriate to use algorithms in applications such as sentencing.
We could hope for an algorithm that is both well calibrated and equal opportunity, but,
as Kleinberg er al. (2016) show, that is impossible.
If the base classes are different, then
any algorithm that is well calibrated will necessarily not provide equal opportunity, and vice
versa. How can we weigh the two criteria? Equal impact is one possibility. In the case of COMPAS, this means weighing the negative utility of defendants being falsely classified as high risk and losing their freedom, versus the cost to society of an additional crime being
committed, and finding the point that optimizes the tradeoff.
This is complicated because
there are multiple costs to consider. There are individual costs—a defendant who is wrong-
fully held in jail suffers a loss, as does the victim of a defendant who was wrongfully released and re-offends. But beyond that there are group costs—everyone has a certain fear that they will be wrongfully jailed, or will be the victim ofa crime, and all taxpayers contribute to the costs of jails and courts. If we give value to those fears and costs in proportion to the size of a group, then utility for the majority may come at the expense of a minority.
Another problem with the whole idea of recidivism scoring, regardless of the model used, is that we don’t have unbiased ground truth data. The data does not tell us who has committed a crime—all we know is who has been convicted of a crime. If the arresting officers, judge, or jury is biased, then the data will be biased. If more officers patrol some locations, then
the data will be biased against people in those locations. Only defendants who are released
are candidates to recommit, so if the judges making the release decisions are biased, the data
may be biased. If you assume that behind the biased data set there is an underlying, unknown,
unbiased data set which has been corrupted by an agent with biases, then there are techniques
to recover an approximation to the unbiased data. Jiang and Nachum (2019) describe various scenarios and the techniques involved.
One more risk is that machine learning can be used to justify bias. If decisions are made by a biased human after consulting with a machine learning system, the human can say “here is how my interpretation of the model supports my decision, so you shouldn’t question my
decision.” But other interpretations could lead to an opposite decision.
Sometimes fairness means that we should reconsider the objective function, not the data
or the algorithm. For example, in making job hiring decisions, if the objective is to hire candidates with the best qualifications in hand, we risk unfairly rewarding those who have
had advantageous educational opportunities throughout their lives, thereby enforcing class
boundaries. But if the objective is to hire candidates with the best ability to learn on the job, we have a better chance to cut across class boundaries and choose from a broader pool. Many
companies have programs designed for such applicants, and find that after a year of training,
the employees hired this way do as well as the traditional candidates. Similarly, just 18% of computer science graduates in the U.S. are women, but some schools, such as Harvey Mudd University, have achieved 50% parity with an approach that is focused on encouraging and retaining those who start the computer science program, especially those who start with less
programming experience.
A final complication is deciding which classes deserve protection.
In the U.S., the Fair
Housing Act recognized seven protected classes: race, color, religion, national origin, sex,
Section 27.3
disability, and familial status.
995
The Ethics of Al
Other local, state, and federal laws recognize other classes,
including sexual orientation, and pregnancy, marital, and veteran status. Is it fair that these
classes count for some laws and not others? International human rights law, which encom-
passes a broad set of protected classes, is a potential framework to harmonize protections across various groups. Even in the absence of societal bias, sample
disparity can lead to biased results.
Sample size disparity
In most data sets there will be fewer training examples of minority class individuals than
of majority class individuals. Machine learning algorithms give better accuracy with more training data, so that means that members of minority classes will experience lower accuracy.
For example, Buolamwini and Gebru (2018) examined a computer vision gender identifica-
tion service, and found that it had near-perfect accuracy for light-skinned males, and a 33% error rate for dark-skinned females. A constrained model may not be able to simultaneously
fit both the majority and minority class—a linear regression model might minimize average
error by fitting just the majority class, and in an SVM model, the support vectors might all correspond to majority class members.
Bias can also come into play in the software development process (whether or not the software involves machine learning). Engineers who are debugging a system are more likely to notice and fix those problems that are applicable to themselves. For example, it is difficult
to notice that a user interface design won’t work for colorblind people unless you are in fact
colorblind, or that an Urdu language translation is faulty if you don’t speak Urdu.
How can we defend against these biases? First, understand the limits of the data you are using. It has been suggested that data sets (Gebru et al., 2018; Hind er al., 2018) and models (Mitchell er al., 2019) should come with annotations: declarations of provenance, security,
conformity, and fitness for use. This is similar to the data sheets that accompany electronic Data sheet components such as resi they allow designers to decide what components to use. In addition to the data sheets, it is important to train engineers to be aware of issues of fairness
and bias, both in school and with on-the-job training. Having a diversity of engineers from different backgrounds makes it easier for them to notice problems in the data or models.
A
study by the Al Now Institute (West et al., 2019) found that only 18% of authors at leading Al conferences and 20% of Al professors are women. Black Al workers are at less than 4%.
Rates at industry research labs are similar. Diversity could be increased by programs earlier in the pipeline—in college or high school—and by greater awareness at the professional level. Joy Buolamwini founded the Algorithmic Justice League to raise awareness of this issue and
develop practices for accountability.
A second idea is to de-bias the data (Zemel ez al., 2013).
We could over-sample from
minority classes to defend against sample size disparity. Techniques such as SMOTE, the synthetic minority over-sampling technique (Chawla e al., 2002) or ADASYN, the adaptive synthetic sampling approach for imbalanced learning (He ef al., 2008), provide principled ways of oversampling. We could examine the provenance of data and, for example, eliminate examples from judges who have exhibited bias in their past court cases. Some analysts object
to the idea of discarding data, and instead would recommend building a hierarchical model of
the data that includes sources of bias, so they can be modeled and compensated for. Google
and NeurlPS have attempted to raise awareness of this issue by sponsoring the Inclusive Images Competition, in which competitors train a network on a data set of labeled images
collected in North America and Europe, and then test it on images taken from all around the
996
Chapter 27 Philosophy, Ethics, and Safety of AT world. The issue is that given this data set, it is easy to apply the label “bride” to a woman
in a standard Western wedding dress, but harder to recognize traditional African and Indian matrimonial dress. A third idea is to invent new machine learning models and algorithms that are more resistant to bias; and the final idea is to let a system make initial recommendations that may be biased, but then train a second system to de-bias the recommendations of the first one. Bellamy
et al. (2018) introduced the IBM AT FAIRNESS 360 system, which provides a framework for all of these ideas. We expect there will be increased use of tools like this in the future.
How do you make sure that the systems you build will be fair? A set of best practices has
been emerging (although they are not always followed):
* Make sure that the software engineers talk with social scientists and domain experts to understand the issues and perspectives, and consider fairness from the start.
+ Create an environment that fosters the development of a diverse pool of software engi-
neers that are representative of society.
+ Define what groups your system will support: different language speakers, different age
groups, different abilities with sight and hearing, etc.
+ Optimize for an objective function that incorporates fairness. + Examine your data for prejudice and for correlations between protected attributes and other attributes. + Understand how any human annotation of data is done, design goals for annotation
accuracy, and verify that the goals are met.
+ Don’t just track overall metrics for your system; make sure you track metrics for sub-
groups that might be victims of bias.
+ Include system tests that reflect the experience of minority group users.
« Have a feedback loop so that when fairness problems come up, they are dealt with. 27.3.4
Trust
Trust and transparency
It is one challenge to make an Al system accurate, fair, safe, and secure; a different chal-
lenge to convince everyone else that you have done so. People need to be able to trust the systems they use. A PwC survey in 2017 found that 76% of businesses were slowing the
adoption of Al because of trustworthiness concerns.
Verification and validation
In Section 19.9.4 we covered some of
the engineering approaches to trust; here we discuss the policy issues.
To earn trust, any engineered systems must go through a verification and validation
(V&V) process.
Verification means that the product satisfies the specifications.
Validation
means ensuring that the specifications actually meet the needs of the user and other affected
parties. We have an elaborate V&V methodology for engineering in general, and for traditional software development done by human coders; much of that is applicable to Al systems.
But machine learning systems are different and demand a different V&V process, which has
not yet been fully developed.
We need to verify the data that these systems learn from; we
need to verify the accuracy and fairness of the results, even in the face of uncertainty that makes an exact result unknowable;
Certification
and we need to verify that adversaries cannot unduly
influence the model, nor steal information by querying the resulting model.
One instrument of trust is certification; for example, Underwriters Laboratories (UL) was founded in 1894 at a time when consumers were apprehensive about the risks of electric
Section 27.3
997
The Ethics of Al
power. UL certification of appliances gave consumers increased trust, and in fact UL is now considering entering the business of product testing and certification for AL Other industries have long had safety standards.
For example, ISO 26262 is an interna-
tional standard for the safety of automobiles, describing how to develop, produce, operate, and service vehicles in a safe way. The Al industry is not yet at this level of clarity, although there are some frameworks
in progress, such as IEEE P7001, a standard defining ethical de-
sign for artificial intelligence and autonomous systems (Bryson and Winfield, 2017). There is ongoing debate about what kind of certification is necessary, and to what extent it should be
done by the government, by professional organizations like IEEE, by independent certifiers such as UL, or through self-regulation by the product companies. Another aspect of trust is transparency:
consumers want to know what is going on
inside a system, and that the system is not working against them, whether due to intentional
Transparency
malice, an unintentional bug, or pervasive societal bias that is recapitulated by the system. In
some cases this transparency is delivered directly to the consumer. In other cases their are
intellectual property issues that keep some aspects of the system hidden to consumers, but
open to regulators and certification agencies. When an Al system turns you down for a loan, you deserve an explanation. In Europe,
the GDPR enforces this for you. An Al system that can explain itself is called explainable AT
(XAI). A good explanation has several properties: it should be understandable and convincing Explainable Al (XAl) to the user, it should accurately reflect the reasoning of the system, it should be complete, and it should be specific in that different users with different conditions or different outcomes
should get different explanations.
It is quite easy to give a decision algorithm access to its own deliberative processes,
simply by recording them and making them available as data structures. This means that machines may eventually be able to give better explanations of their decisions than humans can. Moreover, we can take steps to certify that the machine’s explanations are not deceptions (intentional or self-deception), something that is more difficult with a human.
An explanation is a helpful but not sufficient ingredient to trust. One issue is that expla-
nations are not decisions:
they are stories about decisions. As discussed in Section 19.9.4, we
say that a system is interpretable if we can inspect the source code of the model and see what
itis doing, and we say it is explainable if we can make up a story about what it is doing—even if the system itself is an uninterpretable black box. To explain an uninterpretable black box, we need to build, debug, and test a separate explanation system, and make sure it is in sync with the original system. And because humans love a good story, we are all too willing to be swayed by an explanation that sounds good. Take any political controversy of the day, and
you can always find two so-called experts with diametrically opposed explanations, both of which are internally consistent.
A final issue is that an explanation about one case does not give you a summary over
other cases. If the bank explains, “Sorry, you didn’t get the loan because you have a history
of previous financial problems,” you don’t know if that explanation is accurate or if the bank is
secretly biased against you for some reason. In this case, you require not just an explanation, but also an audit of past decisions, with aggregated statistics across various demographic
groups, to see if their approval rates are balanced.
Part of transparency is knowing whether you are interacting with an Al system or a hu-
man. Toby Walsh (2015) proposed that “an autonomous system should be designed so that
998
Chapter 27 Philosophy, Ethics, and Safety of AT itis unlikely to be mistaken for anything besides an autonomous system, and should identify itself at the start of any interaction.”” He called this the “red flag” law, in honor of the UK’s
1865 Locomotive Act, which required any motorized vehicle to have a person with a red flag walk in front of i, to signal the oncoming danger.
In 2019, California enacted a law stating that “It shall be unlawful for any person to use a bot to communicate or interact with another person in California online, with the intent to
mislead the other person about its artificial identity.”
27.3.5
The future of work
From the first agricultural revolution (10,000 BCE)
to the industrial revolution (late 18th
century) to the green revolution in food production (1950s), new technologies have changed
the way humanity works and lives. A primary concern arising from the advance of Al is that human labor will become obsolete.
point quite clearly:
Aristotle, in Book I of his Politics, presents the main
For if every instrument could accomplish its own work, obeying or anticipating the will of others ....if, in like manner, the shuttle would weave and the plectrum touch the lyre withouta hand to guide them, chief workmen would not want servants, nor masters slaves. Everyone agrees with Aristotle’s observation that there is an immediate reduction in employ-
ment when an employer finds a mechanical method to perform work previously done by a
person. The issue is whether the so-called compensation effects that ensue—and that tend to
increase employment—will eventually make up for this reduction. The primary compensa-
tion effect is the increase in overall wealth from greater productivity, which leads in turn to
greater demand for goods and tends to increase employment. For example, PwC (Rao and Verweij, 2017) predicts that Al contribute $15 trillion annually to global GDP by 2030. The healthcare and automotive/transportation industries stand to gain the most in the short term. However, the advantages of automation have not yet taken over in our economy: the current
rate of growth in labor productivity is actually below historical standards. Brynjolfsson ef al.
(2018) attempt to explain this paradox by suggesting that the lag between the development of
Technological unemployment
basic technology and its implementation in the economy is longer than commonly supposed. Technological innovations have historically put some people out of work. Weavers were replaced by automated looms in the 1810s, leading to the Luddite protests. The Luddites were not against technology per se; they just wanted the machines to be used by skilled workers paid a good wage to make high-quality goods, rather than by unskilled workers to make poorquality goods at low wages. The global destruction of jobs in the 1930s led John Maynard Keynes to coin the term technological unemployment.
employment levels eventally recovered.
In both cases, and several others,
The mainstream economic view for most of the 20th century was that technological em-
ployment was at most a short-term phenomenon. Increased productivity would always lead to increased wealth and increased demand, and thus net job growth. A commonly cited example is that of bank tellers: although ATMs replaced humans in the job of counting out cash for withdrawals, that made it cheaper to operate a bank branch, so the number of branches
increased, leading to more bank employees overall. The nature of the work also changed, becoming less routine and requiring more advanced business skills. The net effect of automation seems to be in eliminating fasks rather than jobs.
Section 27.3
The Ethics of Al
The majority of commenters predict that the same will hold true with AI technology, at
least in the short run. Gartner, McKinsey, Forbes, the World Economic Forum, and the Pew
Research Center each released reports in 2018 predicting a net increase in jobs due to Al-
driven automation. But some analysts think that this time around, things will be different. In 2019, IBM predicted that 120 million workers would need retraining due to automation by 2022, and Oxford Economics predicted that 20 million manufacturing jobs could be lost to
automation by 2030.
Frey and Osborne (2017) survey 702 different occupations, and estimate that 47% of them
are at risk of being automated, meaning that at least some of the tasks in the occupation can
be performed by machine. For example, almost 3% of the workforce in the U.S. are vehicle drivers, and in some districts, as much as 15% of the male workforce are drivers. As we saw in Chapter 26, the task of driving is likely to be eliminated by driverless cars/trucks/buses/taxis.
It is important to distinguish between occupations and the tasks within those occupations.
McKinsey estimates that only 5% of occupations are fully automatable, but that 60% of occupations can have about 30% of their tasks automated.
For example, future truck drivers
will spend less time holding the steering wheel and more time making sure that the goods are picked up and delivered properly; serving as customer service representatives and salespeople at either end of the journey; and perhaps managing convoys of, say, three robotic trucks. Replacing three drivers with one convoy manager implies a net loss in employment, but if transportation costs decrease, there will be more demand,
which wins some of the
jobs back—but perhaps not all of them. As another example, despite many advances in applying machine learning to the problem of medical imaging, radiologists have so far been
augmented, not replaced, by these tools. Ultimately, there is a choice of how to make use of automation: do we want to focus on cutting cost, and thus see job loss as a positive; or do we
want to focus on improving quality, making life better for the worker and the customer? It is difficult to predict exact timelines for automation, but currently, and for the next
few years, the emphasis is on automation of structured analytical tasks, such as reading x-ray images, customer relationship management (e.g., bots that automatically sort customer comBusiness process plaints and respond with suggested remedies), and business process automation that com- automation bines text documents and structured data to make business decisions and improve workflow. Over time, we will see more automation with physical robots, first in controlled warehouse
environments, then in more uncertain environments, building to a significant portion of the
marketplace by around 2030.
As populations in developed countries grow older, the ratio between workers and retirees
changes. In 2015 there were less than 30 retirees per 100 workers; by 2050 there may be over 60 per 100 workers. Care for the elderly will be an increasingly important role, one that can partially be filled by AL Moreover, if we want to maintain the current standard of living, it will also be necessary to make the remaining workers more productive; automation seems
like the best opportunity to do that.
Even if automation has a multi-trillion-dollar net positive impact, there may still be prob-
lems due to the pace of change. Consider how change came to the farming industry: in 1900, Pace of change
over 40% of the U.S. workforce was in agriculture, but by 2000 that had fallen to 2%.> That
In 2010, although only 2% of the U.S. workforce were actual farmers, over 25% of the population (80 million people) played the FARMVILLE game at least once.
999
1000
Chapter 27 Philosophy, Ethics, and Safety of AT is a huge disruption in the way we work, but it happened over a period of 100 years, and thus across generations, not in the lifetime of one worker.
Workers whose jobs are automated away this decade may have to retrain for a new profession within a few years—and then perhaps see their new profession automated and face
yet another retraining period. Some may be happy to leave their old profession—we see that
as the economy improves, trucking companies need to offer new incentives to hire enough
drivers—but workers will be apprehensive about their new roles. To handle this, we as a soci-
ety need to provide lifelong education, perhaps relying in part on online education driven by artificial intelligence (Martin, 2012). Bessen (2015) argues that workers will not see increases
Income inequality
in income until they are trained to implement the new technologies, a process that takes time. Technology tends to magnify income inequality.
In an information economy marked
by high-bandwidth global communication and zero-marginal-cost replication of intellectual property (what Frank and Cook (1996) call the “Winner-Take-All Society”), rewards tend to be concentrated.
If farmer Ali is 10% better than farmer Bo, then Ali gets about 10% more
income: Ali can charge slightly more for superior goods, but there is a limit on how much can be produced on the land, and how far it can be shipped. But if software app developer
Cary is 10% better than Dana, it may be that Cary ends up with 99% of the global market. AT
increases the pace of technological innovation and thus contributes to this overall trend, but
Al also holds the promise of allowing us to take some time off and let our automated agents
handle things for a while. Tim Ferriss (2007) recommends using automation and outsourcing to achieve a four-hour work week. Before the industrial revolution, people worked as farmers or in other crafts, but didn’t
report to a job at a place of work and put in hours for an employer.
But today, most adults
in developed countries do just that, and the job serves three purposes: it fuels the production
of the goods that society needs to flourish, it provides the income that the worker needs to live, and it gives the worker a sense of purpose, accomplishment, and social integration. With increasing automation, it may be that these three purposes become disaggregated—society’s
needs will be served in part by automation, and in the long run, individuals will get their sense of purpose from contributions other than work. Their income needs can be served by
social policies that include a combination of free or inexpensive access to social services and education, portable health care, retirement, and education accounts, progressive tax rates,
earned income tax credits, negative income tax, or universal basic income.
27.3.6
Robot rights
The question of robot consciousness, discussed in Section 27.2, is critical to the question of
what rights, if any, robots should have. would argue that they deserve rights.
If they have no consciousness, no qualia, then few
But if robots can feel pain, if they can dread death, if they are considered “persons,” then the argument can be made (e.g., by Sparrow (2004)) that they have rights and deserve to have their rights recognized, just as slaves, women, and other historically oppressed groups have
fought to have their rights recognized. The issue of robot personhood is often considered in fiction: from Pygmalion to Coppélia to Pinocchio to the movies A7 and Centennial Man, we have the legend of a doll/robot coming to life and striving to be accepted as a human with human rights. In real life, Saudi Arabia made headlines by giving honorary citizenship to Sophia, a human-looking puppet capable of speaking preprogrammed lines.
Section 27.3
1001
The Ethics of Al
If robots have rights, then they should not be enslaved, and there is a question of whether
reprogramming them would be a kind of enslavement. Another ethical issue involves voting rights: a rich person could buy thousands of robots and program them to cast thousands of votes—should
those votes count?
If a robot clones itself, can they both vote?
What is
the boundary between ballot stuffing and exercising free will, and when does robotic voting violate the “one person, one vote” principle? Ernie Davis argues for avoiding the dilemmas of robot consciousness by never building robots that could possibly be considered conscious. This argument was previously made by
Joseph Weizenbaum in his book Computer Power and Human Reason (1976), and before that by Julien de La Mettrie in L’Homme Machine (1748). Robots are tools that we create, to do
the tasks we direct them to do, and if we grant them personhood, we are just declining to take
responsibility for the actions of our own property: “I'm not at fault for my self-driving car crash—the car did it itself.” This issue takes a different turn if we develop human-robot hybrids. Of course we already
have humans enhanced by technology such as contact lenses, pacemakers, and artificial hips. But adding computational protheses may blur the lines between human and machine. 27.3.7
Al Safety
Almost any technology has the potential to cause harm in the wrong hands, but with AT and robotics, the hands might be operating on their own. Countless science fiction stories have warned about robots or cyborgs running amok. Early examples include Mary Shelley’s Frankenstein, or the Modern Prometheus (1818) and Karel Capek’s play R.U.R. (1920), in which robots conquer the world. In movies, we have The Terminator (1984) and The Matrix
(1999), which both feature robots trying to eliminate humans—the robopocalypse (Wilson, Robopocalypse 2011). Perhaps robots are so often the villains because they represent the unknown, just like the witches and ghosts of tales from earlier eras.
We can hope that a robot that is smart
enough to figure out how to terminate the human race is also smart enough to figure out that
that was not the intended utility function; but in building intelligent systems, we want to rely not just on hope, but on a design process with guarantees of safety.
It would be unethical to distribute an unsafe Al agent.
We require our agents to avoid
accidents, to be resistant to adversarial attacks and malicious abuse, and in general to cause
benefits, not harms. That is especially true as Al agents are deployed in safety-critical appli-
cations, such as driving cars, controlling robots in dangerous factory or construction settings, and making life-or-death medical decisions.
There is a long history of safety engineering in traditional engineering fields. We know
how to build bridges, airplanes, spacecraft, and power plants that are designed up front to
behave safely even when components of the system fail. The first technique is failure modes
and effect analysis (FMEA): analysts consider each component of the system, and imagine
every possible way the component could go wrong (for example, what if this bolt were to snap?), drawing on past experience and on calculations based on the physical properties of the component.
Safety engineering Failure modes and effect analysis
(FMEA)
Then the analysts work forward to see what would result from the failure.
If the result is severe (a section of the bridge could fall down) then the analysts alter the design to mitigate the failure. (With this additional cross-member, the bridge can survive the
failure of any 5 bolts; with this backup server, the online service can survive a tsunami taking
tree analysis out the primary server.) The technique of fault tree analysis (FTA) is used to make these Fault (FTA)
1002
Chapter 27 Philosophy, Ethics, and Safety of AT determinations:
analysts build an AND/OR
tree of possible failures and assign probabilities
to each root cause, allowing for calculations of overall failure probability. These techniques can and should be applied to all safety-critical engineered systems, including AI systems. The field of software engineering is aimed at producing reliable software, but the em-
phasis has historically been on correctness, not safety. Correctness means that the software
faithfully implements the specification. But safety goes beyond that to insist that the specification has considered any feasible failure modes, and is designed to degrade gracefully even in the face of unforeseen failures. For example, the software for a self-driving car wouldn’t
be considered safe unless it can handle unusual situations. For example, what if the power to
the main computer dies? A safe system will have a backup computer with a separate power
supply. What if a tire is punctured at high speed? A safe system will have tested for this, and will have software to correct for the resulting loss of control.
Unintended side effect
An agent designed as a utility maximizer, or as a goal achiever, can be unsafe if it has the wrong objective function. Suppose we give a robot the task of fetching a coffee from the kitchen. We might run into trouble with unintended side effects—the robot might rush
to accomplish the goal, knocking over lamps and tables along the way. In testing, we might
notice this kind of behavior and modify the utility function to penalize such damage, but it is
Low impact
difficult for the designers and testers to anticipate all possible side effects ahead of time.
One way to deal with this is to design a robot to have low impact (Armstrong and Levin-
stein, 2017): instead of just maximizing utility, maximize the utility minus a weighted summary of all changes to the state of the world.
In this way, all other things being equal, the
robot prefers not to change those things whose effect on utility is unknown; so it avoids knocking over the lamp not because it knows specifically that knocking the lamp will cause
it to fall over and break, but because it knows in general that disruption might be bad. This
can be seen as a version of the physician’s creed “first, do no harm,” or as an analog to regularization in machine learning: we want a policy that achieves goals, but we prefer policies that take smooth, low-impact actions to get there.
The trick is how to measure impact.
It
is not acceptable to knock over a fragile lamp, but perfectly fine if the air molecules in the room are disturbed a little, or if some bacteria in the room are inadvertently killed. It is cer-
tainly not acceptable to harm pets and humans in the room.
We need to make sure that the
robot knows the differences between these cases (and many subtle cases in between) through
a combination of explicit programming, machine learning over time, and rigorous testing.
Utility functions can go wrong due to externalities, the word used by economists for
factors that are outside of what is measured and paid for. The world suffers when green-
house gases are considered as externalities—companies and countries are not penalized for producing them, and as a result everyone suffers. Ecologist Garrett Hardin (1968) called the exploitation of shared resources the tragedy of the commons. We can mitigate the tragedy by internalizing the externalities—making them part of the utility function, for example with
a carbon tax—or by using the design principles that economist Elinor Ostrom identified as being used by local people throughout the world for centuries (work that won her the Nobel Prize in Economics in 2009):
Clearly define the shared resource and who has access.
+ Adapt to local conditions.
« Allow all parties to participate in decisions.
Section 27.3
The Ethics of Al
Monitor the resource with accountable monitors.
+ Sanctions, proportional to the severity of the violation. + Easy conflict resolution procedures. Hierarchical control for large shared resources.
Victoria Krakovna (2018) has cataloged examples of Al agents that have gamed the system, figuring out how to maximize utility without actually solving the problem that their designers
intended them to solve. To the designers this looks like cheating, but to the agents, they
are just doing their job.
Some agents took advantage of bugs in the simulation (such as
floating point overflow bugs) to propose solutions that would not work once the bug was
fixed. Several agents in video games discovered ways to crash or pause the game when they
were about to lose, thus avoiding a penalty. And in a specification where crashing the game
was penalized, one agent learned to use up just enough of the game’s memory so that when it
was the opponent’s turn, it would run out of memory and crash the game. Finally, a genetic
algorithm operating in a simulated world was supposed to evolve fast-moving creatures but
in fact produced creatures that were enormously tall and moved fast by falling over. Designers of agents should be aware of these kinds of specification failures and take steps
to avoid them. To help them do that, Krakovna was part of the team that released the Al Safety Gridworlds environments (Leike ef al., 2017), which allows designers to test how well their
agents perform.
The moral is that we need to be very careful in specifying what we want, because with
alignment utility maximizers we get what we actually asked for. The value alignment problem is the Value problem problem of making sure that what we ask for is what we really want; it is also known as the
King Midas problem, as discussed on page 33. We run into trouble when a utility function fails to capture background societal norms about acceptable behavior. For example, a human
who is hired to clean floors, when faced with a messy person who repeatedly tracks in dirt,
Kknows that it is acceptable to politely ask the person to be more careful, but it is not acceptable to kidnap or incapacitate said person.
A robotic cleaner needs to know these things too, either through explicit programming or
by learning from observation. Trying to write down all the rules so that the robot always does
the right thing is almost certainly hopeless. We have been trying to write loophole-free tax laws for several thousand years without success. Better to make the robot want to pay taxes,
50 to speak, than to try to make rules to force it to do so when it really wants to do something else. A sufficiently intelligent robot will find a way to do something else.
Robots can learn to conform better with human preferences by observing human behavfor. This is clearly related to the notion of apprenticeship learning (Section 22.6). The robot
may learn a policy that directly suggests what actions to take in what situations; this is often
a straightforward supervised learning problem if the environment is observable. For example, a robot can watch a human playing chess: each state-action pair is an example for the learning process.
Unfortunately, this form of imitation learning means that the robot will
repeat human mistakes. Instead, the robot can apply inverse reinforcement learning to
cover the utility function that the humans must be operating under. Watching even terrible chess players is probably enough for the robot to learn the objective of the game. Given just this information, the robot can then go on to exceed human performance—as, for example, ALPHAZERO
did in chess—by computing optimal or near-optimal policies from the objec-
1003
1004
Chapter 27 Philosophy, Ethics, and Safety of AT tive. This approach works not just in board games, but in real-world physical tasks such as helicopter aerobatics (Coates et al., 2009).
In more complex settings involving, for example, social interactions with humans, it is very unlikely that the robot will converge to exact and correct knowledge of each human’s individual preferences. (After all, many humans never quite learn what makes other humans tick, despite a lifetime of experience, and many of s are unsure of our own preferences t00.) It will be necessary, therefore, for machines to function appropriately when it is uncertain
about human preferences. In Chapter 18, we introduced assistance games, which capture
exactly this situation.
Solutions to assistance games include acting cautiously, so as not to
disturb aspects of the world that the human might care about, and asking questions. For example, the robot could ask whether turning the oceans into sulphuric acid is an acceptable
solution to global warming before it puts the plan into effect.
In dealing with humans, a robot solving an assistance game must accommodate human
imperfections. If the robot asks permission, the human may give it, not foreseeing that the
robot’s proposal is in fact catastrophic in the long term.
Moreover, humans do not have
complete introspective access to their true utility function, and they don’t always act in a way
that is compatible with it. Humans sometimes lie or cheat, or do things they know are wrong.
They sometimes take self-destructive actions like overeating or abusing drugs. AI systems need not learn to adopt these problematic tendencies, but they must understand that they exist when interpreting human behavior to get at the underlying human preferences. Despite this toolbox of safeguards, there is a fear, expressed by prominent technologists such as Bill Gates and Elon Musk and scientists such as Stephen Hawking and Martin Rees,
that AT could evolve out of control. They warn that we have no experience controlling powerful nonhuman entities with super-human capabilities. However, that’s not quite true; we have
centuries of experience with nations and corporations; non-human entities that aggregate the power of thousands or millions of people. Our record of controlling these entities is not very
encouraging: nations produce periodic convulsions called wars that kill tens of millions of human beings, and corporations are partly responsible for global warming and our inability
to confront it. Al systems may present much greater problems than nations and corporations because of
Ultraintelligent machine
Technological singularity
their potential to self-improve at a rapid pace, as considered by L. J. Good (1965b): Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines: there would then unquestionably be an “intelligence explosion,” and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control.
Good’s “intelligence explosion” has also been called the technological singularity by math-
ematics professor and science fiction author Vernor Vinge, who wrote in 1993: “Within thirty
years, we will have the technological means to create superhuman intelligence. Shortly after, the human era will be ended.”
In 2017, inventor and futurist Ray Kurzweil predicted the
singularity would appear by 2045, which means it got 2 years closer in 24 years. (At that
rate, only 336 years to go!) Vinge and Kurzweil correctly note that technological progress on many measures is growing exponentially at present.
Summary
1005
It is, however, quite a leap to extrapolate all the way from the rapidly decreasing cost
of computation to a singularity.
So far, every technology has followed an S-shaped curve,
where the exponential growth eventually tapers off. Sometimes new technologies step in
when the old ones plateau, but sometimes it is not possible to keep the growth going, for
technical, political, or sociological reasons. For example, the technology of flight advanced
dramatically from the Wright brothers” flight in 1903 to the moon landing in 1969, but has
had no breakthroughs of comparable magnitude since then.
Another obstacle in the way of ultraintelligent machines taking over the world is the
world. More specifically, some kinds of progress require not just thinking but acting in the
physical world. (Kevin Kelly calls the overemphasis on pure intelligence thinkism.) An ul- Thinkism traintelligent machine tasked with creating a grand unified theory of physics might be capable
of cleverly manipulating equations progress, it would still need to raise and run physical experiments over analyzing the data and theorizing.
a billion times faster than Einstein, but to make any real millions of dollars to build a more powerful supercollider the course of months or years. Only then could it start Depending on how the data turn out, the next step might
require raising additional billions of dollars for an interstellar probe mission that would take
centuries to complete. The “ultraintelligent thinking™ part of this whole process might actu-
ally be the least important part. As another example, an ultraintelligent machine tasked with
bringing peace to the Middle East might just end up getting 1000 times more frustrated than
a human envoy. As yet, we don’t know how many of the big problems are like mathematics and how many are like the Middle East.
While some people fear the singularity, others relish it. The transhumanism social Transhumanism movement looks forward to a future in which humans are merged with—or replaced by—
robotic and biotech inventions. Ray Kurzweil writes in The Singularity is Near (2005):
The Singularity will allow us to transcend these limitations of our biological bodies and brain. We will gain power over our fates. ... We will be able to live as long as we want .. We will fully understand human thinking and will vastly extend and expand its reach. By the end of this century, the nonbiological portion of our intelligence will be trillions of trillions of times more powerful than unaided human intelligence. Similarly, when asked whether robots will inherit the Earth, Marvin Minsky said “yes, but they will be our children.” These possibilities present a challenge for most moral theorists,
who take the preservation of human life and the human species to be a good thing. Kurzweil also notes the potential dangers, writing “But the Singularity will also amplify the ability to act on our destructive inclinations, so its full story has not yet been written.”
We humans
would do well to make sure that any intelligent machine we design today that might evolve into an ultraintelligent machine will do so in a way that ends up treating us well.
As Eric
Brynjolfsson puts it, “The future is not preordained by machines. It’s created by humans.”
Summary This chapter has addressed the following issues:
+ Philosophers use the term weak Al for the hypothesis that machines could possibly behave intelligently, and strong AI for the hypothesis that such machines would count
as having actual minds (as opposed to simulated minds).
1006
Chapter 27 Philosophy, Ethics, and Safety of AT « Alan Turing rejected the question “Can machines think?” and replaced it with a behavioral test. He anticipated many objections to the possibility of thinking machines.
Few Al researchers pay attention to the Turing test, preferring to concentrate on their
systems’ performance on practical tasks, rather than the ability to imitate humans. « Consciousness remains a mystery. « Al s a powerful technology, and as such it poses potential dangers, through lethal autonomous weapons, security and privacy breaches, unintended side effects, uninten-
tional errors, and malignant misuse. Those who work with AI technology have an
ethical imperative to responsibly reduce those dangers. + Al systems must be able to demonstrate they are fair, trustworthy, and transparent.
+ There are multiple aspects of fairness, and it is impossible to maximize all of them at
once. So a first step is to decide what counts as fair.
« Automation is already changing the way people work. As a society, we will have to deal with these changes. Bibliographical and Historical Notes Weak AL: When Alan Turing (1950) proposed the possibility of AL he also posed many of the key philosophical questions, and provided possible replies. But various philosophers had raised similar issues long before Al was invented. Maurice Merleau-Ponty’s Phenomenology of Perception (1945) stressed the importance of the body and the subjective interpretation of reality afforded by our senses, and Martin Heidegger’s Being and Time (1927) asked what it means to actually be an agent. In the computer age, Alva Noe (2009) and Andy Clark (2015) propose that our brains form a rather minimal representation of the world, use the world itself
on a just-in-time basis to maintain the illusion of a detailed internal model, and use props
in the world (such as paper and pencil as well as computers) to increase the capabilities of the mind. Pfeifer ef al. (2006) and Lakoff and Johnson (1999) present arguments for how the body helps shape cognition. Speaking of bodies, Levy (2008), Danaher and McArthur the issue of robot sex. (2017), and Devlin (2018) address Strong AL René Descartes is known for his dualistic view of the human mind, but ironi-
cally his historical influence was toward mechanism and physicalism. He explicitly conceived
of animals as automata, and he anticipated the Turing test, writing “it is not conceivable [that
a machine] should produce different arrangements of words so as to give an appropriately
meaningful answer to whatever is said in its presence, as even the dullest of men can do”
(Descartes,
1637).
Descartes’s spirited defense of the animals-as-automata viewpoint ac-
tually had the effect of making it easier to conceive of humans as automata as well, even though he himself did not take this step. The book L’Homme Machine (La Mettrie, 1748) did explicitly argue that humans are automata. As far back as Homer (circa 700 BCE), the Greek legends envisioned automata such as the bronze giant Talos and considered the issue of biotechne, or life through craft (Mayor, 2018).
The Turing test (Turing, 1950) has been debated (Shieber, 2004), anthologized (Epstein et al., 2008), and criticized (Shieber, 1994; Ford and Hayes, 1995). Bringsjord (2008) gives advice for a Turing test judge, and Christian (2011) for a human contestant. The annual Loebner Prize competition is the longest-running Turing test-like contest; Steve Worswick’s
Bibliographical and Historical Notes MITSUKU won four in a row from 2016 to 2019. The Chinese room has been debated endlessly (Searle, 1980; Chalmers, 1992; Preston and Bishop, 2002). Herndndez-Orallo (2016)
gives an overview of approaches to measuring Al progress, and Chollet (2019) proposes a measure of intelligence based on skill-acquisition efficiency. Consciousness remains a vexing problem for philosophers, neuroscientists, and anyone who has pondered their own existence. Block (2009), Churchland (2013) and Dehaene (2014) provide overviews of the major theories. Crick and Koch (2003) add their expertise in biology and neuroscience to the debate, and Gazzaniga (2018) shows what can be learned from
studying brain disabilities in hospital cases. Koch (2019) gives a theory of consciousness—
“intelligence is about doing while experience is about being”—that includes most animals,
but not computers. Giulio Tononi and his colleagues propose integrated information theory
(Oizumi et al., 2014). Damasio (1999) has a theory based on three levels: emotion, feeling, and feeling a feeling. Bryson (2012) shows the value of conscious attention for the process of learning action selection. The philosophical literature on minds, brains, and related topics is large and jargon-filled. The Encyclopedia of Philosophy (Edwards, 1967) is an impressively authoritative and very useful navigation aid. The Cambridge Dictionary of Philosophy (Audi, 1999) is shorter and more accessible, and the online Stanford Encyclopedia of Philosophy offers many excellent articles and up-to-date references. The MIT Encyclopedia of Cognitive Science (Wilson and Keil, 1999) covers the philosophy, biology, and psychology of mind. There are multiple intro-
ductions to the philosophical “Al question” (Haugeland, 1985; Boden, 1990; Copeland, 1993; McCorduck, 2004; Minsky, 2007). The Behavioral and Brain Sciences, abbreviated BBS, is
amajor journal devoted to philosophical and scientific debates about Al and neuroscience.
Science fiction writer Isaac Asimov (1942, 1950) was one of the first to address the issue
of robot ethics, with his laws of robotics:
0. A robot may not harm humanity, or through inaction, allow humanity to come to harm.
1. A robot may not injure a human being or, through inaction, allow a human being to come to harm. 2. A robot must obey orders given to it by human beings, except where such orders would conflict with the First Law. 3. A robot must protect its own existence as long as such protection does not conflict with
the First or Second Law.
At first glance, these laws seem reasonable. But the trick is how to implement them. Should
a robot allow a human to cross the street, or eat junk food, if the human might conceivably
come to harm? In Asimov’s story Runaround (1942), humans need to debug a robot that is
found wandering in a circle, acting “drunk > They work out that the circle defines the locus of points that balance the second law (the robot was ordered to fetch some selenium at the center
of the circle) with the third law (there is a danger there that threatens the robot’s existence).* This suggests that the laws are not logical absolutes, but rather are weighed against each other, with a higher weight for the carlier laws. As this was 1942, before the emergence of
4 Science fiction writers are in broad agreement that robot e very b HAL 9000 computer becomes homicidal due to a con Captain Kirk tells an enemy robot that “Everything Harry el youis smoke comes out of the robot’s head and it shuts down.
at resolving contradictions. In 2001, the the Star Trek episode 1, Mudd,” Tam lying.” At this,
1007
1008
Chapter 27 Philosophy, Ethics, and Safety of AT digital computers, Asimov was probably thinking of an architecture based on control theory via analog computing. Weld and Etzioni (1994) analyze Asimov’s laws and suggest some ways to modify the planning techniques of Chapter 11 to generate plans that do no harm. Asimov has considered many of the ethical issues around technology; in his 1958 story The Feeling of Power he tackles the issue of automation leading to a lapse of human skill—a technician rediscovers
the lost art of multiplication—as well as the dilemma of what to do when the rediscovery is
applied to warfare.
Norbert Wiener’s book God & Golem,
Inc.
(1964) correctly predicted that computers
would achieve expert-level performance at games and other tasks, and that specifying what it is that we want would prove to be difficult. Wiener writes:
While it is always possible to ask for something other than we really want, this
possibility is most serious when the process by which we are to obtain our wish
is indirect, and the degree to which we have obtained our wish is not clear until
the very end. Usually we realize our wishes, insofar as we do actually realize them, by a feedback process, in which we compare the degree of attainment of intermediate goals with our anticipation of them.
In this process, the feedback
goes through us, and we can turn back before it is too late.
If the feedback is
built into a machine that cannot be inspected until the final goal is attained, the
possibilities for catastrophe are greatly increased. I should very much hate to ride
on the first trial of an automobile regulated by photoelectric feedback devices, unless there were somewhere a handle by which I could take over control ifT
found myself driving smack into a tree.
We summarized codes of ethics in the chapter, but the list of organizations that have is-
sued sets of principles is growing rapidly, and now includes Apple, DeepMind, Facebook, Google, IBM, Microsoft, the Organisation for Economic Co-operation and Development (OECD), the United Nations Educational, Scientific and Cultural Organization (UNESCO),
the U.S. Office of Science and Technology Policy the Beijing Academy of Artificial Intelligence (BAAI), the Institute of Electrical and Electronics Engineers (IEEE), the Association
of Computing Machinery (ACM), the World Economic Forum, the Group of Twenty (G20),
OpenAl, the Machine Intelligence Research Institute (MIRI), Al4People, the Centre for the Study of Existential Risk, the Center for Human-Compatible Al the Center for Humane Tech-
nology, the Partnership on Al the AI Now Institute, the Future of Life Institute, the Future
of Humanity Institute, the European Union, and at least 42 national governments. We have the handbook on the Ethics of Computing (Berleur and Brunnstein, 2001) and introductions to the topic of Al ethics in book (Boddington, 2017) and survey (Etzioni and Etzioni, 2017a)
form. The Journal of Artificial Intelligence and Law and Al and Society cover ethical issues. ‘We’ll now look at some of the individual issues.
Lethal autonomous weapons: P. W. Singer’s Wired for War (2009) raised ethical, legal,
and technical issues around robots on the battlefield.
Paul Scharre’s Army of None (2018),
written by one of the authors of current US policy on autonomous weapons, offers a balanced
and authoritative view.
Etzioni and Etzioni (2017b) address the question of whether artifi-
cial intelligence should be regulated; they recommend a pause in the development of lethal autonomous weapons, and an international discussion on the subject of regulation.
Bibliographical and Historical Notes Privacy: Latanya Sweeney (Sweeney, 2002b) presents the k-anonymity model and the idea of generalizing fields (Sweeney, 2002a). Achieving k-anonymity with minimal loss of data is an NP-hard problem, but Bayardo and Agrawal (2005) give an approximation algorithm. Cynthia Dwork (2008) describes differential privacy, and in subsequent work gives practical examples of clever ways to apply differential privacy to get better results than the naive approach (Dwork et al., 2014). Guo et al. (2019) describe a process for certified data
removal: if you train a model on some data, and then there is a request to delete some of the
data, this extension of differential privacy lets you modify the model and prove that it does not make use of the deleted data. Ji ef al. (2014) gives a review of the field of privacy. Etzioni (2004) argues for a balancing of privacy and security; individual rights and community. Fung etal. (2018), Bagdasaryan et al. (2018) discuss the various attacks on federated learning protocols. Narayanan e al. (2011) describe how they were able to de-anonymize the obfuscated
connection graph from the 2011 Social Network Challenge by crawling the site where the data was obtained (Flickr), and matching nodes with unusually high in-degree or out-degree between the provided data and the crawled data. This allowed them to gain additional information to win the challenge, and it also allowed them to uncover the true identity of nodes
in the data. Tools for user privacy are becoming available; for example, TensorFlow provides
modules for federated learning and privacy (McMahan and Andrew, 2018). Fairness: Cathy O’Neil's book Weapons of Math Destruction (2017) describes how various black box machine learning models influence our lives, often in unfair ways. She calls
on model builders to take responsibility for fairness, and for policy makers to impose appropriate regulation. Dwork er al. (2012) showed the flaws with the simplistic “fairness through
unawareness” approach. Bellamy er al. (2018) present a toolkit for mitigating bias in machine learning systems. Tramer et al. (2016) show how an adversary can “steal” a machine learning model by making queries against an API, Hardt ef al. (2017) describe equal opportunity as
a metric for fairness. Chouldechova and Roth (2018) give an overview of the frontiers of fairness, and Verma and Rubin (2018) give an exhaustive survey of fairness definitions. Kleinberg et al. (2016) show that, in general, an algorithm cannot be both well-calibrated
and equal opportunity. Berk ef al. (2017) give some additional definitions of types of fairness, and again conclude that it is impossible to satisfy all aspects at once. Beutel et al. (2019) give advice for how to put fairness metrics into practice.
Dressel and Farid (2018) report on the COMPAS recidivism scoring model. Christin et al. (2015) and Eckhouse ef al. (2019) discuss the use of predictive algorithms in the le-
gal system. Corbett-Davies e al. (2017) show that that there is a tension between ensuring
fairness and optimizing public safety, and Corbett-Davies and Goel (2018) discuss the differences between fairness frameworks. Chouldechova (2017) advocates for fair impact: all classes should have the same expected utility. Liu er al. (2018a) advocate for a long-term
measure of impact, pointing out that, for example, if we change the decision point for ap-
proving a loan in order to be more fair in the short run, this could have negative effect in the
long run on people who end up defaulting on a loan and thus have their credit score reduced.
Since 2014 there has been an annual conference on Fairness, Accountability, and Trans-
parency in Machine Learning. Mehrabi et al. (2019) give a comprehensive survey of bias and fairness in machine learning, cataloging 23 kinds of bias and 10 definitions of fairness.
Trust:
Explainable Al was an important topic going back to the early days of expert
systems (Neches ef al., 1985), and has been making a resurgence in recent years (Biran and
1009
1010
Chapter 27 Philosophy, Ethics, and Safety of AT Cotton, 2017; Miller er al., 2017; Kim, 2018).
Barreno et al. (2010) give a taxonomy
of
the types of security attacks that can be made against a machine learning system, and Tygar
(2011) surveys adversarial machine learning. Researchers at IBM have a proposal for gaining trust in AT systems through declarations of conformity (Hind et al., 2018). DARPA requires
explainable decisions for its battlefield systems, and has issued a call for research in the area
(Gunning, 2016). Al safety: The book Arrificial Intelligence Safety and Security (Yampolskiy, 2018) collects essays on Al safety, both recent and classic, going back to Bill Joy’s Why the Future Doesn’t Need Us (Joy, 2000). The “King Midas problem” was anticipated by Marvin Minsky, who once suggested that an Al program designed to solve the Riemann Hypothesis might end up taking over all the resources of Earth to build more powerful supercomputers. Similarly, Omohundro (2008) foresees a chess program that hijacks resources, and Bostrom (2014) de-
scribes the runaway paper clip factory that takes over the world. Yudkowsky (2008) goes into
more detail about how to design a Friendly AL Amodei er al. (2016) present five practical safety problems for Al systems.
Omohundro (2008) describes the Basic Al Drives and concludes, “Social structures which
cause individuals to bear the cost of their negative externalities would go a long way toward
ensuring a stable and positive future.” Elinor Ostrom’s Governing the Commons (1990) de-
scribes practices for dealing with externalities by traditional cultures. Ostrom has also applied
this approach to the idea of knowledge as a commons (Hess and Ostrom, 2007). Ray Kurzweil (2005) proclaimed The Singularity is Near, and a decade later Murray Shanahan (2015) gave an update on the topic.
Microsoft cofounder Paul Allen countered
with The Singularity isn’t Near (2011). He didn’t dispute the possibility of ultraintelligent machines; he just thought it would take more than a century to get there. Rod Brooks is a
frequent critic of singularitarianism; he points out that technologies often take longer than
predicted to mature, that we are prone to magical thinking, and that exponentials don’t last forever (Brooks, 2017). On the other hand, for every optimistic singularitarian there is a pessimist who fears
new technology. The Web site pessimists.co shows that this has been true throughout
history: for example, in the 1890s people were concerned that the elevator would inevitably
cause nausea, that the telegraph would lead to loss of privacy and moral corruption, that the subway would release dangerous underground air and disturb the dead, and that the bicycle— especially the idea of a woman riding one—was the work of the devil. Hans
Moravec
(2000) introduces
some of the ideas of transhumanism,
and Bostrom
(2005) gives an updated history. Good’s ultraintelligent machine idea was foreseen a hun-
dred years earlier in Samuel Butler’s Darwin Among
the Machines (1863).
years after the publication of Charles Darwin’s On the Origins of Species
Written four
and at a time when
the most sophisticated machines were steam engines, Butler’s article envisioned “the ultimate development of mechanical consciousness” by natural selection. The theme was reiterated by
George Dyson (1998) in a book of the same title, and was referenced by Alan Turing, who
wrote in 1951 “At some stage therefore we should have to expect the machines to take control in the way that is mentioned in Samuel Butler’s Erewhon™ (Turing, 1996).
Robot rights: A book edited by Yorick Wilks (2010) gives different perspectives on how we should deal with artificial companions, ranging from Joanna Bryson’s view that robots should serve us as tools, not as citizens, to Sherry Turkle’s observation that we already per-
Bibliographical and Historical Notes sonify our computers and other tools, and are quite willing to blur the boundaries between
machines and life. Wilks also contributed a recent update on his views (Wilks, 2019). The philosopher David Gunkel’s book Robot Rights (2018) considers four possibilities: can robots have rights or not, and should they or not? The American Society for the Prevention of Cruelty to Robots (ASPCR) proclaims that “The ASPCR is, and will continue to be, exactly as
serious as robots are sentient.” The future of work:
In 1888, Edward Bellamy published the best-seller Looking Back-
ward, which predicted that by the year 2000, technological advances would led to a utopia where equality is achieved and people work short hours and retire early. Soon after, E. M. Forster took the dystopian view in The Machine Stops (1909), in which a benevolent machine
takes over the running of a society; things fall apart when the machine inevitably fails. Nor-
bert Wiener’s prescient book The Human Use of Human Beings (1950) argues for the benefits
of automation in freeing people from drudgery while offering more creative work, but also
discusses several dangers that we recognize as problems today, particularly the problem of value alignment. The book Disrupting Unemployment (Nordfors et al., 2018) discuss some of the ways that work is changing, opening opportunities for new careers. Erik Brynjolfsson and Andrew McAfee address these themes and more in their books Race Against the Machine (2011) and The Second Machine Age (2014). Ford (2015) describes the challenges of increasing automation, and West (2018) provides recommendations to mitigate the problems, while MIT’s
Thomas Malone (2004) shows that many of the same issues were apparent a decade earlier, but at that time were attributed to worldwide communication networks, not to automation.
1011
TR 28
THE FUTURE OF Al In which we try to see a short distance ahead.
In Chapter 2, we decided to view Al as the task of designing approximately rational agents. A variety of different agent designs were considered, ranging from reflex agents to knowledge-
based decision-theoretic agents to deep learning agents using reinforcement learning. There is also variety in the component technologies from which these designs are assembled: logical, probabilistic, or neural reasoning; atomic, factored, or structured representations of states; various learning algorithms from various types of data; sensors
and actuators to interact with
the world. Finally, we have seen a variety of applications, in medicine, finance, transportation,
communication,
and other fields.
There has been progress on all these fronts, both in our
scientific understanding and in our technological capabilities.
Most experts are optimistic about continued progress; as we saw on page 28, the median
estimate is for approximately human-level Al across a broad variety of tasks somewhere in
the next 50 to 100 years. Within the next decade, Al is predicted to add trillions of dollars to
the economy each year. But as we also saw, there are some critics who think general Al is centuries off, and there are numerous ethical concerns about the fairness, equity, and lethality of AL In this chapter, we ask: where are we headed and what remains to be done?
We do
that by asking whether we have the right components, architectures, and goals to make Al a successful technology that delivers benefits to the world.
28.1
Al Components
This section examines the components of Al systems and the extent to which each of them might accelerate or hinder future progress.
Sensors and actuators For much of the history of Al, direct access to the world has been glaringly absent. With a
few notable exceptions, Al systems were built in such a way that humans had to supply the inputs and interpret the outputs.
Meanwhile, robotic systems focused on low-level tasks in
which high-level reasoning and planning were largely ignored and the need for perception
was minimized. This was partly due to the great expense and engineering effort required to
get real robots to work at all, and partly because of the lack of sufficient processing power and sufficiently effective algorithms to handle high-bandwidth visual input.
The situation has changed rapidly in recent years with the availability of ready-made programmable robots. These, in turn, have benefited from compact reliable motor drives and
improved sensors. The cost of lidar for a self-driving car has fallen from $75,000 to $1,000,
Section 28.1 Al Components and a single-chip version may reach $10 per unit (Poulton and Watts, 2016). Radar sensors, once capable of only coarse-grained detection, are now sensitive enough to count the number
of sheets in a stack of paper (Yeo ef al., 2018). The demand for better image processing in cellphone cameras has given us inexpensive
high-resolution cameras for use in robotics. MEMS (micro-electromechanical systems) tech-
nology has supplied miniaturized accelerometers, gyroscopes, and actuators small enough to fit in artificial flying insects (Floreano et al., 2009; Fuller et al., 2014). It may be possible to combine millions of MEMS devices to produce powerful macroscopic actuators. 3-D printing (Muth ez al., 2014) and bioprinting (Kolesky ef al., 2014) have made it easier to experiment with prototypes.
Thus, we see that Al systems are at the cusp of moving from primarily software-only sys-
tems to useful embedded robotic systems. The state of robotics today is roughly comparable
to the state of personal computers in the early 1980s: at that time personal computers were
becoming available, but it would take another decade before they became commonplace. It is
likely that flexible, intelligent robots will first make strides in industry (where environments are more controlled,
tasks are more repetitive, and the value of an investment
is easier to
measure) before the home market (where there is more variability in environment and tasks). Representing the state of the world
Keeping track of the world requires perception as well as updating of internal representations.
Chapter 4 showed how to keep track of atomic state representations; Chapter 7 described
how to do it for factored (propositional) state representations; Chapter 10 extended this to
first-order logic; and Chapter 14 described probabilistic reasoning over time in uncertain environments.
Chapter 21 introduced recurrent neural networks, which are also capable of
maintaining a state representation over time.
Current filtering and perception algorithms can be combined to do a reasonable job of recognizing objects (“that’s a cat”) and reporting low-level predicates (“the cup is on the table”). Recognizing higher-level actions, such as “Dr. Russell is having a cup of tea with Dr. Norvig while discussing plans for next week,” is more difficult. Currently it can sometimes be done
(see Figure 25.17 on page 908) given enough training examples, but future progress will require techniques that generalize to novel situations without requiring exhaustive examples
(Poppe, 2010; Kang and Wildes, 2016). Another problem is that although the approximate filtering algorithms from Chapter 14 can handle quite large environments, they are still dealing with a factored representation—
they have random variables, but do not represent objects and relations explicitly. Also, their notion of time is restricted to step-by-step change; given the recent trajectory of a ball, we
can predict where it will be at time 7+ 1, but it is difficult to represent the abstract idea that
what goes up must come down. Section 15.1 explained how probability and first-order logic can be combined to solve
these problems; Section 15.2 showed how we can handle uncertainty about the identity of
objects; and Chapter 25 showed how recurrent neural networks enable computer vision to track the world; but we don’t yet have a good way of putting all these techniques together.
Chapter 24 showed how word embeddings and similar representations can free us from the
strict bounds of concepts defined by necessary and sufficient conditions. It remains a daunting
task to define general, reusable representation schemes for complex domains.
1013
1014
Chapter 28 The Future of AT Selecting actions
The primary difficulty in action selection in the real world is coping with long-term plans—
such as graduating from college in four years—that consist of billions of primitive steps.
Search algorithms that consider sequences of primitive actions scale only to tens or perhaps hundreds of steps. It is only by imposing hierarchical structure on behavior that we humans
cope at all. We saw in Section 11.4 how to use hierarchical representations to handle problems
of this
scale; furthermore, work in hierarchical reinforcement learning has succeeded in
combining these ideas with the MDP formalism described in Chapter 17. As yet, these methods have not been extended to the partially observable case (POMDPs).
Moreover, algorithms for solving POMDPs are typically using the same atomic state repre-
sentation we used for the search algorithms of Chapter 3. There is clearly a great deal of
work to do here, but the technical foundations are largely in place for making progress. The main missing element is an effective method for constructing the hierarchical representations
of state and behavior that are necessary for decision making over long time scales. Deciding what we want
Chapter 3 introduced search algorithms to find a goal state. But goal-based agents are brittle when the environment is uncertain, and when there are multiple factors to consider. In princi-
ple, utility-maximization agents address those issues in a completely general way. The fields of economics and game theory, as well as AI, make use of this insight: just declare what you want to optimize, and what each action does, and we can compute the optimal action.
In practice, however, we now realize that the task of picking the right utility function is a
challenging problem in its own right. Imagine, for example, the complex web of interacting preferences that must be understood by an agent operating as an office assistant for a human
being. The problem is exacerbated by the fact that each human is different, so an agent just “out of the box™ will not have enough experience with any one individual to learn an accurate
preference model; it will necessarily need to operate under preference uncertainty.
Further
complexity arises if we want to ensure that our agents are acting in a way that is fair and equitable for society, rather than just one individual.
We do not yet have much experience with building complex real-world preference mod-
els, let alone probability distributions over such models.
Although there are factored for-
malisms, similar to Bayes nets, that are intended to decompose preferences over complex states, it has proven difficult to use these formalisms in practice.
One reason may be that
preferences over states are really compiled from preferences over state histories, which are
described by reward functions (see Chapter 17). Even if the reward function is simple, the
corresponding utility function may be very complex.
This suggests that we take seriously the task of knowledge engineering for reward func-
tions as a way of conveying to our agents what we want them to do. The idea of inverse
reinforcement learning (Section 22.6) is one approach to this problem when we have an
expert who can perform a task, but not explain it. We could also use better languages for
expressing what we want. For example, in robotics, linear temporal logic makes it easier to say what things we want to happen in the near future, what things we want to avoid, and what
states we want to persist forever (Littman er al., 2017). We need better ways of saying what
we want and better ways for robots to interpret the information we provide.
Section 28.1 Al Components The computer industry as a whole has developed a powerful ecosystem for aggregating user preferences. When you click on something in an app, online game, social network, or shopping site, that serves as a recommendation that you (and your similar peers) would like to see similar things in the future. (Or it might be that the site is confusing and you clicked on the wrong thing—the data are always noisy.) The feedback inherent in this system makes
it very effective in the short run for picking out ever more addictive games and videos.
But these systems often fail to provide an easy way of opting out—your device will auto-
play a relevant video, but it is less likely to tell you “maybe it is time to put away your devices and take a relaxing walk in nature.” A shopping site will help you find clothes that match your style, but will not address world peace or ending hunger and poverty. To the extent that the menu of choices is driven by companies trying to profit from a customer’s attention, the menu
will remain incomplete.
However, companies do respond to customers’ interests, and many customers have voiced
the opinion that they are interested in a fair and sustainable world. Tim O’Reilly explains why profitis not the only motive with the following analogy: “Money is like gasoline during a road trip. You don’t want to run out of gas on your trip, but you’re not doing a tour of gas stations. You have to pay attention to money, but it shouldn’t be about the money.”
Tristan Harris’s time well spent movement at the Center for Humane Technology is a Time well spent step towards giving us more well-rounded choices (Harris, 2016). The movement addresses an issue that was recognized by Herbert Simon in 1971: “A wealth of information creates a poverty of attention.” Perhaps in the future we will have personal agents that stick up for Personal agent our true long-term interests rather than the interests of the corporations whose apps currently
fill our devices. It will be the agent’s job to mediate the offerings of various vendors, protect
us from addictive attention-grabbers, and guide us towards the goals that really matter to us.
Learning Chapters 19 to 22 described how agents can learn. Current algorithms can cope with quite large problems, reaching or exceeding human capabilities in many tasks—as long as we have sufficient training examples and we are dealing with a predefined vocabulary of features and concepts. But learning can stall when data are sparse, or unsupervised, or when we are dealing with complex representations.
Much of the recent resurgence of Al in the popular press and in industry is due to the
success of deep learning (Chapter 21). On the one hand, this can be seen as the incremental
maturation of the subfield of neural networks.
On the other hand, we can see it as a rev-
olutionary leap in capabilities spurred by a confluence of factors: the availability of more
training data thanks to the Internet, increased processing power from specialized hardware,
and a few algorithmic tricks, such as generative adversarial networks (GANs), batch normalization, dropout, and the rectified linear (ReLU) activation function.
The future should see continued emphasis on improving deep learning for the tasks it excels at, and also extending it to cover other tasks. The brand name “deep learning” has proven to be so popular that we should expect its use to continue, even if the mix of techniques. that fuel it changes considerably.
‘We have seen the emergence of the field of data science as the confluence of statistics,
programming, and domain expertise. While we can expect to see continued development in
the tools and techniques necessary to acquire, manage, and maintain big data, we will also
1015
1016
Chapter 28 The Future of AT need advances in transfer learning so that we can take advantage of data in one domain to
improve performance on a related domain. The vast majority of machine learning research today assumes a factored representation, learning a function & : R" — R for regression and i : R" — {0, 1} for classification. Machine
learning has been less successful for problems that have only a small amount of data, or problems that require the construction of new structured, hierarchical representations. Deep learning, especially with convolutional networks applied to computer vision problems, has demonstrated some success in going from low-level pixels to intermediate-level concepts like Eye and Mouth, then to Face, and finally to Person or Cat.
A challenge for the future is to more smoothly combine learning and prior knowledge.
If we give a computer a problem it has
not encountered before—say, recognizing different
models of cars—we don’t want the system to be powerless until it has been fed millions of
labeled examples.
The ideal system should be able to draw on what it already knows: it should already have
amodel of how vision works, and how the design and branding of products in general work; now it should use transfer learning to apply that to the new problem of car models. It should
be able to find on its own information about car models, drawing from text, images, and
video available on the Internet. It should be capable of apprenticeship learning: having a conversation with a teacher, and not just asking “may I have a thousand images of a Corolla,” but rather being able to understand advice like “the Insight is similar to the Prius, but the
Insight has a larger grille.” It should know that each model comes in a small range of possible
colors, but that a car can be repainted, so there is a chance that it might see a car in a color
that was not in the training set. (If it didn’t know that, it should be capable of learning it, or being told about it.)
All this requires a communication and representation language that humans and computers can share; we can’t expect a human analyst to directly modify a model with millions of
weights. Probabilistic models (including probabilistic programming languages) give humans
some ability to describe what we know, but these models are not yet well integrated with
Differentiable programming
other learning mechanisms. The work of Bengio and LeCun (2007) is one step towards this integration. Recently Yann LeCun has suggested that the term “deep learning” should be replaced with the more
general differentiable programming (Siskind and Pearlmutter, 2016; Li ef al., 2018); this
suggests that our general programming languages and our machine learning models could be merged together. Right now, it is common to build a deep learning model that is differentiable, and thus can
be trained to minimize loss, and retrained when circumstances change. But that deep learning
model is only one part of a larger software system that takes in data, massages the data, feeds
it to the model, and figures out what to do with the model’s output. All these other parts of the
larger system were written by hand by a programmer, and thus are nondifferentiable, which means that when circumstances change, it is up to the programmer to recognize any problems and fix them by hand. With differentiable programming, the hope is that the entire system is subject to automated optimization.
The end goal is to be able to express what we know in whatever form is convenient to us:
informal advice given in natural language, a strong mathematical law like F = ma, a statistical
model accompanied by data, or a probabilistic program with unknown parameters that can
Section 28.1 Al Components
1017
be automatically optimized through gradient descent. Our computer models will learn from conversations with human experts as well as by using all the available data. Yann LeCun, Geoffrey Hinton, and others have suggested that the current emphasis on supervised learning (and to a lesser extent reinforcement learning) is not sustainable—that computer models will have to rely on weakly supervised learning, in which some supervi-
sion is given with a small number of labeled examples and/or a small number of rewards, but
most of the learning is unsupervised, because unannotated data are so much more plentiful.
LeCun uses the term predictive learning for an unsupervised learning system that can Predictive learning
model the world and learn to predict aspects of future states of the world—not just predict
labels for inputs that are independent and identically distributed with respect to past data, and
not just predict a value function over states. He suggests that GANs (generative adversarial networks) can be used to learn to minimize the difference between predictions and reality.
Geoffrey Hinton stated in 2017 that “My view is throw it all away and start again,” mean-
ing that the overall idea of learning by adjusting parameters in a network is enduring, but the
specifics of the architecture of the networks and the technique of back-propagation need to be
rethought. Smolensky (1988) had a prescription for how to think about connectionist models; his thoughts remain relevant today. Resources
Machine learning research and development has been accelerated by the increasing availabil-
ity of data, storage, processing power, software, trained experts, and the investments needed to
support them. Since the 1970s, there has been a 100,000-fold speedup in general-purpose processors and an additional 1,000-fold speedup due to specialized machine learning hardware.
The Web has served as a rich source of images, videos, speech, text, and semi-structured data,
currently adding over 10'8 bytes every day.
Hundreds of high-quality data sets are available for a range of tasks in computer vision,
speech recognition, and natural language processing. If the data you need is not already
available, you can often assemble it from other sources, or engage humans to label data for
you through a crowdsourcing platform. Validating the data obtained in this way becomes an important part of the overall workflow (Hirth et al., 2013).
An important recent development is the shift from shared data to shared models.
The
major cloud service providers (e.g., Amazon, Microsoft, Google, Alibaba, IBM, Salesforce) have begun competing to offer machine learning APIs with pre-built models for specific tasks such as visual object recognition, speech recognition, and machine translation. These models
can be used as is, or can serve as a baseline to be customized with your particular data for
your particular application.
‘We expect that these models will improve over time, and that it will become unusual to
start a machine learning project from scratch, just as it is now unusual to do a Web develop-
ment project from scratch, with no libraries. It is possible that a big jump in model quality
will occur when it becomes economical to process all the video on the Web; for example, the YouTube platform alone adds 300 hours of video every minute.
Moore’s law has made it more cost effective to process data;
a megabyte of storage cost $1
million in 1969 and less then $0.02 in 2019, and supercomputer throughput has increased by a
factor of more than 10'” in that time. Specialized hardware components for machine learning
such as graphics processing units (GPUS), tensor cores, tensor processing units (TPUs), and
Shared model
1018
Chapter 28 The Future of AT field programmable gate arrays (FPGAs) are hundreds of times faster than conventional CPUs for machine learning training (Vasilache ef al., 2014; Jouppi et al., 2017). In 2014 it took a full day to train an ImageNet model; in 2018 it takes just 2 minutes (Ying ef al., 2018).
The OpenAl Institute reports that the amount of compute power used to train the largest machine learning models doubled every 3.5 months from 2012 to 2018, reaching over an exaflop/second-day for ALPHAZERO (although they also report that some very influential work used 100 million times less computing power (Amodei and Hernandez, 2018)). The
same economic trends that have made cell-phone cameras cheaper and better also apply to processors—we will see continued progress in low-power, high-performance computing that benefits from economies of scale.
There is a possibility that quantum computers could accelerate Al Currently there are
some fast quantum algorithms for the linear algebra operations used in machine learning (Harrow et al., 2009; Dervovic et al., 2018), but no quantum computer capable of running them. We have some example applications of tasks such as image classification (Mott et al., 2017) where quantum algorithms are as good as classical algorithms on small problems. Current quantum computers handle only a few tens of bits, whereas machine learning algorithms often handle inputs with millions of bits and create models with hundreds of mil-
lions of parameters. So we need breakthroughs in both quantum hardware and software to make quantum computing practical for large-scale machine learning. Alternatively, there may
be a division of labor—perhaps a quantum algorithm to efficiently search the space of hyperparameters while the normal training process runs on conventional computers—but we don’t know how to do that yet. Research on quantum algorithms can sometimes inspire new and better algorithms on classical computers (Tang, 2018). We have also seen exponential growth in the number of publications, people, and dollars
in Al/machine learning/data science. Dean er al. (2018) show that the number of papers about “machine learning” on arXiv doubled every two years from 2009 to 2017. Investors are
funding startup companies in these fields, large companies are hiring and spending as they
determine their Al strategy, and governments are investing to make sure their country doesn’t fall too far behind.
28.2
Al Architectures
It is natural to ask, “Which of the agent architectures in Chapter 2 should an agent use?” The answer is, “All of them!” Reflex responses are needed for situations in which time is of the
essence, whereas knowledge-based deliberation allows the agent to plan ahead. Learning is convenient when we have lots of data, and necessary when the environment is changing, or when human designers have insufficient knowledge of the domain.
AI has long had a split between symbolic systems (based on logical and probabilistic
inference) and connectionist systems (based on loss minimization over a large number of
uninterpreted parameters).
A continuing challenge for Al is to bring these two together,
to capture the best of both. Symbolic systems allow us to string together long chains of
reasoning and to take advantage of the expressive power of structured representations, while connectionist systems can recognize patterns even in the face of noisy data. One line of
research aims to combine probabilistic programming with deep learning, although as yet the various proposals are limited in the extent to which the approaches are truly merged.
Section 28.2
Al Architectures
1019
Agents also need ways to control their own deliberations. They must be able to use the
available time well, and cease deliberating when action is demanded. For example, a taxidriving agent that sees an accident ahead must decide in a split second whether to brake or
swerve. It should also spend that split second thinking about the most important questions,
such as whether the lanes to the left and right are clear and whether there is a large truck close
behind, rather than worrying about where to pick up the next passenger. These issues are usually studied under the heading of real-time AL As Al systems move into more complex Real-time Al domains, all problems will become real-time, because the agent will never have long enough to solve the decision problem exactly.
Clearly, there is a pressing need for general methods of controlling deliberation, rather
than specific recipes for what to think about in each situation.
The first useful idea is the
anytime algorithms (Dean and Boddy, 1988; Horvitz, 1987): an algorithm whose output Anytime algorithm quality improves gradually over time, so that it has a reasonable decision ready whenever it is
interrupted. Examples of anytime algorithms include iterative deepening in game-tree search and MCMC in Bayesian networks. The second technique for controlling deliberation is decision-theoretic metareasoning (Russell and Wefald, 1989; Horvitz and Breese, 1996; Hay ez al., 2012). This method, which
was mentioned briefly in Sections 3.6.5 and 5.7, (Chapter
Decision-theoretic metareasoning
applies the theory of information value
16) to the selection of individual computations (Section 3.6.5).
The value of a
computation depends on both its cost (in terms of delaying action) and its benefits (in terms
of improved decision quality).
Metareasoning techniques can be used to design better search algorithms and to guarantee
that the algorithms have the anytime property. Monte Carlo tree search is one example: the
choice of leaf node at which to begin the next playout is made by an approximately rational metalevel decision derived from bandit theory.
Metareasoning is more expensive than reflex action, of course, but compilation methods can be applied so that the overhead is small compared to the costs of the computations being controlled.
Metalevel reinforcement learning may provide another way to acquire effective
policies for controlling deliberation: in essence, computations that lead to better decisions are
reinforced, while those that turn out to have no effect are penalized. This approach avoids the
myopia problems of the simple value-of-information calculation.
Metareasoning is one specific example of a reflective architecture—that is, an architec-
ture that enables deliberation about the computational entities and actions occurring within
the architecture itself.
A theoretical foundation for reflective architectures can be built by
defining a joint state space composed from the environment state and the computational state
of the agent itself. Decision-making and learning algorithms can be designed that operate
over this joint state space and thereby serve to implement and improve the agent’s compu-
tational activities. Eventually, we expect task-specific algorithms such as alpha—beta search,
regression planning, and variable elimination to disappear from Al systems, to be replaced
by general methods that direct the agent’s computations toward the efficient generation of
high-quality decisions.
Metareasoning and reflection (and many other efficiency-related architectural and algo-
rithmic devices explored in this book) are necessary because making decisions
is hard. Ever
since computers were invented, their blinding speed has led people to overestimate their ability to overcome complexity, or, equivalently, to underestimate what complexity really means.
Reflective architecture
1020
Chapter 28 The Future of AT The truly gargantuan power of today’s machines tempts one to think that we could bypass all the clever devices and rely more on brute force. So let’s try to counteract this tendency.
We begin with what physicists believe to be the speed of the ultimate 1kg computing device: about 105! operations per second, or a billion trillion trillion times faster than the fastest supercomputer as of 2020 (Lloyd, 2000)." Then we propose a simple task: enumerating strings of English words, much as Borges proposed in The Library of Babel. Borges stipulated books of 410 pages.
Would that be feasible? Not quite.
In fact, the computer running for a year
could enumerate only the 11-word strings. Now consider the fact that a detailed plan for a human life consists of (very roughly) twenty trillion potential muscle actuations (Russell, 2019), and you begin to see the scale of the problem.
A computer that is a billion trillion trillion times more powerful than the
human brain is much further from being rational than a slug is from overtaking the starship Enterprise traveling at warp nine.
With these considerations in mind, it seems that the goal of building rational agents is
perhaps a little too ambitious.
Rather than aiming for something that cannot possibly exist,
we should consider a different normative target—one that necessarily exists.
Chapter 2 the following simple idea:
Recall from
agent = architecture + program .
Now fix the agent architecture (the underlying machine capabilities, perhaps with a fixed software layer on top) and allow the agent program to vary over all possible programs that the architecture can support. In any given task environment, one of these programs (or an equiv-
alence class of them) delivers the best possible performance—perhaps not close to perfect
Bounded optimality
rationality, but still better than any other agent program.
the criterion of bounded optimality.
We say that this program satisfies
Clearly it exists, and clearly it constitutes a desirable
goal. The trick is finding it, or something close to it. For some elementary classes of agent programs in simple real-time environments, it is possible to identify bounded-optimal agent programs (Etzioni, 1989; Russell and Subramanian, 1995). The success of Monte Carlo tree search has revived interest in metalevel decision
making, and there is reason to hope that bounded optimality within more complex families of
agent programs can be achieved by techniques such as metalevel reinforcement learning. It should also be possible to develop a constructive theory of architecture, beginning with theorems on the bounded optimality of suitable methods of combining different bounded-optimal
components such as reflex and action-value systems. General Al
Much of the progress in Al in the 21st century so far has been guided by competition on narrow tasks, such as the DARPA Grand Challenge for autonomous cars, the ImageNet object recognition competition, or playing Go, chess, poker, or Jeopardy! against a world champion. For each separate task, we build a separate Al system, usually with a separate machine learning model trained from scratch with data collected specifically for this task. But a truly
intelligent agent should be able to do more than one thing. Alan Turing (1950) proposed his list (page 982) and science fiction author Robert Heinlein (1973) countered with:
! We gloss over the fact that this device consumes the entire energy output of a star and operates at a billion degrees cei
Section 28.2
Al Architectures
A human being should be able to change a diaper, plan an invasion, butcher a hog, conn a ship, design a building, write a sonnet, balance accounts, build a wall, set a bone, comfort the dying, take orders, give orders, cooperate, act alone, solve equations, analyse a new problem, pitch manure, program a computer, cook a tasty meal, fight efficiently, die gallantly. Specialization is for insects.
So far, no Al system measures up to either of these lists, and some proponents of general or human-level Al (HLAI) insist that continued work on specific tasks (or on individual com-
ponents) will not be enough to reach mastery on a wide variety of tasks; that we will need a
fundamentally new approach. It seems to us that numerous new breakthroughs will indeed be necessary, but overall, Al as a field has made a reasonable exploration/exploitation tradeoff, assembling a portfolio of components, improving on particular tasks, while also exploring promising and sometimes far-out new ideas.
It would have been a mistake to tell the Wright brothers in 1903 to stop work on their
single-task airplane and design an “artificial general flight” machine that can take off vertically, fly faster than sound, carry hundreds of passengers, and land on the moon. It also would have been a mistake to follow up their first flight with an annual competition to make spruce wood biplanes incrementally better.
We have seen that work on components can spur new ideas; for example, generative adversarial networks (GANs) and transformer language models each opened up new areas of research.
We have also seen steps towards “diversity of behaviour.”
For example, machine
translation systems in the 1990s were built one at a time for each language pair (such as
French to English), but today a single system can identifying the input text as being one of a hundred languages, and translate it into any of 100 target languages. Another natural language system can perform five distinct tasks with one joint model (Hashimoto et al., 2016).
Al engineering
The field of computer programming started with a few extraordinary pioneers. But it didn’t reach the status of a major industry until a practice of software engineering was developed,
with a powerful collection of widely available tools, and a thriving ecosystem of teachers, students, practitioners, entrepreneurs, investors, and customers.
The Al industry has not yet reached that level of maturity. We do have a variety of pow-
erful tools and frameworks, such as TensorFlow, Keras, PyTorch, CAFFE, Scikit-Learn and
ScIPY. But many of the most promising approaches, such as GANs and deep reinforcement learning, have proven to be difficult to work with—they require experience and a degree of
fiddling to get them to train properly in a new domain. We don’t have enough experts to do this across all the domains where we need it, and we don’t yet have the tools and ecosystem to let less-expert practitioners succeed. Google’s Jeff Dean sees a future where we will want machine learning to handle millions
of tasks; it won’t be feasible to develop each of them from scratch, so he suggests that rather
than building each new system from scratch, we should start with a single huge system and, for each new task, extract from it the parts that are relevant to the task. We have seen some steps
in this direction, such as the transformer language models (e.g., BERT,
GPT-2) with
billions of parameters, and an “outrageously large” ensemble neural network architecture
that scales up to 68 billion parameters in one experiment (Shazeer et al., 2017). Much work remains to be done.
1021
1022
Chapter 28 The Future of AT The future ‘Which way will the future go? Science fiction authors seem to favor dystopian futures over utopian ones, probably because they make for more interesting plots. So far, Al seems to fit
in with other powerful revolutionary technologies
such as printing, plumbing, air travel, and
telephony. All these technologies have made positive impacts, but also have some unintended
side effects that disproportionately impact disadvantaged classes. We would do well to invest in minimizing the negative impacts.
Alis also different from previous revolutionary technologies. Improving printing, plumb-
ing, air travel, and telephony to their logical limits would not produce anything to threaten
human supremacy in the world. Improving Al to its logical limit certainly could.
In conclusion, Al has made great progress in its short history, but the final sentence of
Alan Turing’s (1950) essay on Computing Machinery and Intelligence is still valid today: We can see only a short distance ahead,
but we can see that much remains to be done.
ESEGTA MATHEMATICAL BACKGROUND A.1
Complexity Analysis and O() Notation
Computer scientists are often faced with the task of comparing algorithms to see how fast
they run or how much memory they require. There are two approaches to this task. The first is benchmarking—running the algorithms on a computer and measuring speed in seconds
and memory consumption in bytes. Ultimately, this is what really matters, but a benchmark can be unsatisfactory because it is so specific:
it measures the performance of a particular
Benchmarking
program written in a particular language, running on a particular computer, with a particular compiler and particular input data. From the single result that the benchmark provides, it can be difficult to predict how well the algorithm would do on a different compiler, computer, or of data set. The second approach relies on a mathematical analysis of algorithms, independent Analysis algorithms of the particular implementation and input, as discussed below.
A.1.1
Asymptotic analysis
We will consider algorithm analysis through the following example, a program to compute the sum of a sequence of numbers: function SUMMATION(sequence) returns a number sum 0 for i = 1 to LENGTH(sequence) do
sum & sum + sequenceli] return sum
The first step in the analysis is to abstract over the input, in order to find some parameter or parameters that characterize the size of the input. In this example, the input can be charac-
terized by the length of the sequence, which we will call n. The second step is to abstract over the implementation, to find some measure that reflects the running time of the algorithm
but is not tied to a particular compiler or computer. For the SUMMATION program, this could be just the number of lines of code executed, or it could be more detailed, measuring the
number of additions, assignments, array references, and branches executed by the algorithm.
Either way gives us a characterization of the total number of steps taken by the algorithm as a function of the size of the input. We will call this characterization T (n). If we count lines
of code, we have T (1) =2n+2 for our example.
If all programs were as simple as SUMMATION,
the analysis of algorithms would be a
trivial field. But two problems make it more complicated. First, it is rare to find a parameter
like n that completely characterizes the number of steps taken by an algorithm. Instead,
the best we can usually do is compute the worst case Tyors (1) O the average case Ty ().
Computing an average means that the analyst must assume some distribution of inputs.
1024
Appendix A
Mathematical Background
The second problem is that algorithms tend to resist exact analysis. In that case, it is
necessary to fall back on an approximation. We say that the SUMMATION algorithm is O(n),
meaning that its measure is at most a constant times n, with the possible exception of a few small values of n. More formally, Asymptotic analysis
T(n) is O(f(n)) if T(n) < kf(n) for some k, for all n > no.
The O() notation gives us what is called an asymptotic analysis. We can say without question that, as n asymptotically approaches infinity, an O(n) algorithm is better than an O(n?)
algorithm. A single benchmark figure could not substantiate such a claim.
The O() notation abstracts over constant factors, which makes it easier to use, but less
precise, than the T() notation. For example, an O(n2) algorithm will always be worse than an O(n) in the long run, but if the two algorithms are T'(n* + 1) and 7100+
O(n?) algorithm is actually better for n < 110. Despite this drawback, asymptotic anal;
1000), then the
is the most widely used tool for analyzing
algorithms. It is precisely because the analysis abstracts over both the exact number of operations (by ignoring the constant factor ) and the exact content of the input (by considering
only its size n) that the analysis becomes mathematically feasible. The O() notation is a good compromise between precision and ease of analysis. A.1.2
Complexity analysis
NP and inherently hard problems
The analysis of algorithms and the O() notation allow us to talk about the efficiency of a particular algorithm. However, they have nothing to say about whether there could be a better algorithm for the problem at hand. The field of complexity analysis analyzes problems rather than algorithms. The first gross division is between problems that can be solved in polynomial time and problems that cannot be solved in polynomial time, no matter what algorithm is used. The class of polynomial problems—those which can be solved in time O(n) for some k—is called P. These are sometimes called “easy” problems, because the class contains those
problems with running times like O(logn) and O(n). But it also contains those with time
NP
O(n'), 50 the name “easy™ should not be taken too literally.
Another important class of problems is NP, the class of nondeterministic polynomial
problems. A problem is in this class
if there is some algorithm that can guess
a solution and
then verify whether a guess is correct in polynomial time. The idea is that if you have an
arbitrarily large number of processors so that you can try all the guesses at once, or if you are
very lucky and always guess right the first time, then the NP problems become P problems.
One of the biggest open questions in computer science is whether the class NP is equivalent
to the class P when one does not have the luxury of an infinite number of processors or
omniscient guessing. Most computer scientists are convinced that P 7 NP; that NP problems
are inherently hard and have no polynomial-time algorithms. But this has never been proven.
NP-complete
Those who are interested in deciding whether P = NP look at a subclass of NP called the NP-complete problems. The word “complete” is used here in the sense of “most extreme” and thus refers to the hardest problems in the class NP. It has been proven that either all the NP-complete problems are in P or none of them is. This makes the class theoretically
interesting, but the class is also of practical interest because many important problems are known to be NP-complete. An example is the satisfiability problem: given a sentence of propositional logic, is there an assignment of truth values to the proposition symbols of the sentence that makes it true? Unless a miracle occurs and P = NP, there can be no algorithm
Section A.2
1025
Vectors, Matrices, and Linear Algebra
that solves all satisfiability problems in polynomial time. However, Al is more interested in whether there are algorithms that perform efficiently on typical problems drawn from a pre-
determined distribution; as we saw in Chapter 7, there are algorithms such as WALKSAT that
do quite well on many problems. The class of NP-hard problems consists of those problems that are reducible (in poly- NP-hard nomial time) to all the problems in NP, so if you solved any NP-hard problem, you could solve all the problems in NP. The NP-complete problems are all NP-hard, but there are some NP-hard problems that are even harder than NP-complete.
The class co-NP is the complement of NP, in the sense that, for every decision problem in NP, there is a corresponding problem in co-NP with the “yes” and “no” answers reversed.
‘We know that P is a subset of both NP and co-NP, and it is believed that there are problems in co-NP that are not in P. The co-NP-complete problems are the hardest problems in co-NP. The class #P (pronounced “number P™ according to Garey and Johnson (1979), but often
CoNP Co-NP-complete
pronounced “sharp P”) is the set of counting problems corresponding to the decision problems. in NP. Decision problems have a yes-or-no answer: is there a solution to this 3-SAT formula?
Counting problems have an integer answer: how many solutions are there to this 3-SAT formula? In some cases, the counting problem is much harder than the decision problem. For example, deciding whether a bipartite graph has a perfect matching can be done in time O(VE) (where the graph has V vertices and E edges), but the counting problem “how many perfect matches does this bipartite graph have” is #P-complete, meaning that it is hard as any problem in #P and thus at least as hard as any NP problem. Another class is the class of PSPACE problems—those that require a polynomial amount
of space, even on a nondeterministic machine. It is believed that PSPACE-hard problems are
worse than NP-complete problems, although it could turn out that NP = PSPACE, just as it could turn out that P = NP.
A.2
Vectors,
Matrices,
and
Linear
Algebra
Mathematicians define a vector as a member of a vector space, but we will use a more con-
crete definition: a vector is an ordered sequence of values. For example, in two-dimensional
space, we have vectors such as x= (3,4) and y = (0,2). We follow the convention of boldface
characters for vector names, although some authors use arrows or bars over the names: X or
. The elements of a vector can be accessed using subscripts: 2= (1,22,...,2,). One confusing point: this book is synthesizing work from many subfields, which variously call their sequences vectors, lists, or tuples, and variously use the notations (1,2), [1, 2], or (1, 2). The two fundamental operations on vectors are vector addition and scalar multiplication.
The vector addition x +y is the elementwise sum: x+y=(3 +0,4 +2)=(3,6). Scalar multiplication multiplies each element by a constant: 5x= (5 x 3,5 x 4) = (15,20). The length of a vector is denoted [x| and is computed by taking the square root of the sum of the squares of the elements:
[x|=+/(3%+4%)=5.
The dot product x-y (also called
scalar product) of two vectors is the sum of the products of corresponding elements, that is, X-y= ¥ X,yi, or in our particular case, X-y=3 x 0+4 x 2=8. Vectors are often interpreted as directed line segments (arrows) in an n-dimensional Eu-
clidean space. Vector addition is then equivalent to placing the tail of one vector at the head
of the other, and the dot product x-y is [x| |y| cosf, where is the angle between x and y.
Vector
1026
Appendix A
Mathematical Background
Matrix
A matrix is a rectangular array of values arranged into rows and columns. Here is a matrix A of size 3 x 4: A Axr Asr
Az Az Arg Azz Ags Agy Asz Asz Az
The first index of A
specifies the row and the second the column. In programming lan-
guages, A, is often written A[1, 3] or A[1] [].
The sum of two matrices is defined by adding their corresponding elements; for example (A+B);j=A;;+B;. (The sum is undefined if A and B have different sizes.) We can also define the multiplication of a matrix by a scalar: (cA);;=cA,;. Matrix multiplication (the product of two matrices) is more complicated. The product AB is defined only if A is of size
ax band B is of size b x ¢ (i.e., the second matrix has the same number of rows as the first
has columns); the result is a matrix of size a x . If the matrices are of appropriate size, then the result is
(AB)= YA, By )
Matrix multiplication is not commutative, even for square matrices: AB # BA in general. It
is, however, associative: (AB)C = A(BC). Note that the dot product can be expressed in
has the property that AL=A for all A. The transpose of A, written A is formed by turning rows into columns and vice versa, or, more formally, by A™; ;= A ;. The inverse ofa square matrix A is another square matrix A~' such that A~'A =1. For a singular matrix, the inverse
does not exist. For a nonsingular matrix, it can be computed in O(#3) time.
Matrices are used to solve systems of linear equations in O(n?) time; the time is domi-
nated by inverting a matrix of coefficients. Consider the following set of equations, for which we want a solution in x, y, and z:
+2x+y—z = 8 “3x—y+2z = —11 “2ty+2 = 3. We can represent this system as the matrix equation Ax = b, where 21 -1 x 8 =[-3-1 2], v, —11 2 1 2 z -3
To solve Ax = b we multiply both sides by A~', yielding A~'Ax = A~'b, which simplifies 2 3
I
R
tox=A""b. After inverting A and multiplying by b, we get the answer I
Singular
The identity matrix I has elements I ; equal to 1 when i= j and equal to 0 otherwise. It
=
Identity matrix Transpose Inverse
terms of a transpose and a matrix multiplication: x-y =x"y.
-1
A few more miscellaneous points: we use log(x) for the natural logarithm, log,(x). We use argmax, f(x) for the value ofx for which £(x) is maximal.
Section A.3
A.3
Probability Distribu
1027
Probability Distributions
ns
A probability is a measure over a set of events that satisfies three axioms:
1. The measure of each event is between 0 and 1. We write this as 0 < P(X=x;) < I, where X is a random variable representing an event and x; are the possible values of X. In general, random variables are denoted by uppercase letters and their values by lowercase letters. 2. The measure of the whole set is 1: that is, ¥/ P(X =x;) = 1. 3. The probability of a union of disjoint events is the sum of the probabilities of the individual events; that is, P(X =x; VX =x;) = P(X =x;) + P(X =x2), in the case where x| and x; are disjoint. A probabilistic model con:
of a sample space of mutually exclusive possible outcomes,
together with a probability measure for each outcome.
For example, in a model of the
weather tomorrow, the outcomes might be sun, cloud, rain, and snow.
A subset of these
outcomes constitutes an event. For example, the event of precipitation is the subset consist-
ing of {rain, snow}.
We use P(X) to denote the vector of values (P(X =x1),...,
P(X =x,)). We also use P(x;)
as an abbreviation for P(X =x;) and ¥, P(x) for T2 P(X =x;). The conditional probability P(B|A) is defined as P(BNA)/P(A). A and B are conditionally independent if P(B| A) = P(B) (or equivalently, P(A|B) = P(A)).
For continuous variables, there are an infinite number of values, and unless there are
point spikes, the probability of any one exact value is 0.
So it makes more sense to talk
about the value being within a range. We do that with a probability density function, which has a slightly different meaning from the discrete probability function. Since P(X =x)—the
Probability density function
probability that X has the value x exactly—is zero, we instead measure measure how likely it is that X falls into an interval around x, compared to the the width of the interval, and take
the limit as the interval width goes to zero:
P(x)= Jim0 P(x