Professional Documents
Culture Documents
Articles
W
e’ve all been hearing a lot about con- from a company database. These software
vergence between telephone, televi- agents are on their way, and they’re going to
sion, and computer technology. The be getting a lot of things accomplished by
basic idea is that the networks that constitute interacting with each other. The question is,
our telephone infrastructure, our television How will these agents be cooperating with
(particularly cable) infrastructure, and our each other, competing with each other, and
computer infrastructure will be coalescing into negotiating with each other?
one harmonious whole. Then the user, sitting
in his/her home or office, has some kind of Machines Controlling and
information appliance that can handle the
wealth of resources that are available.
Sharing Resources
Now, AI has a role to play in how the com- There are actually a lot of different environ-
puter works in this brave new world of infor- ments in which machines are making more
and more decisions that affect our daily lives, tion lines, switching nodes, or short- and
and they’re making these decisions not in long-term storage. In each situation, the com-
isolation but in concert with other machines. puters that control these resources can do
Computers that control electric grid networks their own job better by reaching agreements
and need to balance their power require- with other computers.
ments can share electricity with other com-
puters controlling other networks. If there’s a
drop or rise in power consumption, one com-
Heterogeneous,
puter can sell or buy excess power from other Self-Motivated Agents
utilities to which it’s connected. Now, the agents that we are interested in
Another example is routing among looking at are heterogeneous, self-motivated
telecommunication networks. Information, or agents. The systems are not assumed to be
packets, can pass over a network controlled by centrally designed. For example, if you have a
one company onto another network con- personal digital assistant, you might have one
trolled by another company, or it can pass that was built by IBM, but the next person
through one country on through another. over might have one that was built by Apple.
Computers that control a telecommunications They don’t necessarily have a notion of glob-
network might find it beneficial to enter into al utility. Each personal digital assistant or
agreements with other computers that control each agent operating from your machine is
other networks about routing packets more interested in what your idea of utility is and
efficiently from source to destination. The in how to further your notion of goodness.
point is that it might make sense for both They’re dynamic; for example, agents might
machines to be able to exploit the resources enter and leave the system in an unpre-
controlled by the other so that they can both dictable way. The system as a whole is flexi-
… the agents get their jobs done more effectively. ble; new personal assistants are coming in,
We’re also seeing the emergence of tools even new types of personal assistants are
that we such as personal digital assistants, small, hand- being built and coming in to the system and
are interested held computers (for example, the Newton). have to interact with other agents.
in looking These personal assistants are going to assume In particular, the personal assistants do not
the roles of a number of machines or tools act benevolently unless it’s in their interest to
at are involved with managing our daily lives, such do so. They do not necessarily share informa-
heterogeneous, as notebooks, communicators, fax machines, tion, they do not necessarily do things that
telephones, and automated schedulers. We’re other agents ask them to do unless they have
self-motivated going to have some kind of agent software on a good reason for doing so. Thus, imagine
agents. the personal assistant, and it’s ultimately lots of agents, each one residing in your per-
going to get part of its work done by interact- sonal computer or on your personal digital
ing with other agents on other personal assis- assistant, trying to carry out tasks, and inter-
tants. acting with other agents.
Another example is the proliferation of
shared databases, where there’s information
spread all over the world. They have sprung
The Aim of the Research
up with a vengeance in the last decade. You’ve The aim of the research that we’ve been
already got agents, such as Etzioni’s softbots, doing can really be thought of as a kind of
that are going out there to gather informa- social engineering for communities of
tion—for example, a person’s Internet machines. With communities of people,
address—from a shared database. Finally, even social engineering means setting up laws or
something such as traffic control, coordina- setting up an environment that causes people
tion of vehicular traffic, or air traffic control to act in a certain beneficial way or causes
illustrates situations where software agents people to act in certain ways that we decided
will be making decisions based on communi- ahead of time are desirable. At the least, it
cation and agreements with other agents. constrains their behavior.
Each of these situations is an instance Similarly, we’re interested in the creation of
where computers are controlling some interaction environments that foster a certain
resource and might be able to help them- kind of social behavior among machines. We
selves by strategically sharing this resource want to develop conventions, rules of
with other computers. With personal digital encounter, for these software agents that will
assistants, the resource might be a person’s cause them to act in certain ways. Another
time, whereas with a telecommunications way to look at this research is that we’re trying
network, the resource might be communica- to exploit formal tools, in this case, game the-
30 AI MAGAZINE
Articles
ory tools, for high-level protocol design. We’re we show that protocol A has a certain
looking at protocols for interactions among attribute, the designers of these personal digi-
agents, and we would like to design protocols tal assistants might decide to choose it, and
that cause, for example, these personal digital they might not; they might choose some-
assistant agents to act in certain ways. thing else. However, the idea of the research
is to come forward and elucidate and elabo-
rate on the possibilities. Protocol A has this
Broad Working Assumption set of attributes, and protocol B has this
These agents are obviously still pursuing their slightly different set of attributes. Which do
own utility, they’re still pursuing their own you want to choose when you design your
goals, but they’re going to be constrained agents? It’s up to you. Part of the research is
somehow by the environment that we to make it clear to those designers what the
design. However, who’s “we”? “We” means options are.
the designers of the network or the designers
of the system. Let’s say, as a working assump-
tion, that designers from IBM, Apple, Toshi-
Attributes of Standards
ba, and Sony all come together and say, “OK, What are we looking at when we try to
we’ve got this domain. The domain is person- design a standard? Well, we might look at the
al digital assistants doing scheduling, time efficiency of the system, such as Pareto opti-
scheduling. Given this domain, we would like mality; that is, the agreement that is reached
to set up the rules for the way in which by the agents can be made no better for any … what we’re
schedules will be set. We want to set up the one of the them without making it worse for trying to do is
rules, a high-level protocol, that determines one of the other agents. Thus, we have what
the kind of deals, let’s say, that agents can might be considered a broadly acceptable come forward,
make among themselves.” doctrine of general goodness. elucidate a
These designers then come together and Stability: An agent has no incentive to
try to agree on standards for how their auto- deviate from a particular strategy. This idea is
variety of
mated agents interact. An important part of related to the game theory notion of equilib- protocol
this process is that it takes place in a given rium.
domain. They might decide on different pro- Simplicity: This quality is important for
decisions,
tocols for different domains, but given the computer science but much less important and show
domain, like scheduling among personal digi- for game theory. We certainly want our pro- how
tal assistants, the designers are going to tocols to have low computation and commu-
decide on standards for how the agents reach nication costs. This attribute is one that we particular
agreements. might propose, and those designers might protocols
Now, what are the designers actually doing say, “This attribute is absolutely important;
in this meeting? It is a standards committee it’s important that our small processor not
have certain
meeting: They sit around a big table and dis- have a heavy load in doing negotiation.” desirable
cuss the trade-offs of different decisions. One We’d like the protocol to be distributed,
of them says, “You know, if we have this kind typically because if we’re going to have a lot
attributes
of protocol, the agents will be able to come to of distributed agents, it would be nice not to
agreement quickly. The agreements probably have a central decision maker. It would be
won’t be optimal, but we’ll be able to get to nice to have it be symmetric, so that no agent
them fast.” Somebody else at the table says, plays a special role. When the agents come
“Yes, but it’s important to us at IBM that together, they don’t have to decide who’s
these protocols not allow agents to manipu- going to be the master and who’s going to be
late one another. You know, we don’t want the slave or who’s going to do what.
any manipulative, exploitative agents getting The idea is to design protocols for specific
into the system and taking advantage.” Yet classes of domains that satisfy some or all of
another designer says, “Well, that’s not so these attributes. As we present protocols, or
important to me. The most important thing design decisions, it might be the case that it’s
is that the average utility among all the simple but not distributed or that it’s stable
agents be as high as possible.” but not efficient or that it’s efficient but not
We are not trying to impose a group of stable. These are classic trade-offs.
decisions that these company representatives
might make. Instead, what we’re trying to do
is come forward, elucidate a variety of proto-
Distributed AI
col decisions, and show how particular proto- How does this work relate to other research
cols have certain desirable attributes. Once being done in the field? It fits into the broad
FALL 1994 31
Articles
32 AI MAGAZINE
Articles
FALL 1994 33
Articles
34 AI MAGAZINE
Articles
FALL 1994 35
Articles
36 AI MAGAZINE
Articles
Deal and Utility in Figure 6. State-Oriented Domain and Worth-Oriented Domain Examples.
Two-Agent Task-Oriented Domain
First, we would like to have a definition of
the deal, the utility, and the conflict deal. as the cost of its original work minus the cost
Here we have it for a two-agent task-oriented of its new work given the deal. The difference
domain: A deal δ is a pair (D1, D2) such that is how much it has gained from the deal. It
D1 ∪ D2 = T1 ∪ T2. The conflict deal is defined used to have to walk five miles; now it only
as Θ [ (T1, T2). Utilityi(δ) = Cost(Ti) 2 Cost(Di). has to walk three miles. The utility is two.
A deal in a two-agent task-oriented domain
is a pair, D1, D2, such that their union is equal
Negotiation Protocols
to the union of the original task sets. Think As far as the protocol that the agents are
about those postmen with their letters. They going to use, there are lots of good choices.
come together with T1 and T2, their original For the purposes of this discussion, we’re
sacks of letters. A deal is a new distribution, going to assume that the agents are using
such that all the letters are taken care of. some kind of product-maximizing negotia-
The conflict deal in this case is simply the tion protocol, such as in Nash bargaining the-
original sets of letters. If you don’t reach agree- ory. For our purposes, it really doesn’t matter
ment, you deliver your original sack of letters. which one they use as long as it’s symmetric;
The utility of a deal for an agent is defined that is, it maximizes the product of the utili-
FALL 1994 37
Articles
1
e their own tasks are, there’s a well-defined
2 agreement point. This well-defined point tells
them when they have maximized the product
g c of their utilities. They can just move to this
point in one way or another. It doesn’t really
Agents will flip a coin to matter how,” which is true when the agents
f e d decide who delivers all have complete information.
the letters.
/ / Negotiation with
1 2 a Incomplete Information
What about the case where the agents don’t
Hiding Letters have complete information? What about that
situation where the postmen show up in the
morning, and the sacks of letters are opaque,
Post Office
and they have to decide how to negotiate.
1 Let’s assume that the first agent has to carry
f b letters to b and f, and the second agent has to
h a b / go to e. Each agent knows its own letters but
doesn’t know the other’s letters, so they can’t
(1) e compute the agreement for the two of them.
(hidden) 2 Instead, they need some other kind of mecha-
g c nism that will allow them to come to an
agreement. One simple, straightforward tech-
nique is to set up a 21 phase game, a sort of
f e d They then agree that pregame exchange of information (figure 7).
agent 2 delivers to f The two agents broadcast their tasks and
/ / and e. then continue as before, computing where
1 2 the agreement is going to be. Agent 2
b announces that it has to go to e, agent 1 says
“I have to go to b and f,” and they decide
who’s going to do what. Now, just to carry
Figure 7. Dealing with (and Exploiting) Incomplete Information.
out its own tasks, each one would have to go
a distance of eight units. Agent 1 would cer-
tainly go all the way around, and agent 2
ties. There are all sorts of examples of differ- might go half-way around and back, but it’s
ent protocols, different rules, that will bring equivalent to going all the way around. In
the agents to an agreement that maximizes this particular case, because of the structure
the product of their utilities. You can even of the problem, the agents eventually agree to
have a one-step protocol; if they know every- flip a coin, and one of them travels all the
thing, they can compute the agreement way around while the other one stays at the
point, which will be the point they jump to. post office. Assuming it’s a fair coin toss, they
Instead, you can have some kind of a mono- have divided up the work equally.
tonic concession protocol, where each agent first
starts with a deal that’s best for it and then Hiding Letters
iteratively makes compromises to the other However, our intrepid agent has been built by
agent. There are a lot of different product- a smart group of designers, and it makes the
maximizing protocols. following claim: Agent 2 honestly says, “I
Now we return to the last item on our have to go to e”; agent 1 says, “I have to go to
building blocks list. Given that we have a f,” and it hides its letter to b (see figure 7b).
task-oriented–domain specification, a defini- Now, the negotiation situation has changed
38 AI MAGAZINE
Articles
/
from this deal because it still travels eight to decide who goes to b
1, 2 and who goes to c.
units, but it isn’t harmed either. Thus, the
deal where agent 2 does all the work is the 1, 2
only rational, Pareto-optimal deal. In the
a
meantime, agent 1 runs off and delivers its
hidden letter to b at a cost of two units.
Agent 1 has really made off well with this Phantom Letter
manipulation, guaranteeing itself two units
of work instead of eight if it were alone or
Post Office
four units of expected work if it were honest
in its deal making.
b, c, d 1
Phantom Letters
a
Let’s look at another possibility for deception
(figure 8). Let’s say our agents both have to
b, c 2
deliver letters to nodes b and c. It’s an entire- c
ly symmetric situation. Obviously, it makes
sense for one agent to go to b and one to go
to c. However, b is relatively far away, and c is
b
/
1, 2 They agree that
relatively close; each agent would prefer to be
the one to go to c. If they tell the truth and
/ agent 1 goes to c.
declare their true tasks, they’ll flip a coin to
1, 2
/ d
decide which one goes to which node. 1 (phantom)
Agent 1 again decides to manipulate the b
agreement. It announces that it has a letter to
deliver to node d, which is a long way off in
Figure 8. Creating a Phantom Task.
the direction of node c. It only makes sense
for agent 1 to go to c and, presumably, con-
tinue to d. If agent 2 were given the right side
of this route, it would have to do more work for deception. In other words, you have a cer-
than when it’s alone, which wouldn’t be tain limited number of ways to divide up
acceptable to agent 2. They have to agree that tasks between agents. Depending on the par-
agent 2 goes to the left side and that agent 1 ticular encounter, an agent might be able to
goes to the right side. Of course, the letter to maneuver its way to a certain deal that’s bet-
d doesn’t exist; agent 1 just goes to c and ter for it, exploiting the fact that there are
comes back and benefits from its manipula- only certain ways to divide the tasks.
tion.
Part of what’s going on here is that the
Negotiation over Mixed Deals
form of a deal that we defined has con- One straightforward way of getting rid of
strained the kinds of agreement the agents some deception is to make the deal space
can come to. Remember, a deal is just a divi- continuous. We can redefine what a deal is,
sion of tasks, and we get certain discontinu- what a division of tasks is, to include proba-
ities when we define a deal this way. Howev- bility. A mixed deal is defined to be a division
er, it’s really that our deal space is discrete of tasks (D1, D2) with an associated probabili-
that gives rise to some of these possibilities ty p, so that with probability p, agent 1 does
FALL 1994 39
Articles
1 (phantom) Subadditive
b
Task-Oriented Domains
Figure 9. Mixed All-or-Nothing Deals Discouraging Deception. All the examples we have given to this point
have been illustrations of what are called sub-
additive task-oriented domains. A task-oriented
domain is subadditive if for all finite sets of
task set D1, and with probability 1 2 p, it does tasks, the cost of the union of tasks is less
task set D2 (and vice versa for agent 2). This than or equal to the sum of the costs of the
deal definition results in a continuous space separate sets (for finite X,Y in T, c(X ∪ Y) ¶
of deals. In addition, because of the way that c(X) + c(Y)). In other words, if you have two
our class of task-oriented domains is defined, sets of tasks, and you put them together,
if the agents use mixed deals, they can always you’ll never increase the cost, and perhaps
restrict their agreements to the so-called all- you’ll decrease it. You can see the general idea
or-nothing deal. In the all-or-nothing deal, all in figure 10. Putting the sets of tasks together
tasks are put into one big set, and a weighted lowers the overall cost of the combined set.
coin is tossed to decide which agent does all All our examples have been subadditive,
the tasks. In our examples, this kind of agree- but not all task-oriented domains are subad-
ment will always be a potential deal, but ditive. For example, consider a minor varia-
there can be others. By adding this probabili- tion on the postmen domain; we call it the
ty into the deal definition, we make the space delivery domain, where agents go out and
of deals continuous instead of discrete, the deliver their packages but are not required to
way it was originally. return to the post office at the end of the day.
40 AI MAGAZINE
Articles
FALL 1994 41
Articles
a
Concave Task-Oriented
h a b /
(1)
Domains
c
(hidden)
g c b
/ Concave task-oriented domains are a subset of
Sub-Additive /
1, 2 subadditive task-oriented domains that satisfy
f e d 1, 2
Hidden Phantom / d the following condition: Imagine that we
/ / 1 (phantom)
have two sets of tasks, X and Y, where X is a
1 2
Pure L L subset of Y, and we come along with some
A/N T T/P other set of tasks, Z. The cost Z adds to X, the
subset, is greater than or equal to the cost Z
Mix L T/P adds to Y, the superset: c(X ∪ Z) 2 c(X) >
c(Y ∪ Z) 2 c(Y). If a domain is concave, then
Theorem : For all encounters in all sub-additive TODs, when it is also subadditive (figure 12).
using a PMM over all-or-nothing deals, no agent has an Figure 12 is actually a bit misleading
incentive to hide a task. a because it appears like such an obvious prop-
erty that it ought to hold for all reasonable
task-oriented domains, but it doesn’t (see fig-
Decoy Tasks ure 12b).
Of the three task-oriented domains we
introduced originally, only the database
domain and the fax domain are concave. The
Decoy tasks, however, 1 / general postmen domain is not concave
can be beneficial even / 2
unless graphs are restricted to trees. To see an
with all-or-nothing deals 1 / 1
example of a nonconcave encounter in the
general postmen domain, consider the exam-
/ 2
ple we gave for a beneficial decoy task. Agent
Sub-Additive 1 has to travel around the left nodes; let’s call
/ / it X. Agent 2 has to travel around the right
Hidden Phantom Decoy nodes; let’s call Y the set of all dark-gray
1 1
Pure L L L nodes (that is, excluding the middle node
marked Z). Thus, X, the left nodes marked
A/N T T/P L with a 1, is a subset of Y, all the dark-gray
Mix L T/P L nodes. Agent 1 lies with a decoy task to the
node in the middle. This decoy task is set Z.
b The amount of work that Z adds to the set X
is 0. The agent visits Z on the way at no extra
cost. However, the amount of work that Z
Figure 11. Tables Summarizing Strategies, Given Protocols.
adds to Y, a superset of X, is 2. An agent
would have to make a special trip to visit all
Y, then visit Z; so, this example is not con-
when all-or-nothing protocols are used. Look cave.
at figure 11, where agent 1 would prefer to Three-Dimensional Incentive-Compati-
just deliver its own letters and return to the ble Mechanism Table: If we return to the
post office. By creating this decoy letter in the table that we’ve been setting up, we can
middle node, it claims that it would have to examine an entirely new set of possibilities
do a lot of extra work to carry out agent 2’s (figure 13). The table is now three dimension-
delivery. It deceptively locks itself into its al; the black arrows show that relationships
original path. Again, notice the white arrows. exist among the concave and the subadditive
Because lying can be beneficial using an all- dimensions of the table. For example, if hid-
or-nothing protocol, it must also be benefi- ing letters is sometimes beneficial in concave
cial sometimes when using mixed deals, domains when a pure deal protocol is used,
which are a superset. Thus, having filled out then it will also sometimes be beneficial in
one entry in the table, other entries can be subadditive domains because a concave
implied automatically. domain is also always a subadditive domain.
Subadditive task-oriented domains are an The main thing to notice here is that con-
important domain class, but we can use oth- cave domains are considerably better behaved
42 AI MAGAZINE
Articles
FALL 1994 43
Articles
44 AI MAGAZINE
Articles
FALL 1994 45
Articles
46 AI MAGAZINE