reaching agreements: voting

124
Reaching Reaching Agreements: Voting Agreements: Voting 7-1

Upload: chaeli

Post on 02-Feb-2016

34 views

Category:

Documents


0 download

DESCRIPTION

Reaching Agreements: Voting. Voting. Truthful voters vote for the candidate they think is best. Why would you vote for something you didn’t want? (run off election – want to pick competition) (more than two canddiates , figure your candidate doesn’t have a chance) - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Reaching Agreements: Voting

Reaching Agreements: Reaching Agreements: VotingVoting

7-1

Page 2: Reaching Agreements: Voting

2

Truthful voters vote for the candidate they think is best.Truthful voters vote for the candidate they think is best.Why would you vote for something you didn’t want? (run off Why would you vote for something you didn’t want? (run off

election – want to pick competition) (more than two election – want to pick competition) (more than two canddiates, figure your candidate doesn’t have a chance)canddiates, figure your candidate doesn’t have a chance)

We vote in awarding scholarships, teacher of the year, person We vote in awarding scholarships, teacher of the year, person to hire.to hire.

Rank feasible social outcomes based on agents' individual ranking of those outcomes

A - set of n agents O - set of m feasible outcomes Each agent i has a preference relation >i : O x O, asymmetric

and transitive

2

VotingVoting

Page 3: Reaching Agreements: Voting

3

Social choice rule (good for society) Input:Input: the agent preference relations (>1, …, >n)

Output:Output: elements of O sorted according the input - gives the social preference relation <* of the agent group

In other words – creates ordering for the group

3

Page 4: Reaching Agreements: Voting

4

Desirable properties of the social choice rule:

A social preference ordering >* should exist for all possible inputs (Note, I am using >* to mean “is preferred to.)

>* should be defined for every pair (o, o')O >* should be asymmetric and transitive over O The outcomes should be Pareto efficient:

if i A, o >i o' then o >* o‘ (not misorder if all agree) The scheme should be independent of irrelevant alternatives

(if all agree on relative ranking of two, should retain ranking in social choice):

No agent should be a dictator in the sense that

o >i o' implies o >* o' for all preferences of the other agents

4

Page 5: Reaching Agreements: Voting

5

Arrow's impossibility theoremArrow's impossibility theorem No social choice rule satisfies all of the six conditions Must relax desired attributes

May not require >* to always be defined We may not require that >* is asymmetic and transitiveUse plurality protocol: all votes are cast simultaneously and

highest vote count wins. Introducing an irrelevant alternative may split the majority

causing the old majority and the new irrelevant to drop out of favor (The Ross Perot effect).

A binary protocol involves voting pairwise – single eliminationThe order of the pairing can totally change the results

Page 6: Reaching Agreements: Voting

One voter ranks c > d > b > aOne voter ranks a > c > d > bOne voter ranks b > a > c > dNotice, just rotates preferences.

winner (c, (winner (a, winner(b,d)))=awinner (d, (winner (b, winner(c,a)))=d

winner (d, (winner (c, winner(a,b)))=c

winner (b, (winner (d, winner(c,a)))=b

surprisingly, order of pairing yields different winner!

Page 7: Reaching Agreements: Voting

7

Borda protocol (used if binary protocol is too slow) = assigns an alternative |O| points for the highest preference, |O|-1 points for the second, and so on

The counts are summed across the voters and the alternative with the highest count becomes the social choice

Winner turns loser and loser turns winner if the lowest ranked alternative is removed (does this surprise you?)

7

Page 8: Reaching Agreements: Voting

8

Borda Paradox – remove loser, winner changes(notice, c is always ahead of removed item)

a > b > c >d b > c > d >a c > d > a > b a > b > c > d b > c > d> a c >d > a >b a <b <c < da=18, b=19, c=20,

d=13

a > b > c b > c >a c > a > b a > b > c b > c > a c > a >b a <b <c

a=15,b=14, c=13

When loser is removed, next loser becomes winner!

Page 9: Reaching Agreements: Voting

9

Strategic (insincere) voters Suppose your choice will likely come in second place.

If you rank the first choice of rest of group very low, you may lower that choice enough so yours is first.

True story. Dean’s selection. Each committee member told they had 5 points to award and could spread out any way among the candidates. The recipient of the most points wins. I put all my points on one candidate. Most split their points. I swung the vote! What was my gamble?

Want to get the results as if truthful voting were done.

Page 10: Reaching Agreements: Voting

10

Typical Competition Mechanisms

Auction: allocate goods or tasks to agents through market. Need a richer technique for reaching agreements

Negotiation: reach agreements through interaction.

Argumentation: resolve confliction through debates.

Page 11: Reaching Agreements: Voting

Reaching Agreements: Reaching Agreements: VotingVoting

7-11

Page 12: Reaching Agreements: Voting

12

Negotiation

May involve:

Exchange of information

Relaxation of initial goals

Mutual concession

Page 13: Reaching Agreements: Voting

13

Mechanisms, Protocols, Strategies

Negotiation is governed by a mechanism or a

protocol:

defines the ”rules of encounter” between the agents

the public rules by which the agents will come to

agreements.

The deals that can be made

The sequence of offers and counter-offers that can be made

Given a particular protocol, how can a particular strategy be

designed that individual agents can use?

Page 14: Reaching Agreements: Voting

Negotiation is the process of reaching agreements on matters of common interest. It usually proceeds in a series of rounds, with every agent making a proposal at every round.

Negotiation Mechanism

Issues in negotiation process:• Negotiation Space: All possible deals that agents can make, i.e.,

the set of candidate deals. • Negotiation Protocol: – A rule that determines the process of a

negotiation: how and when a proposal can be made, when a deal has been struck, when the negotiation should be terminated, and so.

• Negotiation Strategy: When and what proposals should be made.

7-14

Page 15: Reaching Agreements: Voting

15

Protocol Means kinds of deals that can be made Means sequence of offers and counter-offers Protocol is like rules of chess game, whereas

strategy is way in which player decides which move to make

Do we even understand what is up for grabs? We may ask for a raise without considering bigger office, different appointment, 4-day work week, etc.

Page 16: Reaching Agreements: Voting

16

Negotiation Protocol Who begins Take turns Build off previous offers Give feed back (or not). Tell what utility is (or not) Obligations – requirements for later Privacy Allowed proposals you can make as a result of

negotiation history

Page 17: Reaching Agreements: Voting

17

Negotiation Process 1

Negotiation usually proceeds in a series of rounds,

with every agent making a proposal at every round.

Communication during negotiation:

Proposal

Counter Proposal

Agenti concedes

Agenti Agentj

Page 18: Reaching Agreements: Voting

18

Negotiation Process 2

Another way of looking at the negotiation

process is (can talk about 50/50 or 90/10

depending on who ”moves” the farthest):

Proposals by AjProposals by AiPoint of

Acceptance/aggreement

Page 19: Reaching Agreements: Voting

19

Jointly Improving Direction methodIterate over Mediator helps players criticize a tentative agreement

(could be status quo) Generates a compromise direction (where each of

the k issues is a direction in k-space) Mediator helps players to find a jointly preferred

outcome along the compromise direction, and then proposes a new tentative agreement.

Page 20: Reaching Agreements: Voting

20

Example: list of things to be done. Assigned to individuals already Who does what What is order to do What is paid for tasks

Page 21: Reaching Agreements: Voting

21

Goals of Negotiation (in many cases) Efficiency – not waste utility. Pareto Opt Stability – no agent have incentive to deviate

from agreed-upon strategy (as in one-shot negotiation).

Simplicity – low computational demands on agents

Distribution – interaction rules not require a central decision maker

Symmetry – (in some cases) may not want agents to play different roles.

Page 22: Reaching Agreements: Voting

22

Example: Planes need to be assigned landing time. Rule could be that airplanes with less fuel

land first. Any disadvantage?

Page 23: Reaching Agreements: Voting

23

Slotted Blocks world Like blocks world, only a fixed number of slots

on table. Forces need to coordinate Ex: Need to share car. Has side effects Ex: Schedule classes/professors – no side

effect.

Page 24: Reaching Agreements: Voting

Various Domains

Task Oriented Domain

State Oriented Domain

Worth Oriented Domain

Page 25: Reaching Agreements: Voting

25

Typical Negotiation ProblemsTask-Oriented Domains(TOD): an agent's activity can be defined in terms of a set of tasks that it has to achieve. The target of a negotiation is to minimize the cost of completing the tasks.

State Oriented Domains(SOD): each agent is concerned with moving the world from an initial state into one of a set of goal states. The target of a negotiation is to achieve a common goal. Main attribute: actions have side effects (positive/negative). TOD is a subset of SOD. Most classical AI domains are instances of SOD.

Main attribute of SOD – actions have side effects. Agents can unintentionally achieve one another’s goals. Negative interactions can also occur.

Worth Oriented Domains(WOD): agents assign a worth to each potential state, which captures its desirability for the agent. The target of a negotiation is to maximize mutual worth (rather than worth to individual). Superset of SOD.

Rates the acceptability of final states. Allows agents to compromise on their goals.

Page 26: Reaching Agreements: Voting

26

The simplest plan to achieve On(White,Gray) has the side effect of achieving Clear(black)

Page 27: Reaching Agreements: Voting

27

Single issue negotiation Like money Symmetric (If roles were reversed, I would benefit the

same way you would) If one task requires less travel, both would benefit equally

by having less travel utility for a task is experienced the same way by whomever

is assigned to that task. Non-symmetric – we would benefit differently if roles

were reversed if you delivered the picnic table, you could just throw it in the

back of your van. If I delivered it, I would have to rent a U-haul to transport it (as my car is small).

Page 28: Reaching Agreements: Voting

28

Multiple Issue negotiation Could be hundreds of issues (cost, delivery date,

size, quality) Some may be inter-related (as size goes down, cost

goes down, quality goes up?) Not clear what a true concession is (larger may be

cheaper, but harder to store or spoils before can be used)

May not even be clear what is up for negotiation (I didn’t realize not having any test was an option) (on the job…Ask for stock options, bigger office, work from home.)

Page 29: Reaching Agreements: Voting

29

How many agents are involved? One to one One to many (auction is an example of one

seller and many buyers) Many to many (could be divided into buyers

and sellers, or all could be identical in role) n(n-1)/2 number of pairs

Page 30: Reaching Agreements: Voting

30

Negotiation Domains:Task-oriented

”Domains in which an agent’s activity can be defined in terms

of a set of tasks that it has to achieve”, (Rosenschein & Zlotkin,

1994)

An agent can carry out the tasks without interference (or

help) from other agents – such as ”who will deliver the

mail”

All resources are available to the agent

Tasks redistributed for the benefit of all agents

Page 31: Reaching Agreements: Voting

31

Task-oriented Domain: Definition

How can an agent evaluate the utility of a specific deal?

Utility represents how much an agent has to gain from the deal. (it is always

based on change from original allocation)

Since an agent can achieve the goal on its own, it can compare the cost of

achieving the goal on its own to the cost of its part of the deal.

If utility<0, it is worse off than performing tasks on its own.

Conflict deal: (stay with status quo) if agents fail to reach an agreement:

where no agent agrees to execute tasks other than its own.

utlity = 0

Page 32: Reaching Agreements: Voting

32

Formalization of TOD

A Task Oriented Domain(TOD) is a triple <T, Ag, c>

where: T is a finite set of all possible tasks; Ag={A1, A2,…, An} is a list of participant agents; c:(T)R+ defines cost of executing each subset of tasks.

Assumptions on cost function:1. c() = 0.2. The cost of a subset of tasks does not depend on who carries out

them. (Idealized situation)3. Cost function is monotonic, which means that more tasks, more

cost. (It can’t cost less to take on more tasks.) i. T1 T2 implies c(T1) c(T2)

Page 33: Reaching Agreements: Voting

33

Redistribution of TasksGiven a TOD <T, {A1,A2}, c>, T is original assignment, output is

D: assignment after the “deal” An encounter (instance) within the TOD is an ordered list

(T1, T2) such that for all k, Tk T. This is an original allocation of tasks that they might want to reallocate.

A pure deal on an encounter is the redistribution of tasks among agents: (D1, D2), such that all tasks are reassigned

D1 D2= T1 T2

Specifically, : (D1, D2)=(T1, T2) is called the conflict deal. For each deal =(D1, D2), the cost of such a deal to agent k

is Costk()=c(Dk) (i.e, cost to k of deal is cost of Dk, k’s part of deal)

Page 34: Reaching Agreements: Voting

34

Examples of TOD

Parcel Delivery:

Several couriers have to deliver sets of parcels to different cities. The target of negotiation is to reallocate deliveries so that the cost of travel to each courier is minimal. Database Queries:

Several agents have access to a common database, and each has to carry out a set of queries. The target of negotiation is to arrange queries so as to maximize efficiency of database operations (Join, Projection, Union, Intersection, …) . You are doing a join as part of another operation, so please save the results for me.

Page 35: Reaching Agreements: Voting

35

Possible Deals

Consider an encounter from the Parcel Delivery Domain. Suppose we have two agents. Both agents have parcels to deliver to city a and only agent 2 has parcels to deliver to city b. There are nine distinct pure deals in this encounter:

1. ({a}, {b})

2. ({b}, {a})

3. ({a,b}, )

4. (, {a,b})

5. ({a}, {a,b})

6. ({b}, {a,b})

7. ({a,b}, {a})

8. ({a,b}, {b})

9. ({a,b}, {a,b})

the conflict deal

Page 36: Reaching Agreements: Voting

36

Figure deals knowing union must be {ab} Choices for first agent: {a} {b} {ab} {} Second agent must “pick up the slack” {a} for agent 1 b|ab (for agent 2) {b} for agent 1a|ab {ab} for agent 1 a|ab|b|{} {} for agent 1 ab

Page 37: Reaching Agreements: Voting

37

Utility Function for AgentsGiven an encounter (T1, T2), the utility function for each agent is just the difference of costs and is defined as follow:

Utilityk()=c(Tk)-Costk() = c(Tk)- c(Dk)

where =(D1, D2) is a deal;

c(Tk) is the stand-alone cost to agent k (the cost of achieving its original goal with no help)

Costk() is the cost of its part of the deal.

Note that the utility of the conflict deal is always 0.

Page 38: Reaching Agreements: Voting

38

Parcel Delivery Domain (assuming do not have to return home – like Uhaul)Distribution Point

city a city b

1 1

Cost function:c()=0c({a})=1c({b})=1c({a,b)}=3

Utility for agent 1 (org {a}):

1. Utility1({a}, {b}) = 0

2. Utility1({b}, {a}) = 0

3. Utility1({a, b}, ) = -2

4. Utility1(, {a, b}) = 1

Utility for agent 2 (org {ab}):

1. Utility2({a}, {b}) = 2

2. Utility2({b}, {a}) = 2

3. Utility2({a, b}, ) = 3

4. Utility2(, {a, b}) = 0

2

Page 39: Reaching Agreements: Voting

39

Dominant Deals Deal dominates deal ' if is better for at least one agent and not

worse for the other, i.e., is at least as good for every agent as ':

k{1,2}, Utilityk() Utilityk(') is better for some agent than ':

k{1,2}, Utilityk()> Utilityk(') Deal weakly dominates deal ' if at least the first condition holds

(deal isn’t worse for anyone).

Any reasonable agent would prefer (or go along with) over ' if

dominates or weakly dominates '.

Page 40: Reaching Agreements: Voting

40

Negotiation Set: Space of Negotiation

A deal is called individual rational if weakly dominates the conflict deal. (no worse than what you have already)

A deal is called Pareto optimal if there does not exist another deal ' that dominates . (best deal for x without disadvantaging y)

The set of all deals that are individual rational and Pareto optimal is called the negotiation set (NS).

Page 41: Reaching Agreements: Voting

41

Utility Function for Agents (example from previous slide)

1. Utility1({a}, {b}) =0

2. Utility1({b}, {a})=0

3. Utility1({a,b}, )=-2

4. Utility1(, {a,b})=1

5. Utility1({a}, {a,b})=0

6. Utility1({b}, {a,b})=0

7. Utility1({a,b}, {a})=-2

8. Utility1({a,b}, {b})=-2

9. Utility1({a,b}, {a,b})=-2

1.Utility2({a}, {b}) =2

2.Utility2 ({b}, {a})=2

3.Utility2 ({a,b}, )=3

4.Utility2 (, {a,b})=0

5.Utility2 ({a}, {a,b})=0

6.Utility2 ({b}, {a,b})=0

7.Utility2 ({a,b}, {a})=2

8.Utility2 ({a,b}, {b})=2

9.Utility2 ({a,b}, {a,b})=0

Page 42: Reaching Agreements: Voting

42

Individual Rational for Both(eliminate any choices that are negative for either)

1. ({a}, {b})

2. ({b}, {a})

3. ({a,b}, )

4. (, {a,b})

5. ({a}, {a,b})

6. ({b}, {a,b})

7. ({a,b}, {a})

8. ({a,b}, {b})

9. ({a,b}, {a,b})

individualrational

({a}, {b})

({b}, {a})

(, {a,b})

({a}, {a,b})

({b}, {a,b})

Page 43: Reaching Agreements: Voting

43

Pareto Optimal Deals

1. ({a}, {b})

2. ({b}, {a})

3. ({a,b}, )

4. (, {a,b})

5. ({a}, {a,b})

6. ({b}, {a,b})

7. ({a,b}, {a})

8. ({a,b}, {b})

9. ({a,b}, {a,b})

ParetoOptimal

({a}, {b})

({b}, {a})

({a,b}, )

(, {a,b})Beaten by ({a}{b}) deal

is (-2,3), but nothing beats 3 for agent 2

Page 44: Reaching Agreements: Voting

44

Negotiation Set

Negotiation Set

({a}, {b})

({b}, {a})

(, {a,b})

Individual Rational Deals

({a}, {b})

({b}, {a})

(, {a,b})

({a}, {a,b})

({b}, {a,b})

Pareto Optimal Deals

({a}, {b})

({b}, {a})

({a,b}, )

(, {a,b})

Page 45: Reaching Agreements: Voting

45

Negotiation Set illustrated

Create a scatter plot of the utility for i over the utility for j Only those where both is positive are individually rational (for

both) (origin is conflict deal) Which are pareto optimal?

Utility for i

Utility for j

Page 46: Reaching Agreements: Voting

46

Negotiation Set in Task-oriented Domains

AC

B

D

E

Utility for agent i

Utility for agent j

Utility of conflict Deal for agent i

Utility of conflict Deal for agent j

Conflict deal

The circle delimits the space of all possible deals

Negotiation set:

(pareto optimal+

Individual rational)

Page 47: Reaching Agreements: Voting

47

Negotiation Protocol () – Product of the two agent utilities from product maximizing negotiation protocol One step protocol

Concession protocol

At time t >= 0, A offers (A,t) and B offers (B,t), such that Both deals are from the negotiation set i andt >0, Utilityi((i,t)) <= Utilityi((i,t-1)) I propose something less desirable for me

Negotiation ending Conflict - Utilityi((i,t)) = Utilityi((i,t-1)) Agreement, j !=i Utilityj((i,t)) >= Utilityj((j,t))

Only A => agree (B,t) either agrees with proposal of other Only B => agree (A,t) either agrees with proposal of other Both A,B => agree (k,t) such that ((k))=max{((A)),((B))} Both A,B and ((A))=((B)) => flip a coin (product is the same, but may not

be the same for each agent – flip coin to decide which deal to use)

Pure deals

Mixeddeal

Page 48: Reaching Agreements: Voting

48

The Monotonic Concession Protocol – One direction, move towards middleRules of this protocol are as follows. . . Negotiation proceeds in rounds. On round 1, agents simultaneously propose a deal from the negotiation set

[individually rational, pareto optimal). Can re-propose same deal. Agreement is reached if one agent finds that the deal proposed by the other

is at least as good or better than its proposal. If no agreement is reached, then negotiation proceeds to another round of

simultaneous proposals. An agent is not allowed to offer the other agent less (in term of utility ) than

it did in the previous round. It can either stand still or make a concession. Assumes we know what the other agent values.

If neither agent makes a concession in some round, then negotiation terminates, with the conflict deal.

Meta data may be present: explanation or critique of deal.

Page 49: Reaching Agreements: Voting

49

Condition to Consent an Agreement If both of the agents finds that the deal proposed by the other is at least as good or better than the proposal it made.

Utility1(2) Utility1(1)and

Utility2(1) Utility2(2)

Page 50: Reaching Agreements: Voting

50

The Monotonic Concession Protocol

Advantages:

Symmetrically distributed (no agent plays a special role)

Ensures convergence

It will not go on indefinitely

Disadvantages:

Agents can run into conflicts

Inefficient – no quarantee that an agreement will be

reached quickly

Page 51: Reaching Agreements: Voting

51

Negotiation Strategy

Given the negotiation space and the Monotonic Concession Protocol, a strategy of negotiation is an answer to the following questions: What should an agent’s first proposal be? On any given round, who should concede? If an agent concedes, then how much should it concede?

Page 52: Reaching Agreements: Voting

52

The Zeuthen Strategy – a refinement of monotonic protocolQ: What should my first proposal be?

A: the best deal for you among all possible deals in the negotiation set. (Is a way of telling others what you value.)

Agent 1's best deal agent 2's best deal

Page 53: Reaching Agreements: Voting

53

Example of Zeuthan In interviewing for Women’s center

director, the candidate we were most interested in was approached.

She started by asking for $10K more money Job for husband Tenured full professor in academic department Gold parking pass for terrace

Page 54: Reaching Agreements: Voting

54

What was her strategy? Clearly Zeuthan Advantages: she had something to concede

and we knew what she valued Disadvantage: could be thought of as too

much so that the committee removes her from the pool.

Have had students make “initial request” that backfired as seemed totally off-base.

If you realize someone is using this strategy, you might NOT be offended.

Page 55: Reaching Agreements: Voting

55

The Zeuthen Strategy

Q: I make a proposal in every round, but may be the same as last time. Do I need to make a concession in this round?

A: If you are not willing to risk a conflict, you should make a concession.

How much am I willing to risk a

conflict?

Agent 1's best deal agent 2's best deal

How much am I willing to risk a

conflict?

Page 56: Reaching Agreements: Voting

56

Willingness to Risk Conflict

Suppose you have conceded a lot. Then:– You have lost some of your expected utility (closer to

zero).– In case conflict occurs, you are not much worse off.– An agent will be more willing to risk conflict if the

difference in utility between the loss in making an concession is greater than the loss in taking a conflict deal with respect to the current offer.

If both are equally willing to risk, both concede.

Page 57: Reaching Agreements: Voting

57

Risk Evaluation

riski= utility agent i loses by conceding and accepting agent j's offer

utility agent i loses (from current offer, not ideal) if conflict

You have to calculate?• How much you will lose if you make a concession and

accept your opponent's offer?• How much you will lose if you stand still which causes a

conflict?

=Utilityi (i )-Utilityi (j )

Utilityi (i )

where i and i are the current offer of agent i and j, respectively.

risk is willingness to risk conflict (1 is perfectly willing to risk)risk is willingness to risk conflict (1 is perfectly willing to risk)

Page 58: Reaching Agreements: Voting

58

Risk Ratio -

Page 59: Reaching Agreements: Voting

59

Risk Ratio

Page 60: Reaching Agreements: Voting

60

Risk Evaluation risk measures the fraction you have left to gain. If it is

close to one, you have gained little (and are more willing to risk).

This assumes you know what others utility is. Which is not always the case. If we have such perfect knowledge, we can compute the deal immediately. More likely, we base our decisions on what we THINK the other’s utility is.

You may not agree with their “concept” of utility – as they “claim” certain concessions or expectations.

In the car example, your initial price would be important. What one sets as initial goal affects risk. If I set an

impossible goal, my willingness to risk is always higher.

Page 61: Reaching Agreements: Voting

61

The Zeuthen Strategy

Q: If I concedes, then how much should I concede?

A: Enough to change the balance of risk (who has more to lose). (Otherwise, it will just be your turn to concede again at the next round) Not so much that you give up more than you needed to

Q: What if both have equal risk?

A: Both concede.

Page 62: Reaching Agreements: Voting

62

About MCP and Zeuthen Strategies

Advantages:

Simple and reflects the way human negotiations work.

Stability – in Nash equilibrium – if one agent is using the strategy,

then the other can do no better than using it him/herself.

Disadvantages:

Computationally expensive – players need to compute the entire

negotiation set.

Communication burden – negotiation process may involve several

steps.

Page 63: Reaching Agreements: Voting

63

A one-shot Negotiation Protocol(like dividing a candy bar) Protocol: both agents suggest an agreement; the one giving ahigher product of utilities wins (flip a coin in case of a tie)

Page 64: Reaching Agreements: Voting

64

A one-shot Negotiation Protocol(like dividing a candy bar) Protocol: both agents suggest an agreement; the one giving ahigher product of utilities wins (flip a coin in case of a tie) Obvious strategy: amongst the set of agreements with maximalproduct of utilities, propose the one that is best for you Properties: This mechanism is: – efficient: outcomes have maximal Nash product and are Pareto

optimal (like MCP with Zeuthen Strategy) – stable: no agent has an incentive to deviate from the strategy (like

MCP with extended Zeuthen Strategy) In addition, the one-shot protocol is also: – simple: only one round is required But why should anyone accept to use such a protocol? (There is no

motivation to be less than fair.)

Page 65: Reaching Agreements: Voting

65

Recap: How did we get to this point? Both agents making several small concessions until anagreement is reached is the most intuitive approach toone-to-one negotiation.

The Monotonic Concession Protocol (MCP) is a straightforward formalization of the above intuition. Both propose at every round.

The Zeuthen Strategy is also motivated by intuition (“willingness to risk conflict”) and constitutes a stable and (almost) efficient strategy for the MCP.

The one-shot protocol (together with the obvious strategy)produces similar outcomes as MCP/Zeuthen, but it is a muchsimpler mechanism.

Page 66: Reaching Agreements: Voting

66

Parcel Delivery Domain: Example 2 (don’t return to dist point)

Distribution Point

a d

7 7

Cost function:c()=0c({a})=c({d})=7c({b})=c({c})=c({a,b})=c({c,d})=8c({b,c})=c({a,b,c})=c({b,c,d})=9c({a,d})=c({a,b,d})=c({a,c,d})=c({a,b,c,d})=10

b c1 1 1

Negotiation Set: ({a,b,c,d}, ) ({a,b,c), {d}) ({a,b}, {c,d}) ({a}, {b,c,d}) (, {a,b,c,d})

Conflict Deal: ({a,b,c,d}, {a,b,c,d})

All choices are IR, as can’t do worse ({ac}{bd}) is dominated by ({a}{bcd})

Page 67: Reaching Agreements: Voting

67

Parcel Delivery Domain: Example 2 (Zeuthen works here both concede on equal risk)

No Pure Deal Agent 1's Utility Agent 2's Utility

5 ({a,b,c,d}, ) 0 10

4 ({a,b,c), {d}) 1 3

3 ({a,b}, {c,d}) 2 2

2 ({a}, {b,c,d}) 3 1

1 (, {a,b,c,d}) 10 0

Conflict deal 0 0

agent 1 agent 25 4 3 2 1

Page 68: Reaching Agreements: Voting

68

What bothers you about the previous agreement?

Decide to both get (2,2) utility, rather than the expected utility of (0,10) for another choice.

Is there a solution? Fair versus higher global utility. Restrictions of this method (no promises for

future or sharing of utility)

Page 69: Reaching Agreements: Voting

69

State Oriented Domain Goals are acceptable final states (superset of TOD) Have side effects - agent doing one action might hinder or help another agent.

Example in blocks world, on(white,gray) has side effect of clear(black). Negotiation : develop joint plans (what they each do) and schedules for the agents,

to help and not hinder other agents Example – Slotted blocks world -blocks cannot go anywhere on table – only in

slots (restricted resource) Note how this simple change (slots) makes it so two workers get in each other’s

way even if goals are unrelated.

Page 70: Reaching Agreements: Voting

70

Assumptions of SOD Agents will maximize expected utility (will prefer 51%

chance of getting $100 than a sure $50) Agent cannot commit himself (as part of current

negotiation) to behavior in future negotiation. No explicit utility transfer (no “money” that can be used to

compensate one agent for a disadvantageous agreement) Interagent comparison of utility: common utility units Symmetric abilities (all can perform tasks, and cost is

same regardless of agent performing) Binding commitments

Page 71: Reaching Agreements: Voting

71

Achievement of Final State Goal of each agent is represented as a set of states that they

would be happy with. Looking for a state in intersection of goals Possibilities:

(GREAT) Both can be achieved, at gain to both (e.g. travel to same location and split cost)

(IMPOSSIBLE) Goals may contradict, so no mutually acceptable state (e.g., both need a car)

(NEED ALT) Can find common state, but perhaps it cannot be reached with the primitive operations in the domain (could both travel together, but may need to know how to pickup another)

(NOT WORTH IT) Might be a reachable state which satisfies both, but may be too expensive – unwilling to expend effort (i.e., we could save a bit if we car-pooled, but is too complicated for so little gain).

Page 72: Reaching Agreements: Voting

72

Examples: CooperativeEach is helped by joint plan

Slotted blocks world: initially white block is at 1 and black block at 2. Agent 1 wants black in 1. Agent 2 wants white in 2. (Both goals are compatible.)

Assume pick up is cost 1 and set down is one. Mutually beneficial – each can pick up at the same

time, costing each 2 – Win – as didn’t have to move other block out of the way!

If done by one, cost would be four – so utility to each is 2.

Page 73: Reaching Agreements: Voting

73

Examples: CompromiseBoth succeed, but worse for both than if other agent gone

Slotted blocks world: initially white block is at 1 and black block at 2, two gray blocks at 3. Agent 1 wants black in 1, but not on table. Agent 2 wants white in 2, but not directly on table.

Alone, agent 1 could just pick up black and place on white. Similarly, for agent 2. But would undo others goal.

But together, all blocks must be picked up and put down. Best plan: one agent picks up black, while other agent rearranges (cost 6 for one, 2 for other)

Can both be happy, but unequal roles.

Page 74: Reaching Agreements: Voting

74

Example: conflict I want black on white (in slot 1) You want white on black (in slot 1) Can’t both win. Could flip a coin to decide who wins.

Better than both losing. Weightings on coin needn’t be 50-50.

May make sense to have person with highest worth get his way – as utility is greater. (Would accomplish his goal alone) Efficient but not fair?

What if we could transfer half of the gained utility to the other agent? This is not normally allowed, but could work out well.

Page 75: Reaching Agreements: Voting

75

Negotiation Domains: Worth-oriented

”Domains where agents assign a worth to each potential

state (of the environment), which captures its desirability for

the agent”, (Rosenschein & Zlotkin, 1994)

agent’s goal is to bring about the state of the environment with

highest value

we assume that the collection of agents have available a set of

joint plans – a joint plan is executed by several different agents

Note – not ”all or nothing” – but how close you got to goal.

Page 76: Reaching Agreements: Voting

76

Worth Oriented Domain

Rates the acceptability of final states Allows partially completed goals Negotiation : a joint plan, schedules, and goal relaxation. May

reach a state that might be a little worse that the ultimate objective

Example – Multi-agent Tile world (like airport shuttle) – isn’t just a specific state, but the value of work accomplished

Page 77: Reaching Agreements: Voting

77

How can we calculate Utility? Weighting each attribute

Utility = {Price*60 + quality*15 + support*25} Rating/ranking each attribute

Price : 1, quality 2, support 3 Using constraints on an attribute

Price[5,100], quality[0-10], support[1-5] Try to find the pareto optimum

Page 78: Reaching Agreements: Voting

78

What if choices don’t benefit others fairly? Suppose there are two states that satisfy both

agents. State 1: one has a utility of 6 for one agent

and 3 for the other. State 2: utility of both agents 4. State 1 is better (overall), but state 2 is more

equal. How can we get cooperation (as why should one agent agree to do more)?

Page 79: Reaching Agreements: Voting

79

Mixed Deal If = (J1, J2:p) is a deal, then

utili() = p*util(J)i + (1-p)*util(J)k where k is i’s opponent -the role i plays with (1-p) probability

An “all or nothing” form of a mixed deal simply means one set of tasks is everything.

Page 80: Reaching Agreements: Voting

80

Parcel Delivery Domain (assuming noreturn)

Distribution Point

city a city b

1 1

Cost function:c()=0c({a})=1c({b})=1c({a,b)}=3

Utility for agent 1 (org {a}):

1. Utility1({a}, {b}) = 0

2. Utility1({b}, {a}) = 0

3. Utility1({a, b}, ) = -2

4. Utility1(, {a, b}) = 1

Utility for agent 2 (org {ab}):

1. Utility2({a}, {b}) = 2

2. Utility2({b}, {a}) = 2

3. Utility2({a, b}, ) = 3

4. Utility2(, {a, b}) = 0

2

Page 81: Reaching Agreements: Voting

81

At seats… For the parcel delivery example above, show

what a mixed deal does for the following deals:

1. deal “1”, each does one 2. deal “3”, all or nothing (notice a “pure”

deal3 is not even individually rational)

Page 82: Reaching Agreements: Voting

82

Consider deal 3 with probability ({},{ab}):p means agent 1 does {} with p probabilty

and {ab} with (1-p) probabilty. What should p be to be fair to both (equal utility) (1-p)(-2) + p1 = utility for agent 1 (1-p)(3) + p0 = utility for agent 2 (1-p)(-2) + p1= (1-p)(3) + p0 -2+2p+p = 3-3p => p=5/6 If agent 1 does no deliveries 5/6 of the time, it is fair.

Page 83: Reaching Agreements: Voting

83

Try again with other choice in negotiation (Deal 1) Utility1({a}, {b}) = 0

({a},{b}):p means agent 1 does {a} with p probabilty and {b} with (1-p) probabilty.

What should p be to be fair to both (equal utility) (1-p)(0) + p0 = utility for agent 1 (1-p)(2) + p2 = utility for agent 2 0=2 no solution Can you see why we can’t use a p to make this

fair?

Page 84: Reaching Agreements: Voting

84

Incomplete Information Don’t know tasks of others in TOD. Solution

Exchange missing information Penalty for lie

Possible lies (notice: reduce possibilities in order to be able to solve)

False information Hiding letters (don’t admit part of your job) Lie about letters (claim work that isn’t required)

decoy – produce if needed phantom – can’t produce, caught in lie

Not carry out a commitment

Page 85: Reaching Agreements: Voting

85

Subadditive Task Oriented DomainCost of whole is ≤ cost of parts for finite X,Y in T, c(X U Y) <= c(X) + c(Y)). Example of subadditive:

Deliver to one, saves distance to other (in a tree arrangement) Example of subadditive TOD (= rather than <)

deliver in opposite directions

–doing both saves nothing

Not subadditive: doing both actually costs more than the sum of the pieces. Say electrical power costs, where I get above a threshold and have to buy new equipment.

Page 86: Reaching Agreements: Voting

86

Decoy task We call producible phantom tasks decoy tasks (no

risk of being discovered). Only unproducible phantom tasks are called phantom tasks.

Example: Need to pick something up at store. (Can think of something for them to pick up, but if you are the one assigned, you won’t bother to make the trip.)

Need to deliver empty letter (no good, but deliverer won’t discover lie)

Page 87: Reaching Agreements: Voting

87

Incentive compatible MechanismAre the rules (in terms of allowable deals) we establish sufficient to produce truth telling? L there exists a

beneficial lie in some encounter

T There exists no beneficial lie.

T/P Truth is dominant if the penalty for lying is stiff enough.

Example indicates a case where lying helps. Can you see it? Who lies? What is lie?

Page 88: Reaching Agreements: Voting

88

Explanation of arrow If it is never beneficial in a mixed deal

encounter to use a phantom lie (with penalties), then it is certainly never beneficial to do so in an all-or-nothing mixed deal encounter (which is just a subset of the mixed deal encounters).

Page 89: Reaching Agreements: Voting

89

Concave Task Oriented Domain We have 2 tasks X and Y, where X is a subset of Y Another set of task Z is introduced

c(YU Z) –c(Y) ≤c(XU Z) –c(X)

Page 90: Reaching Agreements: Voting

Concave:

•c(YU Z) –c(Y) ≤c(XU Z) –c(X)•The cost of tasks Z adds to set of tasks Y cannot be greater than the cost Z add to a subset of Y •Expect it to add more to subset (as is smaller)

•At seats – is postmen doman concave (no, unless restricted to trees)

Example: Y is in pacman shape, X is nodes in polygon.adding Z adds 0 to X (as was going that way anyway) but adds 2 to its superset Y (as was going around loop)

• Concavity implies sub-additivity•Modularity implies concavity

y

Modularity: c(X U Y) = c(X) + c(Y) - c(X Y).Notice modular encourages truth telling, more than others

Page 91: Reaching Agreements: Voting

91

Explanation of Previous Chart Arrows show reasons we know this fact (diagonal arrows

are between domains). For example, What is true of a phantom task, may be true for a decoy task in same domain as a phantom is just a decoy task we don’t have to create.

Similarly, what is true for a mixed deal may be true for an all or nothing deal (in the same domain) as a mixed deal is an all or nothing deal where one choice is empty. The direction of the relationship may depend on truth (never helps) or lie (sometimes helps).

The relationships can also go between domains as sub-additive is a superclass of concave and a super class of modular.

Page 92: Reaching Agreements: Voting

92

Modular TOD c(X U Y) = c(X) + c(Y) - c(X Y). Notice modular encourages truth telling, more than others

Page 93: Reaching Agreements: Voting

93

Implied relationship between cells Implied relationship between domains (slanted arrows).L means lying may be beneficialT means telling the truth is always beneficialT/P Truth telling is beneficial if penalty for being caught is great

Page 94: Reaching Agreements: Voting

94

Attributes-Modularity

c(XU Y) = c(X) + c(Y) –c(X∩Y) •The cost of the combination of 2 sets of tasks

is exactly the sum of their individual costs minus the cost of their intersection

Only Fax Domain is modular (as costs are independent)

Modularity implies concavity

Page 95: Reaching Agreements: Voting

Incentive Compatible Facts (return home)Fact1: in SubadditiveTOD, any Optimal

Negotiation Mechanism (ONM) over A-or-N deals, “hiding” lies are not beneficial

Ex:A1hides letter to c, his utility doesn’t increase.

If he tells truth : p=1/2 Expected util ({abc}{})1/2 = 5 Lie: p=1/2 (as utility is same) Expected util (for 1) ({abc}{})1/2 = ½(0) +

½(2) = 1 (as has to deliver the lie)

1

44

1

Page 96: Reaching Agreements: Voting

96

Fact2 in SubadditiveTOD, any ONM over Mixed deals, every “phantom” lie has a positive probability of being discovered. (as if other person delivers phantom, you are found out)

Fact3 in Concave TOD, any ONM over Mixed deals, no “decoy” lie is beneficial. (as less increased cost is assumed so probabilities would be assigned to reflect the assumed extra work)

Fact4 in Modular TOD, any ONM over Pure deals, no “decoy” lie is beneficial. (modular tends to add exact cost – hard to win)

Page 97: Reaching Agreements: Voting

Fact4 Modular, all or nothing, decoy

Both deliver to e and b. Suppose agent 2 lies about having a delivery to c.

Under Lie – benefits are shown

Under Truth, p would be 1/2

If we assign p

({ebc}, ) p

agent 1 utility -2*p + 6(1-p)

Agent 2 (under lie) 8p+0(1-p)

-2*p + 6(1-p)= 8p+0(1-p)

-8p+6 = 8p p=6/16 (so 2 is worse off)

1 U(1) 2 U(2)

Seems

U(2)

(act)

eb 0 ebc 0 0

ebc -2 8 6

6 ebc 0 0

Page 98: Reaching Agreements: Voting

98

Fact5: in Concave TOD, any ONM over Pure deals, “Phantom” lies can be beneficial.

Example from next slide:A1creates Phantom letter at node c, his utility has risen from 3 to 4

Truth: p = ½ so utility for agent 1 is ({a}{b}) ½ = ½(4) + ½(2) = 3

Lie: ({bc}{a}) is logical division as no percent Util for agent 1 is 6 (org cost) – 2(deal cost) = 4

Page 99: Reaching Agreements: Voting

99

Fact6: in SubadditiveTOD, any ONM over A-or-N deals, “Decoy” lies can be beneficial (not harmful). (as it changes the probability. If you deliver, I make you deliver to h)

Ex2 (from next slide):A1lies with decoy letter to h (trying to make agent 2 think picking up bc is worse for agent 1 than it is), his utility has rised from 1.5 to 1.72. (If I deliver, I don’t deliver h)

If tells truth, p (of agent 1 delivering all) = 9/14 as p(-1) + (1-p)6 = p(4) + (1-p)(-3) 14p=9 If invents task h, p=11/18 as p(-3) + (1-p)6 = p(4) + (1-p)(-5) Utility(p=9/14) is p(-1) + (1-p)6 = -9/14 +30/14 = 21/14 = 1.5 Utility(p=11/18) is p(-1) + (1-p)6 = -11/18 +42/18 = 31/18 = 1.72 SO – lying helped!

Page 100: Reaching Agreements: Voting

100

Postmen – return to postoffice

Concave

Subadditive(h is decoy)

Phantom

Page 101: Reaching Agreements: Voting

101

Fact7: in Modular TOD, any ONM over Pure deals, “Hide” lie can be beneficial. (as you think I have less, so increase load will cost more than it realy does)

Ex3 (from next slide): A1 hides his letter node b ({e}{b}) = utility for A1 (under lie) is 0 = utility for A2 (under lie) is 4 UNFAIR (under lie)

({b}{e}) = utility for A1 (under lie) is 2 = utility for A2 (under lie) is 2 So I get sent to b, but I really needed to go there anyway, so my

utility is actually 4. (as I don’t go to e)

Page 102: Reaching Agreements: Voting

102

Fact8:in Modular TOD, any ONM over Mixed deals, “Hide” lies can be beneficial.

A1 hides his letter to node a A1’s Utility is 4.5 > 4 (Utility of telling the truth) Under truth Util({fae}{bcd})1/2 = 4 (save going to two) Under lie divide as ({efd}{cab})p (you always win and I

always lose. Since work is same, swapping cannot help. In a mixed deal, the choices must be unbalanced.

Try again, under lie ({ab}{cdef})p p(4) + (1-p)(0) = p(2) + (1-p)(6) 4p = -4p + 6 p = 3/4 Utility is actually 3/4(6) + 1/4(0) = 4.5

Note, when I get assigned {cdef} ¼ of the time, I STILL have to deliver to node a (after completing by agreed upon deliveries). So I end up going 5 places (which is what I was assigned originally). Zero utility to that.

Page 103: Reaching Agreements: Voting

103

Conclusion

In order to use Negotiation Protocols, it is necessary to know when protocols are appropriate

TOD’s cover an important set of Multi-agent interaction

Page 104: Reaching Agreements: Voting

104

MAS Compromise: Negotiation process for conflicting goals Identify potential interactions Modify intentions to avoid harmful interactions or

create cooperative situations

Techniques required Representing and maintaining belief models Reasoning about other agents beliefs Influencing other agents intentions and beliefs

Page 105: Reaching Agreements: Voting

105

PERSUADER – case study Program to resolve problems in labor relations domain Agents

Company Union Mediator

Tasks Generation of proposal Generation of counter proposal based on feedback from

dissenting party Persuasive argumentation

Page 106: Reaching Agreements: Voting

106

Negotiation Methods: Case Based Reasoning Uses past negotiation experiences as guides to present

negotiation (like in court of law – cite previous decisions) Process

Retrieve appropriate precedent cases from memory Select the most appropriate case Construct an appropriate solution Evaluate solution for applicability to current case Modify the solution appropriately

Page 107: Reaching Agreements: Voting

107

Case Based Reasoning Cases organized and retrieved according to conceptual

similarities. Advantages

Minimizes need for information exchange Avoids problems by reasoning from past failures. Intentional

reminding. Repair for past failure is used. Reduces computation.

Page 108: Reaching Agreements: Voting

108

Negotiation Methods: Preference Analysis From scratch planning method Based on multi attribute utility theory Gets a overall utility curve out of individual ones. Expresses the tradeoffs an agent is willing to make. Property of the proposed compromise

Maximizes joint payoff Minimizes payoff difference

Page 109: Reaching Agreements: Voting

109

Persuasive argumentation Argumentation goals

Ways that an agent’s beliefs and behaviors can be affected by an argument

Increasing payoff Change importance attached to an issue Changing utility value of an issue

Page 110: Reaching Agreements: Voting

110

Narrowing differences Gets feedback from rejecting party

Objectionable issues Reason for rejection Importance attached to issues

Increases payoff of rejecting party by greater amount than reducing payoff for agreed parties.

Page 111: Reaching Agreements: Voting

111

Experiments Without Memory – 30% more proposals Without argumentation – fewer proposals and better

solutions No failure avoidance – more proposals with

objections No preference analysis – Oscillatory condition No feedback – communication overhead increased

by 23%

Page 112: Reaching Agreements: Voting

112

Multiple Attribute: Example

2 agents are trying to set up a meeting. The first agent wishes to

meet later in the day while the second wishes to meet earlier in the

day. Both prefer today to tomorrow. While the first agent assigns

highest worth to a meeting at 16:00hrs, s/he also assigns

progressively smaller worths to a meeting at 15:00hrs, 14:00hrs….

By showing flexibility and accepting a sub-optimal time, an agent

can accept a lower worth which may have other payoffs, (e.g.

reduced travel costs).

Worth function for first agent

0

100

9 12 16

Ref: Rosenschein & Zlotkin, 1994

Page 113: Reaching Agreements: Voting

113

Utility Graphs - convergence

Each agent concedes in every round of negotiation

Eventually reach an agreement

time

Utility

No. of negotiations

Agentj

Agenti

Point of acceptance

Page 114: Reaching Agreements: Voting

114

Utility Graphs - no agreement

•No agreement

Agentj finds offer unacceptable

time

Utility

Agentj

Agenti

No. of negotiations

Page 115: Reaching Agreements: Voting

115

Argumentation

The process of attempting to convince others of

something.

Why argument-based negotiations:game-theoretic

approaches have limitations

Positions cannot be justified – Why did the agent pay so

much for the car?

Positions cannot be changed – Initially I wanted a car

with a sun roof. But I changed preference during the

buying process.

Page 116: Reaching Agreements: Voting

116

4 modes of argument (Gilbert 1994):

1. Logical - ”If you accept A and accept A implies B,

then you must accept that B”

2. Emotional - ”How would you feel if it happened to

you?”

3. Visceral - participant stamps their feet and show the

strength of their feelings

4. Kisceral - Appeals to the intuitive – doesn’t this

seem reasonable

Page 117: Reaching Agreements: Voting

117

Logic Based Argumentation Basic form of argumentation

Database ├ (Sentence,Grounds)Where

Database: is a (possibly inconsistent) set of logical formulae

Sentence is a logical formula know as the conclusion

Grounds is a set of logical formula

grounds database

sentence can be proved from grounds

(we give reason for our conclusions)

Page 118: Reaching Agreements: Voting

118

Attacking Arguments Milk is good for you Cheese is made from milk Cheese is good for you

Two fundamental kinds of attack: Undercut (invalidate premise): milk isn’t good

for you if fatty Rebut (contradict conclusion): Cheese is bad

for bones

Page 119: Reaching Agreements: Voting

119

Attacking arguments

Derived notions of attack used in Literature:

A attacks B = A u B or A r B

A defeats B = A u B or (A r B and not B u A)

A strongly attacks B = A a B and not B u A

A strongly undercuts B = A u B and not B u A

Page 120: Reaching Agreements: Voting

Proposition: Hierarchy of attacks

Undercuts = u

Strongly undercuts = su = u - u -1

Strongly attacks = sa = (u r ) - u -1

Defeats = d = u ( r - u -1)

Attacks = a = u r

Page 121: Reaching Agreements: Voting

121

Abstract Argumentation Concerned with the overall structure of the argument (rather than

internals of arguments) Write x y indicates

“argument x attacks argument y” “x is a counterexample of y” “x is an attacker of y”

where we are not actually concerned as to what x, y are An abstract argument system is a collection or arguments

together with a relation “” saying what attacks what An argument is out if it has an undefeated attacker, and in if all

its attackers are defeated. Assumption – true unless proven false

Page 122: Reaching Agreements: Voting

122

Admissible Arguments – mutually defensible1. argument x is attacked if no member attacks y and

yx2. argument x is acceptable if every attacker of x is

attacked3. argument set is conflict free if none attack each

other4. set is admissible if conflict free and each argument

is acceptable (any attackers are attacked)

Page 123: Reaching Agreements: Voting

a

b

cd

Which sets of arguments can be true? c is always attacked.

d is always acceptable

Page 124: Reaching Agreements: Voting

124

An Example Abstract Argument System