Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

57
Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor
  • date post

    21-Dec-2015
  • Category

    Documents

  • view

    212
  • download

    0

Transcript of Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Page 1: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Learning in Games

Scott E PageRussell Golman & Jenna Bednar

University of Michigan-Ann Arbor

Page 2: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.
Page 3: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Overview

Analytic and Computational Models

Nash Equilibrium– Stability– Basins of Attraction

Cultural Learning

Agent Models

Page 4: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Domains of Interest

Analytic

Computable

Page 5: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Learning/Adaptation

Diversity

Interactions/Epistasis

Networks/Geography

Page 6: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Equilibrium Science

We can start by looking at the role that learning rules play in equilibrium systems. This will give us some insight into whether they’ll matter in complex systems.

Page 7: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Players

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Page 8: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Actions

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Cooperate: C Defect: D

Page 9: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Payoffs

4,4 0,6

6,0 2,2

C D

C

D

Page 10: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best Responses

4,4 0,6

6,0 2,2

C D

C

D

Page 11: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best Responses

4,4 0,6

6,0 2,2

C D

C

D

Page 12: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Nash Equilibrium

4,4 0,6

6,0 2,2

C D

C

D 2,2

Page 13: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

“Equilibrium” Based Science

Step 1: Set up gameStep 2: Solve for equilibriumStep 3: Show how equilibrium depends

on parameters of modelStep 4: Provide empirical support

Page 14: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Is Equilibrium Enough?

Existence: Equilibrium exists

Stability: Equilibrium is stable

Attainable: Equilibrium is attained by a learning rule.

Page 15: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Examples

Best respond to current state

Better respond

Mimic best

Mimic better

Include portions of best or better

Random with death of the unfit

Page 16: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Stability

Stability can only be defined relative to a learning dynamic. In dynamical systems, we often take that dynamic to be a best response function, but with human actors we need not assume people best respond.

Page 17: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Battle of Sexes Game

3,1 0,0

0,0 1,3

EF CG

EF

CG

Page 18: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Three Equilibria

3,1 0,0

0,0 1,3

1/4 3/4

3/4

1/4

Page 19: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Unstable Mixed?

3,1 0,0

0,0 1,3

1/4 +e 3/4 - e

EF

CG

3/4 + 3e

3/4 - 3e

Page 20: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Note the Implicit Assumption

Our stability analysis assumed that Player 1 would best respond to Player 2’s tremble. However, the learning rule could be go to the mixed strategy equilibrium. If so, Player 1 would sit tight and Player 2 would return to the mixed strategy equilibrium.

Page 21: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Empirical Foundations

We need to have some understanding of how people learn and adapt to say anything about stability.

Page 22: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Classes of Learning Rules

Belief Based Learning Rules: People best respond given their beliefs about how other people play.

Replicator Learning Rules: People replicate successful actions of others.

Page 23: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Stability Results

An extensive literature provides conditions (fairly week) under which the two learning rules have identical stability property.

Synopsis: Learning rules do not matter

Page 24: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Basins Question

Do games exist in which best response dynamics and replicator dynamics produce different basins of attraction?

Question: Does learning matter?

Page 25: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best Response Dynamics

x = mixed strategy of Player 1

y = mixed strategy of Player 2

dx/dt = BR(y) - x

dy/dt = BR(x) - y

Page 26: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Replicator Dynamics

dxi/dt = xi( i - ave)

dyi/dt = yi( i - ave)

Page 27: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Symmetric Matrix Game

60 60 30

30 70 20

50 25 25

A

B

C

A B C

Page 28: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

A

BC

A

B

B

Page 29: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best Response Basins

60 60 30

30 70 20

50 25 25

A

B

C

A B C

A > B iff 60pA + 60pB + 30pC > 30pA + 70pB + 20pC

A > C iff 60pA + 60pB + 30pC > 50pA + 25pB + 25pC

Page 30: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Barycentric Coordinates

A B

C

.7C + . 3B

.5A + . 5B

.4A + . 2B + .4C

Page 31: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best Responses

A

A

B

C

C

BC B

Page 32: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Stable Equilibria

A

A

B

C

C

BC B

Page 33: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best Response Basins

A

A

B

C

C

BA B

Page 34: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Replicator Dynamics Basins

A B

C

A

C

C

B? B

Page 35: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Replicator Dynamics

A

A

B

C

C

B

ABB

Page 36: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Recall: Basins Question

Do games exist in which best response dynamics and replicator dynamics produce very different basins of attraction?

Question: Does learning matter?

Page 37: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Conjecture

For any > 0, There exists a symmetric matrix game such that the basins of attraction for distinct equilibria under continuous time best response dynamics and replicator dynamics overlap by less than

Page 38: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Collective Action Game

SI Coop Pred Naive

2 2 2 2

1 N+1 1 1

0 0 0 N2

0 0 -N2 0

SI

Coop

Pred

Naive

Page 39: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Naïve Goes Away

SI Coop

Pred

Page 40: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Basins on Face

SI Coop

Pred

Coop

SI

Page 41: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Lots of Predatory

SI Coop

Pred

Coop

SI

Page 42: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Best ResponsePred

Coop

SI

Page 43: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

ReplicatorPred

Coop

SI

Page 44: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

The Math

dxi/dt = xi( i - ave)

ave = 2xS + xC (1+NxC)

dxc/dt = xc[(1+ NxC- - 2xS - xC (1+NxC)]

dxc/dt = xc[(1+ NxC)(1- xC) - 2xS]

Page 45: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Choose N > 1/

Assume xc >

dxc/dt = xc[(1+ NxC)(1- xC) - 2xS]

dxc/dt > [2(1- ) - 2(1- )] = 0

Therefore, xc always increases.

Page 46: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Collective Action Game

SI Coop Pred Naive

2 2 2 2

1 N+1 1 1

0 0 0 N2

0 0 -N2 0

SI

Coop

Pred

Naive

Page 47: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Theorem: In any symmetric matrix game for which best response and replicator dynamics attain different equilibria with probability one, there exists an action A that is both an initial best response with probability one and not an equilibrium.

Page 48: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Convex Combinations

Suppose the population learns using the following rule

a (Best Response) + (1-a) Replicator

Page 49: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Claim: There exists a class of games in which Best Response, Replicator Dynamics, and any fixed convex combination select distinct equilibria with probability one.

Best Response -> A

Replicator -> B

a BR + (1-a) Rep -> C

Page 50: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Learning/Adaptation

Diversity

Interactions/Epistasis

Networks/Geography

Page 51: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Diversity

EWA (Wilcox) and Quantal Response (Golman) learning models are misspecified for heterogenous agents.

EWA = “hybrid of response and belief based learning”

Page 52: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Interactions

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

7,7 4,14

14,4 5,5

7,7 3,11

11,3 5,5

Page 53: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

GameS Theory

E O

M

E O

M

E O

M

E O

M

Bednar (2008), Bednar and Page (2007)

Page 54: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Learning Spillovers

A

A

B

C

C

BC B

Page 55: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Initial Conditions

A

A

B

C

C

BC B

Page 56: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

Networks/Geography

A B B

B A

B C B

?

A

A

B

C

B

B

C

A

C

AC

BB

B

A

A

A

A

B

C

B

B AC BA

C CBB AB

C

A

A

B

C

B

C

A

Page 57: Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.

ABM complement Math