基础博弈论大学英文讲义
lecture_2(博弈论讲义GameTheory(MIT))

Last Time:Defined knowledge, common knowledge, meet (of partitions), and reachability.Reminders:• E is common knowledge at ω if ()I K E ω∞∈.• “Reachability Lemma” :'()M ωω∈ if there is a chain of states 01,,...m 'ωωωωω== such that for each k ω there is a player i(k) s.t. ()()1()(i k k i k k h h )ωω+=:• Theorem: Event E is common knowledge at ωiff ()M E ω⊆.How does set of NE change with information structure?Suppose there is a finite number of payoff matrices 1,...,L u u for finite strategy sets 1,...,I S SState space Ω, common prior p, partitions , and a map i H λso that payoff functions in state ω are ()(.)u λω; the strategy spaces are maps from into . i H i SWhen the state space is finite, this is a finite game, and we know that NE is u.h.c. and generically l.h.c. in p. In particular, it will be l.h.c. at strict NE.The “coordinated attack” game8,810,11,100,0A B A B-- 0,010,11,108,8A B A B--a ub uΩ= 0,1,2,….In state 0: payoff functions are given by matrix ; bu In all other states payoff functions are given by . a upartitions of Ω1H : (0), (1,2), (3,4),… (2n-1,2n)... 2H (0,1),(2,3). ..(2n,2n+1)…Prior p : p(0)=2/3, p(k)= for k>0 and 1(1)/3k e e --(0,1)ε∈.Interpretation: coordinated attack/email:Player 1 observes Nature’s choice of payoff matrix, sends a message to player 2.Sending messages isn’t a strategic decision, it’s hard-coded.Suppose state is n=2k >0. Then 1 knows the payoffs, knows 2 knows them. Moreover 2 knows that 1knows that 2 knows, and so on up to strings of length k: . 1(0n I n K n -Î>)But there is no state at which n>0 is c.k. (to see this, use reachability…).When it is c.k. that payoff are given by , (A,A) is a NE. But.. auClaim: the only NE is “play B at every information set.”.Proof: player 1 plays B in state 0 (payoff matrix ) since it strictly dominates A. b uLet , and note that .(0|(0,1))q p =1/2q >Now consider player 2 at information set (0,1).Since player 1 plays B in state 0, and the lowest payoff 2 can get to B in state 1 is 0, player 2’s expected payoff to B at (0,1) is at least 8. qPlaying A gives at most 108(1)q q −+−, and since , playing B is better. 1/2q >Now look at player 1 at 1(1,2)h =. Let q'=p(1|1,2), and note that '1(1)q /2εεεε=>+−.Since 2 plays B in state 1, player 1's payoff to B is at least 8q';1’s payoff to A is at most -10q'+8(1-q) so 1 plays B Now iterate..Conclude that the unique NE is always B- there is no NE in which at some state the outcome is (A,A).But (A,A ) is a strict NE of the payoff matrix . a u And at large n, there is mutual knowledge of the payoffs to high order- 1 knows that 2 knows that …. n/2 times. So “mutual knowledge to large n” has different NE than c.k.Also, consider "expanded games" with state space . 0,1,....,...n Ω=∞For each small positive ε let the distribution p ε be as above: 1(0)2/3,()(1)/3n p p n ee e e -==- for 0 and n <<∞()0p ε∞=.Define distribution by *p *(0)2/3p =,. *()1/3p ∞=As 0ε→, probability mass moves to higher n, andthere is a sense in which is the limit of the *p p εas 0ε→.But if we do say that *p p ε→ we have a failure of lower hemi continuity at a strict NE.So maybe we don’t want to say *p p ε→, and we don’t want to use mutual knowledge to large n as a notion of almost common knowledge.So the questions:• When should we say that one information structure is close to another?• What should we mean by "almost common knowledge"?This last question is related because we would like to say that an information structure where a set of events E is common knowledge is close to another information structure where these events are almost common knowledge.Monderer-Samet: Player i r-believes E at ω if (|())i p E h r ω≥.()r i B E is the set of all ω where player i r- believesE; this is also denoted 1.()ri B ENow do an iterative definition in the style of c.k.: 11()()rr I i i B E B E =Ç (everyone r-believes E) 1(){|(()|())}n r n ri i I B E p B E h r w w -=³ ()()n r n rI i i B E B =ÇEE is common r belief at ω if ()rI B E w ¥ÎAs with c.k., common r-belief can be characterized in terms of public events:• An event is a common r-truism if everyone r -believes it when it occurs.• An event is common r -belief at ω if it is implied by a common r-truism at ω.Now we have one version of "almost ck" : An event is almost ck if it is common r-belief for r near 1.MS show that if two player’s posteriors are common r-belief, they differ by at most 2(1-r): so Aumann's result is robust to almost ck, and holds in the limit.MS also that a strict NE of a game with knownpayoffs is still a NE when payoffs are "almost ck” - a form of lower hemi continuity.More formally:As before consider a family of games with fixed finite action spaces i A for each player i. a set of payoff matrices ,:l I u A R ->a state space W , that is now either finite or countably infinite, a prior p, a map such that :1,,,L l W®payoffs at ω are . ()(,)()w u a u a l w =Payoffs are common r-belief at ω if the event {|()}w l w l = is common r belief at ω.For each λ let λσ be a NE for common- knowledgepayoffs u .lDefine s * by *(())s l w w s =.This assigns each w a NE for the corresponding payoffs.In the email game, one such *s is . **(0)(,),()(,)s B B s n A A n ==0∀>If payoffs are c.k. at each ω, then s* is a NE of overall game G. (discuss)Theorem: Monder-Samet 1989Suppose that for each l , l s is a strict equilibrium for payoffs u λ.Then for any there is 0e >1r < and 1q < such that for all [,1]r r Î and [,1]q q Î,if there is probability q that payoffs are common r- belief, then there is a NE s of G with *(|()())1p s s ωωω=>ε−.Note that the conclusion of the theorem is false in the email game:there is no NE with an appreciable probability of playing A, even though (A,A) is a strict NE of the payoffs in every state but state 0.This is an indirect way of showing that the payoffs are never ACK in the email game.Now many payoff matrices don’t have strictequilibria, and this theorem doesn’t tell us anything about them.But can extend it to show that if for each state ω, *(s )ω is a Nash (but not necessarily strict Nash) equilibrium, then for any there is 0e >1r < and 1q < such that for all [,1]r r Î and [,1]q q Î, if payoffs are common r-belief with probability q, there is an “interim ε equilibria” of G where s * is played with probability 1ε−.Interim ε-equilibria:At each information set, the actions played are within epsilon of maxing expected payoff(((),())|())((',())|())i i i i i i i i E u s s h w E u s s h w w w w e-->=-Note that this implies the earlier result when *s specifies strict equilibria.Outline of proof:At states where some payoff function is common r-belief, specify that players follow s *. The key is that at these states, each player i r-believes that all other players r-believe the payoffs are common r-belief, so each expects the others to play according to s *.*ΩRegardless of play in the other states, playing this way is a best response, where k is a constant that depends on the set of possible payoff functions.4(1)k −rTo define play at states in */ΩΩconsider an artificial game where players are constrained to play s * in - and pick a NE of this game.*ΩThe overall strategy profile is an interim ε-equilibrium that plays like *s with probability q.To see the role of the infinite state space, consider the"truncated email game"player 2 does not respond after receiving n messages, so there are only 2n states.When 2n occurs: 2 knows it occurs.That is, . {}2(0,1),...(22,21,)(2)H n n =−−n n {}1(0),(1,2),...(21,2)H n =−.()2|(21,2)1p n n n ε−=−, so 2n is a "1-ε truism," and thus it is common 1-ε belief when it occurs.So there is an exact equilibrium where players playA in state 2n.More generally: on a finite state space, if the probability of an event is close to 1, then there is high probability that it is common r belief for r near 1.Not true on infinite state spaces…Lipman, “Finite order implications of the common prior assumption.”His point: there basically aren’t any!All of the "bite" of the CPA is in the tails.Set up: parameter Q that people "care about" States s S ∈,:f S →Θ specifies what the payoffs are at state s. Partitions of S, priors .i H i pPlayer i’s first order beliefs at s: the conditional distribution on Q given s.For B ⊆Θ,1()()i s B d =('|(')|())i i p s f s B h s ÎPlayer i’s second order beliefs: beliefs about Q and other players’ first order beliefs.()21()(){'|(('),('))}|()i i j i s B p s f s s B h d d =Îs and so on.The main point can be seen in his exampleTwo possible values of an unknown parameter r .1q q = o 2qStart with a model w/o common prior, relate it to a model with common prior.Starting model has only two states 12{,}S s s =. Each player has the trivial partition- ie no info beyond the prior.1122()()2/3p s p s ==.example: Player 1 owns an asset whose value is 1 at 1θ and 2 at 2θ; ()i i f s θ=.At each state, 1's expected value of the asset 4/3, 2's is 5/3, so it’s common knowledge that there are gains from trade.Lipman shows we can match the players’ beliefs, beliefs about beliefs, etc. to arbitrarily high order in a common prior model.Fix an integer N. construct the Nth model as followsState space'S ={1,...2}N S ´Common prior is that all states equally likely.The value of θ at (s,k) is determined by the s- component.Now we specify the partitions of each player in such a way that the beliefs, beliefs about beliefs, look like the simple model w/o common prior.1's partition: events112{(,1),(,2),(,1)}...s s s 112{(,21),(,2),(,)}s k s k s k -for k up to ; the “left-over” 12N -2s states go into 122{(,21),...(,2)}N N s s -+.At every event but the last one, 1 thinks the probability of is 2/3.1qThe partition for player 2 is similar but reversed: 221{(,21),(,2),(,)}s k s k s k - for k up to . 12N -And at all info sets but one, player 2 thinks the prob. of is 1/3.1qNow we look at beliefs at the state 1(,1)s .We matched the first-order beliefs (beliefs about θ) by construction)Now look at player 1's second-order beliefs.1 thinks there are 3 possible states 1(,1)s , 1(,2)s , 2(,1)s .At 1(,1)s , player 2 knows {1(,1)s ,2(,1)s ,(,}. 22)s At 1(,2)s , 2 knows . 122{(,2),(,3),(,4)}s s s At 2(,1)s , 2 knows {1(,2)s , 2(,1)s ,(,}. 22)sThe support of 1's second-order beliefs at 1(,1)s is the set of 2's beliefs at these info sets.And at each of them 2's beliefs are (1/3 1θ, 2/3 2θ). Same argument works up to N:The point is that the N-state models are "like" the original one in that beliefs at some states are the same as beliefs in the original model to high but finite order.(Beliefs at other states are very different- namely atθ or 2 is sure the states where 1 is sure that state is2θ.)it’s1Conclusion: if we assume that beliefs at a given state are generated by updating from a common prior, this doesn’t pin down their finite order behavior. So the main force of the CPA is on the entire infinite hierarchy of beliefs.Lipman goes on from this to make a point that is correct but potentially misleading: he says that "almost all" priors are close to a common. I think its misleading because here he uses the product topology on the set of hierarchies of beliefs- a.k.a topology of pointwise convergence.And two types that are close in this product topology can have very different behavior in a NE- so in a sense NE is not continuous in this topology.The email game is a counterexample. “Product Belief Convergence”:A sequence of types converges to if thesequence converges pointwise. That is, if for each k,, in t *i t ,,i i k n k *δδ→.Now consider the expanded version of the email game, where we added the state ∞.Let be the hierarchy of beliefs of player 1 when he has sent n messages, and let be the hierarchy atthe point ∞, where it is common knowledge that the payoff matrix is .in t ,*i t a uClaim: the sequence converges pointwise to . in t ,*i t Proof: At , i’s zero-order beliefs assignprobability 1 to , his first-order beliefs assignprobability 1 to ( and j knows it is ) and so onup to level n-1. Hence as n goes to infinity, thehierarchy of beliefs converges pointwise to common knowledge of .in t a u a u a u a uIn other words, if the number of levels of mutual knowledge go to infinity, then beliefs converge to common knowledge in the product topology. But we know that mutual knowledge to high order is not the same as almost common knowledge, and types that are close in the product topology can play very differently in Nash equilibrium.Put differently, the product topology on countably infinite sequences is insensitive to the tail of the sequence, but we know that the tail of the belief hierarchy can matter.Next : B-D JET 93 "Hierarchies of belief and Common Knowledge”.Here the hierarchies of belief are motivated by Harsanyi's idea of modelling incomplete information as imperfect information.Harsanyi introduced the idea of a player's "type" which summarizes the player's beliefs, beliefs about beliefs etc- that is, the infinite belief hierarchy we were working with in Lipman's paper.In Lipman we were taking the state space Ω as given.Harsanyi argued that given any element of the hierarchy of beliefs could be summarized by a single datum called the "type" of the player, so that there was no loss of generality in working with types instead of working explicitly with the hierarchies.I think that the first proof is due to Mertens and Zamir. B-D prove essentially the same result, but they do it in a much clearer and shorter paper.The paper is much more accessible than MZ but it is still a bit technical; also, it involves some hard but important concepts. (Add hindsight disclaimer…)Review of math definitions:A sequence of probability distributions converges weakly to p ifn p n fdp fdp ®òò for every bounded continuous function f. This defines the topology of weak convergence.In the case of distributions on a finite space, this is the same as the usual idea of convergence in norm.A metric space X is complete if every Cauchy sequence in X converges to a point of X.A space X is separable if it has a countable dense subset.A homeomorphism is a map f between two spaces that is 1-1, and onto ( an isomorphism ) and such that f and f-inverse are continuous.The Borel sigma algebra on a topological space S is the sigma-algebra generated by the open sets. (note that this depends on the topology.)Now for Brandenburger-DekelTwo individuals (extension to more is easy)Common underlying space of uncertainty S ( this is called in Lipman)ΘAssume S is a complete separable metric space. (“Polish”)For any metric space, let ()Z D be all probability measures on Borel field of Z, endowed with the topology of weak convergence. ( the “weak topology.”)000111()()()n n n X S X X X X X X --=D =´D =´DSo n X is the space of n-th order beliefs; a point in n X specifies (n-1)st order beliefs and beliefs about the opponent’s (n-1)st order beliefs.A type for player i is a== 0012(,,,...)()n i i i i n t X d d d =¥=δD0T .Now there is the possibility of further iteration: what about i's belief about j's type? Do we need to add more levels of i's beliefs about j, or is i's belief about j's type already pinned down by i's type ?Harsanyi’s insight is that we don't need to iterate further; this is what B-D prove formally.Coherency: a type is coherent if for every n>=2, 21marg n X n n d d --=.So the n and (n-1)st order beliefs agree on the lower orders. We impose this because it’s not clear how to interpret incoherent hierarchies..Let 1T be the set of all coherent typesProposition (Brandenburger-Dekel) : There is a homeomorphism between 1T and . 0()S T D ´.The basis of the proposition is the following Lemma: Suppose n Z are a collection of Polish spaces and let021201...1{(,,...):(...)1, and marg .n n n Z Z n n D Z Z n d d d d d --´´-=ÎD ´"³=Then there is a homeomorphism0:(nn )f D Z ¥=®D ´This is basically the same as Kolmogorov'sextension theorem- the theorem that says that there is a unique product measure on a countable product space that corresponds to specified marginaldistributions and the assumption that each component is independent.To apply the lemma, let 00Z X =, and 1()n n Z X -=D .Then 0...n n Z Z X ´´= and 00n Z S T ¥´=´.If S is complete separable metric than so is .()S DD is the set of coherent types; we have shown it is homeomorphic to the set of beliefs over state and opponent’s type.In words: coherency implies that i's type determines i's belief over j's type.But what about i's belief about j's belief about i's type? This needn’t be determined by i’s type if i thinks that j might not be coherent. So B-D impose “common knowledge of coherency.”Define T T ´ to be the subset of 11T T ´ where coherency is common knowledge.Proposition (Brandenburger-Dekel) : There is a homeomorphism between T and . ()S T D ´Loosely speaking, this says (a) the “universal type space is big enough” and (b) common knowledge of coherency implies that the information structure is common knowledge in an informal sense: each of i’s types can calculate j’s beliefs about i’s first-order beliefs, j’s beliefs about i’s beliefs about j’s beliefs, etc.Caveats:1) In the continuity part of the homeomorphism the argument uses the product topology on types. The drawbacks of the product topology make the homeomorphism part less important, but theisomorphism part of the theorem is independent of the topology on T.2) The space that is identified as“universal” depends on the sigma-algebra used on . Does this matter?(S T D ´)S T ×Loose ideas and conjectures…• There can’t be an isomorphism between a setX and the power set 2X , so something aboutmeasures as opposed to possibilities is being used.• The “right topology” on types looks more like the topology of uniform convergence than the product topology. (this claim isn’t meant to be obvious. the “right topology” hasn’t yet been found, and there may not be one. But Morris’ “Typical Types” suggests that something like this might be true.)•The topology of uniform convergence generates the same Borel sigma-algebra as the product topology, so maybe B-D worked with the right set of types after all.。
博弈论讲义入门 slides2

学生的成绩就是 其中, 是学生报出的数字。
VNM公理
• 公理A2(独立性):对任意p,q,r∈P,和 任意a∈(0,1],
VNM公理
• 公里A3(连续性):对任意p,q,r∈P,如 果 p > q > r,那么存在a,b∈(0,1) 使
定理—VNM表示
• P上的关系≥可用VNM效用函数u:Z→R 表示的充分必要条件是它满足公理A1-A3。 • u和v表示≥的充分必要条件是 其中a>0,b∈R
习题
• 考虑一正实数上可用VNM效用函数 u(x)=x2表示的关系≥。 这种关系可用VNM函数 表 示吗? 用 表示又如何呢?
对风险的态度
• 公平赌博:
• 参与者是风险中性的 iff 他对所有公平赌博无 所谓。 • 他是(严格)厌恶风险的 iff 他从不想参与公平赌 博。 • 他是(严格)追逐风险的 iff 他总是想参与公平赌 博。
基本概念:偏好 • 关系≥ (X上)是X×X上的任一子集 • 例如, • T≥C≡(T,C)∈≥ • ≥是完全的iff 要么X≥Y,要么Y≥X • ≥是传递的iff [X≥Y和Y≥Z]
偏好关系
• 定义:一种关系是偏好关系的充分必要 条件是,它是完全和传递的。
实例
• 在本班学生中定义下列关系 • xTy表示x至少和y一样高; • xMy表示x在课程14.04的期末成绩至少和 y一样高; • xHy表示x和y 上的同一所中学; • xYy表示x比y年龄小; • xSy表示x与y 同龄;
实例
可由函数u**表示,其中
练习
• 设想一群学生围绕一张圆桌坐下。定义 一种关系R,写成xRy,表示x坐在y的右 边。你能用一效用函数表示R吗? • 考虑一在正实数上以u表示的关系≥,其 中u(x)=x2。 • 这种关系可表示为 吗? • u**(x)=1/x 又 如何呢?
lecture10(博弈论讲义(Carnegie Mellon University))

Theorem 1 (property of mixed Nash equilibrium) A pair of mixed strategies ((r*, 1-r*), (q*, 1-q*)) is a Nash equilibrium if and only if
v1((r*, 1-r*), (q*, 1-q*)) v1((r*, 1-r*), (q*, 1-q*)) v2((r*, 1-r*), (q*, 1-q*)) v2((r*, 1-r*), (q*, 1-q*))
பைடு நூலகம்
Mixed strategy Nash equilibrium: A pair of mixed strategies ((r*, 1-r*), (q*, 1-q*)) is a Nash equilibrium if (r*,1-r*) is a best response to (q*, 1-q*), and (q*, 1-q*) is a best response to (r*,1-r*). That is,
Static (or SimultaneousMove) Games of Complete Information
Mixed Strategy Nash Equilibrium
June 2, 2003
73-347 Game Theory--Lecture 10
1
Outline of Static Games of Complete Information
June 2, 2003 73-347 Game Theory--Lecture 10 2
Today’s Agenda
Review of previous class Example: Rock, paper and scissors (Exercise
耶鲁大学公开课英文讲义—博弈论第一节

Game Theory: Lecture 1 TranscriptProfessor Ben Polak: So this is Game Theory Economics 159. If you're here for art history, you're either in the wrong room or stay anyway, maybe this is the right room; but this is Game Theory, okay. You should have four handouts; everyone should have four handouts. There is a legal release form--we'll talk about it in a minute--about the videoing. There is a syllabus, which is a preliminary syllabus: it's also online. And there are two games labeled Game 1 and Game 2. Can I get you all to look at Game 1 and start thinking about it. And while you're thinking about it, I am hoping you can multitask a bit. I'll describe a bit about the class and we'll get a bit of admin under our belts. But please try and look at--somebody's not looking at it, because they're using it as a fan here--so look at Game 1 and fill out that form for me, okay?So while you're filling that out, let me tell you a little bit about what we're going to be doing here. So what is Game Theory? Game Theory is a method of studying strategic situations. So what's a strategic situation? Well let's start off with what's not a strategic situation. In your Economics - in your Intro Economics class in 115 or 110, you saw some pretty good examples of situations that were not strategic. You saw firms working in perfect competition. Firms in perfect competition are price takers: they don't particularly have to worry about the actions of their competitors. You also saw firms that were monopolists and monopolists don't have any competitors to worry about, so that's not a particularly strategic situation. They're not price takers but they take the demand curve. Is this looking familiar for some of you who can remember doing 115 last year or maybe two years ago for some of you? Everything in between is strategic. So everything that constitutes imperfect competition is a strategic setting. Think about the motor industry, the motor car industry. Ford has to worry about what GM is doing and what Toyota is doing, and for the moment at least what Chrysler is doing but perhaps not for long. So there's a small number of firms and their actions affect each other.So for a literal definition of what strategic means: it's a setting where the outcomes that affect you depend on actions, not just on your own actions, but on actions of others. All right, that's as much as I'm going to say for preview right now, we're going to come back and see plenty of this over the course of the next semester.So what I want to do is get on to where this applies. It obviously applies in Economics, but it also applies in politics, and in fact, this class will count as a Political Science class if you're a Political Science major. You should go check with the DUS in Political Science. It count - Game Theory is very important in law these days. So for those of you--for the half of you--that aregoing to end up in law school, this is pretty good training. Game Theory is also used in biology and towards the middle of the semester we're actually going to see some examples of Game Theory as applied to evolution. And not surprisingly, Game Theory applies to sport.So let's talk about a bit of admin. How are you doing on filling out those games? Everyone managing to multitask: filling in Game 1? Keep writing. I want to get some admin out of the way and I want to start by getting out of the way what is obviously the elephant in the room. Some of you will have noticed that there's a camera crew here, okay. So as some of you probably know, Yale is undergoing an open education project and they're videoing several classes, and the idea of this, is to make educational materials available beyond the walls of Yale. In fact, on the web, internationally, so people in places, maybe places in the U.S. or places miles away, maybe in Timbuktu or whatever, who find it difficult to get educational materials from the local university or whatever, can watch certain lectures from Yale on the web.Some of you would have been in classes that do that before. What's going to different about this class is that you're going to be participating in it. The way we teach this class is we're going to play games, we're going to have discussions, we're going to talk among the class, and you're going to be learning from each other, and I want you to help people watching at home to be able to learn too. And that means you're going to be on film, at the very least on mike.So how's that going to work? Around the room are three T.A.s holding mikes. Let me show you where they are: one here, one here, and one here. When I ask for classroom discussions, I'm going to have one of the T.A.s go to you with a microphone much like in "Donahue" or something, okay. At certain times, you're going to be seen on film, so the camera is actually going to come around and point in your direction.Now I really want this to happen. I had to argue for this to happen, cause I really feel that this class isn't about me. I'm part of the class obviously, but it's about you teaching each other and participating. But there's a catch, the catch is, that that means you have to sign that legal release form.So you'll see that you have in front of you a legal release form, you have to be able to sign it, and what that says is that we can use you being shown in class. Think of this as a bad hair day release form. All right, you can't sue Yale later if you had a bad hair day. For those of you who are on the run from the FBI, your Visa has run out, or you're sitting next to your ex-girlfriend, now would be a good time to put a paper bag over your head.All right, now just to get you used to the idea, in every class we're going to have I think the same two people, so Jude is the cameraman; why don't you all wave to Jude: this is Jude okay. And Wes is our audio guy: this is Wes. And I will try and remember not to include Jude and Wes in the classroom discussions, but you should be aware that they're there. Now, if this is making you nervous, if it's any consolation, it's making me very nervous. So, all right, we'll try and make this class work as smoothly as we can, allowing for this extra thing. Let me just say, no one's making any money off this--at least I'm hoping these guys are being paid--but me and the T.A.s are not being paid. The aim of this, that I think is a good aim, it's an educational project, and I'm hoping you'll help us with it. The one difference it is going to mean, is that at times I might hold some of the discussions for the class, coming down into this part of the room, here, to make it a little easier for Jude.All right, how are we doing now on filling out those forms? Has everyone filled in their strategy for the first game? Not yet. Okay, let's go on doing a bit more admin. The thing you mostly care about I'm guessing, is the grades. All right, so how is the grade going to work for this class? 30% of the class will be on problem sets, 30% of the grade; 30% on the mid-term, and 40% on the final; so 30/30/40.The mid-term will be held in class on October 17th; that is also in your syllabus. Please don't anybody tell me late - any time after today you didn't know when the mid-term was and therefore it clashes with 17 different things. The mid-term is on October 17th, which is a Wednesday, in class. All right, the problem sets: there will be roughly ten problem sets and I'll talk about them more later on when I hand them out. The first one will go out on Monday but it will be due ten days later. Roughly speaking they'll be every week.The grade distribution: all right, so this is the rough grade distribution. Roughly speaking, a sixth of the class are going to end up with A's, a sixth are going to end up with A-, a sixth are going to end up with B+, a sixth are going to end up with B, a sixth are going to end up with B-, and the remaining sixth, if I added that up right, are going to end up with what I guess we're now calling the presidential grade, is that right?That's not literally true. I'm going to squeeze it a bit, I'm going to curve it a bit, so actually slightly fewer than a sixth will get straight A's, and fewer than a sixth will get C's and below. We'll squeeze the middle to make them be more B's. One thing I can guarantee from past experience in this class, is that the median grade will be a B+. The median will fall somewhere in the B+'s. Just as forewarning for people who have forgotten what a median is,that means half of you--not approximately half, it means exactly half of you--will be getting something like B+ and below and half will get something like B+ and above.Now, how are you doing in filling in the forms? Everyone filled them in yet? Surely must be pretty close to getting everyone filled in. All right, so last things to talk about before I actually collect them in - textbooks. There are textbooks for this class. The main textbook is this one, Dutta'sbook Strategy and Games. If you want a slightly tougher book, more rigorous book, try Joel Watson's book, Strategies. Both of those books are available at the bookstore.But I want to warn everybody ahead of time, I will not be following the textbook. I regard these books as safety nets. If you don't understand something that happened in class, you want to reinforce an idea that came up in class, then you should read the relevant chapters in the book and the syllabus will tell you which chapters to read for each class, or for each week of class, all right. But I will not be following these books religiously at all. In fact, they're just there as back up.In addition, I strongly recommend people read, Thinking Strategically. This is good bedtime reading. Do any of you suffer from insomnia? It's very good bedtime reading if you suffer from insomnia. It's a good book and what's more there's going to be a new edition of this book this year and Norton have allowed us to get advance copies of it. So if you don't buy this book this week, I may be able to make the advance copy of the new edition available for some of you next week. I'm not taking a cut on that either, all right, there's no money changing hands.All right, sections are on the syllabus sign up - sorry on the website, sign up as usual. Put yourself down on the wait list if you don't get into the section you want. You probably will get into the section you want once we're done. All right, now we must be done with the forms. Are we done with the forms? All right, so why don't we send the T.A.s, with or without mikes, up and down the aisles and collect in your Game #1; not Game #2, just Game #1.Just while we're doing that, I think the reputation of this class--I think--if you look at the course evaluations online or whatever, is that this class is reasonably hard but reasonably fun. So I'm hoping that's what the reputation of the class is. If you think this class is going to be easy, I think it isn't actually an easy class. It's actually quite a hard class, but I think I can guarantee it's going to be a fun class. Now one reason it's a fun class, is the nice thing about teaching Game Theory - quieten down folks--one thing about teaching Game Theory is, you get to play games, and that's exactlywhat we've just been doing now. This is our first game and we're going to play games throughout the course, sometimes several times a week, sometimes just once a week.We got all these things in? Everyone handed them in? So I need to get those counted. Has anyone taken the Yale Accounting class? No one wants to - has aspirations to be - one person has. I'll have a T.A. do it, it's all right,we'll have a T.A. do it. So Kaj, can you count those for me? Is that right? Let me read out the game you've just played."Game 1, a simple grade scheme for the class. Read the following carefully. Without showing your neighbor what you are doing, put it in the box below either the letter Alpha or the letter Beta. Think of this as a grade bid. I will randomly pair your form with another form and neither you nor your pair will ever know with whom you were paired. Here's how the grades may be assigned for the class. [Well they won't be, but we can pretend.] If you put Alpha and you're paired with Beta, then you will get an A and your pair a C. If you and your pair both put Alpha, you'll both get B-. If you put Beta and you're paired with Alpha, you'll get a C and your pair an A. If you and your pair both put Beta, then you'll both get B+."So that's the thing you just filled in.Now before we talk about this, let's just collect this information in a more useful way. So I'm going to remove this for now. We'll discuss this in a second, but why don't we actually record what the game is, that we're playing, first. So this is our grade game, and what I'm going to do, since it's kind of hard to absorb all the information just by reading a paragraph of text, I'm going to make a table to record the information. So what I'm going to do is I'm going to put me here, and my pair, the person I'm randomly paired with here, and Alpha and Beta, which are the choices I'm going to make here and on the columns Alpha and Beta, the choices my pair is making.In this table, I'm going to put my grades. So my grade if we both put Alpha is B-, if we both put Beta, was B+. If I put Alpha and she put a Beta, I got an A, and if I put Beta and she put an Alpha, I got a C. Is that correct? That's more or less right? Yeah, okay while we're here, why don't we do the same for my pair? So this is my grades on the left hand table, but now let's look at what my pair will do, what my pair will get.So I should warn the people sitting at the back that my handwriting is pretty bad, that's one reason for moving forward. The other thing I should apologize at this stage of the class is my accent. I will try and improve the handwriting, there's not much I can do about the accent at this stage.So once again if you both put Alpha then my pair gets a B-. If we both put Beta, then we both get a B+; in particular, my pair gets a B+. If I put Alpha and my pair puts Beta, then she gets a C. And if I put Beta and she puts Alpha, then she gets an A. So I now have all the information that was on the sheet of paper that you just handed in.Now there's another way of organizing this that's standard in Game Theory, so we may as well get used to it now on the first day. Rather then drawing two different tables like this, what I'm going to do is I'm going to take the second table and super-impose it on top of the first table. Okay, so let me do that and you'll see what I mean. What I'm going to do is draw a larger table, the same basic structure: I'm choosing Alpha and Beta on the rows, my pair is choosing Alpha and Beta on the columns, but now I'm going to put both grades in. So the easy ones are on the diagonal: you both get B- if we both choose Alpha; we both get B+ if we both choose Beta. But if I choose Alpha and my pair chooses Beta, I get an A and she gets a C. And if I choose Beta and she chooses Alpha, then it's me who gets the C and it's her who gets the A.So notice what I did here. The first grade corresponds to the row player, me in this case, and the second grade in each box corresponds to the column player, my pair in this case. So this is a nice succinct way of recording what was in the previous two tables. This is an outcome matrix; this tells us everything that was in the game.Okay, so now seems a good time to start talking about what people did. So let's just have a show of hands. How many of you chose Alpha? Leave your hands up so that Jude can catch that, so people can see at home, okay. All right and how many of you chose Beta? There's far more Alphas - wave your hands the Beta's okay. All right, there's a Beta here, okay. So it looks like a lot of - well we're going to find out, we're going to count--but a lot more Alpha's than Beta's. Let me try and find out some reasons why people chose.So let me have the Alpha's up again. So, the woman who's in red here, can we get a mike to the - yeah, is it okay if we ask you? You're not on the run from the FBI? We can ask you why? Okay, so you chose Alpha right? So why did you choose Alpha?Student: [inaudible] realized that my partner chose Alpha, therefore I chose [inaudible].Professor Ben Polak: All right, so you wrote out these squares, you realized what your partner was going to do, and responded to that. Any otherreasons for choosing Alpha around the room? Can we get the woman here? Try not to be intimidated by these microphones, they're just mikes. It's okay.Student: The reason I chose Alpha, regardless of what my partner chose, I think there would be better outcomes than choosing Beta.Professor Ben Polak: All right, so let me ask your names for a second-so your name was?Student: Courtney.Professor Ben Polak: Courtney and your name was?Student: Clara Elise.Professor Ben Polak: Clara Elise. So slightly different reasons, same choice Alpha. Clara Elise's reason - what did Clara Elise say? She said, no matter what the other person does, she reckons she'd get a better grade if she chose Alpha. So hold that thought a second, we'll come back to - is it Clara Elise, is that right? We'll come back to Clara Elise in a second. Let's talk to the Beta's a second; let me just emphasize at this stage there are no wrong answers. Later on in the class there'll be some questions that have wrong answers. Right now there's no wrong answers. There may be bad reasons but there's no wrong answers. So let's have the Beta's up again. Let's see the Beta's. Oh come on! There was a Beta right here. You were a Beta right? You backed off the Beta, okay. So how can I get a mike into a Beta? Let' s stick in this aisle a bit. Is that a Beta right there? Are you a Beta right there? Can I get the Beta in here? Who was the Beta in here? Can we get the mike in there? Is that possible? In here - you can leave your hand so that - there we go. Just point towards - that's fine, just speak into it, that's fine. Student: So the reason right?Professor Ben Polak: Yeah, go ahead.Student: I personally don't like swings that much and it's the B-/B+ range, so I'd much rather prefer that to a swing from A to C, and that's my reason. Professor Ben Polak: All right, so you're saying it compresses the range.I'm not sure it does compress the range. I mean if you chose Alpha, you're swinging from A to B-; and from Beta, swinging from B+ to C. I mean those are similar kind of ranges but it certainly is a reason. Other reasons for choosing? Yeah, the guy in blue here, yep, good. That's all right. Don't hold the mike; just let it point at you, that's fine.Student: Well I guess I thought we could be more collusive and kind of work together, but I guess not. So I chose Beta.Professor Ben Polak: There's a siren in the background so I missed the answer. Stand up a second, so we can just hear you.Student: Sure.Professor Ben Polak: Sorry, say again.Student: Sure. My name is Travis. I thought we could work together, but I guess not.Professor Ben Polak: All right good. That's a pretty good reason. Student: If you had chosen Beta we would have all gotten B+'s but I guess not.Professor Ben Polak: Good, so Travis is giving us a different reason, right? He's saying that maybe, some of you in the room might actually care about each other's grades, right? I mean you all know each other in class. You all go to the same college. For example, if we played this game up in the business school--are there any MBA students here today? One or two. If we play this game up in the business school, I think it's quite likely we're going to get a lot of Alpha's chosen, right? But if we played this game up in let's say the Divinity School, all right and I'm guessing that Travis' answer is reflecting what you guys are reasoning here. If you played in the Divinity School, you might think that people in the Divinity School might care about other people's grades, right? There might be ethical reasons--perfectly good, sensible, ethical reasons--for choosing Beta in this game. There might be other reasons as well, but that's perhaps the reason to focus on. And perhaps, the lesson I want to draw out of this is that right now this is not a game. Right now we have actions, strategies for people to take, and we know what the outcomes are, but we're missing something that will make this a game. What are we missing here?Student: Objectives.Professor Ben Polak: We're missing objectives. We're missing payoffs. We're missing what people care about, all right. So we can't really start analyzing a game until we know what people care about, and until we know what the payoffs are. Now let's just say something now, which I'll probably forget to say in any other moment of the class, but today it's relevant.Game Theory, me, professors at Yale, cannot tell you what your payoff should be. I can't tell you in a useful way what it is that your goals in life should be or whatever. That's not what Game Theory is about. However, once we know what your payoffs are, once we know what your goals are, perhaps Game Theory can you help you get there.So we've had two different kinds of payoffs mentioned here. We had the kind of payoff where we care about our own grade, and Travis has mentioned the kind of payoff where you might care about other people's grades. And what we're going to do today is analyze this game under both those possible payoffs. To start that off, let's put up some possible payoffs for the game. And I promise we'll come back and look at some other payoffs later. We'll revisit the Divinity School later.All right, so here once again is our same matrix with me and my pair, choosing actions Alpha and Beta, but this time I'm going to put numbers in here. And some of you will perhaps recognize these numbers, but that's not really relevant for now. All right, so what's the idea here? Well the first idea is that these numbers represent utiles or utilities. They represent what these people are trying to maximize, what they're to achieve, their goals.The idea is - just to compare this to the outcome matrix - for the person who's me here, (A,C) yields a payoff of--(A,C) is this box--so (A,C) yields a payoff of three, whereas (B-,B-) yields a payoff of 0, and so on. So what's the interpretation? It's the first interpretation: the natural interpretation that a lot of you jumped to straight away. These are people--people with these payoffs are people--who only care about their own grades. They prefer an A to a B+, they prefer a B+ to a B-, and they prefer a B- to a C. Right, I'm hoping I the grades in order, otherwise it's going to ruin my curve at the end of the year. So these people only care about their own grades. They only care about their own grades.What do we call people who only care about their own grades? What's a good technical term for them? In England, I think we refer to these guys - whether it's technical or not - as "evil gits." These are not perhaps the most moral people in the universe. So now we can ask a different question. Suppose, whether these are actually your payoffs or not, pretend they are for now. Suppose these are all payoffs. Now we can ask, not what did you do, but what should you do? Now we have payoffs that can really switch the question to a normative question: what should you do? Let's come back to - was it Clara Elise--where was Clara Elise before? Let's get the mike on you again. So just explain what you did and why again.Student: Why I chose Alpha?Professor Ben Polak: Yeah, stand up a second, if that's okay.Student: Okay.Professor Ben Polak: You chose Alpha; I'm assuming these were roughly your payoffs, more or less, you were caring about your grades.Student: Yeah, I was thinking -Professor Ben Polak: Why did you choose Alpha?Student: I'm sorry?Professor Ben Polak: Why did you choose Alpha? Just repeat what you said before.Student: Because I thought the payoffs - the two different payoffs that I could have gotten--were highest if I chose Alpha.Professor Ben Polak: Good; so what Clara Elise is saying--it's an important idea--is this (and tell me if I'm paraphrasing you incorrectly but I think this is more or less what you're saying): is no matter what the other person does, no matter what the pair does, she obtains a higher payoff by choosing Alpha. Let's just see that. If the pair chooses Alpha and she chooses Alpha, then she gets 0. If the pair chooses Alpha and she chose Beta, she gets -1. 0 is bigger than -1. If the pair chooses Beta, then if she chooses Alpha she gets 3, Beta she gets 1, and 3 is bigger than 1. So in both cases, no matter what the other person does, she receives a higher payoff from choosing Alpha, so she should choose Alpha. Does everyone follow that line of reasoning? That's a stronger line of reasoning then the reasoning we had earlier. So the woman, I have immediately forgotten the name of, in the red shirt, whose name was -Student: Courtney.Professor Ben Polak: Courtney, so Courtney also gave a reason for choosing Alpha, and it was a perfectly good reason for choosing Alpha, nothing wrong with it, but notice that this reason's a stronger reason. It kind of implies your reason.So let's get some definitions down here. I think I can fit it in here. Let's try and fit it in here.Definition: We say that my strategy Alpha strictly dominates my strategy Beta, if my payoff from Alpha is strictly greater than that from Beta, [and this is the key part of the definition], regardless of what others do.Shall we just read that back? "We say that my strategy Alpha strictly dominates my strategy Beta, if my payoff from Alpha is strictly greater than that from Beta, regardless of what others do." Now it's by no means my main aim in this class to teach you jargon. But a few bits of jargon are going to be helpful in allowing the conversation to move forward and this is certainly one. "Evil gits" is maybe one too, but this is certainly one.Let's draw out some lessons from this. Actually, so you can still read that, let me bring down and clean this board. So the first lesson of the class, and there are going to be lots of lessons, is a lesson that emerges immediately from the definition of a dominated strategy and it's this. So Lesson One of the course is:do not play a strictly dominated strategy. So with apologies to Strunk and White, this is in the passive form, that's dominated, passive voice. Do not play a strictly dominated strategy. Why? Somebody want to tell me why? Do you want to get this guy? Stand up - yeah.Student: Because everyone's going to pick the dominant outcome and then everyone's going to get the worst result - the collectively worst result.Professor Ben Polak: Yeah, that's a possible answer. I'm looking for something more direct here. So we look at the definition of a strictly dominated strategy. I'm saying never play one. What's a possible reason for that? Let's - can we get the woman there?Student: [inaudible]Professor Ben Polak: "You'll always lose." Well, I don't know: it's not about winning and losing. What else could we have? Could we get this guy in the pink down here?Student: Well, the payoffs are lower.Professor Ben Polak: The payoffs are lower, okay. So here's an abbreviated version of that, I mean it's perhaps a little bit longer. The reason I don't want to play a strictly dominated strategy is, if instead, I play the strategy that dominates it, I do better in every case. The reason I never want to play a strictly dominated strategy is, if instead I play the strategy that dominates it, whatever anyone else does I'm doing better than I would have done. Now that's a pretty convincing argument. That sounds like a convincing argument. It sounds like too obvious even to be worth stating in class, so let me now try and shake your faith a little bit in this answer.。
lecture3(博弈论讲义(Carnegie Mellon University))

for all si Si . That is, si* solves Maximize Subject to
* * * * ui ( s1 ,..., si 1, si , si 1,..., sn ) si Si
May 21, 2003 73-347 Game Theory--Lecture 3 2
Today’s Agenda
Review of previous classes Nash equilibrium Best response function
Use best response function to find Nash
(Confess, Confess) is a Nash equilibrium. Prisoner 1
May 21, 2003
Prisoner 2
Mum
Mum Confess -1 , 0 , -1 -9
Confess
-9 , -6 , 0 -6
10
73-347 Game Theory--Lecture 3
equilibria Examples
May 21, 2003
73-347 Game Theory--Lecture 3
3
Review
The normal-form (or strategic-form) representation of a
game G specifies: A finite set of players {1, 2, ..., n}, players’ strategy spaces S1 S2 ... Sn and their payoff functions u1 u2 ... un where ui : S1 × S2 × ...× Sn→R.
博弈论基础知识.ppt

The theory of rational choice
–
The action chosen by a decision-maker is at least as good, according to her preferences, as every other available action.
第一讲( 第一讲(续) 博弈论基础知识
The theory of rational choice
Actions Preferences and payoff functions
epresents a decision-maker’s preferences if, for any actions a in A and b in A, u(a)>u(b) if and only if the decision-maker prefers a to b. – A decision-maker’s preferences convey only ordinal information.
Conditional probability and Bayes’s rule
– –
Conditional probability and Bayes’s rule
条件概率: P ( B | A) = P ( AB ) P ( A)
全概率公式:设试验 E的样本空间为 S, A为 E的事件, B1, B2, Bn为 S的一个划分, P ( Bi ) > ( i = 1, 2,⋯ , n), 0 则: P ( A) = P ( A | B1 ) P ( B1 ) + P ( A | B2 ) P ( B2 ) + ⋯ + P ( A | Bn ) P ( Bn ) 贝叶斯公式: P ( Bi | A) = P ( A | Bi ) P ( Bi )
section2(博弈论讲义(Harvard University))

Section 2: Externalities
Alexis Diamond adiamond@
Agenda
• Key terms and definitions • Complementarity and cross-partial derivatives)
• πi = 2(i + j + cij) - i2 • πj = 2(i + j + cij) - j2
• Best response functions:
– d (πi)/di = 2+2cj - 2i – • i*=1+cj … This is our BR function for agent i
Nash Eqm (TR & DC)
strategy profiles
Pareto Optima
Set of Rationalizable Strategies: {T,D} x {C,R} Set of Weakly Congruent Strategies: {T,D} x {C,R} Set of Best Response Complete: {T,D} x {C, R} or {T,D} x {L,C,R} Set of Congruent Strategies: {T,D} x {C, R}
Cournot Oligopoly: Explanation
Positive Cross-Partial Derivatives = Complementarity
π1 = (1000 - q1 - q2)q1 - 100q1 d (π1)/dq1 = 1000 - 2q1 - q2 - 100 d ((π1)/dq1)/dq2 = -1 < 0 • Why? What does this mean?
lecture5(博弈论讲义(Carnegie Mellon University))

9
Bertrand model of duopoly (homogeneous products)
Two firms: firm 1 and firm 2. Each firm chooses the price for its product without
knowing the other firm has chosen. The prices are denoted by p1 and p2, respectively. The quantity that consumers demand from firm 1: q1(p1, p2) = a – p1 if p1 < p2 ; = (a – p1)/2 if p1 = p2 ; =0, ow. The quantity that consumers demand from firm 2: q2(p1, p2) = a – p2 if p2 < p1 ; = (a – p2)/2 if p1 = p2 ; =0, ow. The cost to firm i of producing quantity qi is Ci(qi)=cqi.
May 23, 2003
73-347 Game Theory--Lecture 5
4
Bertrand model of duopoly (differentiated products)
The normal-form representation:
Set of players: { Firm 1, Firm 2} Sets of strategies: S1=[0, +∞), S2=[0, +∞) Payoff functions: u1(p1, p2)=(a – p1 + bp2 )(p1 – c) u2(p1, p2)=(a – p2 + bp1 )(p2 – c)
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
Week 11: Game TheoryRequired Reading: Schotter pp. 229 - 260Lecture Plan1. The Static Game TheoryNormal Form GamesSolution Techniques for Solving Static Games Dominant StrategyNash Equilibrium2. Prisoner’s Dilemma3. Decision AnalysisMaximim CriteriaMinimax Criteria4. Dynamic One-Off GamesExtensive Form GamesThe Sub-Game Perfect Nash Equilibrium1. The static Game TheoryStatic games: the players make their move in isolation without knowing what other players have done1.1 Normal Form GamesIn game theory there are two ways in which a game can be represented.1st) The normal form game or strategic form game2nd) The extensive form gameA normal form game is any game where we can identity the following three things:1. Players:2. The strategies available to each player.3. The Payoffs. A payoff is what a player will receive at the endof the game contingent upon the actions of all players in the game.Suppose that two people (A and B) are playing a simple game. A will write one of two words on a piece of paper, “Top” or “Bottom”. At the same time, B will independently write “left” or “right” on a piece of paper. After they do this, the papers will be examined and they will get the payoff depicted in Table 1.Table 1If A says top and B says left, then we examine the top-left corner of the matrix. In this matrix, the payoff to A(B) is the first(Second) entry in the box. For example, if A writes “top” and B writes “left” payoff of A = 1 payoff of B = 2.What is/are the equilibrium outcome(s) of this game?1.2Nash Equilibrium Approach to Solving Static GamesNash equilibrium is first defined by John Nash in 1951 based on the work of Cournot in 1893.A pair of strategy is Nash equilibrium if A's choice is optimal given B's choice, and B's choice is optimal given A's choice. When this equilibrium outcome is reached, neither individual wants to change his behaviour.Finding the Nash equilibrium for any game involves two stages.1) identify each optimal strategy in response to what the other players might do.Given B chooses left, the optimal strategy for A isGiven B chooses right, the optimal strategy for A isGiven A chooses top, the optimal strategy for B isGiven A chooses bottom, the optimal strategy for B isWe show this by underlying the payoff element for each case.2) a Nash equilibrium is identified when all players are player their optimal strategies simultaneouslyIn the case of Table 2,If A chooses top, then the best thing for B to do is to choose left since the payoff to B from choosing left is 1 and the payoff from choosing right is 0. If B chooses left, then the best thing for A to do is to choose top as A will get a payoff of 2 rather than 0.Thus if A chooses top B chooses left. If B chooses left, A chooses top. Therefore we have a Nash equilibrium: each person is making optimal choice, given the other person's choice.If the payoff matrix changes as:Table 2then what is the Nash equilibrium?Table 3If the payoffs are changed as shown in Table 32. Prisoners’ dilemm aPareto Efficiency: An allocation is Pareto efficient if goods cannot be reallocated to make someone better off without making someone else worse off.Two prisoners who were partners in a crime were being questioned in separate rooms. Each prisoner has a choice of confessing to the crime (implicating the other) or denying. If only one confesses, then he would go free and his partner will spend 6 months in prison. If both prisoners deny, then both would be in the prison for 1 month. If both confess, they would both be held for three months. The payoff matrix for this game is depicted in Table 4.Table 4The equilibrium outcome3. Decision AnalysisLet N=1 to 4 a set of possible states of nature, and let S=1 to 4be a set of strategy decided by you. Now you have to decide which strategy you have to choose given the following payoff matrix.Table 5S=YouN=OpponentIn this case you don't care the payoff of your opponent i.e. nature.3.1 The Maximin Decision Rule or Wald criterionWe look for the minimum pay-offs in each choice and then maximising the minimum pay-offLet us highlight the mimimum payoff for each strategy.3.2 The Minimax Decision Rule or Savage criterionOn this rule we need to compute the losses or regret matrix from the payoff matrix. The losses are defined as the difference between the actual payoff and what that payoff would have been had the correct strategy been chosen.Regret/Loss = Max. payoff in each column – actual payoffFor example of N=1 occurs and S=1 is chosen, the actual gain = 2 from Table 3. However, the best action given N=1 is also to choose S=1 which gives the best gain = 2. For (N=1, S=1) regret = 0.If N=1 occurs but S=2 is chosen, the actual gain = 1. However, the best action given N=1 is also to choose S=1 which gives the best gain = 2. For (N=1, S=2) regret = 2-1.Following the similar analysis, we can compute the losses for each N and S and so can compute the regret matrix.Table 6: Regret MatrixAfter computing the regret matrix, we look for the maximum regret of each strategy and then taking the minimum of these.Minimax is still very cautious but less so than the maximin.4. Dynamic one-off GamesA game can be dynamic for two reasons. First, players may be ableto observe the actions of other players before deciding upon theiroptimal response. Second, one-off game may be repeated a number of times.4.1 Extensive Form GamesDynamic games cannot be represented by payoff matrices we have touse a decision tree (extensive form) to represent a dynamic game.Start with the concept of dynamic one-off game the game can beplayed for only one time but players can condition their optimal actions on what other players have done in the past.Suppose that there are two firms (A and B) that are considering whether or not to enter a new market. If both firms enter the market,then they will make a loss of $10 mil. If only one firm enters the market, it will earn a profit of $50 mil. Suppose also that Firm B observes whether Firm A has entered the market before it decides what to do.Any extensive form game has the following four elements in common:Nodes: This is a position where the players must take a decision.The first position is called the initial node, and each node is labelled so as to identify who is making the decision.Branches: These represent the alternative choices that the person faces and so correspond to available actions.Payoff Vectors: These represent the payoffs for each player, with the payoffs listed in the order of players. When we reach a payoffvector the game ends.In period 1, Firm A makes its decisions. This is observed by Firm B which decides to enter or stay out of the market in period 2. In this extensive form game, Firm B’s decision nodes are the sub-game. This means that firm B observes Firm A’s action before making its own decision.4.2 Subgame Perfect Nash EquilibriumSubgame perfect Nash equilibrium is the predicted solution to a dynamic one-off game. From the extensive form of this game, we can observe that there are two subgames, one starting from each of Firm B’s decision nodes.How could we identify the equilibrium outcomes?In applying this principle to this dynamic game, we start with the last period first and work backward through successive nodes until we reach the beginning of the game.Start with the last period of the game first, we have two nodes. At each node, Firm B decides whether or not entering the market based on what Firm A has already done.For example, at the node of “Firm A enters”, Firm B will either make a loss of –$10mil (if it enters) or receive “0” payoff (if it stays out); these are shown by the payoff vectors (-10, -10) and (50, 0). If Firm B is rational, it will stays outThe node “Firm A enters” can be replaced by the vector (50, 0).At the second node “Firm A stays out”, Firm A h as not entered the market. Thus, Firm B will either make a profit of $50mil (if it enters) or receive “0” payoff (if it stays out); these are shown by the payoff vectors (0, 50) and (0, 0). If Firm B is rational, it will enter thus we could rule out the possibility of both firms stay outWe can now move back to the initial node. Here Firm A has to decide whether or not to enter. If Firm B is rational, it is known that the game will never reach the previously “crossed” vectors. Firm A also knows that if it enters, the game will eventually end at (A enters, B stays out) where A gets 50 and B gets 0. On the other hand, if Firm A stays out, the game will end at (A stays out, B enters) where A gets 0 and B gets 50 Firm A should enter the market at the first stage. The eventual outcome is (A enters, B stays out)How to find a subgame perfect equilibrium of a dynamic one-off game?1. Start with the last period of the game cross out the irrelevant payoff vectors.2. Replace the preceding nodes by the uncrossed payoff vectorsuntil you reach the initial node.3. The only uncrossed payoff vector(s) is the subgame perfect Nash equilibrium.。