In game theory a max-dominated strategy is a strategy which is not a best response to any strategy profile of the other players. This is an extension to the notion of strictly dominated strategies, which are max-dominated as well.
A strategy of player is max-dominated if for every strategy profile of the other players there is a strategy such that . This definition means that is not a best response to any strategy profile , since for every such strategy profile there is another strategy which gives higher utility than for player .
If a strategy is strictly dominated by strategy then it is also max-dominated, since for every strategy profile of the other players , is the strategy for which .
Even if is strictly dominated by a mixed strategy it is also max-dominated.
A strategy of player is weakly max-dominated if for every strategy profile of the other players there is a strategy such that . This definition means that is either not a best response or not the only best response to any strategy profile , since for every such strategy profile there is another strategy which gives at least the same utility as for player .
If a strategy is weakly dominated by strategy then it is also weakly max-dominated, since for every strategy profile of the other players , is the strategy for which .
Even if is weakly dominated by a mixed strategy it is also weakly max-dominated.
A game is said to be max-solvable if by iterated elimination of max-dominated strategies only one strategy profile is left at the end.
More formally we say that is max-solvable if there exists a sequence of games such that:
Obviously every max-solvable game has a unique pure Nash equilibrium which is the strategy profile left in .
As in the previous part one can define respectively the notion of weakly max-solvable games, which are games for which a game with a single strategy profile can be reached by eliminating weakly max-dominated strategies. The main difference would be that weakly max-dominated games may have more than one pure Nash equilibrium, and that the order of elimination might result in different Nash equilibria.
Cooperate | Defect | |
Cooperate | -1, -1 | -5, 0 |
Defect | 0, -5 | -3, -3 |
Fig. 1: payoff matrix of the prisoner's dilemma |
The prisoner's dilemma is an example of a max-solvable game (as it is also dominance solvable). The strategy cooperate is max-dominated by the strategy defect for both players, since playing defect always gives the player a higher utility, no matter what the other player plays. To see this note that if the row player plays cooperate then the column player would prefer playing defect and go free than playing cooperate and serving one year in jail. If the row player plays defect then the column player would prefer playing defect and serve three years in jail rather than playing cooperate and serving five years in jail.
In any max-solvable game, best-reply dynamics ultimately leads to the unique pure Nash equilibrium of the game. In order to see this, all we need to do is notice that if is an elimination sequence of the game (meaning that first is eliminated from the strategy space of some player since it is max-dominated, then is eliminated, and so on), then in the best-response dynamics will be never played by its player after one iteration of best responses, will never be played by its player after two iterations of best responses and so on. The reason for this is that is not a best response to any strategy profile of the other players so after one iteration of best responses its player must have chosen a different strategy. Since we understand that we will never return to in any iteration of the best responses, we can treat the game after one iteration of best responses as if has been eliminated from the game, and complete the proof by induction.
1, 1 | 0, 0 |
1, 0 | 0, 1 |
0, 1 | 1, 0 |
It may come by surprise then that weakly max-solvable games do not necessarily converge to a pure Nash equilibrium when using the best-reply dynamics, as can be seen in the game on the right. If the game starts of the bottom left cell of the matrix, then the following best replay dynamics is possible: the row player moves one row up to the center row, the column player moves to the right column, the row player moves back to the bottom row, the column player moves back to the left column and so on. This obviously never converges to the unique pure Nash equilibrium of the game (which is the upper left cell in the payoff matrix).
In game theory, the Nash equilibrium, named after the mathematician John Nash, is the most common way to define the solution of a non-cooperative game involving two or more players. In a Nash equilibrium, each player is assumed to know the equilibrium strategies of the other players, and no one has anything to gain by changing only one's own strategy. The principle of Nash equilibrium dates back to the time of Cournot, who in 1838 applied it to competing firms choosing outputs.
The game of chicken, also known as the hawk–dove game or snowdrift game, is a model of conflict for two players in game theory. The principle of the game is that while the ideal outcome is for one player to yield, the individuals try to avoid it out of pride for not wanting to look like a "chicken". Each player taunts the other to increase the risk of shame in yielding. However, when one player yields, the conflict is avoided, and the game is for the most part over.
In game theory, the best response is the strategy which produces the most favorable outcome for a player, taking other players' strategies as given. The concept of a best response is central to John Nash's best-known contribution, the Nash equilibrium, the point at which each player in a game has selected the best response to the other players' strategies.
In game theory, the centipede game, first introduced by Robert Rosenthal in 1981, is an extensive form game in which two players take turns choosing either to take a slightly larger share of an increasing pot, or to pass the pot to the other player. The payoffs are arranged so that if one passes the pot to one's opponent and the opponent takes the pot on the next round, one receives slightly less than if one had taken the pot on this round, but after an additional switch the potential payoff will be higher. Therefore, although at each round a player has an incentive to take the pot, it would be better for them to wait. Although the traditional centipede game had a limit of 100 rounds, any game with this structure but a different number of rounds is called a centipede game.
This is an article about the math theory, to see the YouTube series of the same name, see MatPat.
In game theory, grim trigger is a trigger strategy for a repeated game.
In game theory, a solution concept is a formal rule for predicting how a game will be played. These predictions are called "solutions", and describe which strategies will be adopted by players and, therefore, the result of the game. The most commonly used solution concepts are equilibrium concepts, most famously Nash equilibrium.
In game theory, normal form is a description of a game. Unlike extensive form, normal-form representations are not graphical per se, but rather represent the game by way of a matrix. While this approach can be of greater use in identifying strictly dominated strategies and Nash equilibria, some information is lost as compared to extensive-form representations. The normal-form representation of a game includes all perceptible and conceivable strategies, and their corresponding payoffs, for each player.
In game theory, strategic dominance occurs when one strategy is better than another strategy for one player, no matter how that player's opponents may play. Many simple games can be solved using dominance. The opposite, intransitivity, occurs in games where one strategy may be better or worse than another strategy for one player, depending on how the player's opponents may play.
In game theory, rationalizability is a solution concept. The general idea is to provide the weakest constraints on players while still requiring that players are rational and this rationality is common knowledge among the players. It is more permissive than Nash equilibrium. Both require that players respond optimally to some belief about their opponents' actions, but Nash equilibrium requires that these beliefs be correct while rationalizability does not. Rationalizability was first defined, independently, by Bernheim (1984) and Pearce (1984).
In game theory, trembling hand perfect equilibrium is a type of refinement of a Nash equilibrium that was first proposed by Reinhard Selten. A trembling hand perfect equilibrium is an equilibrium that takes the possibility of off-the-equilibrium play into account by assuming that the players, through a "slip of the hand" or tremble, may choose unintended strategies, albeit with negligible probability.
In game theory, folk theorems are a class of theorems describing an abundance of Nash equilibrium payoff profiles in repeated games. The original Folk Theorem concerned the payoffs of all the Nash equilibria of an infinitely repeated game. This result was called the Folk Theorem because it was widely known among game theorists in the 1950s, even though no one had published it. Friedman's (1971) Theorem concerns the payoffs of certain subgame-perfect Nash equilibria (SPE) of an infinitely repeated game, and so strengthens the original Folk Theorem by using a stronger equilibrium concept: subgame-perfect Nash equilibria rather than Nash equilibria.
In game theory, a repeated game is an extensive form game that consists of a number of repetitions of some base game. The stage game is usually one of the well-studied 2-person games. Repeated games capture the idea that a player will have to take into account the impact of his or her current action on the future actions of other players; this impact is sometimes called his or her reputation. Single stage game or single shot game are names for non-repeated games.
In game theory, the purification theorem was contributed by Nobel laureate John Harsanyi in 1973. The theorem aims to justify a puzzling aspect of mixed strategy Nash equilibria: that each player is wholly indifferent amongst each of the actions he puts non-zero weight on, yet he mixes them so as to make every other player also indifferent.
Risk dominance and payoff dominance are two related refinements of the Nash equilibrium (NE) solution concept in game theory, defined by John Harsanyi and Reinhard Selten. A Nash equilibrium is considered payoff dominant if it is Pareto superior to all other Nash equilibria in the game. When faced with a choice among equilibria, all players would agree on the payoff dominant equilibrium since it offers to each player at least as much payoff as the other Nash equilibria. Conversely, a Nash equilibrium is considered risk dominant if it has the largest basin of attraction. This implies that the more uncertainty players have about the actions of the other player(s), the more likely they will choose the strategy corresponding to it.
Proper equilibrium is a refinement of Nash Equilibrium by Roger B. Myerson. Proper equilibrium further refines Reinhard Selten's notion of a trembling hand perfect equilibrium by assuming that more costly trembles are made with significantly smaller probability than less costly ones.
In game theory, an epsilon-equilibrium, or near-Nash equilibrium, is a strategy profile that approximately satisfies the condition of Nash equilibrium. In a Nash equilibrium, no player has an incentive to change his behavior. In an approximate Nash equilibrium, this requirement is weakened to allow the possibility that a player may have a small incentive to do something different. This may still be considered an adequate solution concept, assuming for example status quo bias. This solution concept may be preferred to Nash equilibrium due to being easier to compute, or alternatively due to the possibility that in games of more than 2 players, the probabilities involved in an exact Nash equilibrium need not be rational numbers.
In game theory, Mertens stability is a solution concept used to predict the outcome of a non-cooperative game. A tentative definition of stability was proposed by Elon Kohlberg and Jean-François Mertens for games with finite numbers of players and strategies. Later, Mertens proposed a stronger definition that was elaborated further by Srihari Govindan and Mertens. This solution concept is now called Mertens stability, or just stability.
The Price of Anarchy (PoA) is a concept in game theory and mechanism design that measures how the social welfare of a system degrades due to selfish behavior of its agents. It has been studied extensively in various contexts, particularly in auctions.
The Berge equilibrium is a game theory solution concept named after the mathematician Claude Berge. It is similar to the standard Nash equilibrium, except that it aims to capture a type of altruism rather than purely non-cooperative play. Whereas a Nash equilibrium is a situation in which each player of a strategic game ensures that they personally will receive the highest payoff given other players' strategies, in a Berge equilibrium every player ensures that all other players will receive the highest payoff possible. Although Berge introduced the intuition for this equilibrium notion in 1957, it was only formally defined by Vladislav Iosifovich Zhukovskii in 1985, and it was not in widespread use until half a century after Berge originally developed it.