In game theory, a concave game is a generalization of the normal-form game defined by Rosen. [1] He extended the theorem on existence of a Nash equilibrium, which John Nash originally proved for normal-form games, to concave games.
We will describe the generalization step by step.
1. In a normal-form game, each player i can choose one of mi pure actions. The set of strategies available to each player is the set of lotteries over the pure actions, which is a simplex in Rmi.
2. In a normal-form game, the set of strategies available to each player is independent of the strategies chosen by other players. This means that the set of all possible strategy profiles (denoted S) is a Cartesian product of the sets of strategies available to the players (in other words, the constraints on each player's choices are orthogonal).
3. In a normal-form game, the utility function of each player i (denoted ui - a real-valued function on S) is a linear function in each of its components, for any fixed value of the other components (for each agent j, given any choice of strategies by the other agents, the payoff of agent i is a linear function in the probabilities by which agent j chooses his pure actions).
If the above conditions hold (that is, the space S of possible strategy profiles is convex, and each payoff function ui is continuous in all strategies and of all players concave in the strategy of player i), then an equilibrium exists. [1] : Thm.1 The proof uses the Kakutani fixed-point theorem.
Rosen also proved that, under certain technical conditions which include strict concavity, the equilibrium is unique. The conditions are explained below.
Assume that the space S of allowed strategy profiles is defined by some k functions (representing constraints), that is, . In a normal-form game, the constraint functions are linear functions defining the simpices of the players; in a concave game, the hj can be any concave function of x. For the uniqueness proof, it is also required that each hj has continuous first derivatives at any x in S.
Assume also that, for each player i, the utility function ui has continuous first derivatives at the components of xi (that is, each player's utility is continuously differentiable in the player's own strategy).
For any fixed positive vector r>0 in Rn, define the r-weighted-sum of the players' payoffs:
.
Define the pseudogradient of f:
where is the gradient of ui with respect to the xi components. So is a vector in Rmi and g(x,r) is a vector in Rm.
We say that f is diagonally-strictly-concave at r if for every x,y in S:
A sufficient condition for f being diagonally-strictly-concave at r is that the symmetric matrix G(x,r)+GT(x,r) is negative definite. [1] : Thm.5 See the paper for an example for bimatrix games.
Consider a dynamic model of a concave game in which each player changes his strategy such that the strategy-profile remains in S, and subject to that, his utility increases. Each player i changes his strategy at rate ri in the direction of the gradient defining the maximum utility increase subject to the constraints. This results in a set of n differential equations. For any concave game and any staring point in S, this set of differential equations has a continuous solution x(t) that remains in S for all t>0. [1] : Thm.7
If, in addition, the matrix G(x,r)+GT(x,r) is negative definite for all x in S, then the solution x(t) converges to an equilibrium point as t approaches infinity. [1] : Thm.8
If, in addition, the matrix G(x,r)+GT(x,r) is negative definite for all x in S, then the solution x(t) converges to the unique r-normalized equilibrium point. [1] : Thm.9
Takashi Ui [3] shows that the same condition that guarantees uniqueness of a Nash equilibrium also guarantees uniqueness of a Correlated equilibrium. Moreover, an even weaker condition guarantees the uniqueness of a correlated equilibrium - a generalization of a condition proved by Abraham Neyman. [4]
Based on the above results, it is possible to compute equilibrium points for concave games using gradient methods for convex optimization. [1]
Papadimitriou, Vlatakis-Gkaragkounis and Zampetakis [5] prove that computing an equilibrium in a concave game is PPAD-complete. In fact, they prove that the problem is in PPAD even for general concave games, and it is PPAD-hard even in the special case of strongly-concave utilities that can be expressed as multivariate polynomials of a constant degree with axis-aligned box constraints.
Arrow and Hurwicz [6] presented gradient method s for solving two-player zero-sum game s with non-linear utilities.
Krawczyk and Uryasev [7] studied infinite games with nonlinear utilities and coupled constraints. Starting from an existing Relaxation method, they improved it by adding steepest-descent step-size control and other improvements, and proved that it indeed converges to an equiilbrium. They tested their algorithm numerically on several applications, such as pollution of a river basin, and showed that it converges quickly on a wide range of parameters.
Krawczyk [8] explains numerical methods converging to an equilibrium, focusing on the case of coupled constraints. He presents several application examples using a Matlab suite called NIRA.
Chernov [9] presents two numerical search approaches for computing equilibrium points, that have guaranteed convergence without additional requirements on the objective functions: (1) using the Hooke-Jeeves method for residual function minimization (2) an intermediate between the relaxation algorithm and the Hooke-Jeeves method of configurations. Convergence is proved for one-dimensional sets of players strategies. The approaches are tested using numerical experiments.
Flam and Ruzcynzki [10] define a convex-concave game. This is a game in which the space S of strategy profiles is convex, as in Rosen's definition. But instead of requiring smoothness and other conditions on g(x,r), they allow non-smooth data, and only require that the following function is convex in x and concave in y: .
For such convex-concave games, they present two algorithms for finding equilibria, both using partial regularization s and relaxed subgradient projections. They prove that these algorithms converge.