A consistent, closed-loop solution for infinite-horizon, linear- quadratic, dynamic Stackelberg games (Q1108211)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | A consistent, closed-loop solution for infinite-horizon, linear- quadratic, dynamic Stackelberg games |
scientific article |
Statements
A consistent, closed-loop solution for infinite-horizon, linear- quadratic, dynamic Stackelberg games (English)
0 references
1988
0 references
We work out Kydland's suggestion to derive a (time-) consistent, closed- loop, solution in discrete-time, infinite-horizon, linear-quadratic, dynamic, 2-player, Stackelberg games by varying undetermined coefficients in players' linear, constant-coefficient, feedback policy rules [see \textit{F. Kydland}, J. Econ. Theory 15, 307-324 (1977; Zbl 0375.90089)]. The problem of consistency in linear-quadratic dynamic Stackelberg games was discussed, e.g., by \textit{M. Simaan} and \textit{J. B. Cruz} [J. Optimization Theory Appl. 11, 613-626 (1973; Zbl 0245.90039)], \textit{F. Kydland} [loc. cit. and Int. Econ. Rev. 16, 321-335 (1975; Zbl 0335.90061)] and \textit{F. Kydland} and \textit{E. C. Prescott} [J. Polit. Econ. 85, 473-491 (1977)]. Using a form of matrix-differential calculus [cf., \textit{J. R. Magnus} and \textit{H. Neudecker}, ``Matrix differential calculus with applications in statistics and econometrics'', Wiley, New York, to appear], we derive nonlinear, algebraic (nonrecursive), Riccati-like solution equations for such games. With the same method, we also derive analogous continuous- time results. For both the discrete- and continuous-time cases, we propose analogous, numerical, solution algorithms, which proceed with recursive versions of the algebraic Riccati-like equations. The algorithms are illustrated with a duopoly model. In the discrete-time case, the solution extends the dynamic-programming (d-p) feedback solution in this game. The present solution is more appealing than the d-p solution because in the present solution the leader is better off, yet, like in the d-p solution, the equilibrium is consistent. The leader is better off in the present solution because the players also optimize with respect to what may be called anticipative elements of control. ``Anticipative control'' is the ability of the leader, midstream in the game, to indirectly influence the evolution of the state vector backwards in time, by taking the follower's reactions into account. It is this effect which causes dynamic Stackelberg equilibria to be inconsistent, although, here inconsistency is avoided by assuming that the coefficient matrices in policy rules are independent of the initial state vector. By virtue of the principle of optimality upon which it is based, d-p cannot take such anticipative control effects into account.
0 references
dominant-player dynamic games
0 references
time-consistent solutions
0 references
anticipative non-Markovian control
0 references
solution in discrete-time
0 references
infinite-horizon, linear-quadratic, dynamic, 2-player, Stackelberg games
0 references
consistency
0 references
matrix-differential calculus
0 references
Riccati-like solution equations
0 references
duopoly model
0 references
dynamic-programming
0 references
feedback solution
0 references
Anticipative control
0 references