An adjusted payoff-based procedure for normal form games
From MaRDI portal
Publication:2833112
Abstract: We study a simple adaptive model in the framework of an N -player normal form game. The model consists of a repeated game where the players only know their own action space and their own payoff scored at each stage, not those of the other agents. Each player, in order to update her mixed action, computes the average vector payoff she has obtained by using the number of times she has played each pure action. The resulting stochastic process is analyzed via the ODE method from stochastic approximation theory. We are interested in the convergence of the process to rest points of the related continuous dynamics. Results concerning almost sure convergence and convergence with positive probability are obtained and applied to a traffic game. We also provide some examples where convergence occurs with probability zero.
Recommendations
- ON THE PAYOFF REPRESENTATIONS OF NORMAL FORM GAMES
- Refinements of rationalizability for normal-form games
- The theory of normal form games from the differentiable viewpoint
- The refined best-response correspondence in normal form games
- On the core of normal form games with a continuum of players
- Essential equilibrium in normal-form games with perturbed actions and payoffs
- Decompositions and potentials for normal form games
- Quantal response equilibria for normal form games
- Essential equilibria in normal-form games
- Normal form structures in extensive form games
Cites work
- scientific article; zbMATH DE number 3616736 (Why is no real title available?)
- scientific article; zbMATH DE number 1795161 (Why is no real title available?)
- scientific article; zbMATH DE number 1405930 (Why is no real title available?)
- A Simple Adaptive Procedure Leading to Correlated Equilibrium
- A Stochastic Approximation Method
- A behavioral learning process in games
- A payoff-based learning procedure and its application to traffic games
- Adaptive game playing using multiplicative weights
- Do stochastic algorithms avoid traps?
- Individual Q-Learning in Normal Form Games
- Learning in perturbed asymmetric games
- Learning through reinforcement and replicator dynamics
- Nonconvergence to unstable points in urn models and stochastic approximations
- On the convergence of reinforcement learning
- Potential games
- Revisiting log-linear learning: asynchrony, completeness and payoff-based implementation
- The logit-response dynamics
- The statistical mechanics of strategic interaction
- Urn models, replicator processes, and random genetic drift
Cited in
(5)- Learning and equilibrium transitions: stochastic stability in discounted stochastic fictitious play
- Identifying behaviorally robust strategies for normal form games under varying forms of uncertainty
- Penalty-regulated dynamics and robust learning procedures in games
- Payoff-based dynamics for multiplayer weakly acyclic games
- On the stability of an adaptive learning dynamics in traffic games
This page was built for publication: An adjusted payoff-based procedure for normal form games
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2833112)