scientific article; zbMATH DE number 1461223
zbMATH Open0960.93001MaRDI QIDQ4485809FDOQ4485809
Authors: K. Najim, Eduardo Gomez-Ramirez, Alexander S. Poznyak
Publication date: 19 June 2000
Title of this publication is not available (Why is that?)
Recommendations
stochastic approximationadaptive controlinfinite horizon optimal controllearning automatagradient optimizationlinear inequality constraintscontrolled Markov chaindirect approaches
Stochastic approximation (62L20) Optimal stochastic control (93E20) Stochastic learning and adaptive control (93E35) Research exposition (monographs, survey articles) pertaining to systems and control theory (93-02)
Cited In (25)
- Minimizing the learning loss in adaptive control of Markov chains under the weak accessibility condition
- Handling a Kullback--Leibler divergence random walk for scheduling effective patrol strategies in Stackelberg security games
- Solving the cost to go with time penalization using the Lagrange optimization approach
- A Tikhonov regularization parameter approach for solving Lagrange constrained optimization problems
- Learning Control of Dynamical Systems Based on Markov Decision Processes: Research Frontiers and Outlooks
- Optimization based on a team of automata with binary outputs
- Adapting attackers and defenders patrolling strategies: a reinforcement learning approach for Stackelberg security games
- Learning Machiavellian strategies for manipulation in Stackelberg security games
- Controller exploitation-exploration reinforcement learning architecture for computing near-optimal policies
- Recursive estimation of high-order Markov chains: approximation by finite mixtures
- Optimization problems in chemical reactions using continuous-time Markov chains
- Using the extraproximal method for computing the shortest-path mixed Lyapunov equilibrium in Stackelberg security games
- Observer and control design in partially observable finite Markov chains
- Constructing the Pareto front for multi-objective Markov chains handling a strong Pareto policy approach
- Computing the Stackelberg/Nash equilibria using the extraproximal method: convergence analysis and implementation details for Markov chains games
- Using the Manhattan distance for computing the multiobjective Markov chains problem
- Sparse mean-variance customer Markowitz portfolio optimization for Markov chains: a Tikhonov's regularization penalty approach
- Saddle-point calculation for constrained finite Markov chains
- Computing the strong Nash equilibrium for Markov chains games
- Oja's algorithm for graph clustering, Markov spectral decomposition, and risk sensitive control
- A continuous-time Markov Stackelberg security game approach for reasoning about real patrol strategies
- A Tikhonov regularized penalty function approach for solving polylinear programming problems
- Adaptive policy for two finite Markov chains zero-sum stochastic game with unknown transition matrices and average payoffs
- Computing the strong \(L_p\)-Nash equilibrium for Markov chains games: convergence and uniqueness
- Setting Nash Versus Kalai–Smorodinsky Bargaining Approach: Computing the Continuous-Time Controllable Markov Game
This page was built for publication:
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4485809)