Dynamic programming for ergodic control with partial observations. (Q2574544): Difference between revisions
From MaRDI portal
Latest revision as of 12:29, 11 June 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Dynamic programming for ergodic control with partial observations. |
scientific article |
Statements
Dynamic programming for ergodic control with partial observations. (English)
0 references
29 November 2005
0 references
The paper derives a dynamic programming principle for optimal control of a partially observed Markov process taking values in a Euclidean space. The minimized functional is that of average (ergodic) costs over infinite horizon. The control space is compact. The problem is addressed by approximating the original ergodic cost functional by a family of discounted cost functionals with discount factors converging to unity. The dynamic programming principle inequalities are first derived in discrete time and the result is then carried over to partially observed Markov semimartingales in continuous time. The construction of optimal controls proceeds in the following steps: 1. restating the problem by means of a separation principle which makes the control process adapted to the process of observations, 2. changing the probability measure in order to eliminate variability in the marginal distribution of the observation process, 3. introducing a stability assumption for the state process in a Lyapunov function form, 4. embedding the state process into another one with a ``doubled'' range of values, for which an accessible atom exists. The argument draws on earlier results of the same author concerning optimal ergodic control of partially observed finite Markov chains.
0 references
Markov process
0 references
ergodic cost
0 references
0 references
0 references
0 references
0 references