Approximation and optimality necessary conditions in relaxed stochastic control problems (Q995846)

From MaRDI portal
scientific article
Language Label Description Also known as
English
Approximation and optimality necessary conditions in relaxed stochastic control problems
scientific article

    Statements

    Approximation and optimality necessary conditions in relaxed stochastic control problems (English)
    0 references
    0 references
    0 references
    0 references
    10 September 2007
    0 references
    Summary: We consider a control problem where the state variable is a solution of a stochastic differential equation (SDE) in which the control enters both the drift and the diffusion coefficient. We study the relaxed problem for which admissible controls are measure-valued processes and the state variable is governed by an SDE driven by an orthogonal martingale measure. Under some mild conditions on the coefficients and pathwise uniqueness, we prove that every diffusion process associated to a relaxed control is a strong limit of a sequence of diffusion processes associated to strict controls. As a consequence, we show that the strict and the relaxed control problems have the same value function and that an optimal relaxed control exists. Moreover we derive a maximum principle of the Pontryagin type, extending the well-known Peng stochastic maximum principle to the class of measure-valued controls.
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    stochastic differential equation
    0 references
    Peng maximum principle
    0 references
    maximum principle of Pontryagin type
    0 references