Robust risk-aware reinforcement learning

From MaRDI portal
Publication:5065087

DOI10.1137/21M144640XzbMATH Open1484.91426arXiv2108.10403OpenAlexW4214588671MaRDI QIDQ5065087FDOQ5065087


Authors: Sebastian Jaimungal, Silvana M. Pesenti, Ye Sheng Wang, Hariom Tatsat Edit this on Wikidata


Publication date: 18 March 2022

Published in: SIAM Journal on Financial Mathematics (Search for Journal in Brave)

Abstract: We present a reinforcement learning (RL) approach for robust optimisation of risk-aware performance criteria. To allow agents to express a wide variety of risk-reward profiles, we assess the value of a policy using rank dependent expected utility (RDEU). RDEU allows the agent to seek gains, while simultaneously protecting themselves against downside risk. To robustify optimal policies against model uncertainty, we assess a policy not by its distribution, but rather, by the worst possible distribution that lies within a Wasserstein ball around it. Thus, our problem formulation may be viewed as an actor/agent choosing a policy (the outer problem), and the adversary then acting to worsen the performance of that strategy (the inner problem). We develop explicit policy gradient formulae for the inner and outer problems, and show its efficacy on three prototypical financial problems: robust portfolio allocation, optimising a benchmark, and statistical arbitrage.


Full work available at URL: https://arxiv.org/abs/2108.10403




Recommendations




Cites Work


Cited In (8)

Uses Software





This page was built for publication: Robust risk-aware reinforcement learning

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5065087)