Multi-objective reinforcement learning through continuous Pareto manifold approximation
From MaRDI portal
Publication:2829188
Recommendations
- Multi-objective reinforcement learning using sets of Pareto dominating policies
- Efficient multi-objective reinforcement learning via multiple-gradient descent with iteratively discovered weight-vector sets
- A survey of multi-objective sequential decision-making
- Multi-Objective Decision Making
- scientific article; zbMATH DE number 2079783
Cited in
(11)- Multi-condition multi-objective optimization using deep reinforcement learning
- A framework for controllable Pareto front learning with completed scalarization functions and its applications
- scientific article; zbMATH DE number 5957207 (Why is no real title available?)
- The hard lessons and shifting modeling trends of COVID-19 dynamics: multiresolution modeling approach
- Expected policy gradients for reinforcement learning
- Efficient multi-objective reinforcement learning via multiple-gradient descent with iteratively discovered weight-vector sets
- Computing multiobjective Markov chains handled by the extraproximal method
- Necessary and sufficient Karush-Kuhn-Tucker conditions for multiobjective Markov chains optimality
- Using the Manhattan distance for computing the multiobjective Markov chains problem
- Determinantal reinforcement learning with techniques to avoid poor local optima
- Multi-objective reinforcement learning using sets of Pareto dominating policies
This page was built for publication: Multi-objective reinforcement learning through continuous Pareto manifold approximation
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2829188)