Reward tampering problems and solutions in reinforcement learning: a causal influence diagram perspective
DOI10.1007/S11229-021-03141-4zbMATH Open1529.68309arXiv1908.04734OpenAlexW3165436200MaRDI QIDQ6182771FDOQ6182771
Authors: Tom Everitt, Marcus Hutter, Ramana Kumar, Victoria Krakovna
Publication date: 26 January 2024
Published in: Synthese (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1908.04734
Recommendations
- Reward is enough
- Learning reward machines: a study in partially observable reinforcement learning
- Reward machines: exploiting reward function structure in reinforcement learning
- Framing reinforcement learning from human reward: reward positivity, temporal discounting, episodicity, and performance
- Specification-guided reinforcement learning
Learning and adaptive systems in artificial intelligence (68T05) Causal inference from observational studies (62D20) General considerations in statistical decision theory (62C05) Agent technology and artificial intelligence (68T42)
Cites Work
- Planning and acting in partially observable stochastic domains
- Causality. Models, reasoning, and inference
- Complete identification methods for the causal hierarchy
- Multi-agent influence diagrams for representing and solving games.
- Reinforcement learning. An introduction
- Representing and solving decision problems with limited information
- General time consistent discounting
- A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play
- Artificial general intelligence 2008. Proceedings of the 1st AGI conference.
Cited In (1)
This page was built for publication: Reward tampering problems and solutions in reinforcement learning: a causal influence diagram perspective
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6182771)