Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning

From MaRDI portal
Publication:1606316

DOI10.1016/S0004-3702(99)00052-1zbMath0996.68151OpenAlexW2109910161MaRDI QIDQ1606316

Doina Precup, Richard S. Sutton, Satinder Pal Singh

Publication date: 24 July 2002

Published in: Artificial Intelligence (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1016/s0004-3702(99)00052-1




Related Items (46)

A time aggregation approach to Markov decision processesClustering Based Approximation Procedure for Semi-Markov Decision Processes with Incomplete State InformationOn Efficient Reinforcement Learning for Full-length Game of StarCraft IIMultiple Model-Based Reinforcement LearningProbabilistic inference for determining options in reinforcement learningActor-critic algorithms for hierarchical Markov decision processesA policy gradient method for semi-Markov decision processes with application to call admission controlHybrid MDP based integrated hierarchical Q-learningDetect, understand, act: a neuro-symbolic hierarchical reinforcement learning frameworkComputational Benefits of Intermediate Rewards for Goal-Reaching Policy LearningContinual curiosity-driven skill acquisition from high-dimensional video inputs for humanoid robotsExact decomposition approaches for Markov decision processes: a surveyAutomated Reinforcement Learning (AutoRL): A Survey and Open ProblemsDesigning decentralized controllers for distributed-air-jet MEMS-based micromanipulators by reinforcement learningAUTOMATIC COMPLEXITY REDUCTION IN REINFORCEMENT LEARNINGReward-respecting subtasks for model-based reinforcement learningOverlapping layered learningUnnamed ItemUnnamed ItemFrom Reinforcement Learning to Deep Reinforcement Learning: An OverviewAutomated imbalanced classification via layered learningOffline reinforcement learning with task hierarchiesReinforcement learning algorithms with function approximation: recent advances and applicationsGENERATING EFFECTIVE INITIATION SETS FOR SUBGOAL-DRIVEN OPTIONSHierarchical method for cooperative multiagent reinforcement learning in Markov decision processesUnnamed ItemImproving reinforcement learning by using sequence treesOptimal Curiosity-Driven Modular Incremental Slow Feature AnalysisAbstraction from demonstration for efficient reinforcement learning in high-dimensional domainsLearning and control of exploration primitivesDeep Reinforcement Learning: A State-of-the-Art WalkthroughPlanning and navigation as active inferenceReinforcement learning endowed with safe veto policies to learn the control of linked-multicomponent robotic systemsUnifying temporal and organizational scales in multiscale decision-makingA BIOLOGICALLY INSPIRED HIERARCHICAL REINFORCEMENT LEARNING SYSTEMAlgebraic results and bottom-up algorithm for policies generalization in reinforcement learning using concept latticesTransfer in variable-reward hierarchical reinforcement learningChallenges of real-world reinforcement learning: definitions, benchmarks and analysisUnnamed ItemReinforcement learning in the brainReinforcement Learning in Sparse-Reward Environments With Hindsight Policy GradientsHierarchical clustering optimizes the tradeoff between compositionality and expressivity of task structures for flexible reinforcement learningA Sufficient Statistic for Influence in Structured Multiagent EnvironmentsInduction and Exploitation of Subgoal Automata for Reinforcement LearningModel-based Reinforcement Learning: A SurveyReward Machines: Exploiting Reward Function Structure in Reinforcement Learning







This page was built for publication: Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning