Planning and acting in partially observable stochastic domains
DOI10.1016/S0004-3702(98)00023-XzbMATH Open0908.68165DBLPjournals/ai/KaelblingLC98OpenAlexW2168359464WikidataQ56602944 ScholiaQ56602944MaRDI QIDQ72343FDOQ72343
Authors: Leslie Pack Kaelbling, Michael L. Littman, Anthony R. Cassandra, Leslie Pack Kaelbling, Michael L. Littman, Anthony R. Cassandra
Publication date: May 1998
Published in: Artificial Intelligence (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/s0004-3702(98)00023-x
Recommendations
- Planning in partially-observable switching-mode continuous domains
- scientific article
- Strong planning under partial observability
- Learning and planning in partially observable environments without prior domain knowledge
- Replanning in domains with partial information and sensing actions
- Randomized belief-space replanning in partially-observable continuous spaces
Cites Work
- Fast planning through planning graph analysis
- The complexity of stochastic games
- The complexity of mean payoff games on graphs
- Optimal control of Markov processes with incomplete state information
- A survey of algorithmic methods for partially observed Markov decision processes
- Solving H-horizon, stationary Markov decision problems in time proportional to log (H)
- Title not available (Why is that?)
- The Optimal Search for a Moving Target When the Search Path Is Constrained
- Title not available (Why is that?)
- State of the Art—A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms
- The Optimal Control of Partially Observable Markov Processes over the Infinite Horizon: Discounted Costs
- OPTIMAL CONTROL FOR PARTIALLY OBSERVABLE MARKOV DECISION PROCESSES OVER AN INFINITE HORIZON
- Title not available (Why is that?)
- Title not available (Why is that?)
- The Optimal Control of Partially Observable Markov Processes over a Finite Horizon
- Solution Procedures for Partially Observed Markov Decision Processes
- Title not available (Why is that?)
- Application of Jensen's inequality to adaptive suboptimal design
- A survey of solution techniques for the partially observed Markov decision process
Cited In (only showing first 100 items - show all)
- Probabilistic reasoning about epistemic action narratives
- Counterexample-guided inductive synthesis for probabilistic systems
- A dynamic epistemic framework for reasoning about conformant probabilistic plans
- Enforcing almost-sure reachability in POMDPs
- Approximability and efficient algorithms for constrained fixed-horizon POMDPs with durative actions
- Permissive planning: Extending classical planning to uncertain task domains.
- Computation of weighted sums of rewards for concurrent MDPs
- A reinforcement learning scheme for a partially-observable multi-agent game
- Robust almost-sure reachability in multi-environment MDPs
- Integration of AI and OR Techniques in Constraint Programming for Combinatorial Optimization Problems
- Multi-goal motion planning using traveling salesman problem in belief space
- Analyzing generalized planning under nondeterminism
- Finite-horizon LQR controller for partially-observed Boolean dynamical systems
- An integrated approach to solving influence diagrams and finite-horizon partially observable decision processes
- A reinforcement learning scheme for a partially-observable multi-agent game
- Quantitative controller synthesis for consumption Markov decision processes
- Computer Vision - ECCV 2004
- Tutorial series on brain-inspired computing. IV: Reinforcement learning: machine learning and natural learning
- Myopic bounds for optimal policy of POMDPs: an extension of lovejoy's structural results
- POMDP planning for robust robot control
- Markov decision processes with sequential sensor measurements
- Title not available (Why is that?)
- Reinforcement learning with limited reinforcement: using Bayes risk for active learning in POMDPs
- A sufficient statistic for influence in structured multiagent environments
- Representation and Timing in Theories of the Dopamine System
- A survey of inverse reinforcement learning: challenges, methods and progress
- Integration of reinforcement learning and optimal decision-making theories of the basal ganglia
- Induction and exploitation of subgoal automata for reinforcement learning
- Gradient-descent for randomized controllers under partial observability
- Meeting a deadline: shortest paths on stochastic directed acyclic graphs with information gathering
- Knowledge-based programs as succinct policies for partially observable domains
- Dynamic optimization over infinite-time horizon: web-building strategy in an orb-weaving spider as a case study
- Optimizing active surveillance for prostate cancer using partially observable Markov decision processes
- Algorithms and conditional lower bounds for planning problems
- A Fenchel-Moreau-Rockafellar type theorem on the Kantorovich-Wasserstein space with applications in partially observable Markov decision processes
- Geometric backtracking for combined task and motion planning in robotic systems
- Affect control processes: intelligent affective interaction using a partially observable Markov decision process
- Title not available (Why is that?)
- A semi-Markov decision model for recognizing the destination of a maneuvering agent in real time strategy games
- Simultaneous learning and planning in a hierarchical control system for a cognitive agent
- Learning to steer nonlinear interior-point methods
- Deliberative acting, planning and learning with hierarchical operational models
- Soft rumor control in mobile instant messengers
- Partially observable multistage stochastic programming
- Partially observable environment estimation with uplift inference for reinforcement learning based recommendation
- Evidential Markov decision processes
- Privacy stochastic games in distributed constraint reasoning
- Title not available (Why is that?)
- A logic for specifying stochastic actions and observations
- Reasoning and predicting POMDP planning complexity via covering numbers
- Minimax real-time heuristic search
- Planning and control in artificial intelligence: A unifying perspective
- Large-scale financial planning via a partially observable stochastic dual dynamic programming framework
- Gradient-based mixed planning with symbolic and numeric action parameters
- Simplified risk-aware decision making with belief-dependent rewards in partially observable domains
- A conflict-directed approach to chance-constrained mixed logical linear programming
- A Markovian model for the spread of the SARS-CoV-2 virus
- Title not available (Why is that?)
- Title not available (Why is that?)
- Performance prediction of an unmanned airborne vehicle multi-agent system
- Randomized belief-space replanning in partially-observable continuous spaces
- Using machine learning for decreasing state uncertainty in planning
- Optimal cost almost-sure reachability in POMDPs
- Partially observable game-theoretic agent programming in Golog
- Learning and planning in partially observable environments without prior domain knowledge
- Recursively modeling other agents for decision making: a research perspective
- Reasoning about uncertain parameters and agent behaviors through encoded experiences and belief planning
- Representations for robot knowledge in the \textsc{KnowRob} framework
- Bounded-parameter partially observable Markov decision processes: framework and algorithm
- An evidential approach to SLAM, path planning, and active exploration
- Probabilistic may/must testing: retaining probabilities by restricted schedulers
- Strong planning under partial observability
- Goal-directed learning of features and forward models
- Strong planning under uncertainty in domains with numerous but identical elements (a generic approach)
- Transfer in variable-reward hierarchical reinforcement learning
- Recognizing and learning models of social exchange strategies for the regulation of social interactions in open agent societies
- Open problems in universal induction \& intelligence
- Optimal decision rules in repeated games where players infer an opponent's mind via simplified belief calculation
- The value of information for populations in varying environments
- Optimal speech motor control and token-to-token variability: a Bayesian modeling approach
- Active inference and agency: optimal control without cost functions
- Stochastic dynamic programming with factored representations
- Integrated common sense learning and planning in POMDPs
- Exact decomposition approaches for Markov decision processes: a survey
- Task-structured probabilistic I/O automata
- Exploiting symmetries for single- and multi-agent partially observable stochastic domains
- Probabilistic Reasoning by SAT Solvers
- Bottom-up learning of hierarchical models in a class of deterministic pomdp environments
- Policy iteration for bounded-parameter POMDPs
- A two-state partially observable Markov decision process with three actions
- State observation accuracy and finite-memory policy performance
- Conformant plans and beyond: principles and complexity
- A synthesis of automated planning and reinforcement learning for efficient, robust decision-making
- Solving for Best Responses and Equilibria in Extensive-Form Games with Reinforcement Learning Methods
- POMDPs under probabilistic semantics
- Partially observable Markov decision processes with imprecise parameters
- Abstraction and approximate decision-theoretic planning.
- Markov limid processes for representing and solving renewal problems
- Dynamic multiagent probabilistic inference
- Title not available (Why is that?)
Uses Software
This page was built for publication: Planning and acting in partially observable stochastic domains
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q72343)