Planning and acting in partially observable stochastic domains
From MaRDI portal
(Redirected from Publication:72343)
Recommendations
- Planning in partially-observable switching-mode continuous domains
- scientific article; zbMATH DE number 5547912
- Strong planning under partial observability
- Learning and planning in partially observable environments without prior domain knowledge
- Replanning in domains with partial information and sensing actions
- Randomized belief-space replanning in partially-observable continuous spaces
Cites work
- scientific article; zbMATH DE number 3148886 (Why is no real title available?)
- scientific article; zbMATH DE number 4089320 (Why is no real title available?)
- scientific article; zbMATH DE number 700091 (Why is no real title available?)
- scientific article; zbMATH DE number 1095138 (Why is no real title available?)
- scientific article; zbMATH DE number 795587 (Why is no real title available?)
- A survey of algorithmic methods for partially observed Markov decision processes
- A survey of solution techniques for the partially observed Markov decision process
- Application of Jensen's inequality to adaptive suboptimal design
- Fast planning through planning graph analysis
- OPTIMAL CONTROL FOR PARTIALLY OBSERVABLE MARKOV DECISION PROCESSES OVER AN INFINITE HORIZON
- Optimal control of Markov processes with incomplete state information
- Solution Procedures for Partially Observed Markov Decision Processes
- Solving H-horizon, stationary Markov decision problems in time proportional to log (H)
- State of the Art—A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms
- The Optimal Control of Partially Observable Markov Processes over a Finite Horizon
- The Optimal Control of Partially Observable Markov Processes over the Infinite Horizon: Discounted Costs
- The Optimal Search for a Moving Target When the Search Path Is Constrained
- The complexity of mean payoff games on graphs
- The complexity of stochastic games
Cited in
(only showing first 100 items - show all)- Partially observable game-theoretic agent programming in Golog
- Computing rank dependent utility in graphical models for sequential decision problems
- An evidential approach to SLAM, path planning, and active exploration
- A tutorial on partially observable Markov decision processes
- Probabilistic Reasoning by SAT Solvers
- Recognizing and learning models of social exchange strategies for the regulation of social interactions in open agent societies
- scientific article; zbMATH DE number 1509479 (Why is no real title available?)
- scientific article; zbMATH DE number 1728768 (Why is no real title available?)
- Group sparse optimization for learning predictive state representations
- scientific article; zbMATH DE number 2243394 (Why is no real title available?)
- Performance prediction of an unmanned airborne vehicle multi-agent system
- From knowledge-based programs to graded belief-based programs. I: On-line reasoning
- State observation accuracy and finite-memory policy performance
- Goal-directed learning of features and forward models
- Probabilistic may/must testing: retaining probabilities by restricted schedulers
- Bounded-parameter partially observable Markov decision processes: framework and algorithm
- Computer science and decision theory
- scientific article; zbMATH DE number 5547961 (Why is no real title available?)
- scientific article; zbMATH DE number 2243378 (Why is no real title available?)
- scientific article; zbMATH DE number 2243398 (Why is no real title available?)
- Integrated common sense learning and planning in POMDPs
- Randomized belief-space replanning in partially-observable continuous spaces
- The Concept of Opposition and Its Use in Q-Learning and Q(λ) Techniques
- Partial-order planning with concurrent interacting actions
- Conformant plans and beyond: principles and complexity
- Solving for Best Responses and Equilibria in Extensive-Form Games with Reinforcement Learning Methods
- DESPOT: online POMDP planning with regularization
- Multi-stage classifier design
- Planning for multiple measurement channels in a continuous-state POMDP
- A Bayesian approach for learning and planning in partially observable Markov decision processes
- Bottom-up learning of hierarchical models in a class of deterministic pomdp environments
- Testing probabilistic equivalence through reinforcement learning
- Policy iteration for bounded-parameter POMDPs
- Open problems in universal induction \& intelligence
- Exact decomposition approaches for Markov decision processes: a survey
- Optimal cost almost-sure reachability in POMDPs
- Autonomous agents modelling other agents: a comprehensive survey and open problems
- Using machine learning for decreasing state uncertainty in planning
- Planning in partially-observable switching-mode continuous domains
- Optimal speech motor control and token-to-token variability: a Bayesian modeling approach
- Learning and planning in partially observable environments without prior domain knowledge
- An affective mobile robot educator with a full-time job
- Optimal decision rules in repeated games where players infer an opponent's mind via simplified belief calculation
- Contingent planning under uncertainty via stochastic satisfiability
- POMDPs under probabilistic semantics
- pomdpSolve
- scientific article; zbMATH DE number 5547912 (Why is no real title available?)
- Systems of Bounded Rational Agents with Information-Theoretic Constraints
- Planning in hybrid relational MDPs
- A two-state partially observable Markov decision process with three actions
- Stochastic dynamic programming with factored representations
- Partially observable Markov decision processes with imprecise parameters
- Recursively modeling other agents for decision making: a research perspective
- Reward Maximization Through Discrete Active Inference
- Fast strong planning for fully observable nondeterministic planning problems
- A synthesis of automated planning and reinforcement learning for efficient, robust decision-making
- Decentralized MDPs with sparse interactions
- Partially observable Markov decision process approximations for adaptive sensing
- Transfer in variable-reward hierarchical reinforcement learning
- Task-structured probabilistic I/O automata
- Abstraction and approximate decision-theoretic planning.
- Markov limid processes for representing and solving renewal problems
- Exploiting symmetries for single- and multi-agent partially observable stochastic domains
- Cost-sensitive feature acquisition and classification
- Reasoning about uncertain parameters and agent behaviors through encoded experiences and belief planning
- Dynamic multiagent probabilistic inference
- The value of information for populations in varying environments
- Strong planning under partial observability
- Active inference and agency: optimal control without cost functions
- Strong planning under uncertainty in domains with numerous but identical elements (a generic approach)
- Representations for robot knowledge in the \textsc{KnowRob} framework
- scientific article; zbMATH DE number 3871010 (Why is no real title available?)
- Regression and progression in stochastic domains
- An online multi-agent co-operative learning algorithm in POMDPs
- Reward tampering problems and solutions in reinforcement learning: a causal influence diagram perspective
- Off-policy evaluation in partially observed Markov decision processes under sequential ignorability
- scientific article; zbMATH DE number 1559451 (Why is no real title available?)
- Unsupervised basis function adaptation for reinforcement learning
- Learning where to attend with deep architectures for image tracking
- Bridging commonsense reasoning and probabilistic planning via a probabilistic action language
- An educational management problem with continuous signal space
- scientific article; zbMATH DE number 4174347 (Why is no real title available?)
- Supervisor synthesis of POMDP via automata learning
- Strategy Graphs for Influence Diagrams
- Excursions in first-order logic and probability: infinitely many random variables, continuous distributions, recursive programs and beyond
- Planning with partial observability by SAT
- Planning and learning in partially observable systems via filter stability
- scientific article; zbMATH DE number 1844461 (Why is no real title available?)
- Heuristic anytime approaches to stochastic decision processes
- Probabilistic Planning with Reduced Models
- Planning in artificial intelligence
- Optimality guarantees for particle belief approximation of POMDPs
- The complexity of agent design problems: Determinism and history dependence
- A Bayesian theory of mind approach to modeling cooperation and communication
- Safe robust multi-agent reinforcement learning with neural control barrier functions and safety attention mechanism
- Posterior weighted reinforcement learning with state uncertainty
- Epistemic uncertainty aware semantic localization and mapping for inference and belief space planning
- Exploiting expert knowledge in factored POMDPs
- Risk-aware shielding of partially observable Monte Carlo planning policies
- Behavioral model summarisation for other agents under uncertainty
This page was built for publication: Planning and acting in partially observable stochastic domains
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q72343)