The following pages link to (Q4257216):
Displaying 50 items.
- Optimal cost almost-sure reachability in POMDPs (Q253969) (← links)
- Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems (Q280322) (← links)
- A constrained optimization perspective on actor-critic algorithms and application to network routing (Q286519) (← links)
- Potential-based least-squares policy iteration for a parameterized feedback control system (Q289143) (← links)
- Adaptive importance sampling for control and inference (Q290478) (← links)
- Approximate dynamic programming for stochastic linear control problems on compact state spaces (Q299794) (← links)
- Solving average cost Markov decision processes by means of a two-phase time aggregation algorithm (Q300040) (← links)
- Computable approximations for continuous-time Markov decision processes on Borel spaces based on empirical measures (Q302091) (← links)
- Output-feedback adaptive optimal control of interconnected systems based on robust adaptive dynamic programming (Q311890) (← links)
- Multiscale Q-learning with linear function approximation (Q312650) (← links)
- Value iteration and adaptive dynamic programming for data-driven adaptive optimal control design (Q313259) (← links)
- Modeling and optimization control of a demand-driven, conveyor-serviced production station (Q319221) (← links)
- Solving stochastic resource-constrained project scheduling problems by closed-loop approximate dynamic programming (Q319710) (← links)
- New approximate dynamic programming algorithms for large-scale undiscounted Markov decision processes and their application to optimize a production and distribution system (Q320866) (← links)
- Approximate dynamic programming for the dispatch of military medical evacuation assets (Q323422) (← links)
- A perturbation approach to a class of discounted approximate value iteration algorithms with Borel spaces (Q330284) (← links)
- Perspectives of approximate dynamic programming (Q333093) (← links)
- Low-discrepancy sampling for approximate dynamic programming with local approximators (Q336896) (← links)
- Efficient model-based reinforcement learning for approximate online optimal control (Q340682) (← links)
- Planning for multiple measurement channels in a continuous-state POMDP (Q360261) (← links)
- An approximate dynamic programming framework for modeling global climate policy under decision-dependent uncertainty (Q373210) (← links)
- Q-learning and policy iteration algorithms for stochastic shortest path problems (Q378731) (← links)
- General time consistent discounting (Q391749) (← links)
- Robust adaptive dynamic programming for linear and nonlinear systems: an overview (Q397504) (← links)
- Minimax PAC bounds on the sample complexity of reinforcement learning with a generative model (Q399890) (← links)
- Moneybarl: exploiting pitcher decision-making using reinforcement learning (Q400630) (← links)
- Minimum and worst-case performance ratios of rollout algorithms (Q415370) (← links)
- The optimal unbiased value estimator and its relation to LSTD, TD and MC (Q415609) (← links)
- Model selection in reinforcement learning (Q415618) (← links)
- Finding optimal memoryless policies of POMDPs under the expected average reward criterion (Q418072) (← links)
- Optimal control as a graphical model inference problem (Q420939) (← links)
- The optimal control of just-in-time-based production and distribution systems and performance comparisons with optimized pull systems (Q421584) (← links)
- Network revenue management with inventory-sensitive bid prices and customer choice (Q421775) (← links)
- A framework and a mean-field algorithm for the local control of spatial processes (Q433494) (← links)
- Potentials based optimization with embedded Markov chain for stochastic constrained system (Q437320) (← links)
- An online actor-critic algorithm with function approximation for constrained Markov decision processes (Q438776) (← links)
- Approximate dynamic programming for capacity allocation in the service industry (Q439484) (← links)
- Performance optimization of queueing systems with perturbation realization (Q439492) (← links)
- Fitting piecewise linear continuous functions (Q439615) (← links)
- Asymptotic analysis of value prediction by well-specified and misspecified models (Q448322) (← links)
- Iterative methods for the solution of a singular control formulation of a GMWB pricing problem (Q453330) (← links)
- Optimal tracking control of nonlinear partially-unknown constrained-input systems using integral reinforcement learning (Q458982) (← links)
- Value set iteration for Markov decision processes (Q459022) (← links)
- Depth-based short-sighted stochastic shortest path problems (Q460624) (← links)
- A tutorial on event-based optimization -- a new optimization framework (Q461464) (← links)
- Control: a perspective (Q463779) (← links)
- Integral reinforcement learning and experience replay for adaptive optimal control of partially-unknown constrained-input continuous-time systems (Q463819) (← links)
- Reinforcement \(Q\)-learning for optimal tracking control of linear discrete-time systems with unknown dynamics (Q463893) (← links)
- A sparse collocation method for solving time-dependent HJB equations using multivariate \(B\)-splines (Q466457) (← links)
- Temporal difference-based policy iteration for optimal control of stochastic systems (Q467477) (← links)