Publication:2925454

From MaRDI portal


zbMath1298.90001MaRDI QIDQ2925454

Dimitri P. Bertsekas

Publication date: 22 October 2014




Related Items

Distillation optimization: Parameterized relationship between feed flow rate of a steady‐state distillation column and heat duties of reboiler and condenser, Stability analysis of optimal control problems with time-dependent costs, Markov decision processes with burstiness constraints, A flow based formulation and a reinforcement learning based strategic oscillation for cross-dock door assignment, Nonparametric learning for impulse control problems -- exploration vs. exploitation, A Lyapunov-based version of the value iteration algorithm formulated as a discrete-time switched affine system, Linear quadratic optimal regulation for multiplicative noise systems with special terminal penalty, Dissipativity in infinite horizon optimal control and dynamic programming, An improved method for approximating the infinite-horizon value function of the discrete-time switched LQR problem, Optimal periodic sensor scheduling for minimizing communication rate under LQG constraint, Cost-aware defense for parallel server systems against reliability and security failures, A note on the existence of optimal stationary policies for average Markov decision processes with countable states, A dynamic encryption-decryption scheme for replay attack detection in cyber-physical systems, Compromise policy for multi-stage stochastic linear programming: variance and bias reduction, Optimal decision-making of mutual fund temporary borrowing problem via approximate dynamic programming, DDQN-based optimal targeted therapy with reversible inhibitors to combat the Warburg effect, Bellman filtering and smoothing for state-space models, A Two-Time-Scale Stochastic Optimization Framework with Applications in Control and Reinforcement Learning, Optimal guidance algorithms for parking search with reservations, Continuous‐time stochastic gradient descent for optimizing over the stationary distribution of stochastic differential equations, Unnamed Item, Dynamic Programming Deconstructed: Transformations of the Bellman Equation and Computational Efficiency, Least squares policy iteration with instrumental variables vs. direct policy search: comparison against optimal benchmarks using energy storage, Distributional Robustness in Minimax Linear Quadratic Control with Wasserstein Distance, Predictive stochastic programming, Adaptive park-and-ride choice on time-dependent stochastic multimodal transportation network, Undiscounted control policy generation for continuous-valued optimal control by approximate dynamic programming, Mean-field Markov decision processes with common noise and open-loop controls, Solving average cost Markov decision processes by means of a two-phase time aggregation algorithm, Computational approaches for mixed integer optimal control problems with indicator constraints, Designing parsimonious scheduling policies for complex resource allocation systems through concurrency theory, Continuous-action planning for discounted infinite-horizon nonlinear optimal control with Lipschitz values, Efficiently Breaking the Curse of Horizon in Off-Policy Evaluation with Double Reinforcement Learning, Learning Manipulation Through Information Dissemination, An approximate dynamic programming approach for improving accuracy of lossy data compression by Bloom filters, Heuristics for dynamic and stochastic inventory-routing, Least squares approximate policy iteration for learning bid prices in choice-based revenue management, A model for equilibrium in some service-provider user-set interactions, Tail Optimality and Preferences Consistency for Intertemporal Optimization Problems, A general endogenous grid method for multi-dimensional models with non-convexities and constraints, Technical Note—Static Pricing: Universal Guarantees for Reusable Resources, Dynamic portfolio choice: a simulation-and-regression approach, Stochastic decision diagrams, From Infinite to Finite Programs: Explicit Error Bounds with Applications to Approximate Dynamic Programming, Approximate dynamic programming for the military inventory routing problem, Optimal supervisory control with mean payoff objectives and under partial observation, Deep reinforcement learning for wireless sensor scheduling in cyber-physical systems, Statistical learning for analysis of networked control systems over unknown channels, Least-violating symbolic controller synthesis for safety, reachability and attractivity specifications, Error bounds for stochastic shortest path problems, Throughput maximization of complex resource allocation systems through timed-continuous-Petri-net modeling, An optimal control approach to day-to-day congestion pricing for stochastic transportation networks, Finding multiple Nash equilibria via machine learning-supported Gröbner bases, Deep differentiable reinforcement learning and optimal trading, Dynamic Learning and Decision Making via Basis Weight Vectors, Approximate dynamic programming for lateral transshipment problems in multi-location inventory systems, The stochastic shortest path problem: a polyhedral combinatorics perspective, Regular Policies in Abstract Dynamic Programming, Risk-Sensitive Reinforcement Learning via Policy Gradient Search, Markov models of policy support for technology transitions, Optimal multi-stage allocation of weapons to targets using adaptive dynamic programming, Self-Reflective Model Predictive Control, Unnamed Item, An approach to solving optimal control problems of nonlinear systems by introducing detail-reward mechanism in deep reinforcement learning, Recursively feasible stochastic model predictive control using indirect feedback, Symmetry reduction for dynamic programming, Benchmarking a Scalable Approximate Dynamic Programming Algorithm for Stochastic Control of Grid-Level Energy Storage, A Q-learning predictive control scheme with guaranteed stability, Moving target search optimization -- a literature review, MYOPIC POLICIES FOR NON-PREEMPTIVE SCHEDULING OF JOBS WITH DECAYING VALUE, Linear programming formulation for non-stationary, finite-horizon Markov decision process models, Large-scale unit commitment under uncertainty: an updated literature survey, Precautionary replenishment in financially-constrained inventory systems subject to credit rollover risk and supply disruption, Shape constraints in economics and operations research, Network-Based Approximate Linear Programming for Discrete Optimization, Risk-averse model predictive control, Optimal control of Boolean control networks with average cost: a policy iteration approach, On a multistage discrete stochastic optimization problem with stochastic constraints and nested sampling, Variance minimization of parameterized Markov decision processes, Dynamical systems on weighted lattices: general theory, Is Temporal Difference Learning Optimal? An Instance-Dependent Analysis, Unnamed Item, On finding the optimal BDD relaxation, Exit time risk-sensitive control for systems of cooperative agents, Decentralized and distributed active fault diagnosis: multiple model estimation algorithms, Proximal algorithms and temporal difference methods for solving fixed point problems, An intrinsic material tailoring approach for functionally graded axisymmetric hollow bodies under plane elasticity, Tutorial on risk neutral, distributionally robust and risk averse multistage stochastic programming, Stochastic decomposition applied to large-scale hydro valleys management, Discrete time optimal control with frequency constraints for non-smooth systems, A dynamic programming framework for optimal delivery time slot pricing, Choosing a good toolkit. II: Bayes-rule based heuristics, Experiments with Tractable Feedback in Robotic Planning Under Uncertainty: Insights over a Wide Range of Noise Regimes, Concentration bounds for temporal difference learning with linear function approximation: the case of batch data and uniform sampling, Performance Guarantees and Optimal Purification Decisions for Engineered Proteins, Technical Note—Analysis of Scrip Systems: On an Open Question in Johnson et al. (2014), Data Uncertainty in Markov Chains: Application to Cost-Effectiveness Analyses of Medical Innovations, Bayesian Exploration for Approximate Dynamic Programming, Recomposable restricted finite state machines: definition and solution approaches, Intelligent Human–Robot Interaction Systems Using Reinforcement Learning and Neural Networks, Gradient-bounded dynamic programming for submodular and concave extensible value functions with probabilistic performance guarantees, Transmission scheduling for multi-process multi-sensor remote estimation via approximate dynamic programming, Unnamed Item, Unnamed Item, Decision programming for mixed-integer multi-stage optimization under uncertainty, Optimal stopping for response-guided dosing, An application of approximate dynamic programming in multi-period multi-product advertising budgeting, Robust shortest path planning and semicontractive dynamic programming, Data-driven control of hydraulic servo actuator based on adaptive dynamic programming, Fundamental design principles for reinforcement learning algorithms, Bounded rationality in learning, perception, decision-making, and stochastic games, Incremental constraint projection methods for variational inequalities, Stable Optimal Control and Semicontractive Dynamic Programming, Linear controller design for chance constrained systems, Allocating resources via price management systems: a dynamic programming-based approach, Scenario-based, closed-loop model predictive control with application to emergency vehicle scheduling, Unnamed Item, Large-scale unit commitment under uncertainty, Strategic stiffening/cooling in the Ising game, Inertial-type incremental constraint projection method for solving variational inequalities without Lipschitz continuity