Stability Enforced Bandit Algorithms for Channel Selection in Remote State Estimation of Gauss–Markov Processes

From MaRDI portal
Publication:6200040




Abstract: In this paper we consider state estimation and control problems where a sensor or controller can, at each discrete time instant, transmit on one out of M different communication channels. A key difficulty of the situation at hand is that the channel statistics are unknown. We study the case where both learning of the channel reception probabilities and state estimation / control is carried out simultaneously. Methods for choosing the channels based on techniques for multi-armed bandits are presented, and shown to provide stability. Furthermore, we define the performance notions of estimation and control regret, and derive bounds on how it scales with time for the considered algorithms.










This page was built for publication: Stability Enforced Bandit Algorithms for Channel Selection in Remote State Estimation of Gauss–Markov Processes

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6200040)