TY - CONF AU - Redder, Adrian AU - Ramaswamy, Arunselvan AU - Karl, Holger ID - 30793 T2 - Proceedings of the 14th International Conference on Agents and Artificial Intelligence TI - Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication ER - TY - JOUR AB - In many cyber–physical systems, we encounter the problem of remote state estimation of geo- graphically distributed and remote physical processes. This paper studies the scheduling of sensor transmissions to estimate the states of multiple remote, dynamic processes. Information from the different sensors has to be transmitted to a central gateway over a wireless network for monitoring purposes, where typically fewer wireless channels are available than there are processes to be monitored. For effective estimation at the gateway, the sensors need to be scheduled appropriately, i.e., at each time instant one needs to decide which sensors have network access and which ones do not. To address this scheduling problem, we formulate an associated Markov decision process (MDP). This MDP is then solved using a Deep Q-Network, a recent deep reinforcement learning algorithm that is at once scalable and model-free. We compare our scheduling algorithm to popular scheduling algorithms such as round-robin and reduced-waiting-time, among others. Our algorithm is shown to significantly outperform these algorithms for many example scenario AU - Leong, Alex S. AU - Ramaswamy, Arunselvan AU - Quevedo, Daniel E. AU - Karl, Holger AU - Shi, Ling ID - 15741 JF - Automatica SN - 0005-1098 TI - Deep reinforcement learning for wireless sensor scheduling in cyber–physical systems ER - TY - CONF AU - Shiferaw Heyi, Binyam AU - Karl, Holger ID - 2476 TI - Modelling Time-Limited Capacity of a Wireless Channel as aMarkov Reward Process ER - TY - CONF AU - Demirel, Burak AU - Ramaswamy, Arunselvan AU - Quevedo, Daniel AU - Karl, Holger ID - 3217 TI - DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling ER -