[{"date_updated":"2022-01-06T06:56:34Z","_id":"2476","year":"2018","citation":{"ieee":"B. Shiferaw Heyi and H. Karl, “Modelling Time-Limited Capacity of a Wireless Channel as aMarkov Reward Process,” 2018.","short":"B. Shiferaw Heyi, H. Karl, in: Proc. of IEEE Wireless Communications and Networking Conference (WCNC), 2018.","bibtex":"@inproceedings{Shiferaw Heyi_Karl_2018, title={Modelling Time-Limited Capacity of a Wireless Channel as aMarkov Reward Process}, publisher={Proc. of IEEE Wireless Communications and Networking Conference (WCNC)}, author={Shiferaw Heyi, Binyam and Karl, Holger}, year={2018} }","mla":"Shiferaw Heyi, Binyam, and Holger Karl. Modelling Time-Limited Capacity of a Wireless Channel as AMarkov Reward Process. Proc. of IEEE Wireless Communications and Networking Conference (WCNC), 2018.","chicago":"Shiferaw Heyi, Binyam, and Holger Karl. “Modelling Time-Limited Capacity of a Wireless Channel as AMarkov Reward Process.” Proc. of IEEE Wireless Communications and Networking Conference (WCNC), 2018.","ama":"Shiferaw Heyi B, Karl H. Modelling Time-Limited Capacity of a Wireless Channel as aMarkov Reward Process. In: Proc. of IEEE Wireless Communications and Networking Conference (WCNC); 2018.","apa":"Shiferaw Heyi, B., & Karl, H. (2018). Modelling Time-Limited Capacity of a Wireless Channel as aMarkov Reward Process. Proc. of IEEE Wireless Communications and Networking Conference (WCNC)."},"type":"conference","title":"Modelling Time-Limited Capacity of a Wireless Channel as aMarkov Reward Process","ddc":["000"],"user_id":"15572","status":"public","has_accepted_license":"1","project":[{"_id":"24","name":"Netzgewahre Regelung & regelungsgewahre Netze"}],"date_created":"2018-04-24T08:09:00Z","author":[{"full_name":"Shiferaw Heyi, Binyam","first_name":"Binyam","last_name":"Shiferaw Heyi"},{"first_name":"Holger","full_name":"Karl, Holger","last_name":"Karl","id":"126"}],"publisher":"Proc. of IEEE Wireless Communications and Networking Conference (WCNC)","file_date_updated":"2018-04-24T08:08:50Z","department":[{"_id":"75"}],"file":[{"access_level":"closed","file_name":"p2861-heyi.pdf","date_created":"2018-04-24T08:08:50Z","success":1,"relation":"main_file","date_updated":"2018-04-24T08:08:50Z","content_type":"application/pdf","creator":"tabu","file_id":"2477","file_size":428839}]},{"doi":"10.1109/LCSYS.2018.2847721","_id":"3217","date_updated":"2022-01-06T06:59:05Z","year":"2018","citation":{"short":"B. Demirel, A. Ramaswamy, D. Quevedo, H. Karl, in: 2018.","ieee":"B. Demirel, A. Ramaswamy, D. Quevedo, and H. Karl, “DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling,” 2018.","ama":"Demirel B, Ramaswamy A, Quevedo D, Karl H. DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling. In: ; 2018. doi:10.1109/LCSYS.2018.2847721","apa":"Demirel, B., Ramaswamy, A., Quevedo, D., & Karl, H. (2018). DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling. https://doi.org/10.1109/LCSYS.2018.2847721","chicago":"Demirel, Burak, Arunselvan Ramaswamy, Daniel Quevedo, and Holger Karl. “DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling,” 2018. https://doi.org/10.1109/LCSYS.2018.2847721.","bibtex":"@inproceedings{Demirel_Ramaswamy_Quevedo_Karl_2018, title={DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling}, DOI={10.1109/LCSYS.2018.2847721}, author={Demirel, Burak and Ramaswamy, Arunselvan and Quevedo, Daniel and Karl, Holger}, year={2018} }","mla":"Demirel, Burak, et al. DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling. 2018, doi:10.1109/LCSYS.2018.2847721."},"type":"conference","main_file_link":[{"url":"https://arxiv.org/pdf/1803.02998.pdf"}],"title":"DeepCAS: A Deep Reinforcement Learning Algorithm for Control-Aware Scheduling","ddc":["000"],"user_id":"126","has_accepted_license":"1","status":"public","project":[{"name":"Netzgewahre Regelung & regelungsgewahre Netze","_id":"24"}],"date_created":"2018-06-13T10:26:23Z","author":[{"last_name":"Demirel","first_name":"Burak","full_name":"Demirel, Burak"},{"full_name":"Ramaswamy, Arunselvan","first_name":"Arunselvan","last_name":"Ramaswamy"},{"last_name":"Quevedo","full_name":"Quevedo, Daniel","first_name":"Daniel"},{"full_name":"Karl, Holger","first_name":"Holger","id":"126","last_name":"Karl"}],"department":[{"_id":"75"}],"file_date_updated":"2018-06-13T10:47:57Z","file":[{"file_id":"3218","creator":"tabu","file_size":354166,"success":1,"relation":"main_file","content_type":"application/pdf","date_updated":"2018-06-13T10:47:57Z","file_name":"1803.02998.pdf","date_created":"2018-06-13T10:47:57Z","access_level":"closed"}]},{"language":[{"iso":"eng"}],"date_updated":"2022-01-06T06:52:32Z","doi":"10.1016/j.automatica.2019.108759","department":[{"_id":"7"},{"_id":"34"},{"_id":"3"},{"_id":"75"},{"_id":"57"}],"publication_identifier":{"issn":["0005-1098"]},"publication_status":"published","project":[{"name":"Netzgewahre Regelung & regelungsgewahre Netze","_id":"24"}],"title":"Deep reinforcement learning for wireless sensor scheduling in cyber–physical systems","year":"2019","citation":{"mla":"Leong, Alex S., et al. “Deep Reinforcement Learning for Wireless Sensor Scheduling in Cyber–Physical Systems.” Automatica, 108759, 2019, doi:10.1016/j.automatica.2019.108759.","bibtex":"@article{Leong_Ramaswamy_Quevedo_Karl_Shi_2019, title={Deep reinforcement learning for wireless sensor scheduling in cyber–physical systems}, DOI={10.1016/j.automatica.2019.108759}, number={108759}, journal={Automatica}, author={Leong, Alex S. and Ramaswamy, Arunselvan and Quevedo, Daniel E. and Karl, Holger and Shi, Ling}, year={2019} }","apa":"Leong, A. S., Ramaswamy, A., Quevedo, D. E., Karl, H., & Shi, L. (2019). Deep reinforcement learning for wireless sensor scheduling in cyber–physical systems. Automatica. https://doi.org/10.1016/j.automatica.2019.108759","ama":"Leong AS, Ramaswamy A, Quevedo DE, Karl H, Shi L. Deep reinforcement learning for wireless sensor scheduling in cyber–physical systems. Automatica. 2019. doi:10.1016/j.automatica.2019.108759","chicago":"Leong, Alex S., Arunselvan Ramaswamy, Daniel E. Quevedo, Holger Karl, and Ling Shi. “Deep Reinforcement Learning for Wireless Sensor Scheduling in Cyber–Physical Systems.” Automatica, 2019. https://doi.org/10.1016/j.automatica.2019.108759.","ieee":"A. S. Leong, A. Ramaswamy, D. E. Quevedo, H. Karl, and L. Shi, “Deep reinforcement learning for wireless sensor scheduling in cyber–physical systems,” Automatica, 2019.","short":"A.S. Leong, A. Ramaswamy, D.E. Quevedo, H. Karl, L. Shi, Automatica (2019)."},"type":"journal_article","_id":"15741","article_number":"108759","author":[{"full_name":"Leong, Alex S.","first_name":"Alex S.","last_name":"Leong"},{"id":"66937","last_name":"Ramaswamy","full_name":"Ramaswamy, Arunselvan","orcid":"https://orcid.org/ 0000-0001-7547-8111","first_name":"Arunselvan"},{"last_name":"Quevedo","full_name":"Quevedo, Daniel E.","first_name":"Daniel E."},{"last_name":"Karl","id":"126","first_name":"Holger","full_name":"Karl, Holger"},{"full_name":"Shi, Ling","first_name":"Ling","last_name":"Shi"}],"quality_controlled":"1","file_date_updated":"2020-01-31T15:57:50Z","publication":"Automatica","file":[{"date_updated":"2020-01-31T15:57:50Z","content_type":"application/pdf","relation":"main_file","success":1,"file_size":"675382","creator":"hkarl","file_id":"15743","access_level":"closed","file_name":"leoram20a.pdf","date_created":"2020-01-31T15:57:50Z"}],"has_accepted_license":"1","status":"public","date_created":"2020-01-31T15:55:27Z","abstract":[{"lang":"eng","text":"\r\nIn many cyber–physical systems, we encounter the problem of remote state estimation of geo- graphically distributed and remote physical processes. This paper studies the scheduling of sensor transmissions to estimate the states of multiple remote, dynamic processes. Information from the different sensors has to be transmitted to a central gateway over a wireless network for monitoring purposes, where typically fewer wireless channels are available than there are processes to be monitored. For effective estimation at the gateway, the sensors need to be scheduled appropriately, i.e., at each time instant one needs to decide which sensors have network access and which ones do not. To address this scheduling problem, we formulate an associated Markov decision process (MDP). This MDP is then solved using a Deep Q-Network, a recent deep reinforcement learning algorithm that is at once scalable and model-free. We compare our scheduling algorithm to popular scheduling algorithms such as round-robin and reduced-waiting-time, among others. Our algorithm is shown to significantly outperform these algorithms for many example scenario"}],"ddc":["000"],"user_id":"126"},{"language":[{"iso":"eng"}],"date_updated":"2022-11-18T09:32:14Z","doi":"10.5220/0010845400003116","department":[{"_id":"75"}],"project":[{"_id":"16","name":"SFB 901 - C4: SFB 901 - Subproject C4"},{"name":"NICCI-CN: Netzgewahre Regelung & regelungsgewahre Netze","_id":"24"},{"_id":"1","name":"SFB 901: SFB 901"},{"name":"SFB 901 - C: SFB 901 - Project Area C","_id":"4"}],"publication_status":"published","title":"Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication","year":"2022","citation":{"short":"A. Redder, A. Ramaswamy, H. Karl, in: Proceedings of the 14th International Conference on Agents and Artificial Intelligence, SCITEPRESS - Science and Technology Publications, 2022.","ieee":"A. Redder, A. Ramaswamy, and H. Karl, “Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication,” 2022, doi: 10.5220/0010845400003116.","ama":"Redder A, Ramaswamy A, Karl H. Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication. In: Proceedings of the 14th International Conference on Agents and Artificial Intelligence. SCITEPRESS - Science and Technology Publications; 2022. doi:10.5220/0010845400003116","apa":"Redder, A., Ramaswamy, A., & Karl, H. (2022). Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication. Proceedings of the 14th International Conference on Agents and Artificial Intelligence. https://doi.org/10.5220/0010845400003116","chicago":"Redder, Adrian, Arunselvan Ramaswamy, and Holger Karl. “Multi-Agent Policy Gradient Algorithms for Cyber-Physical Systems with Lossy Communication.” In Proceedings of the 14th International Conference on Agents and Artificial Intelligence. SCITEPRESS - Science and Technology Publications, 2022. https://doi.org/10.5220/0010845400003116.","bibtex":"@inproceedings{Redder_Ramaswamy_Karl_2022, title={Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication}, DOI={10.5220/0010845400003116}, booktitle={Proceedings of the 14th International Conference on Agents and Artificial Intelligence}, publisher={SCITEPRESS - Science and Technology Publications}, author={Redder, Adrian and Ramaswamy, Arunselvan and Karl, Holger}, year={2022} }","mla":"Redder, Adrian, et al. “Multi-Agent Policy Gradient Algorithms for Cyber-Physical Systems with Lossy Communication.” Proceedings of the 14th International Conference on Agents and Artificial Intelligence, SCITEPRESS - Science and Technology Publications, 2022, doi:10.5220/0010845400003116."},"type":"conference","_id":"30793","file":[{"access_level":"closed","date_created":"2022-08-31T07:10:13Z","file_name":"ICCART2022.pdf","relation":"main_file","success":1,"date_updated":"2022-08-31T07:10:13Z","content_type":"application/pdf","creator":"aredder","file_id":"33237","file_size":298926}],"publication":"Proceedings of the 14th International Conference on Agents and Artificial Intelligence","file_date_updated":"2022-08-31T07:10:13Z","author":[{"last_name":"Redder","id":"52265","first_name":"Adrian","orcid":"https://orcid.org/0000-0001-7391-4688","full_name":"Redder, Adrian"},{"last_name":"Ramaswamy","id":"66937","first_name":"Arunselvan","full_name":"Ramaswamy, Arunselvan","orcid":"https://orcid.org/ 0000-0001-7547-8111"},{"first_name":"Holger","full_name":"Karl, Holger","last_name":"Karl","id":"126"}],"publisher":"SCITEPRESS - Science and Technology Publications","date_created":"2022-04-06T07:18:36Z","has_accepted_license":"1","status":"public","user_id":"477","ddc":["006"]}]