Learning a model is paramount for sample efficiency in reinforcement learning control of PDEs

S. Werner, S. Peitz, ArXiv:2302.07160 (2023).

Preprint | English
Author
Werner, Stefan; Peitz, SebastianLibreCat
Abstract
The goal of this paper is to make a strong point for the usage of dynamical models when using reinforcement learning (RL) for feedback control of dynamical systems governed by partial differential equations (PDEs). To breach the gap between the immense promises we see in RL and the applicability in complex engineering systems, the main challenges are the massive requirements in terms of the training data, as well as the lack of performance guarantees. We present a solution for the first issue using a data-driven surrogate model in the form of a convolutional LSTM with actuation. We demonstrate that learning an actuated model in parallel to training the RL agent significantly reduces the total amount of required data sampled from the real system. Furthermore, we show that iteratively updating the model is of major importance to avoid biases in the RL training. Detailed ablation studies reveal the most important ingredients of the modeling process. We use the chaotic Kuramoto-Sivashinsky equation do demonstarte our findings.
Publishing Year
Journal Title
arXiv:2302.07160
LibreCat-ID

Cite this

Werner S, Peitz S. Learning a model is paramount for sample efficiency in reinforcement  learning control of PDEs. arXiv:230207160. Published online 2023.
Werner, S., & Peitz, S. (2023). Learning a model is paramount for sample efficiency in reinforcement  learning control of PDEs. In arXiv:2302.07160.
@article{Werner_Peitz_2023, title={Learning a model is paramount for sample efficiency in reinforcement  learning control of PDEs}, journal={arXiv:2302.07160}, author={Werner, Stefan and Peitz, Sebastian}, year={2023} }
Werner, Stefan, and Sebastian Peitz. “Learning a Model Is Paramount for Sample Efficiency in Reinforcement  Learning Control of PDEs.” ArXiv:2302.07160, 2023.
S. Werner and S. Peitz, “Learning a model is paramount for sample efficiency in reinforcement  learning control of PDEs,” arXiv:2302.07160. 2023.
Werner, Stefan, and Sebastian Peitz. “Learning a Model Is Paramount for Sample Efficiency in Reinforcement  Learning Control of PDEs.” ArXiv:2302.07160, 2023.
All files available under the following license(s):
Copyright Statement:
This Item is protected by copyright and/or related rights. [...]

Link(s) to Main File(s)
Access Level
Restricted Closed Access

Export

Marked Publications

Open Data LibreCat

Sources

arXiv 2302.07160

Search this title in

Google Scholar