@article{57472,
  abstract     = {{In this paper we introduce, in a Hilbert space setting, a second order dynamical system with asymptotically vanishing damping and vanishing Tikhonov regularization that approaches a multiobjective optimization problem with convex and differentiable components of the objective function. Trajectory solutions are shown to exist in finite dimensions. We prove fast convergence of the function values, quantified in terms of a merit function. Based on the regime considered, we establish both weak and, in some cases, strong convergence of trajectory solutions toward a weak Pareto optimal solution. To achieve this, we apply Tikhonov regularization individually to each component of the objective function. This work extends results from single objective convex optimization into the multiobjective setting.}},
  author       = {{Bot, Radu Ioan and Sonntag, Konstantin}},
  journal      = {{Journal of Mathematical Analysis and Applications}},
  keywords     = {{Pareto optimization, Lyapunov analysis, gradient-like dynamical systems, inertial dynamics, asymptotic vanishing damping, Tikhonov regularization, strong convergence}},
  title        = {{{Inertial dynamics with vanishing Tikhonov regularization for multobjective optimization}}},
  year         = {{2025}},
}

@unpublished{51160,
  abstract     = {{We rigorously derive novel and sharp finite-data error bounds for highly
sample-efficient Extended Dynamic Mode Decomposition (EDMD) for both i.i.d. and
ergodic sampling. In particular, we show all results in a very general setting
removing most of the typically imposed assumptions such that, among others,
discrete- and continuous-time stochastic processes as well as nonlinear partial
differential equations are contained in the considered system class. Besides
showing an exponential rate for i.i.d. sampling, we prove, to the best of our
knowledge, the first superlinear convergence rates for ergodic sampling of
deterministic systems. We verify sharpness of the derived error bounds by
conducting numerical simulations for highly-complex applications from molecular
dynamics and chaotic flame propagation.}},
  author       = {{Philipp, Friedrich M. and Schaller, Manuel and Boshoff, Septimus and Peitz, Sebastian and Nüske, Feliks and Worthmann, Karl}},
  booktitle    = {{arXiv:2402.02494}},
  title        = {{{Extended Dynamic Mode Decomposition: Sharp bounds on the sample  efficiency}}},
  year         = {{2024}},
}

@article{46019,
  abstract     = {{We derive efficient algorithms to compute weakly Pareto optimal solutions for smooth, convex and unconstrained multiobjective optimization problems in general Hilbert spaces. To this end, we define a novel inertial gradient-like dynamical system in the multiobjective setting, which trajectories converge weakly to Pareto optimal solutions. Discretization of this system yields an inertial multiobjective algorithm which generates sequences that converge weakly to Pareto optimal solutions. We employ Nesterov acceleration to define an algorithm with an improved convergence rate compared to the plain multiobjective steepest descent method (Algorithm 1). A further improvement in terms of efficiency is achieved by avoiding the solution of a quadratic subproblem to compute a common step direction for all objective functions, which is usually required in first-order methods. Using a different discretization of our inertial gradient-like dynamical system, we obtain an accelerated multiobjective gradient method that does not require the solution of a subproblem in each step (Algorithm 2). While this algorithm does not converge in general, it yields good results on test problems while being faster than standard steepest descent.}},
  author       = {{Sonntag, Konstantin and Peitz, Sebastian}},
  journal      = {{Journal of Optimization Theory and Applications}},
  publisher    = {{Springer}},
  title        = {{{Fast Multiobjective Gradient Methods with Nesterov Acceleration via Inertial Gradient-Like Systems}}},
  doi          = {{10.1007/s10957-024-02389-3}},
  year         = {{2024}},
}

@unpublished{51334,
  abstract     = {{The efficient optimization method for locally Lipschitz continuous multiobjective optimization problems from [1] is extended from finite-dimensional problems to general Hilbert spaces. The method iteratively computes Pareto critical points, where in each iteration, an approximation of the subdifferential is computed in an efficient manner and then used to compute a common descent direction for all objective functions. To prove convergence, we present some new optimality results for nonsmooth multiobjective optimization problems in Hilbert spaces. Using these, we can show that every accumulation point of the sequence generated by our algorithm is Pareto critical under common assumptions. Computational efficiency for finding Pareto critical points is numerically demonstrated for multiobjective optimal control of an obstacle problem.}},
  author       = {{Sonntag, Konstantin and Gebken, Bennet and Müller, Georg and Peitz, Sebastian and Volkwein, Stefan}},
  booktitle    = {{arXiv:2402.06376}},
  title        = {{{A Descent Method for Nonsmooth Multiobjective Optimization in Hilbert Spaces}}},
  year         = {{2024}},
}

@article{40171,
  abstract     = {{We present a convolutional framework which significantly reduces the complexity and thus, the computational effort for distributed reinforcement learning control of dynamical systems governed by partial differential equations (PDEs). Exploiting translational equivariances, the high-dimensional distributed control problem can be transformed into a multi-agent control problem with many identical, uncoupled agents. Furthermore, using the fact that information is transported with finite velocity in many cases, the dimension of the agents’ environment can be drastically reduced using a convolution operation over the state space of the PDE, by which we effectively tackle the curse of dimensionality otherwise present in deep reinforcement learning. In this setting, the complexity can be flexibly adjusted via the kernel width or by using a stride greater than one (meaning that we do not place an actuator at each sensor location). Moreover, scaling from smaller to larger domains – or the transfer between different domains – becomes a straightforward task requiring little effort. We demonstrate the performance of the proposed framework using several PDE examples with increasing complexity, where stabilization is achieved by training a low-dimensional deep deterministic policy gradient agent using minimal computing resources.}},
  author       = {{Peitz, Sebastian and Stenner, Jan and Chidananda, Vikas and Wallscheid, Oliver and Brunton, Steven L. and Taira, Kunihiko}},
  journal      = {{Physica D: Nonlinear Phenomena}},
  pages        = {{134096}},
  publisher    = {{Elsevier}},
  title        = {{{Distributed Control of Partial Differential Equations Using  Convolutional Reinforcement Learning}}},
  doi          = {{10.1016/j.physd.2024.134096}},
  volume       = {{461}},
  year         = {{2024}},
}

@article{33461,
  abstract     = {{Data-driven models for nonlinear dynamical systems based on approximating the underlying Koopman operator or generator have proven to be successful tools for forecasting, feature learning, state estimation, and control. It has become well known that the Koopman generators for control-affine systems also have affine dependence on the input, leading to convenient finite-dimensional bilinear approximations of the dynamics. Yet there are still two main obstacles that limit the scope of current approaches for approximating the Koopman generators of systems with actuation. First, the performance of existing methods depends heavily on the choice of basis functions over which the Koopman generator is to be approximated; and there is currently no universal way to choose them for systems that are not measure preserving. Secondly, if we do not observe the full state, we may not gain access to a sufficiently rich collection of such functions to describe the dynamics. This is because the commonly used method of forming time-delayed observables fails when there is actuation. To remedy these issues, we write the dynamics of observables governed by the Koopman generator as a bilinear hidden Markov model, and determine the model parameters using the expectation-maximization (EM) algorithm. The E-step involves a standard Kalman filter and smoother, while the M-step resembles control-affine dynamic mode decomposition for the generator. We demonstrate the performance of this method on three examples, including recovery of a finite-dimensional Koopman-invariant subspace for an actuated system with a slow manifold; estimation of Koopman eigenfunctions for the unforced Duffing equation; and model-predictive control of a fluidic pinball system based only on noisy observations of lift and drag.}},
  author       = {{Otto, Samuel E. and Peitz, Sebastian and Rowley, Clarence W.}},
  journal      = {{SIAM Journal on Applied Dynamical Systems}},
  number       = {{1}},
  pages        = {{885--923}},
  publisher    = {{SIAM}},
  title        = {{{Learning Bilinear Models of Actuated Koopman Generators from  Partially-Observed Trajectories}}},
  doi          = {{10.1137/22M1523601}},
  volume       = {{23}},
  year         = {{2024}},
}

@article{38031,
  abstract     = {{We consider the data-driven approximation of the Koopman operator for
stochastic differential equations on reproducing kernel Hilbert spaces (RKHS).
Our focus is on the estimation error if the data are collected from long-term
ergodic simulations. We derive both an exact expression for the variance of the
kernel cross-covariance operator, measured in the Hilbert-Schmidt norm, and
probabilistic bounds for the finite-data estimation error. Moreover, we derive
a bound on the prediction error of observables in the RKHS using a finite
Mercer series expansion. Further, assuming Koopman-invariance of the RKHS, we
provide bounds on the full approximation error. Numerical experiments using the
Ornstein-Uhlenbeck process illustrate our results.}},
  author       = {{Philipp, Friedrich and Schaller, Manuel and Worthmann, Karl and Peitz, Sebastian and Nüske, Feliks}},
  journal      = {{Applied and Computational Harmonic Analysis }},
  publisher    = {{Springer }},
  title        = {{{Error bounds for kernel-based approximations of the Koopman operator}}},
  doi          = {{10.1016/j.acha.2024.101657}},
  volume       = {{71}},
  year         = {{2024}},
}

@unpublished{53858,
  author       = {{Akhter, Junaid and Fährmann, Paul David and Sonntag, Konstantin and Peitz, Sebastian}},
  booktitle    = {{arXiv}},
  title        = {{{Common pitfalls to avoid while using multiobjective optimization in machine learning}}},
  year         = {{2024}},
}

@article{32447,
  abstract     = {{We present a new gradient-like dynamical system related to unconstrained convex smooth multiobjective optimization which involves inertial effects and asymptotic vanishing damping. To the best of our knowledge, this system is the first inertial gradient-like system for multiobjective optimization problems including asymptotic vanishing damping, expanding the ideas previously laid out in [H. Attouch and G. Garrigos, Multiobjective Optimization: An Inertial Dynamical Approach to Pareto Optima, preprint, arXiv:1506.02823, 2015]. We prove existence of solutions to this system in finite dimensions and further prove that its bounded solutions converge weakly to weakly Pareto optimal points. In addition, we obtain a convergence rate of order \(\mathcal{O}(t^{-2})\) for the function values measured with a merit function. This approach presents a good basis for the development of fast gradient methods for multiobjective optimization.}},
  author       = {{Sonntag, Konstantin and Peitz, Sebastian}},
  issn         = {{1095-7189}},
  journal      = {{SIAM Journal on Optimization}},
  keywords     = {{multiobjective optimization, Pareto optimization, Lyapunov analysis, gradient-likedynamical systems, inertial dynamics, asymptotic vanishing damping, fast convergence}},
  number       = {{3}},
  pages        = {{2259 -- 2286}},
  publisher    = {{Society for Industrial and Applied Mathematics}},
  title        = {{{Fast Convergence of Inertial Multiobjective Gradient-Like Systems with Asymptotic Vanishing Damping}}},
  doi          = {{10.1137/23M1588512}},
  volume       = {{34}},
  year         = {{2024}},
}

@inproceedings{46649,
  abstract     = {{Different conflicting optimization criteria arise naturally in various Deep
Learning scenarios. These can address different main tasks (i.e., in the
setting of Multi-Task Learning), but also main and secondary tasks such as loss
minimization versus sparsity. The usual approach is a simple weighting of the
criteria, which formally only works in the convex setting. In this paper, we
present a Multi-Objective Optimization algorithm using a modified Weighted
Chebyshev scalarization for training Deep Neural Networks (DNNs) with respect
to several tasks. By employing this scalarization technique, the algorithm can
identify all optimal solutions of the original problem while reducing its
complexity to a sequence of single-objective problems. The simplified problems
are then solved using an Augmented Lagrangian method, enabling the use of
popular optimization techniques such as Adam and Stochastic Gradient Descent,
while efficaciously handling constraints. Our work aims to address the
(economical and also ecological) sustainability issue of DNN models, with a
particular focus on Deep Multi-Task models, which are typically designed with a
very large number of weights to perform equally well on multiple tasks. Through
experiments conducted on two Machine Learning datasets, we demonstrate the
possibility of adaptively sparsifying the model during training without
significantly impacting its performance, if we are willing to apply
task-specific adaptations to the network weights. Code is available at
https://github.com/salomonhotegni/MDMTN.}},
  author       = {{Hotegni, Sedjro Salomon and Berkemeier, Manuel Bastian and Peitz, Sebastian}},
  booktitle    = {{2024 International Joint Conference on Neural Networks (IJCNN)}},
  issn         = {{ 2161-4407}},
  location     = {{Yokohama, Japan}},
  pages        = {{9}},
  publisher    = {{IEEE}},
  title        = {{{Multi-Objective Optimization for Sparse Deep Multi-Task Learning}}},
  doi          = {{10.1109/IJCNN60899.2024.10650994}},
  year         = {{2024}},
}

@article{21199,
  abstract     = {{As in almost every other branch of science, the major advances in data
science and machine learning have also resulted in significant improvements
regarding the modeling and simulation of nonlinear dynamical systems. It is
nowadays possible to make accurate medium to long-term predictions of highly
complex systems such as the weather, the dynamics within a nuclear fusion
reactor, of disease models or the stock market in a very efficient manner. In
many cases, predictive methods are advertised to ultimately be useful for
control, as the control of high-dimensional nonlinear systems is an engineering
grand challenge with huge potential in areas such as clean and efficient energy
production, or the development of advanced medical devices. However, the
question of how to use a predictive model for control is often left unanswered
due to the associated challenges, namely a significantly higher system
complexity, the requirement of much larger data sets and an increased and often
problem-specific modeling effort. To solve these issues, we present a universal
framework (which we call QuaSiModO:
Quantization-Simulation-Modeling-Optimization) to transform arbitrary
predictive models into control systems and use them for feedback control. The
advantages of our approach are a linear increase in data requirements with
respect to the control dimension, performance guarantees that rely exclusively
on the accuracy of the predictive model, and only little prior knowledge
requirements in control theory to solve complex control problems. In particular
the latter point is of key importance to enable a large number of researchers
and practitioners to exploit the ever increasing capabilities of predictive
models for control in a straight-forward and systematic fashion.}},
  author       = {{Peitz, Sebastian and Bieker, Katharina}},
  journal      = {{Automatica}},
  publisher    = {{Elsevier}},
  title        = {{{On the Universal Transformation of Data-Driven Models to Control Systems}}},
  doi          = {{10.1016/j.automatica.2022.110840}},
  volume       = {{149}},
  year         = {{2023}},
}

@unpublished{48502,
  abstract     = {{The prediction of photon echoes is an important technique for gaining an understanding of optical quantum systems. However, this requires a large number of simulations with varying parameters and/or input pulses, which renders numerical studies expensive. This article investigates how we can use data-driven surrogate models based on the Koopman operator to accelerate this process. In order to be successful, we require a model that is accurate over a large number of time steps. To this end, we employ a bilinear Koopman model using extended dynamic mode decomposition and simulate the optical Bloch equations for an ensemble of inhomogeneously broadened two-level systems. Such systems are well suited to describe the excitation of excitonic resonances in semiconductor nanostructures, for example, ensembles of semiconductor quantum dots. We perform a detailed study on the required number of system simulations such that the resulting data-driven Koopman model is sufficiently accurate for a wide range of parameter settings. We analyze the L2 error and the relative error of the photon echo peak and investigate how the control positions relate to the stabilization. After proper training, the dynamics of the quantum ensemble can be predicted accurately and numerically very efficiently by our methods.}},
  author       = {{Peitz, Sebastian and Hunstig, Anna and Rose, Hendrik and Meier, Torsten}},
  title        = {{{Accelerating the analysis of optical quantum systems using the Koopman operator}}},
  year         = {{2023}},
}

@unpublished{51159,
  abstract     = {{Sparsity is a highly desired feature in deep neural networks (DNNs) since it ensures numerical efficiency, improves the interpretability of models (due to the smaller number of relevant features), and robustness. In machine learning approaches based on linear models, it is well known that there exists a connecting path between the sparsest solution in terms of the $\ell^1$ norm,i.e., zero weights and the non-regularized solution, which is called the regularization path. Very recently, there was a first attempt to extend the concept of regularization paths to DNNs by means of treating the empirical loss and sparsity ($\ell^1$ norm) as two conflicting criteria and solving the resulting multiobjective optimization problem. However, due to the non-smoothness of the $\ell^1$ norm and the high number of parameters, this approach is not very efficient from a computational perspective. To overcome this limitation, we present an algorithm that allows for the approximation of the entire Pareto front for the above-mentioned objectives in a very efficient manner. We present numerical examples using both deterministic and stochastic gradients. We furthermore demonstrate that knowledge of the regularization path allows for a well-generalizing network parametrization.}},
  author       = {{Amakor, Augustina Chidinma and Sonntag, Konstantin and Peitz, Sebastian}},
  booktitle    = {{arXiv}},
  title        = {{{A multiobjective continuation method to compute the regularization path of deep neural networks}}},
  year         = {{2023}},
}

@unpublished{51158,
  abstract     = {{Extended Dynamic Mode Decomposition (EDMD) is a popular data-driven method to
approximate the Koopman operator for deterministic and stochastic (control)
systems. This operator is linear and encompasses full information on the
(expected stochastic) dynamics. In this paper, we analyze a kernel-based EDMD
algorithm, known as kEDMD, where the dictionary consists of the canonical
kernel features at the data points. The latter are acquired by i.i.d. samples
from a user-defined and application-driven distribution on a compact set. We
prove bounds on the prediction error of the kEDMD estimator when sampling from
this (not necessarily ergodic) distribution. The error analysis is further
extended to control-affine systems, where the considered invariance of the
Reproducing Kernel Hilbert Space is significantly less restrictive in
comparison to invariance assumptions on an a-priori chosen dictionary.}},
  author       = {{Philipp, Friedrich and Schaller, Manuel and Worthmann, Karl and Peitz, Sebastian and Nüske, Feliks}},
  booktitle    = {{arXiv:2312.10460}},
  title        = {{{Error analysis of kernel EDMD for prediction and control in the Koopman  framework}}},
  year         = {{2023}},
}

@unpublished{46578,
  abstract     = {{Multiobjective optimization plays an increasingly important role in modern applications, where several criteria are often of equal importance. The task in multiobjective optimization and multiobjective optimal control is therefore to compute the set of optimal compromises (the Pareto set) between the conflicting objectives. The advances in algorithms and the increasing interest in Pareto-optimal solutions have led to a wide range of new applications related to optimal and feedback control - potentially with non-smoothness both on the level of the objectives or in the system dynamics. This results in new challenges such as dealing with expensive models (e.g., governed by partial differential equations (PDEs)) and developing dedicated algorithms handling the non-smoothness. Since in contrast to single-objective optimization, the Pareto set generally consists of an infinite number of solutions, the computational effort can quickly become challenging, which is particularly problematic when the objectives are costly to evaluate or when a solution has to be presented very quickly. This article gives an overview of recent developments in the field of multiobjective optimization of non-smooth PDE-constrained problems. In particular we report on the advances achieved within Project 2 "Multiobjective Optimization of Non-Smooth PDE-Constrained Problems - Switches, State Constraints and Model Order Reduction" of the DFG Priority Programm 1962 "Non-smooth and Complementarity-based Distributed Parameter Systems: Simulation and Hierarchical Optimization".}},
  author       = {{Bernreuther, Marco and Dellnitz, Michael and Gebken, Bennet and Müller, Georg and Peitz, Sebastian and Sonntag, Konstantin and Volkwein, Stefan}},
  booktitle    = {{arXiv:2308.01113}},
  title        = {{{Multiobjective Optimization of Non-Smooth PDE-Constrained Problems}}},
  year         = {{2023}},
}

@inproceedings{54838,
  author       = {{Boshoff, Septimus and Stenner, Jan and Weber, Daniel and Meyer, Marvin and Chidananda, Vikas and Peitz, Sebastian and Wallscheid, Oliver}},
  booktitle    = {{IEEE Power and Energy Student Summit (PESS)}},
  isbn         = {{978-3-8007-6318-4}},
  pages        = {{124--129}},
  publisher    = {{VDE}},
  title        = {{{Hybrid control of interconnected power converters using both expert-driven droop and data-driven reinforcement learning approaches}}},
  year         = {{2023}},
}

@inproceedings{54839,
  author       = {{Meyer, Marvin and Weber, Daniel and Chidananda, Vikas and Schweins, Oliver and Stenner, Jan and Boshoff, Septimus and Peitz, Sebastian and Wallscheid, Oliver}},
  booktitle    = {{IEEE Power and Energy Student Summit (PESS)}},
  isbn         = {{978-3-8007-6318-4}},
  pages        = {{112--117}},
  publisher    = {{VDE}},
  title        = {{{ElectricGrid.jl – Automated modeling of decentralized electrical energy grids}}},
  year         = {{2023}},
}

@unpublished{42160,
  abstract     = {{The goal of this paper is to make a strong point for the usage of dynamical models when using reinforcement learning (RL) for feedback control of dynamical systems governed by partial differential equations (PDEs). To breach the gap between the immense promises we see in RL and the applicability in complex engineering systems, the main challenges are the massive requirements in terms of the training data, as well as the lack of performance guarantees. We present a solution for the first issue using a data-driven surrogate model in the form of a convolutional LSTM with actuation. We demonstrate that learning an actuated model in parallel to training the RL agent significantly reduces the total amount of required data sampled from the real system. Furthermore, we show that iteratively updating the model is of major importance to avoid biases in the RL training. Detailed ablation studies reveal the most important ingredients of the modeling process. We use the chaotic Kuramoto-Sivashinsky equation do demonstarte our findings.}},
  author       = {{Werner, Stefan and Peitz, Sebastian}},
  booktitle    = {{arXiv:2302.07160}},
  title        = {{{Learning a model is paramount for sample efficiency in reinforcement  learning control of PDEs}}},
  year         = {{2023}},
}

@article{27426,
  abstract     = {{Regularization is used in many different areas of optimization when solutions
are sought which not only minimize a given function, but also possess a certain
degree of regularity. Popular applications are image denoising, sparse
regression and machine learning. Since the choice of the regularization
parameter is crucial but often difficult, path-following methods are used to
approximate the entire regularization path, i.e., the set of all possible
solutions for all regularization parameters. Due to their nature, the
development of these methods requires structural results about the
regularization path. The goal of this article is to derive these results for
the case of a smooth objective function which is penalized by a piecewise
differentiable regularization term. We do this by treating regularization as a
multiobjective optimization problem. Our results suggest that even in this
general case, the regularization path is piecewise smooth. Moreover, our theory
allows for a classification of the nonsmooth features that occur in between
smooth parts. This is demonstrated in two applications, namely support-vector
machines and exact penalty methods.}},
  author       = {{Gebken, Bennet and Bieker, Katharina and Peitz, Sebastian}},
  journal      = {{Journal of Global Optimization}},
  number       = {{3}},
  pages        = {{709--741}},
  title        = {{{On the structure of regularization paths for piecewise differentiable regularization terms}}},
  doi          = {{10.1007/s10898-022-01223-2}},
  volume       = {{85}},
  year         = {{2023}},
}

@inproceedings{30125,
  abstract     = {{We present an approach for guaranteed constraint satisfaction by means of data-based optimal control, where the model is unknown and has to be obtained from measurement data. To this end, we utilize the Koopman framework and an eDMD-based bilinear surrogate modeling approach for control systems to show an error bound on predicted observables, i.e., functions of the state. This result is then applied to the constraints of the optimal control problem to show that satisfaction of tightened constraints in the purely data-based surrogate model implies constraint satisfaction for the original system.}},
  author       = {{Schaller, Manuel and Worthmann, Karl and Philipp, Friedrich and Peitz, Sebastian and Nüske, Feliks}},
  booktitle    = {{IFAC-PapersOnLine}},
  number       = {{1}},
  pages        = {{169--174}},
  title        = {{{Towards reliable data-based optimal and predictive control using extended DMD}}},
  doi          = {{10.1016/j.ifacol.2023.02.029}},
  volume       = {{56}},
  year         = {{2023}},
}

