@inbook{45895,
  author       = {{Karl, Holger and Maack, Marten and Meyer auf der Heide, Friedhelm and Pukrop, Simon and Redder, Adrian}},
  booktitle    = {{On-The-Fly Computing -- Individualized IT-services in dynamic markets}},
  editor       = {{Haake, Claus-Jochen and Meyer auf der Heide, Friedhelm and Platzner, Marco and Wachsmuth, Henning and Wehrheim, Heike}},
  pages        = {{183--202}},
  publisher    = {{Heinz Nixdorf Institut, Universität Paderborn}},
  title        = {{{On-The-Fly Compute Centers II: Execution of Composed Services in Configurable Compute Centers}}},
  doi          = {{10.5281/zenodo.8068664}},
  volume       = {{412}},
  year         = {{2023}},
}

@inproceedings{32811,
  abstract     = {{The decentralized nature of multi-agent systems requires continuous data exchange to achieve global objectives. In such scenarios, Age of Information (AoI) has become an important metric of the freshness of exchanged data due to the error-proneness and delays of communication systems. Communication systems usually possess dependencies: the process describing the success or failure of communication is highly correlated when these attempts are ``close'' in some domain (e.g. in time, frequency, space or code as in wireless communication) and is, in general, non-stationary. To study AoI in such scenarios, we consider an abstract event-based AoI process $\Delta(n)$, expressing time since the last update: If, at time $n$, a monitoring node receives a status update from a source node (event $A(n-1)$ occurs), then $\Delta(n)$ is reset to one; otherwise, $\Delta(n)$ grows linearly in time. This AoI process can thus be viewed as a special random walk with resets. The event process $A(n)$ may be nonstationary and we merely assume that its temporal dependencies decay sufficiently, described by $\alpha$-mixing. We calculate moment bounds for the resulting AoI process as a function of the mixing rate of $A(n)$. Furthermore, we prove that the AoI process $\Delta(n)$ is itself $\alpha$-mixing from which we conclude a strong law of large numbers for $\Delta(n)$. These results are new, since AoI processes have not been studied so far in this general strongly mixing setting. This opens up future work on renewal processes with non-independent interarrival times.}},
  author       = {{Redder, Adrian and Ramaswamy, Arunselvan and Karl, Holger}},
  booktitle    = {{Proceedings of the 58th Allerton Conference on Communication, Control, and Computing}},
  title        = {{{Age of Information Process under Strongly Mixing Communication -- Moment Bound, Mixing Rate and Strong Law}}},
  year         = {{2022}},
}

@inproceedings{30793,
  author       = {{Redder, Adrian and Ramaswamy, Arunselvan and Karl, Holger}},
  booktitle    = {{Proceedings of the 14th International Conference on Agents and Artificial Intelligence}},
  publisher    = {{SCITEPRESS - Science and Technology Publications}},
  title        = {{{Multi-agent Policy Gradient Algorithms for Cyber-physical Systems with Lossy Communication}}},
  doi          = {{10.5220/0010845400003116}},
  year         = {{2022}},
}

@unpublished{30790,
  abstract     = {{Iterative distributed optimization algorithms involve multiple agents that
communicate with each other, over time, in order to minimize/maximize a global
objective. In the presence of unreliable communication networks, the
Age-of-Information (AoI), which measures the freshness of data received, may be
large and hence hinder algorithmic convergence. In this paper, we study the
convergence of general distributed gradient-based optimization algorithms in
the presence of communication that neither happens periodically nor at
stochastically independent points in time. We show that convergence is
guaranteed provided the random variables associated with the AoI processes are
stochastically dominated by a random variable with finite first moment. This
improves on previous requirements of boundedness of more than the first moment.
We then introduce stochastically strongly connected (SSC) networks, a new
stochastic form of strong connectedness for time-varying networks. We show: If
for any $p \ge0$ the processes that describe the success of communication
between agents in a SSC network are $\alpha$-mixing with $n^{p-1}\alpha(n)$
summable, then the associated AoI processes are stochastically dominated by a
random variable with finite $p$-th moment. In combination with our first
contribution, this implies that distributed stochastic gradient descend
converges in the presence of AoI, if $\alpha(n)$ is summable.}},
  author       = {{Redder, Adrian and Ramaswamy, Arunselvan and Karl, Holger}},
  booktitle    = {{arXiv:2201.11343}},
  title        = {{{Distributed gradient-based optimization in the presence of dependent  aperiodic communication}}},
  year         = {{2022}},
}

@unpublished{30791,
  abstract     = {{We present sufficient conditions that ensure convergence of the multi-agent
Deep Deterministic Policy Gradient (DDPG) algorithm. It is an example of one of
the most popular paradigms of Deep Reinforcement Learning (DeepRL) for tackling
continuous action spaces: the actor-critic paradigm. In the setting considered
herein, each agent observes a part of the global state space in order to take
local actions, for which it receives local rewards. For every agent, DDPG
trains a local actor (policy) and a local critic (Q-function). The analysis
shows that multi-agent DDPG using neural networks to approximate the local
policies and critics converge to limits with the following properties: The
critic limits minimize the average squared Bellman loss; the actor limits
parameterize a policy that maximizes the local critic's approximation of
$Q_i^*$, where $i$ is the agent index. The averaging is with respect to a
probability distribution over the global state-action space. It captures the
asymptotics of all local training processes. Finally, we extend the analysis to
a fully decentralized setting where agents communicate over a wireless network
prone to delays and losses; a typical scenario in, e.g., robotic applications.}},
  author       = {{Redder, Adrian and Ramaswamy, Arunselvan and Karl, Holger}},
  booktitle    = {{arXiv:2201.00570}},
  title        = {{{Asymptotic Convergence of Deep Multi-Agent Actor-Critic Algorithms}}},
  year         = {{2022}},
}

@article{32854,
  author       = {{Redder, Adrian and Ramaswamy, Arunselvan and Karl, Holger}},
  journal      = {{IFAC-PapersOnLine}},
  number       = {{13}},
  pages        = {{133–138}},
  publisher    = {{Elsevier}},
  title        = {{{Practical Network Conditions for the Convergence of Distributed Optimization}}},
  volume       = {{55}},
  year         = {{2022}},
}

@article{24140,
  author       = {{Ramaswamy, Arunselvan and Redder, Adrian and Quevedo, Daniel E.}},
  journal      = {{IEEE Transactions on Automatic Control}},
  pages        = {{1--1}},
  title        = {{{Distributed optimization over time-varying networks with stochastic information delays}}},
  doi          = {{10.1109/TAC.2021.3108492}},
  year         = {{2021}},
}

@inproceedings{13443,
  abstract     = {{This work considers the problem of control and resource allocation in networked
systems. To this end, we present DIRA a Deep reinforcement learning based Iterative Resource
Allocation algorithm, which is scalable and control-aware. Our algorithm is tailored towards
large-scale problems where control and scheduling need to act jointly to optimize performance.
DIRA can be used to schedule general time-domain optimization based controllers. In the present
work, we focus on control designs based on suitably adapted linear quadratic regulators. We
apply our algorithm to networked systems with correlated fading communication channels. Our
simulations show that DIRA scales well to large scheduling problems.}},
  author       = {{Redder, Adrian and Ramaswamy, Arunselvan and Quevedo, Daniel}},
  booktitle    = {{Proceedings of the 8th IFAC Workshop on Distributed Estimation and Control in Networked Systems}},
  keywords     = {{Networked control systems, deep reinforcement learning, large-scale systems, resource scheduling, stochastic control}},
  location     = {{Chicago, USA}},
  title        = {{{Deep reinforcement learning for scheduling in large-scale networked control systems}}},
  year         = {{2019}},
}

