@inproceedings{29220,
  abstract     = {{Modern services often comprise several components, such as chained virtual network functions, microservices, or
machine learning functions. Providing such services requires to decide how often to instantiate each component, where to place these instances in the network, how to chain them and route traffic through them. 
To overcome limitations of conventional, hardwired heuristics, deep reinforcement learning (DRL) approaches for self-learning network and service management have emerged recently. These model-free DRL approaches are more flexible but typically learn tabula rasa, i.e., disregard existing understanding of networks, services, and their coordination. 

Instead, we propose FutureCoord, a novel model-based AI approach that leverages existing understanding of networks and services for more efficient and effective coordination without time-intensive training. FutureCoord combines Monte Carlo Tree Search with a stochastic traffic model. This allows FutureCoord to estimate the impact of future incoming traffic and effectively optimize long-term effects, taking fluctuating demand and Quality of Service (QoS) requirements into account. Our extensive evaluation based on real-world network topologies, services, and traffic traces indicates that FutureCoord clearly outperforms state-of-the-art model-free and model-based approaches with up to 51% higher flow success ratios.}},
  author       = {{Werner, Stefan and Schneider, Stefan Balthasar and Karl, Holger}},
  booktitle    = {{IEEE/IFIP Network Operations and Management Symposium (NOMS)}},
  keywords     = {{network management, service management, AI, Monte Carlo Tree Search, model-based, QoS}},
  location     = {{Budapest}},
  publisher    = {{IEEE}},
  title        = {{{Use What You Know: Network and Service Coordination Beyond Certainty}}},
  year         = {{2022}},
}

@inproceedings{21543,
  abstract     = {{Services often consist of multiple chained components such as microservices in a service mesh, or machine learning functions in a pipeline. Providing these services requires online coordination including scaling the service, placing instance of all components in the network, scheduling traffic to these instances, and routing traffic through the network. Optimized service coordination is still a hard problem due to many influencing factors such as rapidly arriving user demands and limited node and link capacity. Existing approaches to solve the problem are often built on rigid models and assumptions, tailored to specific scenarios. If the scenario changes and the assumptions no longer hold, they easily break and require manual adjustments by experts. Novel self-learning approaches using deep reinforcement learning (DRL) are promising but still have limitations as they only address simplified versions of the problem and are typically centralized and thus do not scale to practical large-scale networks.

To address these issues, we propose a distributed self-learning service coordination approach using DRL. After centralized training, we deploy a distributed DRL agent at each node in the network, making fast coordination decisions locally in parallel with the other nodes. Each agent only observes its direct neighbors and does not need global knowledge. Hence, our approach scales independently from the size of the network. In our extensive evaluation using real-world network topologies and traffic traces, we show that our proposed approach outperforms a state-of-the-art conventional heuristic as well as a centralized DRL approach (60% higher throughput on average) while requiring less time per online decision (1 ms).}},
  author       = {{Schneider, Stefan Balthasar and Qarawlus, Haydar and Karl, Holger}},
  booktitle    = {{IEEE International Conference on Distributed Computing Systems (ICDCS)}},
  keywords     = {{network management, service management, coordination, reinforcement learning, distributed}},
  location     = {{Washington, DC, USA}},
  publisher    = {{IEEE}},
  title        = {{{Distributed Online Service Coordination Using Deep Reinforcement Learning}}},
  year         = {{2021}},
}

@inproceedings{20693,
  abstract     = {{In practical, large-scale networks, services are requested
by users across the globe, e.g., for video streaming.
Services consist of multiple interconnected components such as
microservices in a service mesh. Coordinating these services
requires scaling them according to continuously changing user
demand, deploying instances at the edge close to their users,
and routing traffic efficiently between users and connected instances.
Network and service coordination is commonly addressed
through centralized approaches, where a single coordinator
knows everything and coordinates the entire network globally.
While such centralized approaches can reach global optima, they
do not scale to large, realistic networks. In contrast, distributed
approaches scale well, but sacrifice solution quality due to their
limited scope of knowledge and coordination decisions.

To this end, we propose a hierarchical coordination approach
that combines the good solution quality of centralized approaches
with the scalability of distributed approaches. In doing so, we divide
the network into multiple hierarchical domains and optimize
coordination in a top-down manner. We compare our hierarchical
with a centralized approach in an extensive evaluation on a real-world
network topology. Our results indicate that hierarchical
coordination can find close-to-optimal solutions in a fraction of
the runtime of centralized approaches.}},
  author       = {{Schneider, Stefan Balthasar and Jürgens, Mirko and Karl, Holger}},
  booktitle    = {{IFIP/IEEE International Symposium on Integrated Network Management (IM)}},
  keywords     = {{network management, service management, coordination, hierarchical, scalability, nfv}},
  location     = {{Bordeaux, France}},
  publisher    = {{IFIP/IEEE}},
  title        = {{{Divide and Conquer: Hierarchical Network and Service Coordination}}},
  year         = {{2021}},
}

@article{21808,
  abstract     = {{Modern services consist of interconnected components,e.g., microservices in a service mesh or machine learning functions in a pipeline. These services can scale and run across multiple network nodes on demand. To process incoming traffic, service components have to be instantiated and traffic assigned to these instances, taking capacities, changing demands, and Quality of Service (QoS) requirements into account. This challenge is usually solved with custom approaches designed by experts. While this typically works well for the considered scenario, the models often rely on unrealistic assumptions or on knowledge that is not available in practice (e.g., a priori knowledge).

We propose DeepCoord, a novel deep reinforcement learning approach that learns how to best coordinate services and is geared towards realistic assumptions. It interacts with the network and relies on available, possibly delayed monitoring information. Rather than defining a complex model or an algorithm on how to achieve an objective, our model-free approach adapts to various objectives and traffic patterns. An agent is trained offline without expert knowledge and then applied online with minimal overhead. Compared to a state-of-the-art heuristic, DeepCoord significantly improves flow throughput (up to 76%) and overall network utility (more than 2x) on realworld network topologies and traffic traces. It also supports optimizing multiple, possibly competing objectives, learns to respect QoS requirements, generalizes to scenarios with unseen, stochastic traffic, and scales to large real-world networks. For reproducibility and reuse, our code is publicly available.}},
  author       = {{Schneider, Stefan Balthasar and Khalili, Ramin and Manzoor, Adnan and Qarawlus, Haydar and Schellenberg, Rafael and Karl, Holger and Hecker, Artur}},
  journal      = {{Transactions on Network and Service Management}},
  keywords     = {{network management, service management, coordination, reinforcement learning, self-learning, self-adaptation, multi-objective}},
  publisher    = {{IEEE}},
  title        = {{{Self-Learning Multi-Objective Service Coordination Using Deep Reinforcement Learning}}},
  doi          = {{10.1109/TNSM.2021.3076503}},
  year         = {{2021}},
}

