IEEE Paper Template in A4 (V1) International Journal on Advances in ICT for Emerging Regions 2022 15 (3): December 2022 International Journal on Advances in ICT for Emerging Regions Communication-Affinity Aware Colocation and Merging of Containers Nishadi N. Wickramanayaka, Chamath I. Keppitiyagama, Kenneth Thilakarathna Abstract— Microservice architecture relies on message passing between services. Inter-service communication introduces an overhead to the applications’ overall performance. This overhead depends on the runtime placement of the services bundled in containers and it can be reduced by intelligently deploying the containers by considering the communication affinities between services. Researchers have attempted to colocate microservices and merge containers based on affinities. However, container merging has not been considered up to now. This study shows that the problem of service placement in a microservice application considering communication affinities, constrained by computational resources, can be mapped to an instance of the Binary Knapsack Problem. We propose a container colocation and merging mechanism based on a heuristic solution to the Binary Knapsack Problem. The proposed approach reduced the communication overhead of a benchmark application by 58.5% and as a result, execution time was reduced by approximately 13.4% as well. Keywords— Affinity, Binary Knapsack, Colocation, Container Networks, Communication Overhead, Docker Containers, Microservices, Microservice Architecture, Inter-service Communication I. INTRODUCTION icroservice architecture is used to develop software applications as suites of independently deployable small services that interact with each other [1]. It is often used to decompose an existing system rather than to compose a system anew using services offered by different enterprises. Microservices are typically deployed in containers with each service contained in a dedicated container [2] which are then hosted in multiple hosts. Hence microservices of an application exchange a significant amount of data, creating communication affinities. Affinity is defined as a relation between two microservices [6] which in this study given by the total amount of data exchanged between those two services over time. To place services in different containers, function calls in the monolithic application should be converted to network calls between the containers in the microservice architecture. Those network calls add an extra layer of networking with expensive operations such as packet encapsulation, decapsulation, address translations [3], which ultimately increase the services’ request/response time [2], [4]. Correspondence: Nishadi N. Wickramanayaka #1 (E-mail: nishadinuwa1995@gmail.com) Received: 24-12-2021 Revised:26-08-2022 Accepted: 30-08-2022 Nishadi N. Wickramanayaka, Chamath I. Keppitiyagama and Kenneth Thilakarathna are from University of Colombo School of Computing, Sri Lanka. (e-mail: nishadinuwa1995@gmail.com, chamath@ucsc.cmb.ac.lk, kmt@ucsc.cmb.ac.lk). DOI: http://doi.org/10.4038/icter.v15i3.7251 © 2022 International Journal on Advances in ICT for Emerging Regions Therefore, the resulting communication overhead adversely impacts the overall performance of the microservice application (μApp) despite the benefits of the architecture. Hence what this paper focuses as the main research problem is that the degraded performance of μApps due to the communications between the services of the application. Container networks used to connect containers with each other play a major role in communication complexities of microservice architecture. Overlay networks are used for host-to-host communication when containers/services are deployed in different hosts whereas bridge networks are used when containers/services are deployed within the same host. Processes inside the same container communicate over the loopback interface. Overhead imposed by an overlay network is higher than the overhead imposed by a bridge network [3]. Loopback interface imposes the least overhead since it eliminates the intervention of bridge network as well. Therefore, placement of two services with high communication affinities in different physical nodes makes this situation worse [6]. Thus, it is evident that the container placement decisions of the μApps need to be taken carefully in the deployment time. As long as microservice architecture is used to design an application it is impossible to completely eliminate the communication overhead incurred at runtime. Because even if all the services are located inside a single machine there will still be communication across address spaces. A possible way of addressing the problem is by reducing the communication overhead to a certain extent by making the deployment decisions carefully. Further, if mapping of the exact same design decisions to runtime is not strictly necessary, then alterations may be applicable to further reduce the communication overhead. The motivation behind this study is to increase the performance of a μApp, by reducing the overhead of inter-service communication by carefully analysing the runtime behaviour of a μApp and containers. Therefore, our main objectives were to explore the impact of container networks on μApps, to discover the possibilities of reducing the communication overhead of a μApp without changing the design of the application and finally to measure up to what extent the performance can be increased from the proposed solution. In order to achieve these objectives, we present a novel mechanism of container colocation and container merging. Container colocation is defined as moving the services with high communication affinities into a single host. Colocation is constrained by the resources available on the hosts. We could map this problem to an instance of the Binary Knapsack Problem (BKP). Container merging is the process of executing services that are already colocated, in a single container. Through the colocation process the overlay network is reduced to a bridge network. Merging process reduces the bridge network to communications over the M This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits nrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited mailto:nishadinuwa1995@gmail.com mailto:nishadinuwa1995@gmail.com mailto:chamath@ucsc.cmb.ac.lk http://doi.org/10.4038/icter.v15i3.7251 Communication-Affinity Aware Colocation and Merging of Containers 34 December 2022 International Journal on Advances in ICT for Emerging Regions container’s loopback interface. As a result, communication overhead is reduced, and the performance of the application is improved. Also, this might reduce the number of hosts needed to execute the μApp. The rest of this paper is organized as follows. Section II presents a review of the background related to the study. Section III presents the design and implementation of the solution. Results and the evaluation of the proposed approach are summarized in Section IV. Finally, Section V concludes and outlines some future directions. II. BACKGROUND AND RELATED WORK A. Performance Degradation of μApps and Container Networks With the inclination of the industry towards cloud-based infrastructure, microservice architecture has received massive attention from the academic community. Hence ample amount of studies which compare the μApps with monolithic applications show the performance penalty resulted when using μApps [2], [4], [8]. The performance of microservices in container-based and virtual machine (VM) based environments has also been studied by Salah et al. [9]. Amaral et al. [10] evaluated the performance impact of two models of implementing microservices in a container environment: master-slave and nested-containers. They mention that nested-container model is hardly adopted in real-world applications since there are some trade-offs in terms of network performance. Suo et al. [3] have done a thorough investigation on latency, throughput, scalability, and startup cost of various container networks on a single host and on multiple hosts in a virtualized environment. Out of four networking modes on a single host (none mode, bridge mode, container mode, host mode), bridge mode network incurred 18% and 30% throughput loss in upload and download respectively in facilitating each container to own an isolated network namespace, resulting all inter-container communications to go through the docker0 bridge. Out of four networking modes available on multiple hosts environments (host mode, NAT, Overlay network, Routing), both NAT and Routing incurred considerable performance degradation due to the overhead of address translation and packet routing. However, the overlay network caused a high-performance loss of 82.8% throughput drop and a 55% latency increase compared to the host mode. They explain the reason for the performance degradation of μApps in terms of bridge network and overlay network which are used to connect the containers. This study has done a comparison between several container networks in a single host and multiple hosts separately. They have not compared the single host container networks with multi-host container networks. Further, they have conducted the experiment in VMs where an additional network overhead is introduced to containers through the virtualization. Yang et al. [11] have attempted to bridge the above gap by deploying the containers on both VM and bare-metal environments. The results confirm the overhead of VM environments with a throughput loss compared to bare-metal deployment. In all tests, the multi-hosts control group showed a significant throughput loss compared to the single host control group. Further, Kratzke [12] has analysed the performance impact of the overlay network in terms of encryption to HTTP-based and REST-like services. Even though these analyses show the impact of container networks on imposing communication overhead, they have only considered inter-container communications and none of these studies have considered intra-container communications. Hence, it is identified that intra-container communications should also take into consideration in order to further explore the ways of reducing the communication overhead. B. Container Placement Problem Based on the aforementioned studies, placement of the containers has been identified as one of the major reasons in creating these communication affinities. Hence, it is pertinent to explore the state of the art of container placement process in practice. However, identifying the best placement of containers is not an easy task. Existing container management tools implement several common placement strategies. Kubernetes [13] places a minimum number of microservices per host in the cluster [6]. This is called the Spread strategy. However, it can add latency to communication and lower the μApp’s performance. Also, this does not take resource optimization into consideration during the deployment. Some management tools use the Bin-pack strategy: deploying a μApp in a minimum number of hosts so that it avoids the cluster resource wastage. Besides these commonly used two strategies, the Random strategy is also used where the management tool selects a host to deploy a microservice randomly. All these three strategies are available in Docker Swarm [14]. Irrespective of the strategy, management tools only consider the instantaneous resource usage of the service when they place them on hosts and rarely try to find an optimal setting. However, they do not consider the communication affinities between services resulting in placing microservices with high communication affinities in different hosts. Eventually, the large amount of network traffic that takes place between two services over the network can hinder the overall performance of the application. Sampaio et al. [6] propose REMaP, a runtime microservices placement mechanism. They consider microservices’ resource usage and as well as their affinities when placing the microservices in hosts. This problem is modelled as an instance of the multi-dimensional bin- packing problem. The objective of REMaP is to maximize the affinity score while deploying the microservices in a minimum number of hosts during the runtime. In solving the problem, they have used the First Fit as a heuristic in their approach since a runtime placement needs quick solutions. REMaP instruments the microservices to gather information required to take colocation decisions during the runtime. Though, we noticed the heavy cost of this instrumentation on the microservices. Hence, the benefits derived through colocation are negatively affected due to the instrumentation cost. Further, REMaP cannot handle data synchronization across different hosts after migrating a stateful microservice. Hence, the migration of stateful microservices may lead the μApp in an inconsistent state. In addition, runtime migration cost may not be justifiable compared to the benefits derived due to colocation. REMaP does not use the hints available in the configuration files about the resource usage or the relationships between microservices indicated in them. We 35 N. N. Wickramanayaka#1, C. I. Keppitiyagama*2, K. Thilakarathna# International Journal on Advances in ICT for Emerging Regions December 2022 further noticed that REMaP does not consider container merging at all. Han et al. [15] propose a refinement framework for profiling-based microservices placement to identify and respond to workload characteristics. The resource requirements obtained through profiling has been fed into a greedy-based heuristic algorithm to make microservices placement. However, the main focus of this work is not a placement algorithm but a profiling-based framework for microservices deployment. Hence any placement algorithm can be adapted to their framework. Both aforementioned solutions depend on the data collected at run time. However, we have identified that some essential parameters that are required to take the placement decision are already available in configuration files even before the runtime of the μApp. C. VM Placement Problem Once the containers are mapped to VMs and communications between containers to the communications between VMs, virtual machine placement in physical machines (PMs) can be considered as the closest research area to the container placement problem. Tziritas et al. [7] propose a communication-aware graph-coloring algorithm, placing the VMs in the underlying system in an energy- efficient manner while optimizing the network overhead due to the VM communication inter-dependencies. However, that VM selection process cannot be directly mapped into the container selection process as their study pre-defines the number of servers to place VMs and container placement may not necessarily give the number of host machines to place the containers as the problem is to optimize the placement of services, thus the algorithm itself should be able to identify the minimum number of hosts to locate the services. However, it is possible to map their VM communication graph to a container/service communication graph. Chen et al. [5] propose a different approach for the VM placement problem which is an affinity-aware grouping method for allocation of VMs into PMs based on a heuristic bin-packing algorithm. It groups the VMs based on the affinities among them and then allocates those identified VM groups into a minimum number of PMs using the bin packing strategy. One major limitation of this research is the generation of one large VM affinity group with total resource requests overstepping the PM resource limit. Sonnek et al. [16] present a decentralized affinity-aware migration technique that incorporates heterogeneity and VM communication cost to allocate VMs on the available physical resources. Their technique monitors network affinity between pairs of VMs periodically and triggers the migration if inter-server traffic exceeds intra-server traffic and uses a distributed bartering algorithm, to dynamically adjust VM placement such that communication overhead is minimized. Since the migration also has a cost, they refrain from migrating VMs if it results in only minor benefits. III. METHODOLOGY The proposed solution comprises of two phases: • Container colocation: moving containers that are initially deployed in different hosts into a single host. • Container merging: placing the services that are initially deployed in different containers inside a single container. Containers deployed in multiple hosts are connected through overlay networks and containers in the same host communicate through a bridge network. As mentioned before, overlay networks impose a higher overhead than bridge networks [3], [11], [12]. Hence during the colocation phase services with high communication affinities are identified to deploy on a single host. This is not a trivial task since the colocation is constrained by the processing resources available on the host. In this study, we propose a novel approach to solve the colocation problem by mapping it to an instance of the BKP. Current approaches to reduce the communication cost use only the container colocation process [6] which replaces the overlay network with bridge network. It is evident that the elimination of this bridge network should further reduce the overhead in inter-service communication. This study introduces a novel concept of merging the colocated containers to further reduce the communication overhead by eliminating the bridge network as well. Once two containers are merged, services deployed on them would execute on a single container as two processes. These two processes communicate over the container’s loopback interface. The merging process converts the inter-container communications into intra-container communications. Spread strategy deployment of Docker Swarm is considered as the baseline to this study. This deployment strategy distributes services evenly among the hosts, resulting in a minimum number of services per host. Thus, we consider the deployment of service instance per host as the baseline since there are not any optimizations present in that strategy. Change of the number of containers and hosts throughout the whole process of colocation and merging can be shown as in Table I. TABLE I CHANGE OF THE NUMBER OF CONTAINERS AND HOSTS THROUGH THE PROCESS Initial deployment After colocation After merging n services n services n services n containers n containers m containers (m