Energy efficient VM scheduling and routing in multi-tenant cloud data center

https://doi.org/10.1016/j.suscom.2019.04.004Get rights and content

Highlights

Abstract

Cloud data center hosting many composite applications of multiple tenants consumes a massive amount of energy. Developing energy efficient mechanisms for data center management has become a vital issue for the cloud providers. Most of the works that deal with mechanisms for energy-efficient resource provisioning focus either on reducing the energy consumption of servers or reducing the energy consumption of network elements, but not both. We have formulated the problem of jointly optimizing the energy consumption of servers and network elements by optimal VM scheduling and routing as an integer programming problem. To solve it a phase-wise optimization approach with two ant colony based meta-heuristic algorithms is proposed. The topology features of the data center network and the communication patterns of the applications are considered in the construction of the solution. The solution is tested for three standard data center networks, 3-tier, B-Cube and Hyper-tree of different sizes and compared against two standard algorithms, first-fit and round-robin algorithms. The results showed that the proposed solution improves energy savings by 15% and 20% on an average when compared with first-fit and round-robin respectively.

Introduction

Cloud computing is a buzzing paradigm in the computing industry, and recently there is a massive surge in the use of cloud services to support many sought-after internet applications of different fields like e-commerce, social networking, on-demand video streaming, big-data analytics and so on. The reasons for this growing trend are the alleviation of the users from the burden of owning and maintaining the server resources, reduction in total cost of ownership, ability to access the resources and data from anywhere, and flexibility to scale up and down the resources required as per their dynamically changing needs.

To service this ever-increasing demand for cloud services, cloud providers are deploying large-scale data centers containing thousands of servers and network switches across the world. With the rapid growth in the number and the size of the data centers, the energy consumption and the costs associated with it have increased dramatically. As per the latest study on the US data center energy usage, the data centers in the U.S is estimated to consume 73 billion kWh, which is nearly 2% of the total energy consumed by the US [1]. As the electricity prices are also rapidly increasing with time, the cloud service providers are experiencing substantial energy costs. Apart from the operational costs, exorbitant levels of energy consumption lead to adverse effect on the environment through the large volumes of carbon emissions get released from these data centers. Hence reducing the energy usage has become a key concern for the cloud providers in the design and operation of large-scale data centers.

The energy cost of a data center mainly arises from the energy consumed by physical servers, networking components and cooling equipment. Typically, an idle server in a data center consumes more than 50% of the energy it consumes when it is fully loaded. Based on this finding a popular approach to save energy is server consolidation [2], [3], [4], [5], [6]. Server consolidation is the process of allocating virtual machines that service one or many tasks onto a few servers. The remaining servers are switched off entirely or switched to a low power mode to save energy, as the energy-disproportionate servers in data centers consume a considerable amount of power even when they are idle. Moreover, the network switches and the routers in the data center network contribute to a significant portion of the total power consumed by the data center. The networking components also consume a sizable power even when they are idle, almost equivalent to 30% of the power consumed when they are fully loaded. Like in the case of physical servers, a prudent approach to save energy is allocating the network flows of the virtual machines to a few number networking components and turning off the remaining ones as they do not have any load on them.

At a time, in a cloud data center, many composite applications belonging to different tenants need be deployed. These composite applications are made of several sub-tasks which are logically connected by data and flow dependencies. Generally, a virtual machine in a cloud can serve one or more tasks of an application. Each tenant requests several virtual machines to service the tasks belonging to their application. Virtual machines, belonging to a tenant, communicate and exchange data with each other during their execution. If the behavior of the application belonging to the tenant is known in advance, the communication patterns of the virtual machines can be predicted. If two communicating virtual machines are placed very far in the data center network, then the data that is being exchanged would go through many network elements resulting in higher network energy consumption. Here, the idea is to consolidate the virtual machines onto minimal number of servers in such a way that any two communicating virtual machines are placed very close to each other in the proximity of the data center network. Then the communication data among the virtual machines is routed through minimal number of links and switches to save the energy by turning off the remaining idle links and switches.

Contributions of the paper

  • We model the joint optimization of server and network element energy consumption during VM scheduling and routing as an integer programming problem

  • Two meta heuristic algorithms based on ant colony optimization are proposed as a solution to the problem.

  • The proposed solution is tested for three standard data center network topologies namely 3-Tier, B-Cube and Hyper-Tree of different sizes and its effectiveness is compared with two standard heuristic solutions, first-fit and round robin.

Section snippets

Related work

In recent times, effective energy management in the data centers has become a vital issue for cloud providers. The growing energy costs of the data centers have motivated the researchers to propose energy efficient mechanisms for data center management [3], [7], [8], [9], [10].

Sheikh et al. [11] proposed an evolutionary approach for optimizing the combined objective of performance–energy–temperature for scheduling of parallel tasks on multi-core servers. The solution here is aimed at reducing

Data center model

A heterogeneous single site cloud data center, consisting a set of m heterogeneous servers {H1, H2, …, Hm} is considered here. The capacity of a server Hi in terms of computing power, ram size and storage is given by Zicomp, Zimem and Zistorage respectively. The data center network topology considered here is a hierarchical topology that connects the switches in different layers namely core, aggregate and edge. The set of switches is represented as S{S1, S2, …}. As the switches are

Problem formulation

Given a data center network G(S,E) and virtual machine requests from n number of tenants, with a communication graph CjVˆ,Eˆ for each tenant indicating the communication pattern of the requested virtual machines, our aim is to consolidate the VM requests onto minimal number of servers and switches. The remaining servers and switches are turned off to save energy. So the objective here is to consolidate the virtual machines belonging to multiple tenants onto the physical servers in such a way

Energy efficient VM scheduling and routing

The solution to the problem is divided into two stages. The first stage is finding an optimal schedule of virtual machines belonging to different tenants. Once the virtual machines are placed optimally, the communication flows belonging to different tenants need to be routed through the links in such a way that reduces the total power consumption of network switches. This is termed as the second stage.

Experimental results

This section presents a detailed discussion of the results of the experiments carried out to test the proposed solution. The solution is implemented on a laptop with Intel core i5-4200U CPU with four cores and 6GB RAM. Java is used to implement the algorithms proposed in the solution. The efficiency of the solution is tested for three data centers of different sizes. The results of the above experiments are compared with two standard heuristic algorithms for VM scheduling, first fit and round

Conclusion

In this paper, communication aware energy efficient VM scheduling and routing problem in cloud data center is addressed. To save energy, the proposed two-phase solution first consolidates the virtual machines on to few servers while placing communicating virtual machines in close proximity. Then it consolidates the communication flows of virtual machines on to few switches and links. The solution is tested for three different standard data center network architectures of three different sizes.

References (21)

  • A. Beloglazov et al.

    Energy-aware resource allocation heuristics for efficient management of data centers for cloud computing

    Future Gener. Comput. Syst.

    (2012)
  • T. Baker et al.

    An energy-aware service composition algorithm for multiple cloud-based IOT applications

    J. Netw. Comput. Appl.

    (2017)
  • Y. Li et al.

    An online power metering model for cloud environment

    2012 IEEE 11th International Symposium on Network Computing and Applications

    (2012)
  • J. Koomey et al.

    United States Data Center Energy Usage Report, Technical Report LBNL-1005775

    (2016)
  • N. Quang-Hung et al.

    Epobf: energy efficient allocation of virtual machines in high performance computing cloud

    Transactions on Large-Scale Data-and Knowledge-Centered Systems XVI

    (2014)
  • M.A. Oxley et al.

    Makespan and energy robust stochastic static resource allocation of a bag-of-tasks to a heterogeneous computing system

    IEEE Trans. Parallel Distribut. Syst.

    (2015)
  • L. Wang et al.

    Greendcn: A general framework for achieving energy efficiency in data center networks

    IEEE J. Select. Areas Commun.

    (2014)
  • K. Ye et al.

    Profiling-based workload consolidation and migration in virtualized data centers

    IEEE Trans. Parallel Distribut. Syst.

    (2015)
  • S.U. Khan et al.

    A pure nash equilibrium-based game theoretical method for data replication across multiple servers

    IEEE Trans. Knowledge Data Eng.

    (2009)
  • T. Baker et al.

    Greeaodv: an energy efficient routing protocol for vehicular ad hoc networks

    International Conference on Intelligent Computing

    (2018)
There are more references available in the full text version of this article.

Cited by (20)

  • Multi-search-routes-based methods for minimizing makespan of homogeneous and heterogeneous resources in Cloud computing

    2023, Future Generation Computer Systems
    Citation Excerpt :

    Liu et al. [39] proposed OEMACS combining OEM local search techniques and ACO to resolve VMs deployment in Cloud computing, which reduced the energy consumption and improved the effectiveness of different resources compared with conventional heuristic and other evolutionary-based approaches. Chakravarthy et al. [13] proposed two ant colony-based algorithms (TACO) to address VM scheduling and routing in multi-tenant Cloud data centers aiming at improving the utilization of energy in Cloud computing. GA imitates the process of natural evolution as the search route of the local search algorithm.

  • Deep reinforcement learning-based algorithms selectors for the resource scheduling in hierarchical Cloud computing

    2022, Journal of Network and Computer Applications
    Citation Excerpt :

    Finally, we conclude this paper in Section 6. Frequently used approaches for resource scheduling of Cloud computing contain migration such as VMs migration (Sudarshan Chakravarthy et al., 2019), application migration (Duc et al., 2019), task migration (Miao et al., 2020), and workload migration (Fiandrino et al., 2017); Queuing model such as M/M/S (Ding et al., 2020) and M/G/1 (Li, 2009); multi-phase approach (Laili et al., 2020); as well as scheduling algorithm. The core of these approaches is still the scheduling algorithm including several categories according to the solution way: direct allocation algorithms, search algorithms, and machine learning algorithms.

  • A comparative study on the online shopping willingness of fresh agricultural products between experienced consumers and potential consumers

    2021, Sustainable Computing: Informatics and Systems
    Citation Excerpt :

    Third, enterprises should improve the logistics service system to improve the perceived value of consumers. On the one hand, fresh e-commerce enterprises are required to speed up the research and development of cold chain logistics technology, improve the construction of logistics infrastructure system, and reduce the possible consumption of fresh agricultural products from the place of origin to the hands of consumers to the minimum [24]; on the other hand, they are required to integrate the existing logistics resources of the society and develop high-quality logistics and distribution services to meet the specific needs of consumers in combination with the actual conditions of enterprises so as to improve consumers’ perceived value and satisfaction and enhance their online shopping willingness. Fourth, enterprises should strengthen the credibility of online word-of-mouth and improve the guiding role of online word-of-mouth.

  • An optimal energy efficient cross-layer routing in MANETs

    2020, Sustainable Computing: Informatics and Systems
View all citing articles on Scopus
View full text