iBet uBet web content aggregator. Adding the entire web to your favor.
iBet uBet web content aggregator. Adding the entire web to your favor.



Link to original content: https://doi.org/10.3390/su14106256
A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing
Next Article in Journal
Safety Risk Assessment of Air Traffic Control System Based on the Game Theory and the Cloud Matter Element Analysis
Previous Article in Journal
The Impact of Improving the Quality of the Port’s Infrastructure on the Shippers’ Decisions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing

by
Salil Bharany
1,*,
Sandeep Sharma
1,
Osamah Ibrahim Khalaf
2,
Ghaida Muttashar Abdulsahib
3,
Abeer S. Al Humaimeedy
4,
Theyazn H. H. Aldhyani
5,
Mashael Maashi
6 and
Hasan Alkahtani
7
1
CET Department, Guru Nanak Dev University, Amritsar 143005, India
2
Al-Nahrain Nano Renewable Energy Research Center, Al-Nahrain University, Baghdad 10072, Iraq
3
Department of Computer Engineering, University of Technology, Baghdad 19006, Iraq
4
Software Engineering Department, College of Computer and Information Sciences, King Saud University, P.O. Box 103786, Riyadh 11616, Saudi Arabia
5
Applied College in Abqaiq, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia
6
Software Engineering Department, King Saud University, Riyadh 11543, Saudi Arabia
7
College of Computer Science and Information Technology, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(10), 6256; https://doi.org/10.3390/su14106256
Submission received: 8 April 2022 / Revised: 12 May 2022 / Accepted: 13 May 2022 / Published: 20 May 2022

Abstract

:
Global warming is one of the most compelling environmental threats today, as the rise in energy consumption and CO2 emission caused a dreadful impact on our environment. The data centers, computing devices, network equipment, etc., consume vast amounts of energy that the thermal power plants mainly generate. Primarily fossil fuels like coal and oils are used for energy generation in these power plants that induce various environmental problems such as global warming ozone layer depletion, which can even become the cause of premature deaths of living beings. The recent research trend has shifted towards optimizing energy consumption and green fields since the world recognized the importance of these concepts. This paper aims to conduct a complete systematic mapping analysis on the impact of high energy consumption in cloud data centers and its effect on the environment. To answer the research questions identified in this paper, one hundred nineteen primary studies published until February 2022 were considered and further categorized. Some new developments in green cloud computing and the taxonomy of various energy efficiency techniques used in data centers have also been discussed. It includes techniques like VM Virtualization and Consolidation, Power-aware, Bio-inspired methods, Thermal-management techniques, and an effort to evaluate the cloud data center’s role in reducing energy consumption and CO2 footprints. Most of the researchers proposed software level techniques as with these techniques, massive infrastructures are not required as compared with hardware techniques, and it is less prone to failure and faults. Also, we disclose some dominant problems and provide suggestions for future enhancements in green computing.

1. Introduction and Motivation

Cloud computing is evolving continuously to fulfill the increased demand for computing resources, allowing clients to purchase a specific set of resources and only pay for their use [1]. Software as a Service (SaaS), Platform as a Service (PaaS), infrastructure as a Service (IaaS), and Container as a Service (CaaS) are the different types of well-known services provided by the cloud systems [2]. The Pay-per-Use model allows customers to scale services on demand.
One recent research shows that, if the power consumption rate grows, it will even cross a data center infrastructure hardware expense. This will also lead to more emissions of CO2, which have a direct impact on our environment. Here, the concept of green computing comes into the picture. Green computing ecologically uses computers and related resources as its goal is to optimize the energy consumption of computing machines, servers, data centers, networks, and cooling systems [2,3].
The prominent role of green computing is regarding data centers as they represent a significant percentage of global energy, and thus there is a need for such practices that include the proper deployment of energy-efficient technology in data centers and reduce the infrastructure and carbon emissions. There is a fundamental understating that the more hardware and equipment employed, the more energy requirements [3,4].
Moreover, this also diminishes the cost of operating a data center. Green computing is similar to green chemistry [5]. There is a primary focus is to reduce the use of hazardous materials. It also includes the proper disposal of electronic waste, such as computing devices with a minimal impact on the environment.
This also leads to an improved distribution mechanism for resources and services that will result in a reduction in equipment and in the carbon footprint. As shown in Figure 1, the main task is to decrease the energy drop or loss, which is not consumed by data centers, or energy that does not provide helpful output, such as loss, while running on idle mode [2,3,4,5].
There are millions of servers worldwide; Google is estimated to have over a million servers, and Microsoft’s Chicago data center has almost more than 3 hundred thousand servers, which consume 23.5% of electric power, which is generated by coal in the USA [3,4]. The increase in energy consumption from data centers has raised many environmental issues and economic concerns. According to Cisco 2018, global data center traffic will hit up to 20.6 zettabytes (ZB), increasing by 6 ZB per year since 2016. They forecast 628 hyper-scale data centers worldwide in 2021, whereas there were 338 in 2016.
If we talk about mobile users, these will increase by 70% of users from the year 2017 to the year 2022 [5]. It would be beneficial for our environmental sustainability and internet firms if data centers adopted the concept of green computing. Virtualization is a core aspect of green cloud computing, which has revolutionized cloud computing by optimizing the resource utilization capacity and increasing the dynamic migration of resources. This allows the equipment to be used more effectively and lowers the cost by combining a set of physical servers into one [6].
Excessive heat liberation also occurs due to increased power consumption, increasing cooling device usage. Cooling in data centers is the second most important source of power consumption, as the first is operating the data center [4,5]. Usually, a data center has a 1-megawatt requirement, which consumes about $20,000,000 in energy [5,6] during its lifespan, and 30%–40% of the total cost is used by its cooling systems [7] The cost of operating and cooling an extreme system is equal to the sum of money to purchase new hardware within two years.
The cooling at or below 70 degrees F (21 °C) controls the entire data center’s temperatures, even cooling more than a specific limit can also lead to high humidity, exposing the devices to high moisture, thereby, facilitating salt deposits on conductive filaments in the circuit, which can cause short circuits [8]. Low-temperature maintenance is necessary to assure reliability and maintain the hardware’s life. From the above data, we can say that high-performance computation requires more energy for operations and conclude that performance is directly proportional to energy consumption [9,10,11,12].

1.1. Motivation for Research

This study primarily intends to address the rapidly growing energy consumption and to analyze this issue by presenting a cross-sectional survey of recent developments in energy-efficient cloud technology. In addition to this, we also offer a brief introduction and attempt to explain the importance of efficient energy usage in cloud data centers. Data center power consumption is incredibly high as they require energy for computation and cooling purposes.
We conducted a systematic mapping study to examine and report our empirical findings on the associations between energy consumption activities in data centers and their environmental impacts. We hope this research will reduce certain ambiguities concerning the advantages and limitations of energy-efficiency technologies. The motivation behind our research can be said in broader terms as follows:
  • The needs and demands for understanding existing energy-efficient techniques are proliferating. Power-saving innovations, issues related to CO2 emissions, and proper energy distribution are crucial requirements in the current era. This paper reports a comprehensive systematic mapping study (SMS) to analyze, amalgamate, and present our empirical findings relevant to energy-efficiency techniques.
  • There might be no need to implement more infrastructure using virtualization, as it will operate many machines on a single virtual machine. However, this will not control carbon emissions if the infrastructure of a data center does not work efficiently. Such a virtualized environment will require more energy to operate and cool down the equipment, which prompted a comprehensive systemic mapping procedure to examine and evaluate the growing research in this field.

1.2. Our Contribution

We initially reported 2903 relevant articles from six digital libraries and academic search engines published until February 2022. These papers were later examined regarding the title, abstract, full text, manual search, snowballing references, and other perspectives. A total of 119 primary studies (PSs) (Appendix A) were selected after evaluating the quality of articles. To make the following significant contributions, these selected PSs were further investigated.
  • We provide a novel taxonomical overview of various energy-efficient techniques at the cloud data center level by classifying the existing literature extracted from available research papers.
  • We describe the role of energy-efficient techniques in reducing the ecological and financial impact on cloud data centers.
  • We precisely define the research questions related to effective energy utilization in the cloud environment, further associated with cloud datacenters.
  • We discuss the results generated by various techniques by different simulation software, which helps to predict and evaluate the effects of techniques used to reduce energy consumption. We outline issues, concerns, and recommendations for future research work.
To conclude, most current studies do not give a comprehensive review of all cloud-based energy-saving approaches. Details about the systematic survey technique are not included in the majority of the surveys. The majority of the surveys focused on hardware- and software-level techniques for energy efficiency. Our systematic literature study identifies and finds the significant issues in the current research and presents a taxonomy of various procedures executed on the cloud as energy-saving methods.
Our survey also provides recommendations for the issues in different techniques. None of the previous studies has threats to validity originating from the unintentional contradictions involved while reviewing the literature, selecting the primary study, evaluating the quality assessment, and data extraction from the selected studies. Including conclusion, layout, and internal and external threats, we also discuss these constraints and the actions taken to mitigate these threats. We hope that this SMS outcome will help cloud researchers and professionals to identify the open research fields for further investigation and an up-to-date current state of research in this area.
This research will enable the scientific community to consider the consequences and challenges to be addressed for better energy utilization and resource distribution in the cloud computing industry. Section 2 discuss the evolution of energy-efficiency techniques in the data centers and explains the need and role of energy in data centers. It also focuses on disclosing cloud data centers’ contribution to handling high energy consumption and other energy-related issues. Section 3 elaborates on this article’s review methodology and identifies the open research questions in this domain. This division explores the details of the included search keywords, search strings, and selection criteria (inclusion and exclusion details).
The systematic review results are explained in Section 4, and we provide a detailed description and analysis of various energy-management strategies in the cloud data centers. Different taxonomies were defined to obtain a comprehensive and elaborated classification of current cloud energy-efficiency techniques. Section 5 highlights some issues and recommendations for every category of energy-efficient technique. Section 6 discusses potential threats to the effectiveness of the work. In the end, Section 7 summarizes the conclusions of this SMS, and Section 8 outlines the possible recommendations for further research in this area.

2. Background and Related Work

The need for energy-efficient techniques is discussed in Section 2.1 and Section 2.2, the role of Energy Efficiency in Cloud Data Centers is discussed. In Section 2.3, the history and evolution of energy-efficient techniques are explained. Later, in Section 2.4, related surveys on energy-efficiency techniques are discussed.

2.1. Need of Energy-Efficient Techniques

Energy consumption in data centers was estimated as 1.4% of the total EEC (power consumption), increasing 12% annually [12,13,14]. Green cloud computing is a new area that has attracted researchers’ attention world-wide [15]. Today, we need such technology to solve energy and performance tradeoffs. Proper distribution of workloads and maintaining performance by providing access to services anywhere, anytime to users is one of the cloud’s prime objectives. Minimizing energy usage may cause a violation of SLA (Service Level Agreement); however, SLA breaches may cause a penalty to cloud service providers, and thus we should avoid any SLA violation while considering energy-efficiency systems [12].
A recent report by Accenture shows that small organizations and mid-range corporations can reduce up to 90% CO2 emissions if they shift toward cloud resources, and even for large organizations, they can save at least 40–60% reductions in emissions if they establish the whole infrastructure on the cloud [12,13,14,15]. Most data centers are only interested in cost reductions regarding electricity, not on carbon emissions; therefore, the plan is to eliminate the source of CO2 as well [13].
If data centers are constructed on the green computing principle, the cloud computing benefits are also more prominent for environmental protection. Massive data centers are typically designed for cloud-computing services, connected with several high-speed networks and virtual servers, including facilities, such as temperature control, power systems, and others [16]. In Figure 1, there are four conditions in a system where energy is used inefficiently, i.e., drop or loss, where both “drop” and “loss” apply to insufficient energy use.
Energy drop refers to the energy transmitted to the process but not consumed by any subsystem (e.g., D1), e.g., electricity is wasted due to transfer or conversion. A second cause of drop includes the support system’s overhead costs (D2), such as cooling or lighting in data centers primarily supplied by cloud service providers. Power loss refers to the energy used for its primary purpose but lost due to system abandonment (L1)—for example, when the system is running but in idle mode [16,17,18,19,20]. Another cause of loss is over-the-limit usage or redundant usage (L2), such as keeping the system cool to a maximum at night when the temperature is already low.
Reducing power consumption in cloud data centers and cloud computing systems is a challenge that has raised researcher interest over time and has become a trending research domain today.

2.2. Role of Energy Efficiency in Cloud Data Centers

Cloud computing has become a crucial evolutionary paradigm. It provides a powerful and dynamic ability to use a computer to handle all IT needs, including computer storage and software requirements. A cloud data center can deliver a powerful and efficient computing environment by virtualization. This concept has widely disseminated and gained the interest of organizations in reducing hardware and software investment [15]. It offers advantages in cost when thousands of users share the same resources and pay per use with no capital investment providing high scalability and dynamic provisioning.
The key concepts in cloud computing are grid computing, parallel computing, virtualization, containerization, and distributed computing, and the latest trending areas are microservice architecture and SOA (service-oriented architecture). Cloud technology has recently emerged to meet users’ computationally intensive needs by using virtual infrastructure. The cloud can be determined and accessible globally via the internet [20,21]. Numerous cloud data centers have also been set up, which offer colossal computing infrastructure.
These data centers are home to massive servers that require additional computing tools, such as storage, cooling devices, and backup system support [22,23,24,25,26,27,28,29,30]. Moreover, these thousands of servers in the data center take up enormous space and consume tremendous electricity. These servers dissipate heat and must be maintained in a completely air-conditioned environment specially designed for data centers. The energy consumption of data centers is also expected to increase rapidly [31].
An estimate assumes the proportion of electricity data center consumption worldwide will increase by 1.3% to 8% from 2010 to 2020 [17]. This massive increase in power consumption from data centers makes it very clear that many data centers and servers are responding to cloud services’ growing demand [32,33,34]. This often represents an inefficient use of energy as not all power supplied to a data center is used to perform cloud storage services.
The cooling network consumes a large portion of the energy as a continuous power source. In legacy data centers, up to 50% of the capacity is used by non-server machines [14]. These data centers do not yet operate on the best energy management patterns in their design and operation phases. Still, the industry has attained substantial progress in this field. Therefore, the possibilities for improving data centers’ energy efficiency now rely heavily on finding ways to increase energy efficiency for a server in a data center [31,32,33,34,35].

2.3. Evolution of Energy Efficiency Techniques

The major aim is to reduce the energy requirements of specific products and services, which will automatically lead to reduced temperature as this will limit the carbon emissions, also known as energy efficiency [36,37,38]. Many reasons exist for improving energy efficiency: reduced energy usage decreases expenditure on energy and could result in a financial benefit to customers [39,40]. Another reason is that harmful gasses are released into the environment with more energy consumption, increasing the earth’s temperature. This section shows the evolution of studies on energy efficiency as seen in Figure 2.
This topic deals with energy-management technologies, CO2 emissions from data centers, and maximizing performance with minimum energy consumption. Various authors have suggested different methods for optimizing energy consumption. Ref. [14] presented attributes to aggregate power utilization on massive servers. Simultaneously, a collection of nearly 15,000 points was used for different applications for approximately six months.
The authors used the proposed modeling framework to approximate the power management scheme, thereby, minimizing the peak power load and energy utilization at a data center. A novel software platform for energy-efficient cloud management was proposed by [10], which suggests that the cloud infrastructure has real potential for energy efficiency. It substantially improved the performance regarding the response time and cost reductions in dynamic workload scenarios. Similarly, Ref. [36] recommended a simulation paradigm for energy-aware cloud computing data centers.
This simulator was contrived to explicitly gather information on power usage by data center modules, such as IT equipment, servers, and switches. The workload distribution was handled; however, it also handled packet-level connectivity patterns using different settings. The simulation findings showed the usefulness of applying power control strategies, such as voltage scaling, frequency scaling, and dynamic shutdown equipment. VMeter, an innovative tool for power modeling that expresses direct correlations between online resources and the overall power usage of a data center, was suggested [29].
GRMP-Q protocol is a generic resource allocation gossip protocol introduced by [25]. It aims to reduce power consumption by server consolidation while shifting the load pattern. The results indicate that core efficiency metrics do not change with the machine volume, allowing the resource allocation mechanism to be scalable with around 100,000 servers. Developing an energy reduction model for cloud computing [37,38,39,40,41,42,43,44] offers a practical tool to explore various aspects of the network based on mathematical methods to reduce energy usage.
While performing implementation on AMPL/CPLEX platforms, they also offered an analytical approach to investigate different aspects of the energy reduction network. Ref. [45] developed a Cloud Monitor that can generate data models for cloud-based power forecasting, which also provides insight into the energy costs of activity in the cloud without using any additional external equipment [37]. A new simple energy consumption model was introduced by [34], which also works as an analysis tool for cloud-computing environments to track and calculate the total energy consumption based on different runtime activities and to support static or dynamic system-level optimization [46,47,48,49,50,51,52].
The authors in [53] suggested various concepts for optimizing the system by maintaining the same ratio in processing speed with energy consumption and the volume of CO2 dissipation in the atmosphere. Refs. [53,54,55,56] proposed an approach that reduces energy consumption by disabling idle virtual machines (VMs) and categorizing active VMs into different energy clusters using a fuzzy C-means algorithm. Consequently, the extent of the Service Level Agreement (SLA) violation was minor [57,58,59,60,61,62,63,64,65,66,67,68].
The authors of [57] unveiled a new concept for cloud computing energy efficiency that tracks how the emissions of different greenhouse gasses from data centers have caused, how much ozone depletion, and how eco-friendly our technology can be [69,70,71,72,73,74,75,76,77,78,79]. It also addresses the numerous main problems when these energy-saving strategies are applied in cloud data centers. Ref. [24] suggested an innovative approach for the power management, reconfiguration, and migration of VMs. It involves an integrated monitoring system, where proactive initiatives are implemented for all three domains in a hierarchically ordered fashion [80,81,82,83,84,85,86,87,88,89,90,91,92].
To enhance energy efficiency and control carbon footprints. Ref. [22] developed a conceptual model and realistic recommendations for the integrated management of all infrastructures, including computers, networks, storage, and cooling systems [93,94,95,96,97,98,99]. They also proposed a range of potential research directions and suggested more practical improvements in this field [100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119,120,121,122].

2.4. Related Surveys

Although earlier surveys [18,100,122] have been innovative, there is still a need for a systematic literature survey to determine, update, and expand research objectives as research in the field of green cloud computing is continuously gaining interest. The first two surveys discussed energy-efficient strategies based on the hardware optimization process. In [122], existing energy-efficient technologies were summarized in hardware and software-oriented optimization systems [19]. However, they primarily focused on software-level-based techniques, and some of the latest innovative technologies, such as bio-inspired thermal-aware technologies were also included.
In this current study, we attempt an evaluation process and to extend the taxonomy of energy-efficiency techniques defined by [122], such as power management systems, bio-inspired approaches, and non-technical techniques. Ref. [25] decomposed sensor cloud energy-efficient approaches into multiple categories and evaluated each one using a variety of metrics.
This survey determines how much of each parameter is used by each technique and the average percentage of parameters used by the sensor cloud energy efficiency. According to the observations by the authors, the vast majority of energy-saving approaches ignore the quality of service (quality of service) factors, scalability, and network lifetime. The total energy usage by a cloud system can be reduced by using service allocation mechanisms outlined in [74].
In this survey, these authors developed the service allocation issue and an energy model based on a generalized system architecture. A taxonomy of energy-efficient resource allocation approaches in the literature is also presented in this survey. Ref. [91] looked at all conceivable locations in high energy consumption in cloud computing infrastructures. This study examines the current solutions that have been built to increase energy efficiency in large-scale cloud-computing environments without sacrificing the quality of service or performance.
Various methods for reducing the power consumption of data centers are discussed in the survey by [51]. All strategies used to lower energy-specific hardware components/levels are covered in extreme detail. There is much emphasis on techniques deployed at the hardware-level (network- or server-level) that can lead to energy-efficient or ecologically friendly data centers [122,123,124,125,126,127,128,129,130,131,132].
Ref. [60] presented research to map out the current state of knowledge on energy efficiency’s impact on cloud computing and, in turn, help to build a better picture of the future of this area of study. No emphasis was given to techniques to reduce energy consumption [132,133,134,135,136,137,138,139,140,141,142,143,144,145].
Several different applications of the bi-level optimization model in managing ship emissions are discussed in an article [146]. These applications include the Energy Efficiency Design Index, the Emissions Control Area, the Market-Based Measure, the Carbon Intensity Indicator, and the Vessel Speed Reduction Incentive Program. An extensive taxonomy of sustainable cloud computing was developed in the paper [36]. A taxonomy was designed to explore the existing methodologies for sustainable cloud computing, covering numerous areas, such as application design, sustainability measures, capacity planning, energy management, and other methods.
In the paper, the authors [93] highlight the critical issues related to reducing carbon emission from ships. In this paper, the main focus was bulti on managing the critical measure, the carbon intensity indicator (CII), and the carbon emissions per unit of transport work for each ship. In order to lower the carbon footprint of data centers, this research focused on decreasing brown energy and increasing renewable energy consumption. The authors [7] proposed a self-adaptive architecture based on microservices and renewable energy for interactive and batch workloads [140,141,142,143,144,145,146].
To sum up, we concluded that most state-of-the-art papers do not provide a systematic survey on all techniques for energy efficiency in the cloud. Most of the surveys do provide the systematic survey methodology by which they conducted the survey. This review augments previous surveys and introduces a new systematic literature survey to determine and discover the key challenges in existing studies and provide a taxonomy of different techniques conducted on the cloud as energy mitigation techniques. Table 1 compares our conducted systematic survey with previous related studies and highlights the key features that are only included in our survey.

3. Research Methodology

The research approach in this paper follows the different recommendations submitted for the systematic survey performed by [12,13,31,84,119]. This review was performed in accordance to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines.

3.1. Survey Methodology

In the first step, we start with formulating the research questions, discussing the origins of information, collection criteria, quality assurance, and analyzing the research paper’s findings. The complete process can be seen in Figure 3. The following segment focused on common keywords from a particular domain, and only relevant combinations of these words were used in the final search strings. In the next step, inclusion and exclusion criteria were defined so that the review contains only the related research papers.
Strings were thoroughly examined from different information sources, followed by reading the titles and abstracts intended to remove those not relevant or even linked with the energy domain but not connected with cloud computing. The procedures were replicated up to all sources, and the related articles were studied and evaluated in detail. The obtained documents were thoroughly investigated and categorized carefully in the final process. Consequently, this study contains 119 papers after performing each step mentioned above.

3.2. Source of Information

We thoroughly reviewed research articles and conference papers from Scopus, Web of Science, ACM, Research Gate, Google Scholar, books, magazines, and white papers. In Figure 4, we can see the ratio of papers analyzed from multiple sources.
The following databases were used in our search:

3.3. Research Questions

The research questions as seen in Table 2 were constructed to explain the different parameters used in different energy-efficiency techniques and their effects on the environment. These were recognized in the beginning stages of the literature survey. All previous works done in this field have been focused on identifying the unknown relation between energy efficiency and its role in making a sustainable environment.
We successfully answer and attempt to analyze the research questions listed below based upon the data extracted from the primary studies. We also attempt to identify the most and least explored part in cloud computing, which is high in energy consumption and as well as in carbon emissions. The mapping between the PICOC criteria and SLR RQs can be seen in Table 3.
In RQ1, the parameters on which the whole computing system is evaluated are answered. In RQ2 are the factors that affect power consumption, and in RQ3, an estimation was built for the total energy consumption in data centers. In RQ4, the current status of data centers regarding energy efficiency is discussed. Different energy-efficiency techniques and how these techniques will help in reducing carbon footprints are addressed in RQ5 to RQ10. In R11, a brief discussion is conducted on some of the non-technical mechanisms employed for energy mitigation in data centers.

3.4. Search Keywords

After retrieving all the related papers, the search keywords contained various combinations of words. The aim was to cover multiple fields, including energy-efficient technology related to ICT and its role in a sustainable environment.
The keywords used for searching were mainly focused on our research questions. The final search strings included primary, secondary, and additional “AND” operation keywords as shown in Figure 5. The primary keywords included basic terms related to energy, power, and cloud computing, while the secondary keywords consisted of general words, such as environment, sustainable, techniques, and reduction. To ensure the search strings cover all related papers, we added some new keywords by performing the ‘AND’ operation, which combines different keywords to enhance our search for related papers.

3.5. Search String

To formulate our search string, we complied with certain guidelines [55]. The following steps were used for the preparation of the preliminary string.
  • First, the leading search words were taken from the research questions.
  • The following lists were obtained for abbreviations, synonyms, and alternate orthography for the main search words. Finally, by using Boolean AND operators, relevant search keywords were constructed. A pre-search string was created with all possible spellings, synonyms, and abbreviations. The ensuing preliminary string is: (cloud OR energy OR power OR power consumption OR methods OR techniques OR carbon OR “energy efficiency”) AND (empirical ∗ OR “case study” OR “case studies” OR experiment ∗ OR survey) AND (energy efficiency ∗ OR power-efficiency ∗).
We selected 30 papers manually to validate the search terms with the initial search string (cloud energy consumption and energy-efficiency-related). The search strings were searched by the title, abstract, and article keywords in each chosen electronic databases.
We thus verified the search terms for the title, abstract, and keywords from the 30 publications nominated. We assume this validation will ensure that the related literature published in identified digital repositories is wholly covered. However, only 17 of all 30 selected papers were captured by our original search string, which helped us upgrade our search string’s article distribution to a broader range. After comprehensive experiments, the resulting general search string is as follows:
We specified some search strings for every digital library, although different libraries require distinct search formation/syntax rules (Table 4). We also conducted a pilot study in selected digital libraries before shifting to a data-collection method using formulated search strings. Then, we compared the test sequence set with 30 reference articles that we manually generated. The search string obtained 25 of 30 items, thus, validating our search string.

3.6. Selection Criteria

Once the final search strings were constructed, several articles were identified easily using them. However, a substantial number of articles were especially significant for a survey. A total of 2309 papers were selected using the methods discussed above, of which 119 papers were included in this survey based on the selection criteria and other analysis as seen in Figure 6. The following explanations are as follows:
  • Research papers use the term “energy” in different contexts, as published articles associated with computer networks, wireless sensor networks, and neural networks but not with cloud computing.
  • Some of the research articles were excluded because they mainly concentrated on power consumption in parallel computing, thermal energy, mechanical machines, etc. Our surveys aim to cover only power in the cloud-computing domain only. We included papers from January 2008 to February 2022 in our research.

3.6.1. Criteria for Inclusion and Exclusion

Table 5 lists the inclusion and exclusion criteria used to evaluate articles that do not relate to our SMS research questions. While 88 papers were picked at random through an automated search result at the initial stage by which both the authors individually analyzed the inclusion and exclusion criteria. Cohen Kappa statistics measured the agreement of both authors [123] and in the first trial made Kappa = 0.54 according to [65], which was graded into a moderate category.
This mild consensus between the two authors can be due to differing interpretations of inclusion and exclusion criteria. Therefore, the authors held meetings and diligent consultations to establish a shared interpretation of the refined inclusion criteria and address inconsistencies accordingly. Later we re-applied statistics from Cohen Kappa [123] to a different set of 85 papers randomly chosen (substantial). The significant agreement suggests that both authors were sufficiently perspicuous about the inclusion and exclusion requirements and that these criteria could now be added to the paper screening process for more effective and consistent evaluation.
Before screening research articles, we did not use Chen Kappa’s coefficient again, as, after the first phase, we already had a refined criterion. Regarding the last inclusion criteria, the essential motivation was to consider only the energy topic related to the cloud domain as we did not involve articles covering the energy domain in networking or other computing devices. Instead, our goal was only to assess the influence of high energy consumption in the cloud.

3.6.2. Reference Checking

A manually verified referenced list of the 119 relevant articles collected after refining automated and manual search outcomes was then used to conduct a reference snowballing independently by the authors. Additionally, regards to the previously released cloud computing energy-efficient technologies SLRs [18,76,100,122]. We investigated to reduce the probability that related papers cannot be neglected. The results were obtained after the evaluation done by snowball references, and both authors created a consistent list of research papers. In this step, the probability of missing any related literature was minimized to the lowest.

3.6.3. Article Screening

The selection of the primary studies was made by the article screening process, which consists of five steps. The relevant literature search was performed first in December 2021 and later revised in February 2021, covering the literature published until January 2022. Figure 7 shows the screening of articles, and each stage ends with a retrieved item count. In Stage 1, an automated and manual search was performed in selected journals and conference proceedings integrated with six electronic databases for automatic search.
The search results as 2903 articles were obtained by an automatic search and further extracted via Zotero (Zotero is a free and open-source reference management software to manage bibliographic data and related research materials, was originally created at the Center for History and New Media at George Mason University) which was subsequently exported into an MS Excel spreadsheet. In Stage 2, the cleaning process on auto-search results was performed to filter out immaterial articles, such as bibliography, workshops, and symposium summaries, and we manually abolished 1423 articles. In Stage 3, the authors filtered the remaining search results independently in subsequent stages from an excel spreadsheet based on title, abstract, and full text, complying with the inclusion/exclusion criteria specified in Section 3.6.1.
At each phase, we used Kappa statistics [123] to examine the homogeneity between both the authors, as adumbrated by [55]. In Stage 4, based on their titles, we discarded the articles, and the statistical analysis resulted in a kappa value of (0.74) and (0.71) respectively, for Phase 1 and Phase 2, which reflects significant agreement in both cases, according to [65]. In the event of a disagreement, the authors agreed to include research papers in the next step to reducing the risk of eradicating the relevant research articles suggested by [87].
While following this procedure, we chose 221 (192 + 29) articles and excluded 312 research papers from Phase 1 and Phase 2, and these selected articles were supplementarily used in the next stage as input. The subsequent consistency of agreement between the two authors was considered significant in both phases. In Stage 5, both the authors conducted the inclusion and exclusion procedure independently by reading the complete text, and 110 (90 + 20) final papers were obtained. We discussed 19 articles (10 in Phase 1 and 9 in phase 2) to settle the lack of congruity among the two authors; however, both insisted on an additional 9 out of 19.
Hence, this process included 119 respectively, after complete text-based removal. This resulted in the selection of 1480 (1288 + 192) in Stage 2, 947 from Stage 3 (475 + 58), and 221 (192 + 29) from Stage 4. After all stages, we had 221 articles in the final, and reference snowballing results were further processed based on the abstract and full text. Eventually, the studies selected after Stage 5 (including reference snowballing) were regarded as the final list of 119 primary studies.

3.6.4. Quality Assessment of Study

While performing an SLR, the quality of studies needs to be evaluated to determine the study’s significance during the synthesis of the results. This helps to pick high-quality studies to deliver accurate results, as indicated by [55]. Evaluating research quality in an SMS is not an obligatory activity; however, we attempted to do it as an add-on. However, we decided to integrate our mapping protocol with a quality assessment step based on a previous SMS [122]. In the first instance, we developed a quality evaluation checklist based on [55].
This section involves ten questions covering different research areas of the same domain, including architecture, conduct, and evaluating data and conclusion. Later, to answer the consistency checklist questions, we read the full text of each article. The author evaluated each study on an iterative three-point scale with a rating ‘yes’, ‘partially’, or ‘no’ for each question (the first author evaluated each study’s qualities, and the second author later checked the outcomes of each study).
The numeric values 0, 0.5, and 1 to ‘Yes’, ‘partially’, and ‘no’ were quantified for each question. The study’s final quality score was subsequently obtained by summarizing all questions relating to quality assessments. Studies that achieved over 4.4 were eligible for inclusion in the final list. Additionally, Section 3.6.3 outlines the technique employed to evert the incompatibility of the quality evaluation process. A few of the questions did not apply to a few primary studies, and thus we used a point scale ‘NA’ value for those cases.
Table 6 shows the results of this quality evaluation phase as it indicates the positive responses to most quality evaluation questions (QA1, QA3, QA6, QA9, and QA10). In QA1, we analyzed each study’s abstract and initial sections to see if the research goals were clearly defined or not. The 119 primary studies identified the study goals appropriately (87.6%) or partially (12.4%). Concerning QA2, we examined whether the authors clearly outline the methodology of research used to determine the effect of the cloud’s energy efficiency on the environment.
We discussed the study design and methods sections for this purpose and found that 76.6% of the studies appropriately explained the research methodology. To answer QA3–QA7, we must revise the complete paper composition. Approximately 93.5% of studies have identified the characteristics or metrics for energy efficiency in clouds (QA3), and 65.4% of reviews correctly identified these parameters or measures for energy consumption in cloud systems so that it was easy to evaluate the level of optimization in energy consumption (QA4). Concerning QA5 and QA6, we noted that the dataset’s exact size and language were stated correctly (72% and 94.7%, respectively).
Regarding QA7, very few (only 14.3 percent) SPs could measure the statistical importance of the ensues achieved involving the effect of high energy consumption in the cloud. We reviewed the articles’ discussion, shortcomings, and improvements for QA8 and found no validity threats. We replied ‘Yes’ to this question for studies where validity threats are specifically addressed and marked ‘Partially’ for articles where validity threats were listed without sufficient explanation. To answer QA9 and QA10, we reviewed the papers’ critical determinations, discussion, and conclusions. This helps us decide whether all the questions for analysis are answered and the research results were reported adequately. For most research, positive answers were obtained (99.1% and 96.2%, respectively).

3.6.5. Data Extraction

In order to address the research questions specified in Section 3.3, we extracted the related information from the collection of 119 papers. Data extraction information was documented after reading each article in detail. The points below were extracted from every article.
  • Primary and complete reference information, including title, author, title, and publication year.
  • Determine energy efficiency techniques used in the cloud.
  • Set features such as the name, type of techniques, and language of programming used.
  • Empirical results concerning the influence of high energy consumption on cloud performance and the environment.

4. Results

Only high-quality papers were used for a systematic research approach to classifying the energy-efficiency techniques and their role in cloud computing. A total of 2903 publications were found after searching for energy-management applications in the cloud computing area. One hundred nineteen survey papers were finalized for the survey and distribution of published articles in different years as shown in Figure 8.
There are the following numbers of articles in various categories:
  • A total of 28 papers were extracted for hardware energy-efficient techniques (controlling the frequency and voltage of servers). A significant portion of 32 papers was extracted for software energy-efficiency techniques.
  • Different energy efficiency techniques have been performed by different types of applications categorized as 17 papers for bio-inspired, 11 for consolidation, for power-management techniques 13 papers were selected.
  • For thermal and cooling techniques, 13 research papers and five articles for non-technical and various techniques were finalized for the survey.

4.1. Answer to Research Questions

All 119 PSs were thoroughly reviewed to extract the required information that is demanded to respond the research questions mentioned in Section 3.1. The answers to the research questions are as follows. While answering the research question we are referring to primary studies in format [S and study reference number] to avoid confusion for readers.

4.1.1. RQ1: Which Measuring Parameters Are Considered for Power Consumption in Our Selected Studies?

It is essential to understand the energy consumption pattern for optimizing usage. In cloud environments, servers consume a significant portion of energy, and their energy consumption depends on the type of usage and amount of computation power required [56]. It also depends upon the type of computation the server is performing. For example, data recovery and processing will consume different amounts of energy. Usually, networking equipment, lightning, and cooling devices lend to the overall energy consumption. However, the contribution of each of them corresponds to a few percentage points of the total consumption; however, this is increasing [49].
The PDUs and UPSs consume considerable resources and increase the workload, leading to a rise in consumption. The UPS continuously charges and supplies power until generators can start in the utility’s failure. For a basic understanding of parameters used in power consumption, let us start with the mathematical formula [65] for the total energy consumption of an active server. The active server’s overall energy consumption is the amount of energy consumed in a time frame when the machine is fixed and dynamic, defined as ΔE Total, is prepared as follows:
ΔE Total = ΔE Fix + ΔE Total
Further energy is generated by overhead scheduling, as indicated by ESched. This paper mainly concentrates on energy consumption, which requires various parameters, such as server idle conditions, cooling systems, computing, storage, and the use of communications resources. The following are defined:
(i)
Server idle mode energy consumption is symbolized by EIdle.
(ii)
Cooling system energy consumed is symbolized by Ecool.
(iii)
Computation resource consumed energy is symbolized by ECompu.
(iv)
Storage resource consumed energy is symbolized by EStore.
(v)
Communication resource consumed of energy is symbolized by Ecommu.
Equation (1) above can, therefore, be translated into:
ΔE Total = (Eidle + Ecool + ECommu + EStore + ECompu) + ESched
Metrics are used to understand the working as well as evaluating it, without certain parameters [72], we are not able to validate a technique. Other parameters used to measure energy constraints are given below.
i.
The energy consumption must be considered in accordance with performance metrics on efficient use of energy in infrastructure. Below are some energy consumption assessment metrics [17] as seen in Table 7.
ii.
Some sustainability metrics considered for evaluating the total energy consumption and its environmental effects, which are aligned to international carbon reduction initiatives can be seen in Table 8:
iii.
Rating systems of data centers are dealing with energy efficiency metrics and other, such as operational and regulatory term as global metrics which can be seen in Table 9.

4.1.2. RQ2: Describe the Impact of High Energy Consumption by Cloud Data Centers; Explain the Estimation on Total Energy Consumption by Data Centers

Data centers can be considered the internet’s brain, and their role is to process, store, and transmit data consisting of the information and services on which we rely in our daily lives [68,69]. This could be video streaming, email, social media, online collaboration, or scientific computing. Data centers’ power and energy requirements are very high because they use them for computing and cooling, thus, increasing their expenses and high carbon emissions. The consumption of energy in data centers worldwide is estimated at 1.4% of the total EEC (electricity energy consumption) and grows annually at 12% [28].
Usually, with a data center, several network appliances and equipment are connected to the internet, facilitating data traffic flows that consume vast amounts of energy, progressively transforming into heat. The heat generated must be extracted from the data centers; otherwise, cooling devices are needed that further consume more electricity [85]. As seen in Figure 9 below, most of the power consumption of data centers is by infrastructure and servers. Many of the largest data centers worldwide also have several thousand IT hardware devices that consume more than 100 megawatts (MW) of energy, sufficient to supply about 80,000 U.S. homes (US DOE 2020) [88].
While there has been a rise in the global demand for data centers, the number of users is also increasing day by day, leading to high Global IP traffic, and the volume of internet data increased by more than double from 2010 to 2018 [27], whereas data center storage capacities worldwide decreased by 25% [71]. The number of computing instances operating on the servers worldwide increased by six times. Most of the data on electricity consumption by the data center is not currently officially available. However, we use data obtained on regional levels, and thus we can create some estimation of the energy consumption.
Using some statistical approaches, also called “bottom-up” models, helps to measure the energy consumption accounting for the installed IT equipment in various data centers, and their energy consumption characteristics used to estimate the overall energy consumption [87]. Although bottom-up studies consider many factors for driving an assessment of energy usage, they do not appear very frequently as they are data-dependent and time-intensive. As [57] reported, data centers accounted for approximately 1.1% to 1.5% of the global energy consumption in 2010.
However, with such extrapolation models, we continue to predict significant growth in power usage in data centers, provided that the demand metrics on which they depend often rise rapidly. Some estimations conducted by various authors [8,16] suggested that from 2010 energy used in data centers is doubling and is expected to grow at the same pace in the future. Ref. [46] attempted to create a widespread assumption that a significant increase in demand for data is directly equal to the rapidly rising energy usage of data centers [47,48].
Over the last decade, the global energy usage in data centers has possibly increased by 6 percent between 2010 and 2018 [71]. These findings were centered on integrating various recent datasets that considered account installed stocks, functional properties, and the total energy consumption in data centers. Currently, the community faces a sustainable energy economy because of the high electricity use by CDCs. The CDC’S energy consumption is continuously rising. As shown in Figure 10 most of the energy is consumed by the cooling system and networking equipment. By 2030, the energy consumed will be 8000 Tera Watt hours (TWh) [6].
However, information on the electricity consumed by global data centers is a valuable benchmark to test affirmation regarding the CO2 consequences of data center services. One frequent assumption is that global data centers expel about 900 billion kilograms of CO2 produced by the global aviation industry [86]. The latest claims also indicate that the emissions from 30 min of watching Netflix (1.6 kg CO2) are the same as the emissions from driving nearly four miles. This statement is supported by the assumption that Netflix streaming services in data centers consume approximately 370 TWh per year [50].
However, this is 1.8-times greater than the projected 205 Tera Watt hours (TWh) for all the collective data centers worldwide that provide society with countless former information services other than simply watching Netflix videos. Three main vital factors influence energy performance [58]. First, IT manufacturers must continue to improve technology for energy efficiency in IT devices, especially servers and storage drives [59,60,61,62,63,64]. Second, increasing virtualization software deployment, which empowers many applications to run on a single server, undoubtedly reduces every application’s energy consumption. Third, promote the trend toward the large cloud and hyper-scale class data centers that use ultra-efficient cooling systems to reduce energy consumption [88,89,90,91].

4.1.3. RQ3: The Current Status of Cloud Energy Efficiency

With the advancement of cloud-based technologies and increasing collaborative work among customers, remote connectivity has made this feasible [73,79]. Almost all cloud-based companies [82,83], such as Amazon, Google, Microsoft, Sun, and IBM focus on developing improved data centers, thus, strengthening their structure for global data centers [77].
These data centers currently need enormous power or energy to process network equipment, monitors, cooling fans, machines, air conditioners, etc. The worldwide energy use in data centers is 65% higher than in the previous year. Both the issues are due to high energy usage and greenhouse emissions [49,50]. We can suppose that cloud-based systems are one of the causes of global warming because some gasses produced by cloud-storage infrastructure, such as carbon dioxide, carbon monoxide, have caused significant problems for our environment [70].
Unless data centers’ electricity and power usage are minimized, this will favor global warming and an intelligent economy, significantly contributing to environmental protection. As per the predictions, the estimated cost of cloud activities for the Amazon website and operating its data centers is about 53 percent of the overall budget. Energy-related expenses amount to 42 percent of the overall cost, including the cost of power needed for the system operations and cooling facilities [41].
Data centers use around 2% of the world’s energy; however, by 2030, this is projected to be as high as 8%. According to Hewlett Packard Enterprise’s research, only about 6% of all data generated by us is being used. That means 94% appears to be in a large cyber deposit that would cause significant carbon footprints if it creates a loop [75]. In 2018, the total data available on the planet totaled 33 Zettabytes; however, by an estimation conducted by the International Data Corporation, this will reach 175 zettabytes at the end of 2030. Today, our community generates about 2.5 quintillion bytes of data every day, which will be stored and processed in a data center, leading to gradual hikes in the cloud data center’s energy consumption.
Energy efficiency issues have steadily increased since advances in the computing industry with high energy demand were made [37]. This segment focuses on and aims to improve the energy efficiency in data centers. The effect of increased energy use in data centers should be investigated, and this article analyzes the various energy conservation techniques being implemented in cloud data centers. To reduce inevitable data center energy consumption, a parallel investment in renewable power sources would be required [71].

4.1.4. RQ4: What Kind of Different Techniques Used for Energy Efficiency Proposed by Our Selected Studies?

There are many ways to optimizes energy consumption and to improve resource reliability. However, there is scope for improvement in different energy-saving techniques as well as in energy-aware resource allocation heuristics, ISN, SN, and SI policies using green control algorithms, thermal efficient resource management, bin packing algorithms for virtual machine placement, and dynamic allocation based on the current utilization of resources [S17].
Different countries’ governments are now collecting taxes on greenhouse gases, such as CO2 dissipation from IT industries. The total cost for cloud services increases due to inefficient energy utilization to provide a proper understanding of various techniques. We divide these techniques into seven categories as shown in Figure 11. Hardware-level, software-level (virtualization and consolidation), and power-aware management; thermal-based, bio-inspired, and other various techniques, such as non-technical techniques. These are the fundamental techniques by which energy consumption can be reduced and interconnected with each other. The latest improvements at the hardware level, such as the latest processors, have proven to be more energy-efficient than older ones.
Moreover, improvements to the server’s processing capability are made by regulating the frequencies and voltages of the servers. The software solution has a more significant effect on energy usage [49]. It is mainly concerned with servers’ energy, which involves reducing the number of servers and the operating number of memory nodes. Network techniques that reduce the traffic between servers take less energy while processing [104]. Power management deals with resource allocation and migrating strategies that lead to more efficient energy usage. Finally, non-technical includes building organization and data center designs [62]. This also helps to reduce energy. Even weather and nearby available renewable energy are essential factors for energy-efficient techniques.

4.1.5. RQ5: Describe Various Energy Efficiency Techniques Employed at the Hardware Level

At the hardware level, we will deal with hardware components by which we are attempting to efficiently use energy even with various servers with different computing capabilities. Efficiency can be achieved by controlling the servers’ frequencies and voltages, computation quality, speed of processors, and other hardware components [67]. The processing unit consumes more energy than other hardware components as energy efficiency by processors is improved over time as processors are upgraded in terms of computation, speed, and performance.
Still, more energy optimization needs to be achieved by modifying and upgrading the architecture of microprocessors. The carbon footprint depends on the energy source applied in the data centers [51]. Green computing is incorporated to solve both problems of increasing carbon footprint and energy optimization [52,53,54]. Before discussing hardware-level techniques in detail, we first look at the causes of problems, such as high energy consumption and high carbon footprints.
Moreover, we also attempt to understand the absolute need for cloud and cloud data centers. Today most gadgets, such as smartphones, tablets, smartwatches, health care devices, and sensors connect to clouds for their private data storage purpose. Software applications, such as e-mails, messengers, enterprise apps, social web networks, e-cart apps, audio and video streaming apps, broadcasting, and entertainment services also utilize cloud services to store, process, share, and secure their data.
The most popular search engine giant Google hosts all of its services, such as Gmail, Google Earth, Google Drive, Google Play, and YouTube on its cloud platform (GCF) to offer high-quality services to its worldwide customers. Currently, almost more than 50 percent of mobile users own smartphones and Android phones, which has become prevalent due to their flexibility and usability. Their servers are being hosted in the cloud, including thousands of apps and websites. The most extensive database globally, DNS is also expanding as many domain names are purchased and updated. Cloud computing requires a massive number of data centers or server farms.
Each data center comprises hundreds or thousands of physical machines organized in hundreds of racks that can run without virtual machines (VMs). For providing access to mail, videos, pictures, etc., anytime, data are being distributed and stored in massive data centers. The backup of all data is stored and synchronized at the geographically different data centers to protect data in emergencies and unpredictable natural calamities, such as tsunamis and earthquakes.
Google has at least thirty data centers worldwide and more than a million servers that usually consume 500–681 megawatts. Amazon web services use 38 data centers [34] and 454,400 servers globally. We learned that energy efficiency at the hardware level is achieved from design to the manufacturing and implementation phase of hardware. Intensive computation by processors generates excessive heat, and thus cooling is needed to control the heat and temperature. There is always a need to use some software-level solution for hardware-level techniques as these are very important to efficiently managing hardware resources.
Providing minimum required energy to machines saves energy. However, performance degrades by this, and on the other hand, supplying max required energy increases performance but increases heat liberation and power consumption. In hardware techniques, we include cooling equipment, lighting equipment, power supply, architecture and framework, infrastructure, Virtual Machine Allocation- and Scheduling-Based Techniques. We are treating power supply as a separate domain of techniques. We will discuss different techniques proposed by various researchers based on hardware to reduce energy consumption.
Ref. [S2] demonstrated that the Network Connectivity Proxy (NCP) can be used in an organization to boost computer’s energy efficiency significantly. Authors say that PC idle time in enterprises is roughly more than 70 percent, the energy used by new desktop PCs is 1.5 W while off, 2.5 W in sleep, and 60 W energy consumed while on. If we minimize the idle time by putting it into sleep mode or off mode, the total yearly energy saving would be nearly 400 kWh. If one enterprise consists of 10,000 such PCs, the energy savings would be 4,000,000 kWh.
Ref. [S3] introduced a green cloud architecture, which seeks to reduce data centers’ resource usage and facilitates extensive online monitoring, lives virtual machine migration, and VM location optimization. Evaluation is conducted using the online real-time game Tremulous as a VM program. Ref. [S1] attempted to address the energy issue by analyzing how much energy is required for virtualized environments. Based on the EARI model, the electrical costs associated with virtual machines, such as booting, usage, and migration, were calculated and analyzed quantitatively.
Such application frameworks combine load-balancing solutions with migration facilities and an on/off infrastructure associated with predictive mechanisms. Ref. [S3] proposed a new framework for energy efficiency with a scalable cloud computing architecture feature with minimal performance overhead. The network’s performance is enhanced in the data center through basic power-aware scheduling methods, variable resource management, live migration, and virtual-machine architecture.
Ref. [S4] introduced a display power-management technique that reduces the screen’s energy consumption as battery life is still considered a critical issue. When the user is not present in front of the PC or desktop, it will detect by capturing images via webcam, and then the image is processed. If the user is not looking at a desktop or laptop display device, the display power controller senses the user’s intent and promotes low power activities by up to 50%. With this, it can reduce energy consumption by up to 13%.
Ref. [S20] suggests a DVFS strategy on cluster systems deployed with a different type of processor that can conveniently work at different voltage and frequency scales to incorporate a power-scheduling algorithm with time-limitation tasks and decreased energy usage without breaching the SLA. A new algorithm is being introduced using the Dynamic Voltage and Frequency Scaling (DVFS) technique and a DVS server that controls the voltage supply. Additional hardware DVS servers help to minimize the energy consumption, reduce CO2 emissions, and fulfill certain quality of service to check that the scheduled work completed within its time limit [S5].
Cloud servers in the data centers are only busy from 10% to 30% of the time; on average, 70% of the time, they remain idle. Virtual machines are consolidated into a minimum number of physical machines, and idle machines are placed into lower power states, such as hibernate, sleep, or off. Green algorithms [S17] achieve cost optimization using three N-Control policies: SI, ISN, and SN. Energy and cost are optimized using those strategies.
Switching physical machines into sleep, off, and idle mode frequently needs more power and performance. N-Control policy limits more switching, which reduces both costs and energy. A Green Cloud Architecture by [S6] for system efficiency in a data center was designed to implement modern energy sustainable design, VM system imaging, and image processing modules to discover innovative ways to save energy.
The Integrated Green Cloud Architecture was introduced by [S7], which provides a Green Cloud client-based middleware that allows us to improve cloud infrastructure control and deployment. The middleware intelligently tackles decisions to be inferred by using specific pre-defined parameters, such as quality of service, SLAs, and equipment specification. Energy-efficient approaches were introduced in e-learning to reduce the cloud-based online study’s environmental effects.
The e-learning strategy has significantly transformed the educational systems, decreased the use of paper and documentation, and reduced energy and CO2 emissions. Therefore, cloud-oriented green computer architecture [S8] reduces costs, saves energy, and encourages companies to install and manage apps with limited resources. Ref. [S9] develop a Green Cloud Broker while considering energy-efficient indicators and providing operations to resolve resource procurement problems in an environment-friendly way.
This mechanism can also determine the allocation and payment for the job submitted. Ref. [S10] illustrates the platform firmware’s role in DRAM power optimization and control. They proposed a UEFI-based firmware for memory allocation in an energy-aware manner and accurately estimating DRAM power, optimizes the DRAM locality, and enhances the DRAM power limiting (RAPL). PAPI Performance Analysis Library [S11] to measure energy consumption and workload power, RAPL capabilities, and obtain real-time immediate code analysis. While utilizing PAPI Higher-Level Tools, full power and energy analytics are automatically configured, primarily when used with the new PAPI versions.
A quantitative study by [S12] was conducted at the chip level to calculate power and productivity in five hardware generations using different specifications. The results recommend that while designing and evaluating energy-efficient hardware, it is essential to include native and managed workloads and suggest including on-chip power meters, making it easy for the researcher to optimize power and performance.
Ref. [S13] offers a hardware accelerator designed specifically for neural networks in which particular emphasis is placed on the effects of memory productivity and energy consumption. Microsoft uses the server with Field Programmable Gate Arrays (FPGAs) to build a Convolution Neural Network (CNN) accelerator. Ref. [S14] provides a fully automated process for designing and optimizing DNN hardware accelerators for software, architecture, and circuit stages. These studies are related to overall energy efficiency by a general-purpose processor.
Characterization methods are used to identify energy’s effect on various designs with power-restricted IoT or mobile devices. Heterogeneous Multicore Platform Energy Management was proposed by [S14] to improve performance and efficiency, which is better than the DVFS technique. H-EARTtH algorithm manages and schedules a heterogeneous processor’s optimal point. Users in clouds demand a bundle of virtual machine (VM) instances.
A new strategy was proposed by [S15] to manage available resources properly and increase its profit. To overcome a physical machine resource management problem, the authors proposed a G-PMRM and VCG- PMRM algorithm and even a winner-determination algorithm. G- PMRM was more efficient for deciding allocation faster than the VCG-PMRM algorithm. The performance of G-PMRM improved as the number of users increased.
Ref. [S16] explores a price-based approach for energy-efficient energy allocation in multiuser relay-assisted networks. The authors imposed a network price as a penalty if the power consumption reached a limit value and analyzed its tradeoff effect on the EE and spectral effectiveness (SE). Routers, switches, bridges, and various other network components require energy in a computing environment. The routing algorithm uses minimum greenhouse gas emissions.
Ref. [S27] proposed a new energy-efficient model for wireless storage area networks. The linear power model and low power blade models were used to decrease the energy consumption in the central processing unit, measure the number of tasks performed by the user, and select them based on the user server’s needs. Power-management systems such as voltage scaling, frequency scaling, and dynamic shutdown strategies were optimized for the remaining servers.
Ref. [S21] proposed the NTV concept, promising energy efficiency improvement in magnitude over the current silicon systems range. Intel’s Near threshold voltage of IA-32 processor undertakes to explain the circuit techniques and benefits of NTV design. It addresses critical technologies vital for harnessing NTV’s real potential to create a new ‘green’ computing era. A thin client approach for mobile devices was proposed by [S77], which showed a 57% improvement in performance compared with old traditional self-reliant devices.
Laptops are more energy-efficient than desktops; however, the laptop’s average life is less than a desktop because cooling facilities increase a machine’s life. Regular maintenance of internal cooling system components in laptops and desktops also causes a change in energy consumption. E-waste is another aspect of achieving green computing, as many greenhouse gases liberation, pollution at the time of manufacturing will cause many environmental issues. Proper recycling of hardware components is essential to reduce the carbon footprint and environmental concerns.
Eco-aware online energy management and load scheduling methods were proposed based on renewable energy sources [S117]. Its vital role is to minimize eco-conscious cloud-data-center power consumption costs while ensuring a quality experience (QoE). The method is implemented using Lyapunov optimization theory to develop an online control algorithm to solve an optimization problem known as a constrained stochastic problem. Ref. [S22] aimed at energy-aware scheduling for non-real-time and real-time tasks on the cloud.
During peak hours, priority-based recommendations help with processing requests for keeping performance high. The authors in [S29] addressed physical machine resource management issues in the cloud data center. They attempted to provide a solution using an optimal auction-based setting and design strategy. The winner-determination algorithm was used for selecting users, provisions of VM to PM, allocation of selected users, and payment function for determining the amount to be paid by the user to the service provider [67]. Network software-defined technique (SDN) is a new approach for energy-aware flow scheduling, i.e., time-dimensional scheduling and EXR for every flow. In other words, the flow still only uses its routing path links.
EXR can conserve network resources efficiently relative to standard fair sharing routing (FSR) and, while setting higher routing priorities for lower flows, dramatically minimizes flow completion times. Ref. [S19] proposed an architectural model for energy-efficient cloud computing. The energy allocation heuristics, which offer data center services to consumer systems, boost the data center’s energy performance while satisfying the quality of service.
Ref. [S26] designed a model to catch the intrinsic trade-off among the electricity cost and WAN connectivity costs and formulate the optimal VM placement problem, which is NP-hard because of its binary and quadratic existence. The authors in [S35] introduced new adaptive heuristics for complex VM consolidation, which focused on understanding and analyzing the resources used by VM. The algorithms aimed to reduce energy consumption and maintain high compliance with the agreement by using real-world workload traces from more than a thousand Planet Lab VMs; the authors validated the proposed algorithms’ high efficiency.
Ref. [S24] provides a methodology for analyzing how the conventional integration of the data center with dispatched load impacts grid dynamics and performance. In contrast, adding dispatchable data centers by the power grid reduces stranded power and improves grid cost and stability, even at high RPS. [S25] discussed the complexities of harmonized energy management in data centers and proposed a modern and optimized, unified energy management system framework for data centers developed within the European Commission-funded GENiC project.
This system prototype [S51] implemented joint loads and thermal-control algorithms [3]. Different hardware techniques for energy efficiency are compared in Table 10, and the techniques are evaluated and compared by their main benefits and drawbacks. All hardware-level technique taxonomy is shown in Figure 12 and Figure 13, and the distribution of research articles can be seen.

4.1.6. RQ6: Describe the Various Energy Efficiency Techniques Employed at the Software Level

Mainly, there are two constitutional approaches to eliminating energy usage at the software level. The first is minimizing the resources used by computers (reducing the number of active servers). The other way is to decrease the energy consumed by memory (cutting down the number of active memory ports). This type of technique usually comes under the optimization of scheduling, a standard green cloud approach that is more economical than hardware optimization.
The best way to minimize power consumption [11] at the software level is to map the requests between virtual machines and physical servers correctly. Currently, the focus of virtualization technology at the data center is on energy management; therefore, we consider all virtualization techniques as software techniques. There are two primary issues in the cloud framework: where to position the VMs and where to relocate VM if necessary.
Several proposed techniques are discussed below to solve these issues and reduce energy. The impact of virtualization can be seen on the Amazon EC2 data center performance network, which was assessed by [66]. Virtualization is a concept designed to run several logical (virtual) computers on a single physical computer (hardware device), which enables VM management techniques resulting in high efficiency with low cost as managed by using the abstraction process [42,43,44,45].
Services and resource efficiency and availability were enhanced by dynamic migration and aggregating a physical server’s collections into one server. One of the significant energy efficiency challenges in virtual cloud environments is where new VM requests should be laid on physical servers. Virtualization [37] is defined as multiple virtual machines that will execute a couple of tasks simultaneously running on the same physical server with a dedicated operating system and hosted applications.
A hypervisor is the system software that works as an operating system (abstraction layer) for virtual machines and coordinates with the underlying hardware components according to the virtual machine’s predefined instructions [124,125,126,127]. Virtualization is not a new concept in the IT sector as it has already been implemented with our grand old Main Frames, which belong to second-generation computing devices. Generally, cloud systems are designed with high-end configuration components, such as RAM, processors, disks, routers, and switches.
Traditional processing methods (sequential) will allocate the entire resource set to the running tasks before they begin. The allocated resources of a task would not be able to exchange with any other running tasks. In this way, the allocated resources are underutilized, blocked with some tasks, and the execution takes more time to complete. Hypervisor-based VMs are designed to run multiple jobs in parallel on the same machine with resource sharing facilities to overcome the sequential processing limitations.
Ref. [110] studied and analyzed cloud virtualization for reducing physical hardware by using different virtualization techniques. While analyzing this study, we discovered that hypervisor selection has a significant impact on power consumption. They also attempted to use the optimal number of VMs as, if they cross a specific limit, they will consume more power. The hypervisor is the significant virtualization component of physical servers, which uses only 10% of their capacity [56].
Thus, with virtualization, we have an advantage known as server consolidation. Un-utilized physical machines are migrated into virtual machines that are further combined into a single unit to reduce consumption and the amount of hardware required [138]. Achieving high performance from resources, reducing the frequent investments in infrastructure, and efficient resource utilization are the main advantages of virtualization. Dynamic workload balancing with VMs, resource sharing across VMs, the design of secured VMs, and energy optimization techniques for virtualization are green cloud trending activities.
Ref. [S18] introduced the Ena-Cloud method, contributing to an alive and dynamic application placement policy. Ena-Cloud uses virtual servers to install applications and deployment tools, to reduce the number of machines operating and power consumption. Ref. [S114] develop an online energy-aware supply strategy for consolidated and virtualized computing systems for HPC applications. All similar tasks are grouped and sent to the same host system for execution, while an unused subsystem or new element is excluded by switching off them.
Energy efficiency can be achieved through workload-aware, just-right dynamic provisioning, and the ability to supply host subsystems where VM mapping of VMs is not required. Consolidation excessively reduces the number of servers that remain idle and wastes energy. Ref. [S28] developed a technology that consolidates VMs dynamically depending on the adjustable utilization thresholds to meet SLAs [129]. They recommended methods for modifying consumption levels dynamically by evaluating empirical data obtained over the existence of VMs.
Techniques for optimization must choose the correct candidate for the migration of virtual machines. The selected VMs and the new VMs requested by the user are placed into the physical node. Virtualization is proven to be a solution for energy optimization; however, more improvements in algorithms for VM consolidation and deconsolidation are still required. While considering SLA’s limitations, ref. [S58] proposed energy-efficient technology to provide resources to manage energy wastage problems through scheduling algorithms.
According to SLAs, resource managers can unify virtual machines with physical machines to meet client SLA requests. In [35], the authors proposed a nature-inspired ant colony system to consolidate VMs; however, the probability of a peak situation arising increases as the consolidation occurs. Similar energy and performance-aware policies for deconsolidation are required. Ref. [S30] also proved that optimization is equally significant while working on energy efficiency [130,131,132]. They proposed a method for scheduling VM workflows in hybrid and private clouds in which pre-power techniques and the least-load-first algorithms were used to make it energy efficient.
As with the hybrid pre-power planning strategy, the time needed to respond to the incoming request decreases, and workload balancing problems are solved with the least-loaded first algorithm. An Energy-Efficient Scheduling System (EESS) [S31] for Migration, Cloning, and Regeneration of ideas through the First Come First Out policy using the minimum load distribution. By Hybrid Energy-Efficient Scheduling Algorithms, the requests for incoming VMs are submitted to the appropriate VMs. However, if requests rise above one level, the scheme will distribute the workload using migration.
Ref. [S55] proposed a technique to solve the problem of Static Virtual Machine Allocation (SVMAP), and a genetic power-aware algorithm (GAPA) was introduced. The GAPA and a baseline scheduling algorithm (BFD) overcome the same SVMAP problem by sorting the virtualmachines list at the beginning and using the best fit decreasing algorithm. Consequently, the total energy consumption of the GAPA algorithm was lower than the simulated baseline algorithm. One such technique was developed by [S28] based on the heuristics of reallocation.
They used the live migration technique for VM migration utilizing some heuristics related to reallocation. These heuristics are based on the current requirements by following up on the desired quality of service. Reallocation is performed to minimize the workload of some nodes that carry loads beyond their limits and exclude unused nodes for power saving. Ref. [S36] proposed an algorithm MLS-ONC using a multi-start local search algorithm implemented in the Open-Nebula through which the cloud infrastructure was fragmented and distributed geographically to reduce the energy consumption and carbon dioxide emissions.
Ref. [S32] proposed a Cloud Global Optimization algorithm for allocating resources at runtime in an energy-efficient way by exploiting the VM live migration technique’s ability. The basic idea is to use a unified approach based on consolidation and rearrangement in which it first moves the VM from a server with less load to heavily loaded servers. Next, it changes the VM from the old to a modern server.
A control technique for scheduling VMs and IAS workload in a data center was developed by [S37]. This method offers a better OptSched timer that uses the specified optimization requests time Of VMM. The algorithm aims to reduce maintenance workload requirements as well as the number of servers and the improvement of the cumulative machine uptime (CMU).
An application for energy awareness was proposed by [S38] in which resource scaling and management employ IaaS architectures. The main aim is to ensure that only a minimum number of servers operates beyond their capacity. It also involves the transfer of software from the low-loaded storage server to another server using a mechanism by which unnecessary energy usage can be avoided by shutting off the device, which decreases the operating cost. By reducing the number of active servers, one can reduce the energy consumption. This is usually implemented with an optimized scheduling method [11].
A new, heuristic, energy-efficient approach to deploying VMs in a cloud data center based on a statistical analysis of historical data was proposed by [S39]. The application uses multiple correlation coefficients (MCC) to calculate the intensity of the correlation between two variables and pick a server to provide an acceptable trade-off between power efficiency and SLA violations.
The application uses multiple correlation coefficients (MCC). If the CPU usage coefficient for running VMs is high on a server, then the SLA violations in the data centers will more likely to happen. A solution based on real ants’ actions was suggested by [S40]. It uses the Ant Colony Optimization approach to deploy the VM to reduce the number of active servers, as VM deployment is conducted by dynamically estimating the current server load.
Any ant (server) receives and begins scheduling some of the virtual machines’ requests in this algorithm. After all the ants have formulated their solution, the best solution is chosen for the target function. Energy-efficient allocation strategy and scheduling algorithms were developed in the context of the quality of service requirements and power consumption characteristics [S34]. This random machine is chosen to migrate an overloaded server based on a discreet random variable. MADLVF algorithms [S41] were indeed prepared to resolve inconsistent resource use, high energy consumption, and increased CO2 emissions [135].
The Interior Search Algorithm (ISA) can be used, which decreases the Data Center’s energy consumption. The inefficient use of resources [S42] can be excluded by the Energy-Intensive virtual machine allocation method. The results demonstrate that the energy use of genetic algorithms (GA) and best-fit reductions (BFDs) were 90–95% higher than the estimated EE-IS, which was approximately 65%. The new EE-IS strategy also expanded the average electricity saving by around 30%. Ref. [S43] suggested novel QoS-aware virtual machine allocation algorithms.
The type of allocation depends on the resource history to boost the service efficiency and reduce electricity usage. Ref. [S44] introduced a power-law featured management framework named VMPL for energy efficiency. VMPL predicts the video’s resource utilization based on its popularity, ensures enough resources for upcoming videos, and turns off idle servers for power saving. The results show that the energy consumption and resource utilization relative to the Nash and Best-Fit Algorithms decreased using the proposed algorithm [136].
The energy utilization model was suggested by [S45] based on the mathematical approach. This approach sets the workload threshold for each server, and if the server exceeds the capacity, the VM is moved from an overloaded server to a new one. Ref. [S46] planned to build two VM scheduling algorithms to minimize power consumption and migration costs.
Ref. [S47] introduced three VM policies for placement and migration. If there is a rise in or shortage in the number of VM, a strategy is required for VM migration service wholly based on the policies named FDT, DRT, and DDT. This suggested approach indicates that the minimum migration process and few SLA violations reduce CO2 emissions.
Ref. [S48] addressed the VM placement problem and proposed a heuristic greedy VM and live migration algorithm that will to improve the use of the resources and lower energy use. The heuristic algorithm estimates the workload and maps the storage-sensitive and CPU-intensive workload to the same physical server. Compared with single-objective approaches that focused on CPU utilization only, significant developments were made in energy conservation and workload balancing using multi-objective approaches.
Ref. [S49] proposed a technique for VM deployment that supports VM application planning and live migration to reduce the number of active nodes in virtual databases. authors also developed a statistical, mathematical framework for VM implementation that incorporates the entire virtualization expense into the complex migration process and seeks to decrease active host involvement, reducing power consumption.
Ref. [S50] proposed a general algorithm based on a logistical regression model and a median absolute derivation for host overloading detection. In combination with VM-consolidation techniques, the VM migration [137,138] will help to prevent overloading and reduce the number of active servers, ultimately saving energy. Ref. [S51] suggested a VM placement algorithm, in which a process was mapped to VMs user demand, and then VMs are allocated to the physical machine. The algorithm reduces the number of active VM-supporting servers by minimizing the energy consumption and reducing the rejection rate.
Ref. [S65] suggested an energy-sensitive scheduling algorithm using the ESWCT Consolidation Workload-aware. The algorithm attempts to consolidate the VM to a minimal number of servers based on an optimized resource balance (processor, memory, and network width) concurrently shared between a cloud data center and users. This algorithm aims to minimize energy usage by optimizing resource utilization because heterogeneous loads are different in resource utilization. Ref. [S52] proposed an IaaS Cloud platform-oriented online scheduling algorithm for reducing energy consumption. The algorithm aims to ensure improved quality of operation for heterogeneous computers and various workload scenarios.
An energy-sensitive virtual machine scheduling algorithm is called a dynamic round-robin algorithm [S53]. This algorithm turned off the idle physical servers for a temporary period to minimize energy usage in the cloud data center. It saved 43.7 percent of energy concerning other scheduling algorithms. A virtual machine-aware energy-efficient scheduling algorithm named EMinTRE-LFT was proposed by [S55], based on the idea that a reduction in energy consumption is directly equivalent to a reduction in the completion time of all physical servers.
The Cloud algorithms for scheduling algorithms face multiple problems due to cloud user requests’ complex and volatile nature. Ref. [S99] conducted a statistical study to establish the relation between the machine’s energy consumption and efficiency (system performance). They suggested an algorithm that does not require any previous user request information.
Ref. [S56] suggested an algorithm for real-time distributed systems to achieve an efficient solution by using a polynomial-time algorithm integrating a variety of heuristic rules. Different software-level techniques for energy efficiency are compared in Table 11. In these various techniques, the comparison is made according to their benefits and disadvantages. In Figure 14 and Figure 15, taxonomies of all software-level techniques are described. In Figure 16, the distribution of research papers among different types of software techniques can be seen.

4.1.7. RQ7: Explain the Various Techniques for Virtual Machine Consolidation Applied at Data Centers

In Green Cloud Computing, the concept of consolidation means “the process of deploying different data centers related to data processing applications on a single server with virtualization technology”. Consolidation is a critical feature derived from virtualization. It is committed to implementing at the process level for load balancing, better utilization of virtual systems, and reducing power consumption [56].
In the management and organization of Resource Pool Access, Virtualization plays a vital role via being a Virtual Machine Monitor or hypervisor. It hides physical resource information and offers virtualized tools for applications across different levels. The fundamental feature of a virtual machine is that the software it runs is restricted to virtual machine resources and abstractions only. The virtualization also facilitates VM consolidation methodology, which groups many virtual machines into a single physical server as shown in Figure 17.
VM consolidation can bring substantial benefits to cloud computing by allowing greater use of the available data center resources [72]. The consolidation can be implemented statically or dynamically as The VMM assigns physical resources based on the VM’s highest demand. The case of static consolidation (over provision) leads to the loss of resources because the workload is not always at an extreme level.
On the other hand, the VMM adjusts the VM capacities based on current workload demands when the consolidated dynamic VM is considered (resizing). It also allows better use of the resources at data centers [140]. The VMs can be re-located with a dynamic VM consolidation scheme by doing live migration to understate the number of active physical servers known as hosts while considering the current PM resource demand. The idle hosts are switched to low power modes with short transition times to reduce energy waste and minimize total power consumption [96].
The hosts are re-activated to ensure that service specifications are not ignored and will meet the demands. Dynamic consolidation, threshold-based consolidation, and optimization of the consolidation process are presently trending topics in green cloud virtual machine consolidation. This approach key role was to derogate energy usage and improve the quality of service. Simultaneously, the different methodology used for VM consolidation has specific issues and benefits discussed in this section.
Ref. [S76] proposed a highly energy-efficiency and dynamic virtual machine consolidation (EQVC) method through algorithms that lead to different stages of VM consolidation. They choose redundant VMs from hosts for energy-saving and quality assurance before the overloading and migration of VMs to other hosts. Ref. [S35] thoroughly discussed the need for consolidation, the procedure of dynamic consolidation of virtual machines, and the advantages in detail. They explained how to consolidate a single physical server with multiple virtual machines (one–many approaches) and multiple physical servers with multiple machines (many–many approaches) [142].
The authors proposed online deterministic and non-deterministic algorithms to explain the process of VM migration in the cloud. Another research paper [S57] proposed a threshold-based approach for the IaaS platform to perform VM consolidation to balance the load efficiently and avoid resource underutilization problems. Apart from their former approaches based on a threshold value, they also introduced the determination of threshold value dynamically, based on the present VM need and historical usage statistics.
Ref. [77] raised live problems while doing the virtual machine consolidation process. The consolidation process is a resource incentive and expects the intelligence support to reduce the server downtime to a minimum. To overcome the consolidation process’s limitations, they proposed a DVFS (Dynamic Voltage Frequency Scaling) virtual machine consolidation technique to save energy by running the servers at different voltage frequencies.
Ref. [S59,S69] proposed an approach called Datacenter Energy-Efficient Network-Aware Scheduling (DENS). The authors projected DENS to balance power consumption in data centers while taking network insights and performances into account. It works by disseminating tasks between servers according to the workload, the heat emitted, and communication potential [S23], suggesting a VM consolidation approach whose primary purpose was to reduce the energy usage of a data center providing various SLAs. Thus, they proposed an adaptive energy management policy that includes VM preemption to regulate energy consumption based on the user performance requirements.
Ref. [S60] propose two methods: the DVFS technique is used for energy efficiency and the second one for VM consolidation. The first method identifies performance weaknesses with energy consumption and provides familiar workload management with DVFS, which saves up to 39.14% for dynamic load. The second method is VM consolidation, which determines the dynamic frequency by distributing the load to achieve quality service. Various scheduling algorithms for resource utilization have been proposed.
However, they do not solve the hosts’ machine heterogeneity problems, leading to greater energy consumption; thus, ref. [S61] introduced a Job Consolidation Algorithm (JCA) to overcome these problems, which utilizes cloud resources effectively by including server malfunctions by implementing DVFS techniques. Ref. [S62] suggested a self-adaptive solution called SAVE, which focuses solely on local knowledge and determines the assignment and transfer of VMs by probabilistic processes [143].
A new approach [S63], named VMDC, was proposed to manage network and communications topology in data centers. Using the multi-objective genetic algorithm simultaneously reduces energy consumption and traffic communication, which cause performance problems while fulfilling the SLA migration time constraints. In [S64], a Bayonian-based network (BNEM) model allows a comprehensive solution involving nine factors in an actual data center.
Criteria for migrating virtual machines and VM placement are presented by combining three algorithms corresponding to VMs consolidation methods. They propose a hybrid Bayesian-network-based VMs consolidation (BN-VMC) method. Ref. [S65] designed and implemented C4, Aliyun’s cloud-scale computing cost-effective consolidation service. They analyzed user patterns, resource utilization, and migration costs in Aliyun, showing the need for conventional usage-based consolidation approaches, especially for locally storage computing, which cannot be met in the production environment. The heuristic worst-fit method is used to migrate the load-balancing instances.
The consolidation technique [S66] is built based on historical VM usage data analysis. In choosing a VM for migration, the VM with a minimal migration period was selected compared to other VMs. The algorithm indicated a significant energy reduction, thus, ensuring a high-quality level of operation improves energy consumption standards by making the best use of resources [145]. This is because underutilized resources sit inactive for a long time and do very little work as a node to consume more. As it has more load on it, this degrades the node’s efficiency [S66].
Overutilization causes a node to consume more, as it has more load on it, often degrading the node’s efficiency [17]. Underutilization prevents a node from functioning optimally, leading to idle time. Simultaneously, different energy-efficient techniques were designed to prevent the underuse and over-use of resources, leading to increased energy consumption. Broad strategies based on consolidation are evaluated by various criteria, such as their pros and cons, implementation environment described in Table 12, and taxonomy for consolidation techniques are also shown below in Figure 18. In Figure 19, the distribution of research papers among different types of consolidation-based techniques can be seen.

4.1.8. RQ8: Explain the Various Bio-Inspired Techniques for Energy Efficiency Applied at Data Centers

In computer science, an optimization algorithm can be developed to solve complex problems using meta-heuristics, which simulate or inspire biological behaviors, such as ants colonies, birds’ flocks, swarms of bees, and fish schools. These techniques have brought the attention of computer scientists to figure out a variety of scientific and engineering problems. The story comes with an investigation of the natural coincidence between biological and computer science.
We present an overview of the biologically influenced literature on energy-efficiency technologies developed exclusively on animal behaviors, which are in search of food and match findings. People may also learn from animal behavior to construct optimization algorithms for tackling complicated problems. The Artificial Bee Colony (ABC) algorithm, for example, co-operates with the bee colonization, the Specific Swarm Optimization (PSO) algorithm simulates the biological behavior of the bird and fish schools, the Social Spider Optimization (SSO) algorithm, and the Lion Optimization Algorithm (LOA), which imitates the behavior and co-operation characteristics of the lions.
Now, we discuss some of the techniques being developed based on bio-inspired techniques. Machine placement based on bio-inspired in the IaaS environment was first proposed by [S81], in which the initial VM placement problem was identified as a multi-objective optimization problem. Finally, a multi-objective optimization technique based on Ant Colony Optimization was applied. Ref. [S82] suggest the virtual machine placement problem solution. The intention is to achieve a compelling set of non-dominated solutions that simultaneously reduce total resource and electricity consumption using ant colony system algorithms in a multi-objective placement technique.
The proposed techniques are compared with the current multi-objective genetic algorithm and two other algorithms with a single objective, such as popular bin-packing and MMAS algorithms. Ref. [S83] offers an AVVMC VM consolidation scheme based on the balanced resource usage of servers in various computer resources (CPUs, memory, and the I/O network) as they propose modification and incorporation of the meta-heuristic Ant Colony Optimization (ACO) with a balanced use of vector algebra-based computer resources.
The paper [S84] introduced a simultaneous VM placement and data positioning solution. The key goal is to minimize cross-network traffic and bandwidth used by placing the number of VMs and data needed physically closer to Physical Machines (PMs). Authors introduce and evaluate an Ant Colony Optimization (ACO)-based metaheuristic algorithm, which selects an adjacent PM to position data and VMs. An ACO-based VMP solution, based on global search data, which combines artificial ants with other local exchange and migration policies, was proposed by [S85].
This algorithm was configured globally by assigning VMs to the total active PMs [144]. This solution has been implemented on many VMP problems in homogeneous and heterogeneous cloud environments with differing VM sizes as an effective means to minimize the number of active servers. Ref. [S86] proposed a VMP algorithm based on the recruitment process within ant colonies that enhanced the exploitation of PM resources and maintained a maximum balance of resources. With this new approach, VMP was optimized, and based on it, energy consumption and resource waste are reduced and balanced in the DC Cloud.
Ref. [S88] established an optimal way of delivering virtual machines to DC Cloud physical machines. There are two-stage of the VMP algorithm in this model. First, we introduce an architecture that serves a large community of virtual machines and schedules them. Second, to avoid and reduce the waste of resources and energy consumption in DC, a multifunctional VMP algorithm was proposed, known as the VMP Crow Search Algorithm. A new resource-dependent strategy leads to the optimal process.
Large VMs on minimal DCs meet the SLA and desired quality of service for optimal DC utilization [S87] based on the meta-heuristic crow search based on greedy crow search and traveling salesman problem. Ref. [S89] proposed a requirement predictor technique by the optimal evolutionary placement of a virtual machine placement (EOVMP) algorithm. Using this, standard computation requirement is obtained and used to assign virtual machines by reservation and job request processing plans.
Ref. [S91] took an idea from Google’s massive data processing framework for constructing multifunctional job scheduling using GA. This technique has enormous capabilities to operate on Google Networks and uses many ways to train and interpret networking features. The authors developed a practical individual encoding and decoding method and built ’servers’ overall energy efficiency function as the individual fitness value.
Ref. [S92] suggested a low carbon virtual private cloud (LCVPC) technique that lowers the carbon emission rate of remote data centers interconnected by an extensive network, whether private or public but operates in distributed areas. Smart Live VM migration is used on the WAN network by which both carbon footprints and energy used from the entire network were calculated first by VPC managers, and then the modified GA is applied.
Ref. [S93] proposes an energy efficiency mechanism that assigns cloud resources without violating the ant colony framework’s service-level agreements (SLA). A new algorithm was proposed by [S94]. The workload consolidation-driven algorithm was developed based on the ACO metaheuristic, and a dynamic decision was made about the workload distribution. This also resolves the workload consolidation problem within the application of a single resource and difficulties in assigning workload as a multidimensional bin-package problem.
The bio-inspired techniques developed by [S95] combines both ACO and Cuckoo searches strengths. In this hybrid Algorithm, the Voltage Scaling Factor Approach reduces energy consumption and efficiently schedules the job. The mechanism for using PSOTBM was designed [S96] to switch off an idle or unused server for energy consumption reduction. PSOTBM is the initial assignment of computing, storage, communications resources, and applying a mathematical system to find the best solution by PSO-optimized algorithms.
Based on the Simplified Swarm Optimization method (SSO), a PSO-based technique on which energy consumption in a cloud-based system can be reduced was proposed by [S97]. The optimized energy is consumed without performance degradation using the DVs technique, storing tasks entering the network, and detecting critical pathways for DVS’s successful implementation.
Ref. [S27] proposed a nature-inspired ant colony system to consolidate VM; however, the probability of a peak situation arising increases as the consolidation occurs. Similar energy and performance-aware policies for deconsolidation are required. Various bio-inspired methods are compared by various metrics and advantages, as listed in Table 13. The taxonomy of bio-inspired techniques is shown in Figure 20, and the distribution of research papers among different types of bio-inspired techniques can be seen in Figure 21.

4.1.9. RQ9: Describe the Different Thermal Management and Cooling Technologies Employed at Data Centers

Data centers were previously intended to focus only on IT equipment’s stability instead of saving energy. Data center temperatures and humidity levels were used to exert ideal thermal conditions for the proper working of IT equipment, which is explicitly maintained at 21.5 °C. For years, the IT infrastructure had to comply with the standards for power modules of high density, which include the required equipment cooling and operating condition criteria.
For maintaining a steady temperature and humidity level, a considerable amount of energy is required. ASHRAETC9, the American Society of Heating, Refrigerating, and Air-Conditioning Engineers, specified the IT server temperature between 18–27 °C, and therefore allowable temperature ranges from 15–32 °C (class 1) and 10–35 °C (class 2) with an acceptable relative humidity range from 20–80% [74]. The server’s internal parts are kept steady to minimize server inlet temperature effects by increasing the fan speed.
Consequently, the input air temperature does not explicitly influence the stability of the server components. Many experimental studies have been on data center cooling systems under fault conditions, including thermal energy efficiency [20] and optimizing the system delivery. Using a thermal-aware approach for scheduling, numerous strategies have been developed for homogeneous High-Performance Computing (HPC) data centers. An efficient scheduling strategy minimizes the entire data center’s temperature. In contrast, energy savings are not limited only to the reduction of server energy consumption but also include reducing cooling equipment resources installed in data centers.
To demonstrate the data center cooling system in the case of a power failure and examine the characteristics of the rise in air temperature in IT environments, ref. [S72] developed a transient thermal modeling process. To achieve the necessary temperature control, the method requirement depends on each cooling system’s characteristics. An energy efficiency technique with low first-time cost and ventilation cooling technology (VCT) was proposed by [S100].
A new VCT control system designed to ensure indoor temperature, humidity, and energy savings can be used under specific meteorological conditions for various telecommunication base stations (TBS). When the outside air temperature and humidity satisfy the cooling criteria, fans replace the air conditioners with environmental control.
Ref. [S101] introduced new technology for eliminating the discharge of excess heat from a telecommunications plant. The proposed technique will cool the station using a heat exchanger when the outdoor temperature is low enough. Instead of using air conditions for around 5014 h a year, it uses plate heat exchange, resulting in an energy-saving rate of up to 29%. The aim was to provide a renewable and sustainable energy source for cooling data centers to analyze and demonstrate what amount of heat can be obtained by a data center server blade.
Ref. [S102] proposed 123 servers encapsulated in mineral oil. Using offered technology 10-ton chiller with a design point of 50.2 kWth and the heat quality in the oil discharge line was up to 12.3 °C higher on average from water-cooled experiments. In [S103], a separate microchannel heat pipe (MCSHP), a telecommunications station’s cooling system, investigating the effects of geometric parameters and environmental conditions are implemented.
When a difference from 6 to 8 °C was seen in indoor and outdoor temperatures, the cooling capacity was improved by 135 percent. Ref. [S104] suggested a cooling system consisting of a three-fluid heat exchanger that incorporates two separate loops, the mechanical cooling loop and a thermosyphon loop, to avoid a reliability risk mode due to switching valves.
A cooling system was studied by [S105] in which servers were immersed in a dielectric liquid, and the water was being used for transmitting heat away from the data centers. The research involves a dry air cooler and buffers heat exchanger to dispose of heat from a liquid-cooled stand. The PUE from the cooling system was found to be as low as 1.08. Ref. [S106] focused on energy consumption aspects in a data center. They worked on a direct-fresh-air-cooled container-based data center (CDC), which reported a 20.8 percent decrease in energy usage over by projected use of a CRAC-cooled CDC for one year.
Ref. [S57] developed a latent heat storage unit (LHSU) model to reduce a significant amount of energy consumed in space cooling. Then, the concept of free cooling and its future use in data centers was explored based on the proven performance characteristics of heat-pipe technologies. Data for a common area was implemented and evaluated by [S107].
There was a case study, and the findings suggest the possible energy savings of up to 75% using a heat pipe-based free cooling system. A heat pipe-based thermal management system for energy-efficient data centers was developed [S108] in data center environments where a fluid loop heat exchanger is connected to the rear of the rack server, which further develops a new Computational Fluid Dynamics technique (CFD) [109]. The study compared liquid back door cooling and conventional air-cooling systems and demonstrated the energy-efficient benefits of air-to-liquid heat exchangers [118,119,120].
In [S110], the authors reported the strength and energy performance of the first hot-water-cooled supercomputer system. The system also has an air-cooled part that allows measuring the output of coolants. This proposed system eliminated chillers and permitted heat reuse [S94]. Different thermal and cooling techniques for energy efficiency are compared in Table 14. All thermal and cooling techniques taxonomy is shown in Figure 22, and the distribution of research papers among different types of thermal management techniques can be seen in Figure 23.

4.1.10. RQ10: Describe the Different Power Aware Management Techniques

Most energy-efficient strategies rely on migration and consolidation. They maintain energy efficiency at internal logical levels, which will minimize energy consumption by controlling the power supplied to virtual machines. Different energy scaling techniques can minimize power consumption at the superficial level [100]. Power management is categorized into two categories: static and dynamic power management. In static power management, power is managed at the time of design circuit, architecture, and system-level, focusing on reducing the switching activity of power in circuits [128,129,130,131,132,133,134].
Dynamic power management is applied to running state systems [49]. Dynamic voltage and frequency scaling (DVFS) is one of the methods used in DPM to cut down energy consumption by supplying dynamic voltage and clock frequency when the CPU is in an idle state [26]. Many research studies in the area of power management have been performed. Power management techniques are categorized into two major components as shown in Figure 24.

Static Power Management (SPM)

Evaluation of hardware efficiency is done by employing an SPM [108], which includes CPU, memory, storage, network equipment, and power supply [90]. The power diminution in SPM is a never-ending process that consists of all optimization approaches used in logic design, circuits, and architecture [92,94].
(a)
Logic-level optimization: Seek to cut the energy consumed by the combinational and sequential switching operations.
(b)
Circuit-level optimization: Minimize the power consumed by each logic gate and transistor switching disk operation. This typically use a sophisticated gate architecture and transistor size.
(c)
Architecture-level optimization: The mapping of a high-level problematic architecture to a low-level configuration is a way of saving energy and maintaining a reasonable degree of efficiency [91,92,93,94,95,96,97,98,99].

Dynamic Power Management (DPM)

The main focus and the key role of Dynamic Power Management (DPM) is:
(1)
The resources available and their use in the program.
(2)
To optimize application workloads for reducing energy consumption.
We also believe that the DPM technology helps the system automatically change its power rates depending on the current situation [102,103]. Moreover, the workload requirements can also be estimated, which allows the adjustment of future system behavior to execute adequate action under these requirements [107]. Therefore, reducing DPM in the system is temporary, and depending on the available resources, the current workload will persist indefinitely [121]. Their application level defines these techniques as hardware or software level, as explained below.

Hardware-Level Solutions

Applied at the hardware level, DPM technology seeks to dynamically reconfigure devices for developing techniques to fulfil the required facilities with a limited count of active components or a minimal load [111,112,113,114,115,116]. The hardware-level DPM strategies [20] can selectively disable idle device components or reduce the output of slightly unexploited components.
Some modules, such as the CPU, can also be switched among active and idle modes to save power.
DPM can be classified across the hardware level as:
(a)
Dynamic System Component Deactivation (DSCD): is divided into two additional types:
i.
Partial Dynamic System Component Deactivation (PDSCD): these types of techniques use the idea of gating a clock in a part of an electronic device.
ii.
Complete Dynamic System Component Deactivation (CDSCD): These techniques are founded on the idea of complete disabling the components during periods of inactive states.
(b)
Dynamic System Performance Scaling (DSPS): In this, the dynamic performance of the components is adjusted in a manner proportional to the power consumption.

Software-Level Solutions

In the hardware-level approach, any adjustment or reconfiguration is challenging to enforce. It is, therefore, essential to push toward software solutions. Several approaches for energy reduction have been proposed, such as Advanced Power Management (APM) performed by Simple Input/Output System (BIOS), firmware-based systems, and platform-specific power management systems. These solutions, however, depend on both hardware and software. Now, we discuss the new techniques invented by various researcher using power management technology [139,140,141].
A modern system architecture named FAWN (fast array of wimpy nodes) was introduced by [S67] for intensive computing with low-power CPUs involving limited local flash storage volumes. The authors reported that such a combination provides successful parallel machine data access. Ref. [S113] investigated FAWN intensively on different workloads. The results demonstrated that nodes with low-power CPUs are more efficient from an energy perspective than are traditional high-performance CPUs. A new Gordon architecture for low-power data-centric applications was defined in [S68].
Gordon can reduce power consumption by using low-power processors and flash storage. The Non-dominated Sorting Genetic Algorithm II (NSGA-II), a power-efficient algorithm for the static scheduling of independent jobs on the homogeneous single-core infrastructure of cloud computing data centers implemented by [S69], which is based on a genetic algorithm. Ref. [S70] suggested a genetic algorithm for energy awareness (GAPA) to solve the VM allocation problem. A chromosome of an individual task was associated with a tree structure in which each tree structure instance shows VM to PM allowance.
The PSO-based algorithm that could minimize the energy cost and time required to search for feasible solutions, introduced by [S71], involves a heterogeneous multiprocessor system, similar to the cloud data center situation. The proposed algorithm is designed to find the most feasible solution to minimize energy usage by assigning jobs or exchanging appropriate processors. Ref. [S74] stated an energy-conscious network algorithm for data centers. The author proposes to make sleep only a few components of the network system rather than placing the network system entirely in sleep mode.
Ref. [S118] proposed the Elastic Tree for data center the network, a dynamic power management scheme that dynamically specifies a set of active network components, including switches and links. Elastic Tree on open flow switches was implemented for the experiment. The results revealed that network energy was reduced by up to 50 percent. Ref. [S79] proposed a DVFS-enabled Energy-efficient Workflow Task Scheduling algorithm named DEWTS. Ref. [S73] suggested a scheduling technique for virtual machines using the Dynamic Voltage Frequency Scaling (DVFS) in computing clusters to minimize power consumption.
Generally, we use dynamically scaling in supplied voltages to allocate virtual machines in a DVFS-enabled cluster. Ref. [S75] introduced a new architecture that would include practical green improvements to the distributed cloud-computing environment. The system’s overall efficiency is greatly enhanced in a cloud-based data center with minimized overhead performance through power-aware scheduling techniques, variable resource management, live migration, and minimal virtual machine design [90].
In [18], the authors explore different power-efficient virtual machine provisioning techniques for real-time services. First, they investigated all such different power-aware DVFS-based schemes. They then used a purveying algorithm to improve the energy consumption and system reliability and reduce operating costs. A systematic approach to optimization of the computational in cloud computing proposed by [S78] implemented a private cloud system with facilities supplied by the DAS-4 cluster.
By implementing various high-performance computing workloads on both VMs and hosts [3], the author compared their performance in power, power efficiency, and energy metrics. Additionally, all these schemes help to minimize the effort to cool down the equipment. A Surge-Guard power trade framework [S80] balances idle times with cooling capacity, which reduces total energy consumption [20]. The framework also aims to reduce overall energy effectiveness, optimizing idle time, and reducing response time.
The diminution of SPC requires reducing the leakage of current, and this can be done in three ways [108].
  • Reduction of the voltage level for device components (CPUs and cache memory) is known as Supply Voltage Reduction (SVR).
  • Cutting down the circuit size of the device, either by designing lower-transistor circuits or cutting power supplies into idle components to minimize the effective transistor count.
  • Cooling technologies can reduce power leakage by enabling circuits for faster operation when electrical resistance is reduced at lower temperatures. Reduce the possibility of degradation of the reliability and life of a chip due to high temperatures.
Four methods can reduce the DPC:
  • Minimizing the process of switching.
  • Decreasing physical capacitance, which relies on low-level design parameters, such as the size of transistors.
  • Cutting down the voltage of the supply.
  • Reducing the frequency of the clock.
The power-management technique taxonomy is shown in Figure 25. The techniques are compared according to their operating environment, methods used, and primary advantages and limitations in Table 15. In Figure 26, the distribution of research papers among different types of software techniques can be seen.

4.1.11. RQ11: Explain the Key Aspects That Make the Cloud Favorable for Reducing Carbon Footprints and Better Utilization of Energy; Describe Non-Technical Technologies Employed at Cloud Data Centers for Enhancing Energy Efficiency

  • Reducing clock rate and voltage for decreasing CPU power dissipation
Processor power dissipation significantly increases as the clock frequency increases and the number of transistors expanded. For avoiding an increase in cooling costs, CPU power dissipation must be reduced [45]. Processor consumes energy in the form of electricity for operating various network devices, chips and mainly for its charging purposes. This energy is often emitted into the environment in the form of heat. By using free cooling, the power consumption can be diminished.
ii.
Use of renewable energy sources
Every cloud data center required a diesel generator to supply backup power and ended up in the discharge of CO2 and GHGs from that data center. Rather than using such energy, we can opt for these clean-power options, such as hydro and wind power, which can fulfill operation and colling requirements. Google data centers continuously consume electricity at almost 260 million watts, nearly a quarter of nuclear power production.
Parasol, a solar operated Datacenter was developed by [S111] as a research platform that includes an air conditioning unit, electrical distribution, and monitoring infrastructure that works entirely on renewable energy sources. In [81], they already obtained renewable energy certificates to balance their non-renewable energy consumption and finance renewable energy production from wind and solar sources. AWS 2019 announced that its global infrastructure is focused on creating 100% renewable energy use.
iii.
By using energy-efficient storage
Energy-efficient storage has already been devised to supplement current cloud storage. As the data center has limited age of 9–10 years, it is possible to use energy-efficient memory elements, such as solid-state storage, as they have no moving mechanical component. It will need less energy to cool down than the usual hard disc drive.
iv.
Advanced clock gating mechanism
The clock gate is a hardware switch to transform the clock pulse to on and off modes. The logic unit of the clock can only be triggered if the logic block is running. Moreover, if the logic block is in the perfect model, it needs to be switched off. This famous technique was utilized on many synchronous circuitries and can also reduce dynamic power dissipation on globally asynchronous and locally synchronous circuits [S112].
v.
By using split plane power
Splitting means dividing up; however, division along the horizontal axis is in the processor’s language. Here, the North Bridge does not share the processing device’s power (these are Intel microchips that allow CPU and motherboard communication). The motherboard supports a split plane power supply, and separate power supplies are required for the North Bridge and the processor.
vi.
By improving the energy efficiency of processors
Dynamic voltage scaling and frequency scaling involve adjusting these parameters according to the computer or software specifications. Even the processor voltage or clock rate can be adjusted easily to reduce the CPU power usage. This method minimizes energy consumption through the following techniques in a real-time framework: AMD Cool ’n’ Quiet, AMD Power Now, IBM Energy Scale, Intel Speed Step, Transmeta Longrun, and Longrun2.
vii.
Through energy-efficient computer architecture design
Heterogeneous computers are composed of several processors of a relatively independent type. In using such systems of heterogeneous computers, which usually use processors as the central control unit and other particular purpose processors, such as the accelerator unit, the energy consumption ratio is less effective. Ref. [S113] developed a Fast Array and Wimpy Nodes (FAWN), an energy-efficient cluster system capable of handling large numbers of wimpy nodes built into an integrated cluster. This can function as a classical cluster, however, with only a fraction of the power consumption by having energy-efficient processors and utilizing a limited flash memory.
viii.
By using Nano Data Centers
A new, distributed computing platform called Nano Data Centers (NaDa) increases the energy efficiency in modern and traditional data centers. NaDa offers cloud computing resources to build a decentralized data center architecture using a managed peer-to-peer model. Instead of using many large data centers in different geographical locations, it is recommended to use small-sized, interconnected data centers. This energy consumption involves VoD access model data access from NaDa [78], making 30% less consumption than traditional data centers.
To provide processing and storage facilities, NaDa uses ISP-managed home gateways and follows a controlled peer-to-peer architecture to create a distributed data center infrastructure. They choose Video-on-Demand (VoD) platforms to determine the opportunity for energy savings on the NaDa network. Using NaDa, the electricity is saved from standard data centers at least 20 percent to 30 percent, avoids the cooling cost, and reduces network energy consumption.
Optimizing the total power consumption within a data center is an NP-hard problem. The project for the nano data center was started with the promise of consolidating power and cooling costs. This study aims to reduce the gross energy usage in a nano data center by implementing the efficient power-aware server selection procedure. Usually, it is an NP-hard problem to optimize the overall power consumption in a data center. The energy consumed according to various parts of a data center is shown in Figure 27.
As the current trend is shifting toward energy consumption in data centers, [S119] proposed a new distributed computing platform called Nano Data Centers (NaDa). Ref. [S115] developed two heuristics for selecting nano servers and intermediary routers for that. These heuristics were based on the option of a short path for the number of intermediate routers and the use of the least used server to support a new request. Using the proposed heuristics, the total energy usage in a data center decreases while the optimum solution’s time complexity increases.
ix.
Energy Saving Strategy for optimizing compiling process and application software power consumption
Compiler optimization also plays a significant role in reducing the energy consumption. Better compilation technology also helps to evaluate the application programs’ behavior to reduce the device and processor power usage. Applying specific approaches to the source program structure level will reduce the power consumption by application software. Approaches for application software power consumption concentrate on reducing space and time complexities in the algorithm to reduce energy consumption, including a specific algorithm for compressing data storage space.
The authors in [S116] proposed a model for lowering the execution frequency, which reduces power in the multithreading system structure. Normally programmers do not consider the energy efficiency of code, and even fewer know how to assess and improve the code’s energy efficiency. Hence, ref. [2] attempted a comprehensive study to assess the impact while choosing the programming languages, compiler optimization, and implementation. By examining the results, considerable energy can be preserved to solve the same problem with the same input size by meticulously choosing the correct language, optimization flag, and data structure.
x.
Thermodynamic computing
When we cross-machine constraints to see how researchers are exploring the concept of thermodynamic computing [105] to make computers function more expeditiously, on the premise that thermodynamics contributes to the autonomy and evolution of natural systems. Hence, thermodynamics will guide us towards the self-organization and development of future computing systems, building them to be more adequate, stable, and energy-efficient.
xi.
Nano technology
Presently, the industry is under 10 nm node, moving to 5 and 3 nm can provide some development of efficiencies and optimization [S118]; however, in the future, we will require more than today’s latest transistor architectures, such as nanosheets and nanowires.

5. Issues and Recommendations

From the studies performed in the previous section, we created a graph on the number of studies proposed in each year. By analyzing Figure 28, from the year 2011 to 2015, in this time span, the energy efficiency topic became the hottest research domain. With these software techniques, huge infrastructure is not required as it is with hardware techniques, and it is less prone to failure and faults.

5.1. Power-Consumption Parameters (RQ1)

Issue 1:
The majority of the authors for PSs do not consider the recommendations provided by previous studies to select the parameters. Moreover, no sufficient detail of the parameters used in their research is given. We see that specific techniques significantly gained less attention, such as bio-inspired and thermal-management techniques; however, their results are productive.
Recommendation 1:
It is advised to select such parameters by which we can thoroughly analyze the techniques that can produce more accurate results. Moreover, we attempt to justify their cause by selecting specific parameters. Usually, without concerning all parameters, we cannot predict any specific technique to be effective or not.
Most of the selected PSs have not provided any valid justification while choosing their studies’ energy-consumption parameters. Based on previous studies, while developing any software technique, most of the researchers gave only the most negligible efforts regarding energy consumption as their focus was on boosting the performance.
Issue 2:
Many PSs have not involved industry researchers, and implementation was not performed in any industry settings.
Recommendation 2:
It is suggested to researchers to further conduct empirical study by thoroughly examining the impact of industrial setup techniques.
This study highlights the divergence between academic and industry research in energy-saving techniques. The energy-efficiency techniques used by academics and energy efficiency activities used throughout implementation by industry practitioners are different. Researchers are also advised to include or derive similar information from previous research done by industry professionals. Furthermore, it would be beneficial if the research involving these energy efficiency initiatives were conducted on professional software systems in an industrial environment.

5.2. Impact of High Energy Consumption and Current Trends (RQ2 and RQ3)

Issue 1:
Most of the studies utilize energy-efficiency techniques in data centers; however, they do not greatly consider the carbon footprint, as all the energy being consumed by techniques is generated by non-renewable sources of energy. Moreover, there are not many details about the total carbon emissions from data centers.
Recommendation 1:
It is suggested that researchers propose techniques that involve minimal non-renewable energy sources for operating data centers as all this energy comes from coal or nuclear power stations, which produce harmful gases in the environment. Researchers should also attempt to provide accurate details of the carbon emissions and the other gases emitted from data centers.
Large corporations generally do not share their data; however, mostly all are moving toward green computing technology and attempting to optimize the techniques. Today, ICT can also lessen global pollution by reducing fossil fuel use and utilizing renewable sources. Facebook undertook to use 100% green energy in 2011. Google supplemented in 2012, as did Apple. Almost 20 web corporations did the same in 2017. Google is the top green energy purchaser in the world.

5.3. Energy-Efficient Techniques (RQ4)

Service level agreements (SLAs) are used for customer availability and reliability, which are integral aspects of cloud systems. The provider must pay fines in the case of any breach of the SLA. SLAs must cover data leakage, mismanagement of resources, multi-tenancy problems, and data sharing. Metrics in the data center field are measurable comparative characteristics and important characteristics leading to energy efficiency and sustainability improvements.
Issue 1:
Very few PSs had worked on certain important concepts, such as SLA violation. Most of the studies did not account for SLA violations. There was no explanation about QOS (quality of services) in most PSs.
Recommendation 1:
Researchers are suggested to propose a technique involving minimal SLA violations while operating energy-efficient data techniques. All PSs must discuss the SLA and have proper details about any SLA violations in their articles.

5.4. Hardware Techniques (RQ5)

Issue 1:
Many PSs about hardware techniques ignored the critical point: the cost for implementation. Using hardware techniques, we can teach how to efficiently reduce the carbon footprint; however, with all these, we also increase the cost of construction and operation.
Recommendation 1:
It is suggested to report the cost factor as only high efficiency would not solve all issues. While constructing hardware, we should also consider the costs of the hardware and equipment.
Issue 2:
The majority of the PSs for hardware techniques have faced an issue regarding the construction of such systems. These systems are complex to implement, and the results obtained by these PSs in the academic environment are not very consistent with the results obtained by industrial systems.
Recommendation 2:
It is suggested to execute the studies in both settings as we can eliminate the possibility of inconsistent results. We must validate the proposed studies in both academic and industrial systems for proper analysis of the results. We will have a more robust result and know more about the system’s efficiency by performing such practices.

5.5. Software Techniques (RQ6-RQ7)

Issue 1:
Most of the PSs had been implemented to evaluate the techniques’ impact, mainly on Cloudsim. Sometimes the implementation of such object-oriented software was more complex than in others.
Recommendation: 
It is advised to develop such tools to efficiently address the technique’s impact by employing easy language, such as Python and C#. Currently, some intelligent languages that support AI and data analysis can also be used, and these can show results more efficiently.
Issue 2:
Most of the PSs missed some critical points, such as Network traffic, Load balancing, and optimization is required in software techniques.
Recommendations 2:
We suggest involving these parameters in research if there is high network traffic. It would lead to more power consumption in data centers that manage the traffic. There should also be care for proper load distribution if the load is oddly assigned to a server. This will lead to degradation of the performance.

5.6. Power-Aware Management Techniques (RQ8)

Issue 1:
Most of the studies that utilize the power-aware management technique are considered to be costly and are implemented on the tools that are either obsolete or unavailable. Primarily, these tools do not evaluate the system with all concerned parameters.
Recommendation 1:
The development of the latest tool that employs all attributes to predict or evaluate any system and accurate cost details must be added to studies.

5.7. Bio-Inspired and Thermal Techniques (RQ9–RQ11)

Issue 1:
Mostly, the finding obtained by both techniques are still not implemented in the real-world entities. The techniques were implemented in their simulation, making it challenging to predict the exact results without implementing it in the same environment—there are no details about the resource wastage. Load balancing was found in PSs.
Recommendation 1:
It is advised to develop such a framework or tool capable of executing all implementations only to evaluate a system efficiently by all parameters and requirements.
Issue 2:
Most PSs utilizing techniques improve a particular parameter, whereas another PS with the same technique applying different parameters for evaluation may result in ambiguous results.
Recommendation 2:
It is suggested to develop a unified framework for energy-efficient techniques in the cloud, which will be capable of a more significant measure of techniques.
The metrics help designers, managers, and end-users and are observable for monitoring, evaluation, optimization, cost estimation, and planning criteria. The problem might be to evaluate the work of IT equipment or software, which is subjective because of the device, network and storage diversity, and usage ratios.

6. Threats to Validity

The primary threats to our SMS’s validity originate from the unintentional contradictions involved while reviewing the literature [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111], selecting the primary study, evaluating the quality assessment, and extracting data from the selected studies. Including the conclusion and internal and external threats, we followed the threat description given by [112] to discuss those constraints and the steps taken to mitigate these threats [113,114,115,116,117,118,119,120,121].

6.1. Conclusion Validity

The conclusion’s validity refers to reproducibility, which helps other researchers produce the same results as the original replication study [30]. This research has demonstrated the search string, electronic search databases, manual search, inclusion and exclusion criteria, and details of the systematic methodology used for undertaking this study.
In addition to this, the entire SMS process information was taken from a standard SMS conducted by [122], which is available online. However, other researchers could produce different results because of their different perceptions of the selection stage. To reduce this threat, both authors separately examined the titles, abstracts, and full texts of the selected articles while considering inclusion and exclusion criteria. Additionally, the authors discussed the discrepancies they identified during the selection process in the case of disagreements.

6.2. Convergent and Discriminant Validity

The validity of the construct relates to primary studies identified during the selection process. To select the primary studies unbiased, we adopted a well-known protocol suggested by [55]. First, a peculiar search string was built to select studies relevant to our concept. We further validated our search terms with 20 reference papers to expand our initial collection of keywords or define additional search terms. Paper [55] recommended a methodology for an automatic search for articles conducted most suitably and efficiently.
However, according to the keywords used to search the publication sources, some literature could be missing due to the defined keywords. The findings of the pilot study supported our research string formulations. A pilot study by both the authors was done on 24 sample papers to evaluate the challenges of the automated search and verify our search string. After the automated searching process was completed, a manually searching process was conducted on top publications houses and conference proceeding databases to find any relevant study on energy efficiency to ensure that we do not miss any critical studies.
To enhance the efficiency of our searching process and mitigate the threat of validity of incomplete primary studies, the authors independently evaluated the chosen study sources and previous surveys. The selection stage comprises criteria for inclusion/exclusion and the method for quality evaluation. Due to a significant absence of availability or language barrier, we might have missed out on some critical research related to our domain.

6.3. Internal Validity

The reviewing process of research articles utilizing inclusion and exclusion criteria and quality evaluation processes is prone to revisionist bias to exclude relevant studies. However, the preparation against these threats was taken very carefully. A random collection of 76 documents received through automated search was screened separately by both authors to explain any equivocal views of inclusion and exclusion criteria. The results of the screening were compared, and the variations were resolved.
According to [101], the first author performed the quality evaluation and data extraction from primary studies, and the second author-verified each decision. Furthermore, the excluded studies do not fulfil the quality specified for this study, as we follow the well-defined lineament assessment standards proposed by [55]. In the research classification process, the authors have participated entirely, and any conflicts of opinion were settled through discussion to ensure mutual consistency.

6.4. External Validity

Such research focuses mainly on the benefits and pitfalls of cloud-based energy-efficient technologies. The findings obtained about the environmental effect of high energy consumption will differ from programming paradigms since various techniques occur. Furthermore, the results of this SMS are valid primarily for energy-efficient cloud computing technology. This analysis results may differ for other software devices, such as networking and the operating system’s architecture. This study might act as a framework for many researchers to further recognize and classify our climate and ICT technology’s effect on high power usage.

7. Conclusions

This paper addresses a rigorous systematic mapping analysis to disclose the current state of the art for energy-efficiency techniques and recognize future accessible issues in the empirical literature concerning the effect of high energy consumption on the environment. First, we provided a short overview of energy-efficiency techniques used in the cloud to cut down the energy consumption and further discuss their need and role in a sustainable environment. Later, we addressed the complete procedure for conducting an SMS to perform the mapping analysis.
A two-phase methodology was employed to determine the relevant literature regarding our concern topic. Initially, we searched from six databases, Springer, Scopus, ScienceDirect, IEEE Xplore, Wiley, and ACM Digital library. There were almost ten esteemed journals and six conferences considered for manual research. After the manual search, we also conducted snowballing to enhance our search efficiency and reduce the chance of neglecting any relevant article. After carefully screening our selected studies against the quality assessment criteria, we selected a pool of 119 PSs out of 2903 initial search results. These PSs were further categorized and analyzed due to our research questions.
The mapping of PSs according to their publishing years shows that the determination to find the impact of high energy usage on the environment is continuously thriving and establishes a highly active research field of cloud computing. We hope that this study will support industry professionals in understanding current research related to the effects and impact of high energy consumption in cloud data centers. The limitations and obstacles addressed in this SMS will also help researchers identify critical research gaps that must be addressed.
In recent years, computer systems’ energy consumption has experienced many turning points as it is one of the latest trends in the ICT industry. The authors attempted to highlight both the beneficial and harmful effects of cloud computing and to explain their impact on the ecosystem. According to the World energy outlook report in 2013, annual energy efficiency spending will increase to US$550 billion in 2035 with almost 60 per cent of clean energy investments.
Cloud computing is typically supposed to establish a harmonious relationship with the community because the manufacturers of ICT equipment and the suppliers in this sector have agreed on environmental policy and NGOs’ recommendations about minimizing the harmful impact of energy consumed by operating the data centers. The techniques were grouped into VM-based strategies, consolidation techniques, bio-based optimization processes, thermal control, and non-technical techniques. A future study might examine the effect on results and quality of service by various energy-efficiency techniques.

8. Future Work

Today, non-renewable resources, such as coal, are the primary energy source in the world’s energy market. With the limited supply of non-renewable energy sources, the emphasis has been moved from taking concrete steps to derogate energy usage levels in the data centers. An energy-efficient strategy that can significantly deal with the energy consumption issue and proficiently manage such a massive data volume is now a fundamental requirement. Other relevant areas require extensive investigation to optimize energy efficiency in cloud data centers.
In this study, we attempted to draw various facts about energy-efficient systems into one picture. For green computing, we instituted four sections, manufacturing computers, software strategies, awareness of people, and standardized policies. The output of this paper will be beneficial for all the four categories, as we presented how green computing concepts, various mechanisms, and some non-technical activities in data centers can improve energy savings, time, and costs.
Cloud computing is a modern paradigm that combines existing technologies to improve resource usage and better workload distribution. According to our research, the contribution of cloud computing to ecological sustainability is considered important, and the most relevant elements follow as under: According to green ICT standards, cloud technology should be applied with minimal harmful effects on the environment. The providers offering cloud services should decrease their non-renewable energy usage and substitute this consumption with renewable energy to follow environmental conservation provisions.
A rise in renewable energy consumption would result in lower emissions of CO2. However, because the first indicator is still not planned, it is doubtful that the pollution reductions caused by carbon will meet environmental organizations’ expectations. Reductions of resource usage, CO2 emissions, and e-waste can concern both cloud service providers and consumers while selecting their services. Cloud computing can impact reduction and substitution for the count of equipment demanded by organizations. Cloud service providers should be alert and customize their practices according to all region’s environmental guidelines.
Organizations and NGOs for the environment should verify and appraise the details about the effect of cloud infrastructure on the ecosystem, and these details are readily available online and can be very advantageous for building up new policies to protect our environment. This research is crucial for decision-makers as they can easily predict a requirement for more comprehensive and robust models for future computing technologies and mitigation options.
Researchers can use the output of this study before developing new techniques as they will have a better idea about the limitations and scope of the previous studies conducted on energy efficiency. Moreover, we must develop more techniques for modelling emerging trends, such as the rollout of 5G, AI, and edge and fog computing. Using this paper’s results, policymakers will obtain former insights into their potential energy consumption implications and accelerate their investments in energy-saving technologies that reduce consumption gains.

Author Contributions

Conceptualization: S.B. and S.S.; methodology: S.B.; validation: S.B.; formal analysis: O.I.K., G.M.A., M.M. and S.S.; investigation: S.B., S.S. and O.I.K.; resources: S.B.; data curation: S.S.; writing original draft preparation: S.B.; writing review and editing: S.B., S.S. and G.M.A.; visualization: S.B.; supervision: S.S., O.I.K., M.M. and T.H.H.A.; project administration: S.B. and S.S.; funding acquisition: O.I.K., G.M.A., A.S.A.H., M.M. and H.A. All authors have read and agreed to the published version of the manuscript.

Funding

Research Center of College of Computer and Information Sciences, Deanship of Scientific Research, King Saud University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The study did not report any data.

Acknowledgments

This research project was supported by a grant from the Research Center of College of Computer and Information Sciences, Deanship of Scientific Research, King Saud University.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Primary Studies
  • [S1] Lefevre, L., Orgerie, A. Designing and evaluating an energy efficient Cloud. J Supercomput 51, 352–373 (2010). doi.org/10.1007/s11227010-0414
  • [S2] B. Nordman and L. Berkeley, “Greener PCs for the Enterprise,” no. August, 2009 Doi 10.1109/MITP.2009.71
  • [S3] J. Younge, G. von Laszewski, L. Wang, S. Lopez-Alarcon and W. Car“thers, “Efficient resource management for Cloud computing envir”nments,” International Conference on Green Computing, Chicago, IL, 2010, pp. 357-364, doi: 10.1109/GREENCOMP.2010.5598294.
  • [S4] J. M. Kim, M. Kim, J. Kong, H. B. Jang and S. W.“Chung, “Display Power Management That Detects User”Intent,” in Computer, vol. 44, no. 10, pp. 60-66, Oct. 2011, doi: 10.1109/MC.2011.312.
  • [S5] F. Cao and M. M. Zhu, “Energy efficient workflow job scheduling for green cloud,” Proc. 2013 IEEE 27th Int. Symp. on Parallel and Distributed Processing Workshops and PhD Forum, pp. 2218–2221.
  • [S6] Lin, C. (2012). A Novel Green Cloud Computing Framework for Improving System Efficiency. Physics Procedia, 24, 2326–2333. https://doi.org/10.1016/j.phpro.2012.02.345
  • [S7] Hulkury, M. N., & Doomun, M. R. (2012, November). Integrated Green Cloud Computing Architecture. 2012 International Conference on Advanced Computer Science Applications and Technologies (ACSAT). https://doi.org/10.1109/acsat.2012.16]
  • [S8] K. Palanivel and S. Kuppuswami “A Cloud-Oriented Green Computing Architecture for E-Learning Applications “in International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169 Volume: 2 Issue: 11 3775–3783
  • [S9] Ketankumar, D. C., Verma, G., & Chandrasekaran, K. (2015). A Green Mechanism Design Approach to Automate Resource Procurement in Cloud. Procedia Computer Science, 54, 108–117. https://doi.org/10.1016/j.procs.2015.06.013
  • [S10] Khanna R, Zuhayri F, Nachimuthu M, Le C, Kumar MJ. Unified extensible firmware interface: An innovative approach to DRAM power control. In: Energy Aware Computing (ICEAC), 2011 International Conference on. IEEE; 2011. p. 1±6. 10.1109/ICEAC.2011.6136703
  • [S11]  Weaver VM, Johnson M, Kasichayanula K, Ralph J, Luszczek P, Terpstra D, et al. Measuring energy and power with PAPI. In: Parallel Processing Workshops (ICPPW), 2012 41st International Conference on. IEEE; 2012. p. 262±268 10.1109/ICPPW.2012.39.
  • [S12] Esmaeilzadeh H, Cao T, Xi Y, Blackburn SM, McKinley KS. Looking back on the language and hardware revolutions: measured power, performance, and scaling. In: ACM SIGARCH Computer Architecture News. vol. 39. ACM; 2011. p. 319±332
  • [S13] Chen Y, Chen T, Xu Z, Sun N, Temam O. DianNao family: energy-efficient hardware accelerators for machine learning. Communications of the ACM. 2016; 59(11):105±112. https://doi.org/10.1145/2996864
  • [S14] Rotem, E., Weiser, U. C., Mendelson, A., Ginosar, R., Weissmann, E., & Aizik, Y. (2016). H-EARtH: Heterogeneous Multicore Platform Energy Management. Computer, 49(10), 47–55. https://doi.org/10.1109/mc.2016.309
  • [S15] Mashayekhy, L., Nejad, M. M., & Grosu, D. (2015). Physical Machine Resource Management in Clouds: A Mechanism Design Approach. IEEE Transactions on Cloud Computing, 3(3), 247–260. https://doi.org/10.1109/tcc.2014.2369419
  • [S16] Singh, K., & Ku, M.-L. (2015). Toward Green Power Allocation in Relay-Assisted Multiuser Networks: A Pricing-Based Approach. IEEE Transactions on Wireless Communications, 14(5), 2470–2486.https://doi.org/10.1109/twc.2014.2387165
  • [S17] Chiang, Y.-J., Ouyang, Y.-C., & Hsu, C.-H. (2015). An efficient green control algorithm in cloud computing for cost optimization. IEEE Transactions on Cloud Computing, 3(2), 145–155. https://doi.org/10.1109/tcc.2014.2350492
  • [S18] B. Li, J. Li, J. Huai, T. Wo, Q. Li and L.“Zhong, “EnaCloud: An Energy-Saving Application Live Placement Approach for Cloud Computing Envir”nments,” 2009 IEEE International Conference on Cloud Computing, Bangalore, India, 2009, pp. 17-24, doi: 10.1109/CLOUD.2009.72.
  • [S19] Anton Beloglazov and Rajkumar Buyya. (2012). Optimal Online Deterministic Algorithms and Adaptive Heuristics for Energy and Performance Efficient Dynamic Consolidation of Virtual Machines in Cloud Data Centers. Concurrency Computation: Pracices & Experiments- 2012; 24:1397–1420.
  • [S20] K. H. Kim and J. Kim, “Power aware scheduling of bag-of-tasks applications with deadline constraints on DVS-enabled clusters,” CCGRID, vol. 7, pp. 541–548, May 2007. Doi 10.1109/CCGRID.2007.85
  • [S21] Khare, S., & Jain, S. (2013, January). Prospects of Near-Threshold Voltage Design for Greeben Computing. 2013 26th International Conference on VLSI Design and 2013 26th International Conference on VLSI Design: concurrently with the 12th International Conference on Embedded Systems. https://doi.org/10.1109/vlsid.2013.174
  • [S22] Reddy, S. P., & Chandan, H. K. S. (2014, February). Energy aware scheduling of real-time and non real-time tasks on cloud processors (Green Cloud Computing). International Conference on Information Communication and Embedded Systems (ICICES2014). https://doi.org/10.1109/icices.2014.7033827
  • [S23] Mohsen Amini Salehi, Pisipati Radha Krishna, Krishnamurty Sai Deepak, and Rajkumar Buyya. 2012. Preemption-aware energy management in virtualized data centres cloud. IEEE 5th International Conference on Cloud Computing. 844–851. DOI: http://dx.doi.org/10.1109/CLOUD.2012.147
  • [S24] Kim, K., Yang, F., Zavala, V. M., & Chien, A. A. (2017). Data Centers as Dispatchable Loads to Harness Stranded Power. IEEE Transactions on Sustainable Energy, 8(1), 208–218. https://doi.org/10.1109/tste.2016.2593607
  • [S25] Pesch, D., Rea, S., Torrens, J. I., Zavrel, V., Hensen, J. L. M., Grimes, D., O’Sullivan, B., Scherer, T., Birke, R., Chen, L., Engbersen, T., Lopez, L., Pages, E., Mehta, D., Townley, J., & Tsachouridis, V. (2017). Globally Optimized Energy-Efficient Data Centres.–In ICT-Energy Concepts for Energy Efficiency and Sustainability. InTech. https://doi.org/10.5772/65988
  • [S26] K. Chen, Y. Xu, K. Xi, and H. J. Chao, “Intelligent Virtual Machine Placement for Cost Efficiency in Geo- Distributed Cloud Systems,” pp. 3498-3503, 2013.
  • [S27] Y. E. F. M. S. Vasanthi, “Energy Efficient Mechanism for Green Computing in Wireless Storage Area Networks,” pp. 1311-1314, 2015.
  • [S28] Anton Beloglazov and Rajkumar Buyya. (2010). Adaptive Threshold-Based Approach for Energy-Efficient Consolidation of Virtual Machines in Cloud Data Cent‘rs MGC ‘10 Proceedings of the 8th International Workshop on Middleware for Grids, Clouds and e-Science, Bangalore, India Nove–ber 29-December 03.
  • [S29] Jian-Sheng Liao, Chi-Chung Chang, Yao-Lun Hsu, Xiao-Wei Zhang, Kuan-Chou Lai, and Ching- Hsien Hsu. 2012. Energy-efficient resource provisioning with SLA consideration on cloud computing. 41st International Conference on Parallel Processing Workshops (ICPPW’12), 206–211.
  • [S30] An Energy-efficient Scheduling Approach Based on Private Clouds Jiandun Li, Junjie Peng, Zhou Lei, Wu Zhang Journal of Information & Computational Science 8: 4 (2011) 716–724 Available at http://www.joics.com
  • [S31] Shailesh S. Deore, Ashok N. Patil, and Ruchira Bhargava. 2012. Energy-efficient scheduling scheme for virtual machines in cloud computing. International Journal of Computer Applications 56, 10 (Oct. 2012). DOI: http://dx.doi.org/10.5120/8926–2999
  • [S32] Dang Minh Quan, Andrey Somov, and Corentin Dupont. 2012. Energy usage and carbon emission optimization mechanism for federated data centers. In Energy Efficient Data Centres. Lecture Notes in Computer Science, Vol. 7396 (2012). Springer, Berlin, 129–140. DOI: http://dx.doi.org/10.1007/978-3-642-33645-4_12
  • [S33] Dang Minh Quan, Robert Basmadjian, Hermann De Meer, Ricardo Lent, Toktam Mahmoodi, Domenico Sannelli, Federico Mezza, Luigi Telesca, and Corenten Dupont. 2012. Energy efficient resource allocation strategy for cloud data centres. In Computer and Information Sciences II (2012), 133–141, Springer, London. DOI: http://dx.doi.org/10.1007/978-1-4471-2155-8_16
  • [S34] Anton Beloglazov and Rajkumar Buyya. 2010. Adaptive threshold-based approach for energy-efficient consolidation of virtual machines in cloud data centers. In Proceedings of the 8th International Workshop on Middleware for Grids, Clouds and e-Science. ACM, 4. DOI:http://dx.doi.org/10.1145/1890799.1890803
  • [S35] Beloglazov, J. Abawajy and R.“Buyya, “Energy-aware resource allocation heuristics for efficient management of data centers for Cloud c”mputing”, Future Generation Computer Systems, vol. 28, no. 5, pp. 755-768, 2012.
  • [S36] Yacine Kessaci, Nouredine Melab, and El-Ghazali Talbi. 2012. An energy-aware multi-start local search heuristic for scheduling VMs on the OpenNebula cloud distribution. International Conference on High Performance Computing and Simulation (HPCS). 112–118. DOI: http://dx.doi.org/10.1109/HPCSim.2012.6266899
  • [S37] Thomas Knauth and Christof Fetzer. 2012. Energy-aware scheduling for infrastructure clouds. 4th International Conference on Cloud Computing Technology and Science (CloudCom’12). IEEE, 58–65. DOI:http://dx.doi.org/10.1109/CloudCom.2012.6427569
  • [S38] Ashkan Paya and Dan C.Marinescu. 2013. Energy-Aware Application Scaling On a Cloud. Retrieved October,2013 from arXiv:1307.3306
  • [S39] N. Kord and H. Haghighi, “An energy-efficient approach for virtual machine placement in cloud based data centers,” in Proc. 5th Information and Knowledge Technology Conf., 2013, pp. 44–49
  • [S40] Feller, L. Rilling and C. Morin, “Energy-aware ant colony based workload placement in clouds,” Proc. 2011 IEEE/ACM 12th Int. Conf. on Grid Computing, pp. 26–33.
  • [S41] J. Verma, C. Katti and P. “axena, “MADLVF: An Energy Efficient Resource Utilization Approach for Cloud C”mputing”, International Journal of Information Technology and Computer Science, vol. 6, no. 7, pp. 56-64, 2014. 17
  • [S42] M. J. Usman, A. Samad, Ismail, H. Chizari and A.“Aliyu, “Energy- Efficient virtual machine allocation technique using interior search algorithm for cloud dat”center, 2017 6th ICT International StudentProject Conference (ICT-ISPC), Skudai, 2017, pp. 1-4. Dooi 10.1109/ICT-ISPC.2017.8075327
  • [S43] Horri, M. Mozafari and G. Dastghai“yfard, “Novel resource allocation algorithms to performance and energy efficiency in cloud c”mputing”, The Journal of Supercomputing, vol. 69, no. 3, pp. 1445-1461, 2014.
  • [S44] Tseng, T. Yang, K. Yang and P“ Chen, “An Energy Efficient VM Management Scheme with Power-Law Characteristic in Video Streaming Data”Centers”, IEEE Transactions on Parallel and Distributed Systems, vol. 29, no. 2, pp. 297-311, 2018.
  • [S45] K. Wu, R. Du, L. Chen and “. Yan, “An Energy-Saving Virtual-Machine Scheduling Algorithm of Cloud Computing”System,” 2013 International Conference on Information Science and Cloud Computing Companion, Guangzhou, pp. 219-224, 2013.
  • [S46] Ghribi, M. Hadji and D. Zeg“lache, “Energy Efficient VM Scheduling for Cloud Data Centers: Exact Allocation and Migration Al”orithms”, 2013 13 th IEEE/ACM International Symposium on Cluster Cloud and Grid Computing, pp. 671-678, May 2013
  • [S47] D. Dad and G. B“lalem, “Efficient allocation of VMs in servers of data center to reduce energy cons”mption, 2017 3rd International Conference of Cloud Computing Technologies and Applications (CloudTech), Rabat, pp. 1-6, 2017. 10.1109/CloudTech.2017.8284743
  • [S48] J. Wang, C. Huang, K. He, X. Wang, X. Chen and “. Qin, “An energyaware resource allocation heuristics for VM scheduling ”n cloud”,Proceedings of 2013 IEEE International Conference on High Performance Computing and Communications (HPCC 2013) and 2013IEEE International Conference on Embedded and Ubiquitous Computing (EUC 2013), pp. 587-594, 2013. 20
  • [S49] X. Zheng and “. Cai, “Dynamic Virtual Machine Placement for Cloud Computing Envir”nments,” 2014 43rd International Conference on Parallel Processing Workshops, Minneapolis, MN, pp. 121-128, 2014
  • [S50] Y. Jararweh, M. Issa, M. Daraghmeh, M. Al-Ayyoub and M. Al“mirat, “Energy efficient dynamic resource management in cloud computing based on logistic regression model and median absolute d”viation”, Sustainable Computing: Informatics and Systems, vol. 19, pp. 262-274, 2018.
  • [S51] S. K. Mishra “t al., “Energy-efficient VM-placement in cloud dat” center”, Sustain. Comput. Informat. Syst., pp. 48-50, 2018.
  • [S52] Tchernykh, L. Lozano, P. Bouvry, J. E. Pecero, U. Schwiegelshohn and S. Nesm“chnow, “Energy-aware online scheduling: Ensuring quality of service for IaaS”clouds,” 2014 International Conference on High Performance Computing & Simulation (HPCS), Bologna, 2014, pp. 911- 918.doi 10.1109/HPCSim.2014.6903786
  • [S53] Lin, P. Liu and“J. Wu, “Energy-Aware Virtual Machine Dynamic Provision and Scheduling for Cloud Co”puting,” 2011 IEEE 4th International Conference on Cloud Computing, Washington, DC, 2011,pp. 736-737
  • [S54] N. Kim, J. Cho and “. Seo, “Energy-Based Accounting and Scheduling of Virtual Machines in a Cloud”System,” 2011 IEEE/ACM International Conference on Green Computing and Communications, Sichuan, 2011, pp. 176-181.
  • [S55] N. Quang-Hung and N.“Thoai, “Minimizing Total Busy Time with Application to Energy-Efficient Scheduling of Virtual Machines in IaaS”Clouds,” 2016 International Conference on Advanced Computing and Applications (ACOMP), Can Tho, 2016, pp. 141-148.
  • [S56] F. Juarez, J. Ejarque and R.“Badia, “Dynamic energy-aware scheduling for parallel task-based application in cloud c”mputing”, Future Generation Computer Systems, vol. 78, pp. 257-271, 2018.
  • [S57] Anton Beloglazov and Rajkumar Buyya. 2010. Energy efficient allocation of virtual machines in cloud data centers. 10th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing (CCGrid’10). 577–578. DOI: http://dx.doi.org/10.1109/CCGRID.2010.45
  • [S58] X. Sun, Q. Zhang, M.A. Medina, Y. Liu, S. Liao, A study on the use of phase change materials (PCMs) in combination with a natural cold source for space cooling in telecommunications base stations (TBSs) in China, Appl. Energy 117 (2014) 95–103.
  • [S59] Dzmitry Kliazovich, Pascal Bouvry, and Samee U. Khan. 2013. DENS: Data center energy-efficient network aware scheduling. Cluster Computing 16, 1 (2013), 65–75, 2013. DOI: http://dx.doi.org/10.1007/s10586- 011-0177-4
  • [S60] P. Arroba, J. M. Moya, J. L. Ayala and R.“Buyya, “DVFS-Aware Consolidation for Energy-Efficient”Clouds,” 2015 International Conference on Parallel Architecture and Compilation (PACT), San Francisco, CA, pp. 494-495, 2015.
  • [S61] P. Sanjeevi and P. Viswa“athan, “Towards energy-aware job consolidation scheduling i” cloud,” 2016 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, pp. 1-6, 2016.
  • [S62] W. Guo, P. Kuang, Y. Jiang, X. Xu, and W. Tian, “SAVE: Self-adaptive consolidation of virtual machines for energy ef_ciency of CPU-intensive applications in th’’cloud,” J. Supercomput., pp. 1_25, Jun. 2019. doi: 10.1007/s11227-019-02927-1.
  • [S63] G. Cao, “Topology-aware multi-objective virtual machine dynamic consolidation for cloud dat’’enter,” Sustain. Comput. Inform. Syst., vol. 21, pp. 179_188, Mar. 2019
  • [S64] Z. Li, C. Yan, X. Yu, and N. Yu, “Bayesian network-based virtual machines consolidation’’ethod,” Future Gener. Comput. Syst., vol. 69,no. 3, pp. 75_87, Apr. 2017.
  • [S65] Yan, H.Wang, X. Li, Y.Wang, D. Li, Y. Zhang, Y. Xie, Z. Liu, W. Cao, and F. Yu, “Cost-ef_cient consolidating service fo’ Aliyun’s cloud-scale co’’uting,” IEEE Trans. Services Comput., vol. 12, no. 1, pp. 117_130, Jan./Feb. 2019.
  • [S66] Anton Beloglazov, Rajkumar Buyya, Young C. Lee, and Albert Zomaya. 2011. A taxonomy and survey of energy-efficient data centers and cloud computing systems. Advances in Computers 82, 2 (2011)
  • [S67] D Andersen, J Franklin, M Kaminsky, Phanishayee A, Tan L, Vasudevan V (2009) FAWN: A Fast Array of Wimpy Nodes. In: 22nd ACM Symposium on Operating Systems Principles
  • [S68] Caulfield A, Grupp L, Swanson S (2009) Gordon: using flash memory to build fast, power-efficient clusters for data-intensive applications. In: 14th international conference on Architectural support for programminglanguages and operating systems
  • [S69] Portaluri G, Giordano S, Kliazovich D, Dorronsoro B (2014) A power efficient genetic algorithm for resource allocation in cloud computing data centers. In: IEEE 3rd International conference on cloud networking
  • [S70] Quang-Hung N, Nienz P, Namz N, Tuong N, Thoa N (2013) A genetic algorithm for power-aware virtual machine allocation in private cloud. Lecture Notes in Compr Sci 7804:170–179
  • [S71] Zhang W, Xie H, Cao B, Cheng A (2014) Energy-aware real-time task scheduling for heterogeneous multiprocessors with particle swarm optimization algorithm. Math Probl Eng 2014:Article ID 287475
  • [S72] Lin, M.; Shao, S.; Zhang, X.S.; VanGilder, J.W.; Avelar, V.; Hu, X. Strategies for data center temperature control during a cooling system outage. Energy Build. 2014, 73, 146–152.
  • [S73] Gregor von Laszewski, Lizhe Wang, Andrew J. Younge, and Xi He. 2009. Power-aware scheduling of virtual machines in DVFS-enabled clusters. International Conference on Cluster Computing and Workshops (CLUSTER’09). IEEE, 1–10. DOI: http://dx.doi.org/10.1109/CLUSTR.2009.5289182
  • [S74] P. Mahadevan, P. Sharma, S. Banerjee, and P. Ranganathan, “Energy Aware Network Operations,” IEEE INFOCOM Workshops 2009, pp. 1– 6, Apr. 2009
  • [S75] Andrew J. Younge, Gregor Von Laszewski, Lizhe Wang, Sonia Lopez-Alarcon, and Warren Carithers. 2010. Efficient resource management for cloud-computing environments. In International Green Computing Conference. IEEE, 357–364. DOI: http://dx.doi.org/10.1109/GREENCOMP.2010.5598294
  • [S76] Enhancing Energy-Efficient and QoS Dynamic Virtual Machine Consolidation Method in Cloud Environment Yaqiu Liu; Xinyue Sun; Wei Wei; Weipeng Jing 10.1109/ACCESS.2018.2835670
  • [S77] T. Ghose, V. Namboodiri, and R. Pendse, “Thin is green: Leveraging the thin-client paradigm for sustainable mobile computing q,” Comput. Electr. Eng., vol. 45, no. 2015, pp. 155-168, 2016
  • [S78] Qingwen Chen. 2011. Towards energy-aware VM scheduling in IaaS clouds through empirical studies. MScThesis. University of Amsterdam.
  • [S79] Tang, Z., Qi, L., Cheng, Z., Li, K., Khan, S. U., & Li, K. (2015). An Energy-Efficient Task Scheduling Algorithm in DVFS-enabled Cloud Environment. Journal of Grid Computing, 14(1), 55–74. https://doi.org/10.1007/s10723-015-9334-y
  • [S80] Faraz Ahmad and T. N. Vijaykumar. 2010. Joint optimization of idle and cooling power in data centers while maintaining response time. ACM SIGPLAN Notices 45, 3 (March 2010) 243–256. DOI:http://dx.doi.org/10.1145/1735971.1736048
  • [S81] Ma, F., Liu, F., & Liu, Z. (2012). Multi-objective optimization for initial virtual machine placement in cloud data center.
  • [S82] Gao, Y., Guan, H., Qi, Z., Hou, Y., & Liu, L. (2013). A multi-objective ant colony system algorithm for virtual machine placement in cloud computing. Journal of Computer and System Sciences, 79(8), 1230–1242. https://doi.org/10.1016/j.jcss.2013.02.004
  • [S83] Ferdaus, M. H., Murshed, M., Calheiros, R. N., & Buyya, R. (2014). Virtual Machine Consolidation in Cloud Data Centers Using ACO Metaheuristic. In Lecture Notes in Computer Science (pp. 306–317). Springer International Publishing. https://doi.org/10.1007/978-3-319-09873-9_26
  • [S84] Shabeera, T., Kumar, S.M., Salam, S.M., Krishnan, K.M.: Optimizing VM allocation and data placement for data-intensive applications in the cloud using ACO metaheuristic algorithm. Eng. Sci. Technol. Int. J. 20(2), 616–628 (2017). https://doi.org/10.1016/j.jestch.2016.11.006
  • [S85] Liu, X.-F., Zhan, Z.-H., Deng, J. D., Li, Y., Gu, T., Zhang, J: An energy-efficient ant colony system for virtual machine placement in cloud computing. In: Proceedings of the IEEE Transactions on Evolutionary Computation (2016)
  • [S86] Ben Hassen, F., Brahmi, Z., & Toumi, H. (2016, April). VM placement algorithm based on recruitment process within ant colonies. 2016 International Conference on Digital Economy (ICDEc). 2016 International Conference on Digital Economy (ICDEc). https://doi.org/10.1109/icdec.2016.7563138
  • [S87] Satpathy, A., Addya, S.K., Turuk, A.K., Majhi, B., Sahoo, G.: Crow search based virtual machine placement strategy in cloud data centers with live migration. Comput. Electr. Eng. 69, 334–350 (2018). https://doi.org/10.1016/j.compeleceng.2017.12.032
  • [S88] Kyong Hoon Kim, Anton Beloglazov, and Rajkumar Buyya. 2011. Power-aware provisioning of virtual machines for real-time Cloud services. Concurrency and Computation: Practice and Experience 23, 13 (2011), 1491–1505.
  • [S89] Mark, C.C.T., Niyato, D., Chen-Khong, T.: Evolutionary optimal virtual machine placement and demand forecaster for cloud computing. In: Proceedings of the 2011 IEEE International Conference on Advanced Information Networking and Applications (AINA), pp. 348–355, IEEE (2011)
  • [S90] Yacine Kessaci, Mohand Mezmaz, Nouredine Melab, El-Ghazali Talbi, and Daniel Tuyttens. 2011. Parallel evolutionary algorithms for energy aware scheduling. Intelligent Decision Systems in Large-Scale Distributed Environments, Studies in Computational Intelligence, Vol. 362, Springer, Berlin, 75–100. DOI: http://dx.doi.org/10.1007/978-3-642-21271-0_4
  • [S91] Xiaoli Wang, Yuping Wang, and Hai Zhu. 2012. Energy-efficient multi-job scheduling model for cloud computing and its genetic algorithm. Mathematical Problems in Engineering (2012).
  • [S92] Fereydoun Farrahi Moghaddam, Mohamed Cheriet, and Kim Khoa Nguyen. Low carbon virtual private clouds. 2011. IEEE International Conference on Cloud Computing (CLOUD’11), 259–266 DOI: http://dx.doi.org/10.1109/CLOUD.2011.36
  • [S93] Lskrao Chimakurthi and S. D. Madhukumar. 2011. Power efficient resource allocation for clouds using ant colony framework. arXiv:1102.2608. http://arxiv.org/pdf/1102.2608.pdf
  • [S94] Zhang, W.Z., et al.: Secure and Optimized Load Balancing for Multitier IoT and Edge-Cloud Computing Systems. IEEE Internet Things J. 8, 8119–8132 (2021)
  • [S95] R. G. Babukarthik, R. Raju, and P. Dhavachelvan. 2012. Energy-aware scheduling using Hybrid Algorithm for cloud computing. 3rd International Conference on Computing Communication and Networking Technologies (ICCCNT’12). 1–6. DOI: http://dx.doi.org/10.1109/ICCCNT.2012.6396014
  • [S96] Zhiming Wang, Kai Shuang, Long Yang, and Fangchun Yang. 2012. Energy-aware and revenue-enhancingcombinatorial scheduling in virtualized of cloud datacenter. Journal of Convergence Information Technology, AICIT 7, 1 (2012), 62–70. DOI: http://dx.doi.org/10.4156/jcit.vol7.issue1.8
  • [S97] Neil Bergmann, Yuk Ying Chung, Xiangrui Yang, Zhe Chen, Wei Chang Yeh, Xiangjian He, and Raja Jurdak. 2013. Using swarm intelligence to optimize the energy consumption for distributed systems. Modern Applied Science 7, 6 (2013), 59–66. DOI: http://dx.doi.org/10.5539/mas.v7n6p59
  • [S98] Zavrel, V., Bartak, M., & Hensen, J. L. M. (2014). Simulation of a data center cooling system in an emergency situation. In Proceedings of the 8th IBPSA-CZ conference Simulace Budov a Techniky Prostředí, 6-7 November 2014, Prague, Czech Republic (pp. 57-64). Česká t–chnika-nakladatelství ČVUT.
  • [S99] Y. Chen, C. Lin, J. Huang, X. Xiang and X“ Shen, “Energy Efficient Scheduling and Management for Large-Scale Services Computing”Systems”, IEEE Transactions on Services Computing, vol. 10, no. 2, pp.217-230, 2017.
  • [S100] Y. Chen, Y. Zhang, Q. Meng, Study of ventilation cooling technology for telecommunication base stations: control strategy and application strategy, Energy Build. 50 (no. 7) (2012) 212–218.
  • [S101] L. Bao, J. Wang, L. Kang, The applied effect analysis of heat exchanger installed in a typical communication base station in Beijing of China, Energy Procedia 14 (2012) 620–625.
  • [S102] A.M. Haywood, J. Sherbeck, P. Phelan, G. Varsamopoulos, S.K.S. Gupta, The relationship among CPU utilization, temperature, and thermal power for waste heat utilization, Energy Convers. Manag. 95 (2015) 297–303.
  • [S103] L. Ling, Q. Zhang, Y. Yu, Y. Wu, S. Liao, Study on thermal performance of microchannel separate heat pipe for telecommunication stations: Experiment and simulation Étude de la performance thermique de caloduc séparé à micro-canauxpour les stations de telecommunication: expérienc vol. 59, (2015), pp. 198–209.
  • [S104] H. Zhang, S. Shao, H. Xu, H. Zou, C. Tian, Integrated system of mechanical refrigeration and thermosyphon for free cooling of data centers, Appl. Therm. Eng. 75 (2015) 185–192.
  • [S105] Almaneea, H. Thompson, J. Summers, N. Kapur, Cooling system analysis for a data center using liquid immersed servers, Int. J. Therm. Technol. 4 (no. 3) (2014)200–207
  • [S106] H. Endo, H. Kodama, H. Fukuda, T. Sugimoto, T. Horie, M. Kondo, Effect of climatic conditions on energy consumption in direct fresh-air container data centers 2013, 2013 Int. Green Comput. Conf. Proceedings, IGCC 6 (2013) 17–25.
  • [S107] Deng, X., Wu, D., Shen, J., & He, J. (2016). Eco-Aware Online Power Management and Load Scheduling for Green Cloud Datacenters. In IEEE Systems Journal (Vol. 10, Issue 1, pp. 78–87). Institute of Electrical and Electronics Engineers (IEEE). https://doi.org/10.1109/jsyst.2014.2344028
  • [S108] H. Jouhara and R. Meskimmon, Heat Pipe based Thermal Management Systems for Energy-Efficient Data Centres.
  • [S109] Almoli, A. Thompson, N. Kapur, J. Summers, H. Thompson, G. Hannah, Computational fluid dynamic investigation of liquid rack cooling in data centres, Appl. Energy 89 (1) (2012) 150–155.
  • [S110] S. Zimmermann, I. Meijer, M.K. Tiwari, S. Paredes, B. Michel, D. Poulikakos, Aquasar: a hot water cooled data center with direct energy reuse, Energy 43 (no. 1) (2012) 237–245.
  • [S111] Í. Goiri, W. Katsak, K. Le, T. D. Nguyen and R. Bia“chini, “Designing and Managing Data centers Powered by Renewable”Energy,” in IEEE Micro, vol. 34, no. 3, pp. 8-16, May-June 2014, doi: 10.1109/MM.2014.6.
  • [S112] Jonas Carlsson, Kent Palmkvist, and Lars Wanhammar, “A Clock Gating Circuit for Globally Asynchronous Locally Synchronous Systems”, Norchip Conference, 2006.
  • [S113] FAWN: A Fast Array of Wimpy Nodes David G. Andersen ACM 978-1-60558-752-3/09/10
  • [S114] Ivan Rodero, Juan Jaramillo, Andres Quiroz, Manish Parashar, Francesc Guim, and Stephen Poole. 2010. Energy-efficient application-aware online provisioning for virtualized clouds and data centers. International Green Computing Conference. 31–45. DOI: http://dx.doi.org/10.1109/GREENCOMP.2010.5598283
  • [S115] F. Sheikh, H. Fazal, F. Taqvi and J. “hamsi, “Power-aware server selection in nano data”center,” 2015 IEEE 40th Local Computer Networks Conference Workshops (LCN Workshops), Clearwater Beach, FL, 2015, pp. 776-782, doi: 10.1109/LCNW.2015.7365927.
  • [S116] Zhao, R. et al.,” A Multithreaded Compiler Optimization Technology with Low Power”. Journal of Software, 2002 pp. 1123-1129
  • [S117] S.Abdulsalam, D. Lakomski, Q. Gu, T. Jin and Z“ Zong, “Program energy efficiency: The impact of language, compiler and implementation ”hoices,” International Green Computing Conference, Dallas, TX, 2014, pp. 1-6, doi: 10.1109/IGCC.2014.7039169.
  • [S118] Heller, B.; Seetharaman, S.; Mahadevan, P.; Yiakoumis, Y.; Sharma, P.; Banerjee, S.; McKeown, N. Elastic tree: Saving Energy in Data Center Networks. In Proceedings of the 7th USENIX Conference on Networked Systems Design and Implementation,
  • [S119] San José, CA, USA, 28–30 April 2010; pp. 249–264.Vytautas Valancius, Nikolaos Laoutaris, Laurent Massoulié, Christophe Diot, and Pablo Rodriguez. 2009. Greening the internet with nano data centers. In Proceedings of the 5th international conference on Emerging networking experiments and technologies ‘CoNEXT ‘09). Association for Computing Machinery, New York, NY, USA, 37–48. DOI: https://doi.org/10.1145/1658939.1658944

References

  1. Jain, A.; Mishra, M.; Peddoju, S.K.; Jain, N. Energy efficient computing- Green cloud computing. In Proceedings of the 2013 International Conference on Energy Efficient Technologies for Sustainability (ICEETS), Nagercoil, India, 10–12 April 2013; pp. 978–982. [Google Scholar]
  2. Abdulsalam, S.; Lakomski, D.; Gu, Q.; Jin, T.; Zong, Z. Program energy efficiency: The impact of language, compiler and implementation hoices. In Proceedings of the International Green Computing Conference, Dallas, TX, USA, 3–5 November 2014; pp. 1–6. [Google Scholar] [CrossRef]
  3. Abd-El-Atty, B.; Iliyasu, A.M.; Alaskar, H.; El-Latif, A.A.A. A Robust Quasi-Quantum Walks-based Steganography Protocol for Secure Transmission of Images on Cloud-based E-healthcare Platforms. Sensors 2020, 20, 3108. [Google Scholar] [CrossRef] [PubMed]
  4. Air Transport Action Group Facts and Figures. 2020. Available online: https://www.atag.org/facts-figures.html (accessed on 15 November 2021).
  5. Amazon Web Services. AWS Sustainability. Available online: https://aws.amazon.com/about-aws/sustainability/ (accessed on 15 November 2019).
  6. Andrae, A.S.G.; Edler, T. On global electricity usage of communication technology: Trends to 2030. Challenges 2015, 6, 117–157. [Google Scholar] [CrossRef] [Green Version]
  7. Xu, M.; Toosi, A.N.; Buyya, R. A self-adaptive approach for managing applications and harnessing renewable energy for sustainable cloud computing. IEEE Trans. Sustain. Comput. 2020, 6, 544–558. [Google Scholar] [CrossRef]
  8. Anders, A. Total Consumer Power Consumption Forecast. In Proceedings of the (Huawei) Nordic Digital Business Summit, Helsinki, Finland, 5 October 2017. [Google Scholar]
  9. Kurp, P. Green computing. In Communications of the ACM; Association for Computing Machinery (ACM): New York, NY, USA, 2008; Volume 51, pp. 11–13. [Google Scholar]
  10. Beloglazov, A.; Buyya, R. Energy efficient allocation of virtual machines in cloud data centers. In Proceedings of the 10th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing (CCGrid’10), Melbourne, VIC, Australia, 17–20 May 2010; pp. 577–578. [Google Scholar] [CrossRef]
  11. Gayathri, B. Green cloud computing. In Proceedings of the IET Chennai 3rd International Conference on Sustainable Energy and Intelligent Systems SEISCON 2012, Tiuchengode, India, 27–29 December 2012. [Google Scholar]
  12. Kitchenham, B.; Brereton, O.P.; Budgen, D.; Turner, M.; Bailey, J.; Linkman, S. Systematic literature reviews in software engineering—A systematic literature review. Inf. Softw. Technol. 2009, 51, 7–15. [Google Scholar] [CrossRef]
  13. Kitchenham, B. Procedures for performing systematic reviews. Keele UK Keele Univ. 2004, 33, 1–26. [Google Scholar]
  14. Barroso, L.A.; Clidaras, J.; Hölzle, U. The Datacenter as a Computer: An Introduction to the Design of Warehouse-Scale Machines, Second edition. Synth. Lect. Comput. Archit. 2013, 8, 1–154. [Google Scholar] [CrossRef] [Green Version]
  15. Bawdy, T. Global warming: Data centres to consume three times as much energy in next decade, experts warn. Independent 2016, 23, 276. [Google Scholar]
  16. Belkhir, L.; Elmeligi, A. Assessing ICT global emissions footprint: Trends to 2040 & recommendations. J. Clean. Prod. 2018, 177, 448–463. [Google Scholar]
  17. Beloglazov, A.; Buyya, R.; Lee, Y.C.; Zomaya, A. A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems (Version 2). arXiv 2010, arXiv:1007.0066. [Google Scholar]
  18. Beloglazov, A.; Buyya, R. Optimal online deterministic algorithms and adaptive heuristics for energy and performance efficient dynamic consolidation of virtual machines in Cloud data centers. Concurr. Comput. Pract. Exp. 2011, 24, 1397–1420. [Google Scholar] [CrossRef]
  19. Beloglazov, A.; Abawajy, J.; Buyya, R. Energy—Aware resource allocation heuristics for efficient management of data centers for Cloud computing Futur. Gener. Comput. Syst. 2012, 28, 755–768. [Google Scholar] [CrossRef] [Green Version]
  20. Bharany, S.; Sharma, S.; Badotra, S.; Khalaf, O.I.; Alotaibi, Y.; Alghamdi, S.; Alassery, F. Energy-Efficient Clustering Scheme for Flying Ad-Hoc Networks Using an Optimized LEACH Protocol. Energies 2021, 14, 6016. [Google Scholar] [CrossRef]
  21. Big Power—An Overview of the Electricity Used by the Global Digital Ecosystem. August 2013. Available online: https://www.cepi.org/the-cloud-begins-with-coal-an-overview-of-the-electricity-used-by-the-global-digital-ecosystem (accessed on 7 April 2022).
  22. Rajkumar, B.; Sukhpal, S.G. Sustainable Cloud Computing: Foundations and Future Directions. Bus. Technol. Digit. Transform. Strateg. Cut. Consort. 2018, 21, 1–9. [Google Scholar]
  23. Cisco Global Cloud Index: Forecast and Methodology, 2016–2021. January 2018. Available online: https://www.cisco.com/c/en/us/solutions/collateral/serviceprovider/global-cloud-index-gci/white-paper-c11-738085.html (accessed on 7 April 2022).
  24. Damien, B.; Michael, M.; Georges, D.-C.; Jean-Marc, P.; Ivona, B. Energy-efficient and SLA-aware management of IaaS clouds. In Proceedings of the 3rd International Conference on Future Energy Systems: Where Energy, Computing and Communication Meet, Madrid, Spain, 9–11 May 2012. [Google Scholar]
  25. Das, K.; Das, S.; Darji, R.K.; Mishra, A. Survey of Energy-Efficient Techniques for the Cloud-Integrated Sensor Network. J. Sens. 2018, 2018, 1597089. [Google Scholar] [CrossRef] [Green Version]
  26. Data Center Efficiency Assessment, January 2018 Issue paper Scaling up Energy Efficiency Across the Data Center Industry: Evaluating Key Drivers and Barriers. Available online: https://www.nrdc.org/sites/default/files/data-center-efficiency-assessment-IP.pdf (accessed on 7 April 2022).
  27. Cavdar, D.; Alagoz, F. A survey of research on greening data centers. In Proceedings of the 2012 IEEE Global Communications Conference (GLOBECOM), Anaheim, CA, USA, 3–7 December 2012. [Google Scholar] [CrossRef]
  28. Dybå, T.; Dingsöyr, T.; Hanssen, G.K. Applying systematic reviews to diverse study types: An experience report. In Proceedings of the First International Symposium on Empirical Software Engineering and Measurement, Madrid, Spain, 20–21 September 2007; pp. 225–234. [Google Scholar]
  29. Bohra, A.E.H.; Chaudhary, V. VMeter: Power modelling for virtualized”clouds. In Proceedings of the IEEE International Symposium on Parallel & Distributed Processing, Workshops and Phd Forum (IPDPSW), Atlanta, GA, USA, 19–23 April 2010; pp. 1–8. [Google Scholar]
  30. Easterbrook, S.; Singer, J.; Storey, M.-A.; Damian, D. Selecting empirical methods for software engineering research. In Guide to Advanced Empirical Software Engineering; Springer: London, UK, 2008; pp. 285–311. [Google Scholar]
  31. Kupiainen, E.; Mantyla, M.; Itkonen, J. Using metrics in agile and lean software development–a systematic literature review of industrial studies. Inf. Softw. Technol. 2015, 62, 143–163. [Google Scholar] [CrossRef]
  32. Energy Consumption Comparison of Nano and Centralized Data Centers Fatemeh Jalali, Rob Ayre, Arun Vishwanath, Kerry Hinton, Tansu Alpcan, Rod Tucker. Available online: https://www.cesc.kth.se/polopoly_fs/1.647731.1600689503!/Energy%20Consumption%20Comparison%20of%20Nano%20and%20Centralised%20DCs%20ExtremeGreen%202014.pdf (accessed on 7 April 2022).
  33. Liu, Y.; Sun, X.; Wei, W.; Jing, W. Enhancing Energy-Efficient and QoS Dynamic Virtual Machine Consolidation Method in Cloud Environment. IEEE Access 2018, 6, 31224–31235. [Google Scholar] [CrossRef]
  34. Chen, F.F.; Schneider, J.; Yang, L.Y.; Grundy, J.; He, Q. An energy consumption model and analysis tool for Cloud computing environments. In Proceedings of the 2012 First International Workshop on Green and Sustainable Software (GREENS), Zurich, Switzerland, 3 June 2012; pp. 45–50. [Google Scholar]
  35. Farahnakian, F.; Ashraf, A.; Pahikkala, T.; Liljeberg, P.; Plosila, J.; Porres, I.; Tenhunen, H. Using Ant Colony System to Consolidate VMs for Green Cloud Computing. IEEE Trans. Serv. Comput. 2015, 8, 187–198. [Google Scholar] [CrossRef]
  36. Gill, S.S.; Buyya, R. A taxonomy and future directions for sustainable cloud computing: 360 degree view. ACM Comput. Surv. 2018, 51, 1–13. [Google Scholar] [CrossRef]
  37. Lombardi, F.; di Pietro, R. Secure virtualization for cloud computing. J. Netw. Comput. Appl. 2021, 34, 1113–1122. [Google Scholar] [CrossRef]
  38. Google Data Center Efficiency Best Practices, January 2017. Available online: http://www.cs.uu.nl/docs/vakken/ebu/Laudon-Traver_Ecommerce12_Case3.2_GoogleDataCenter.pdf (accessed on 7 April 2022).
  39. Abdi, H. Multiple Correlation Coefficien; Sage: Thousand Oaks, CA, USA, 2007; pp. 648–651. [Google Scholar]
  40. Akeiber, H.; Nejat, P.; Majid, M.Z.A.; Wahid, M.A.; Jomehzadeh, F.; Famileh, I.Z.; Calautit, J.K.; Hughes, B.R.; Zaki, S.A. A review on phase change material (PCM) for sustainable passive cooling in building envelopes. Renew. Sustain. Energy Rev. 2016, 60, 1470–1497. [Google Scholar] [CrossRef]
  41. Hamilton, J. Cooperative Expendable Micro-Slice Servers (CEMS): Low Cost, Low Power Servers for Internet-Scale Services. In Proceedings of the 4th Biennial Conference Innovative Data Systems Research (CIDR), Asilomar, CA, USA, 4–7 January 2009; Available online: http://slashdot.org/topic/datacenter/new-york-timestakes-aim-at-datacenters (accessed on 7 April 2022).
  42. Ibrahim, A.S.; Hamlyn-Harris, J.; Grundy, J. Emerging security challenges of cloud virtual infrastructure. In Computer Science & Software Engineering, Faculty of Information & Communication Technologies Swinburne University of Technology; Hawthorn: Victoria, Australia, 2016. [Google Scholar]
  43. Liu, L.; Wang, H.; Liu, X.; Jin, X.; He, W.B.; Wang, Q.B.; Chen, Y. GreenCloud: A new architecture for green data center. In Proceedings of the 6th International Conference Industry Session on Autonomic Computing and Communications Industry Session, ICA-INDST ‘09, Barcelona, Spain, 15 June 2009; pp. 29–38. [Google Scholar] [CrossRef]
  44. Intel: Improving Data Center Efficiency. January 2016. Available online: https://www.intel.co.jp/content/dam/doc/technology-brief/efficient-datacenter-highambient-temperature-operation-brief.pdf (accessed on 7 April 2022).
  45. Smith, J.W.; Khajeh-Hosseini, A.; Ward, J.S.; Sommeville, I. CloudMonitor: Profiling Powe” Usage. In Proceedings of the IEEE Fifth International Conference on Cloud Computing, Hawaii, HI, USA; 2012; pp. 947–948. [Google Scholar]
  46. Jones, N. How to stop data centres from gobbling up the world’s electricity. Nature 2018, 561, 163–166. [Google Scholar] [CrossRef] [PubMed]
  47. Toress, J. Green Computing: The next wave in computing. In Green Computing, Communication of the ACM51.10; Andy, H., Ed.; UPC Technical University of Catalonia: Barcelona, Spain, 2018; pp. 11–13. [Google Scholar]
  48. Chen, K.; Xu, Y.; Xi, K.; Chao, H.J. Intelligent Virtual Machine Placement for Cost Efficiency in Geo- Distributed Cloud Systems. In Proceedings of the 2013 IEEE International Conference on Communications (ICC), Budapest, Hungary, 9–13 June 2013; pp. 3498–3503. [Google Scholar]
  49. Cameron, K.W. Energy Oddities, Computing Is Odd. Computer 2013, 46, 90–93. [Google Scholar] [CrossRef]
  50. Kamiya, G. Factcheck: What Is the Carbon Footprint of Streaming Video on Netflix? Carbon Brief, 25 February 2020. Available online: https://www.carbonbrief.org/factcheck-what-is-the-carbon-footprint-of-streaming-video-on-netflix? (accessed on 7 April 2022).
  51. Katal, A.; Dahiya, S.; Choudhury, T. Energy efficiency in cloud computing data center: A survey on hardware technologies. Clust. Comput. 2021, 25, 675–705. [Google Scholar] [CrossRef]
  52. Kaur, M.; Sinh, P. Energy efficient Green Cloud: Underlying structure. In Proceedings of the 2013 International Conference on Energy Efficient Technologies for Sustainability (ICEETS), Nagercoil, India, 10–12 April 2013; pp. 207–212. [Google Scholar]
  53. Talwar, B.; Arora, A.; Bharany, A. Proactive Detection of Deteriorating Node Based Migration for Energy-Aware Fault Tolerance. Think India J. 2019, 22, 2256. [Google Scholar]
  54. Kim, K.; Yang, F.; Zavala, V.M.; Chien, A.A. Data Centers as Dispatchable Loads to Harness Stranded Power. IEEE Trans. Sustain. Energy 2017, 8, 208–218. [Google Scholar] [CrossRef]
  55. Kitchenham, B.; Charters, S. Guidelines for Performing Systematic Literature Eviewsins of Wareengineering; Version 2.3; UKEBSE Technical Report; Keele University: Newcastle-under-Lyme, UK, 2007. [Google Scholar]
  56. Kliazovich, D.; Bouvry, P.; Khan, S.U. Green Cloud: A packet level simulator of energy-aware cloud computing data centers. In Proceedings of the Global Telecommunications Conference (GLOBECOM 2010), Miami, FL, USA, 6–10 December 2010; pp. 1–5. [Google Scholar]
  57. Koomey, J. Growth in Data Center Electricity Use 2005 to 2010; Analytics Press: Oakland, CA, USA, 2011. [Google Scholar]
  58. Arora, A.; Talwar, B.; Bharany, S. Reliability Aware Mechanism to ensure increased Fault Tolerance using Throttle Load Balancer. In Proceedings of the 2021 9th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 3–4 September 2021. [Google Scholar] [CrossRef]
  59. Kotas, C.; Naughton, T.; Imam, N. A comparison of Amazon Web Services and Microsoft Azure cloud platforms for high performance computing. In Proceedings of the IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 12–14 January 2018; pp. 1–4. [Google Scholar]
  60. Kozakiewicz, A.; Lis, A. Energy Efficiency in Cloud Computing: Exploring the Intellectual Structure of the Research Field and Its Research Fronts with Direct Citation Analysis. Energies 2021, 14, 7036. [Google Scholar] [CrossRef]
  61. Kumar Yadav, A.; Garg, M.L.; Ritika, D. An Efficient Approach for Security in Cloud Computing. Int. J. Adv. Stuetflixic Res. 2018, 3, 8. Available online: https://ssrn.com/abstract=3315609 (accessed on 14 January 2019).
  62. Kurp, P. Green computing. Commun. ACM 2008, 51, 11–13. [Google Scholar] [CrossRef]
  63. Kim, K.H.; Beloglazov, A.; Buyya, R. Power-aware provisioning of virtual machines for real-time Cloud services. Concurr. Comput. Pract. Exp. 2011, 23, 1491–1505. [Google Scholar] [CrossRef] [Green Version]
  64. Uchechukwu, K.L.A.; Shen, Y. Energy consumption in cloud computing datanters. Int. J. Cloud Comput. Services Sci. 2014, 3, 3148. [Google Scholar]
  65. Landis, J.R.; Koch, G.G. Measurement of observer agreement for categorical data. Biometrics 1977, 33, 159–174. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Hongyou, L.; Jiangyong, W.; Jian, P.; Junfeng, W.; Tang, L. Energy-aware scheduling scheme using workload-aware consolidation technique in cloud data centres. Commun. China 2013, 10, 114–124. [Google Scholar] [CrossRef]
  67. Li, D.; Shang, Y.; He, W.; Chen, C. EXR: Greening Data Center Network with Software Defined Exclusive Routing. IEEE Trans. Comput. 2015, 64, 2534–2544. [Google Scholar] [CrossRef]
  68. De Vuyst, M.; Kumar, R.; Tullsen, D.M. Exploiting unbalanced thread scheduling for energy and performance on a CMP of SMT processors. In Proceedings of the 20th IEEE International Parallel & Distributed Processing Symposium, Rhodes, Island, 25–29 April 2006. [Google Scholar] [CrossRef] [Green Version]
  69. Mills, M.P. The Cloud Begins With Coal. Big Data, Big Networks, Big Infrastructure. 2013. Available online: https://www.tech-pundit.com/wp-content/uploads/2013/07/Cloud_Begins_With_Coal.pdf (accessed on 7 April 2022).
  70. Markandey, A.; Dhamdhere, P.; Gajmal, Y. Data Access Security in Cloud Computing: A Review. In Proceedings of the 2018 International Conference on Computing, Power and Communication Technologies (GUCON), Greater Noida, Uttar Pradesh, India, 28–29 September 2018; pp. 633–636. [Google Scholar]
  71. Masanet, E.; Shehabi, A.; Lei, N.; Smith, S.; Koomey, J. Recalibrating global data center energy-use estimates. Science 2020, 367, 984–986. [Google Scholar] [CrossRef] [PubMed]
  72. Malone, C.; Belady, C. Metrics to Characterize Data Center & IT Equipment Energy Use. In Proceedings of the 2006 Digital Power Forum, Richardson, TX, USA, 21 September 2006; Available online: https://www.researchgate.net/publication/337801067_Metrics_to_Characterize_Data_Center_IT_Equipment_Energy_Use_Proceedings_of_Digital_Power_Forum_Richardson_TX (accessed on 7 April 2022).
  73. Meisel, M.; Pappas, V.; Zhang, L. A taxonomy of biologically inspired researchin computer networking. Comput. Netw. 2020, 54, 901–916. [Google Scholar] [CrossRef] [Green Version]
  74. Mishra, S.K.; Sahoo, S.; Sahoo, B.; Jena, S.K. Energy-Efficient Service Allocation Techniques in Cloud: A Survey. IETE Tech. Rev. 2019, 37, 339–352. [Google Scholar] [CrossRef]
  75. Mishra, S.K.; Sahoo, B.; Parida, P.P. Load balancing in cloud computing: A big picture. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 149–158. [Google Scholar] [CrossRef]
  76. Bharany, S.; Sharma, S.; Bhatia, S.; Rahmani, M.K.I.; Shuaib, M.; Lashari, S.A. Energy Efficient Clustering Protocol for FANETS Using Moth Flame Optimization. Sustainability 2022, 14, 6159. [Google Scholar] [CrossRef]
  77. Shirvani, M.H.; Rahmani, A.M.; Sahafi, A. A survey study on virtual machine migration and server consolidation techniques in DVFS-enabled cloud datacenter: Taxonomy and challenges. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 267–286. [Google Scholar]
  78. Kocaoglu, M.; Malak, D.; Akan, O.B. Fundamentals of green communications and computing: Modeling and simulation. Computer 2012, 45, 40–44. [Google Scholar] [CrossRef]
  79. Murtazaev, A.; Oh, S. Sercon: Server consolidation algorithm using live migration of virtual machines for green computing. IETE Tech. Rev. 2011, 28, 212–231. [Google Scholar] [CrossRef]
  80. Sabry, N.; Krause, P.A. A digital ecosystem view on cloud computing. In Proceedings of the 6th IEEE International Conference on Digital Ecosystems Technologies (DEST), Campione d’Italia, Italy, 18–20 June 2012; pp. 1–6. [Google Scholar]
  81. Hunt, N. Renewable Energy at Netflix: An Update. June 2017. Available online: https://media.netflix.com/en/company-blog/renewable-energy-at-netflix-an-update (accessed on 7 April 2022).
  82. Naseer Qureshi, K.; Bashir, F.; Iqbal, S. Cloud Computing Model for Vehicular Ad hoc Networks. In Proceedings of the 2018 IEEE 7thInternational Conference on Cloud Networking (CloudNet), Tokyo, Japan, 22–24 October 2018; pp. 1–3. [Google Scholar]
  83. News by EECS. Featured Research Article. Green Computing: Higher Energy Efficiency from Silicon to the Cloud. 2011. Available online: http://www.eecs.umich.edu/eecs/about/articles (accessed on 16 December 2012).
  84. Navimipour, N.J.; Charband, Y. Knowledge sharing mechanisms and techniques in project teams: Literature review, classification, and current trends. Comput. Hum. Behav. 2016, 62, 730–742. [Google Scholar] [CrossRef]
  85. Ovtcharov, K.; Ruwase, O.; Kim, J.Y.; Fowers, J.; Strauss, K.; Chung, E.S. Accelerating deep convolutional neural networks using specialized hardware. Microsoft Res. Whitepaper 2015, 2, 1–4. [Google Scholar]
  86. Pearce, F. Energy Hogs: Can World’s Huge Data Centers Be Made More Efficient? Yale Environment 360, 3 April 2018. Available online: https://e360.yale.edu/features/energy-hogs-can-huge-data-centers-be-made-more-efficient (accessed on 31 January 2021).
  87. Petersen, K.; Ali, N.B. Identifying strategies for study selection in systematic reviews and maps. In Proceedings of the International Symposium on Empirical Software Engzewskiineering and Measurement (ESEM), Banff, AB, Canada, 22–23 September 2011; pp. 351–354. [Google Scholar]
  88. Preist, C.; Shabajee, P. Energy use in the media cloud: Behaviour change, or technofix. In Proceedings of the 2010 IEEE Cloud Computing Technology and Science (CloudCom), Indianapolis, IN, USA, 30 November–3 December 2010; pp. 581–586. [Google Scholar]
  89. Press, Completed at the Request of The New York Times; 2011. Available online: https://www.nytimes.com/2011/08/01/technology/data-centers-using-less-power-than-forecast-report-says.html (accessed on 25 February 2022).
  90. Dinakarrao, S.M.P. Self-aware power management for multi-core microprocessors. Sustain. Comput. Inform. Syst. 2021, 29, 1480. [Google Scholar] [CrossRef]
  91. Puhan, S.; Panda, D.; Mishra, B.K. Energy Efficiency for Cloud Computing Applications: A Survey on the Recent Trends and Future Scopes. In Proceedings of the 2020 International Conference on Computer Science, Engineering and Applications (ICCSEA), Gunupur, India, 13–14 March 2020. [Google Scholar] [CrossRef]
  92. Quang-Hung, N.; Nien, P.D.; Nam, N.H.; Huynh Tuong, N.; Thoai, N. A Genetic Algorithm for Power-Aware Virtual Machine Allocation in Private Cloud. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2013; pp. 183–191. [Google Scholar] [CrossRef] [Green Version]
  93. Wang, S.; Psaraftis, H.N.; Qi, J. Paradox of international maritime organization’s carbon intensity indicator. Commun. Transp. Res. 2021, 1, 100005. [Google Scholar] [CrossRef]
  94. Buyya, R.; Beloglazov, A.; Abawajy, J. Energy-Efficient Management of Data Center Resources for Cloud Computing: A Vision, Architectural Elements, and Open Challenges (Version 1). arXiv 2010, arXiv:1006.0308. [Google Scholar]
  95. Yanggratoke, R.; Wuhib, F.; Sadler, R. Gossip-based resource allocation for green computing in large clouds. In Proceedings of the 7th International Conference on Network and Service Management, CNSM, Paris, France, 24–28 October 2011. [Google Scholar]
  96. Reagen, B.; Whatmough, P.; Adolf, R.; Rama, S.; Lee, H.; Lee, S.K.; Hernández-Lobato, J.M.; Wei, G.Y.; Brooks, D. Minerva: Enabling low-power, highly accurate deep neural network accelerators. In Proceedings of the 43rd International Symposium on Computer Architecture, Seoul, Korea, 18–22 June 2016. pp. 267–278.
  97. Reddy, S.P.; Chandan, H.K.S. Energy aware scheduling of real-time and non-real-time tasks on cloud processors (Green Cloud Computing). In Proceedings of the International Conference on Information Communication and Embedded Systems (ICICES2014), Chennai, India, 27–28 February 2014. [Google Scholar] [CrossRef]
  98. Garg, S.K.; Buyya, R. Green Cloud Computing and Environmental Sustainability, 1st ed.; Murugesan, S., Gangadharan, G.R., Eds.; Wiley-IEEE Press Ebook. 2012, pp. 76–87. Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.221.8835&rep=rep1&type=pdf (accessed on 7 April 2022).
  99. Shehabi, A.; Smith, S.; Sartor, D.; Brown, R.; Herrlin, M.; Koomey, J.; Masanet, E.; Horner, N.; Azevedo, I.; Lintner, W. United States Data Center Energy Usage Report; No. LBNL-1005775; Lawrence Berkeley National Lab (LBNL): Berkeley, CA, USA, 2016. [Google Scholar]
  100. Jing, S.; Ali, S.; She, K.; Zhong, Y. State-of-the-art research study for green cloud computing. J. Supercomput. 2013, 65, 445–468. [Google Scholar] [CrossRef]
  101. Staples, M.; Niazi, M. Experiences using systematic review guidelines. J. Syst. Softw. 2007, 80, 1425–1437. [Google Scholar] [CrossRef]
  102. Khare, S.; Jain, S. Prospects of Near-Threshold Voltage Design for Green Computing. In Proceedings of the 2013 26th International Conference on VLSI Design: Concurrently with the 12th International Conference on Embedded Systems, Pune, India, 5–10 January 2013. [Google Scholar] [CrossRef]
  103. Ghose, T.; Namboodiri, V.; Pendse, R. Thin is green: Leveraging the thin-client paradigm for sustainable mobile computing q. Comput. Electr. Eng. 2016, 45, 155–168. [Google Scholar] [CrossRef]
  104. Yapicioglu, T.; Oktug, S. A traffic-aware virtual machine placement method for cloud data centers. In Proceedings of the Utility and Cloud Computing (UCC) 2013 IEEE/ACM 6th International Conference on Utility and Cloud Computing, Dresden, Germany, 9–12 December 2013; pp. 299–301. [Google Scholar]
  105. Conte, T.; DeBenedictis, E.; Ganesh, N.; Hylton, T.; Still, S.; Strachan, J.W.; Williams, S.; Alemi, A.; Altenberg, L.; Crooks, G.; et al. Thermodynamic Computing (Version 2). arXiv 2019, arXiv:1911.01968. [Google Scholar]
  106. Valliyammai, C.; Uma, S.; Dhivya, B.K.; Sura, P. Efficient energy consumption in green cloud. In Proceedings of the International Conference on Recent Trends in Information Technology, Chennai, India, 10–12 April 2014; pp. 1–4. [Google Scholar]
  107. Vasudevan, V.; Andersen, D.; Kaminsky, M.; Tan, L.; Franklin, J.; Moraru, I. Energy-efficient cluster computing with FAWN: Workloads and implications. In Proceedings of the 1st International Conference on Energy-Efficient Computing and Networking, New York, NY, USA, 13–15 April 2010. [Google Scholar]
  108. Talwar, B.; Arora, A.; Bharany, S. An Energy Efficient Agent Aware Proactive Fault Tolerance for Preventing Deterioration of Virtual Machines within Cloud Environment. In Proceedings of the 2021 9th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 3–4 September 2021. [Google Scholar] [CrossRef]
  109. Venkatachalam, V.; Franz, M. Power reduction techniques for microprocessor systems. ACM Comput. Surv. 2005, 37, 195–237. [Google Scholar] [CrossRef]
  110. Motochi, V.; Barasa, S.; Owoche, P.; Wabwoba, F. The Role of Virtualization towards GreenComputing and Environmental Sustainability. Int. J. Adv. Res. Comput. Eng. Technol. (IJARCET) 2017, 6, 6. [Google Scholar]
  111. Wang, G.; Ng, T.S.E. The Impact of Virtualization on Network Performance of Amazon EC2 Data Center. In Proceedings of the 2010 Proceedings IEEE INFOCOM, San Diego, CA, USA, 14–19 March 2010. [Google Scholar]
  112. Wohlin, C.; Runeson, P.; Höst, M.; Ohlsson, M.C.; Regnell, B.; Wesslén, A. Experimentation in Software Engineering, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  113. Deng, X.; Wu, D.; Shen, J.; He, J. Eco-Aware Online Power Management and Load Scheduling for Green Cloud Data centers. IEEE Syst. J. 2014, 10, 78–87. [Google Scholar] [CrossRef]
  114. Fan, X.; Weber, W.-D.; Brroso, L.A. Power provisioning form a warehouse-sized computer. In Proceedings of the 34th Annual International Symposium on Computer Architecture (ISCA 2007), New York, NY, USA, 9–13 June 2007; pp. 13–23. [Google Scholar]
  115. Chen, Y.; Lin, C.; Huang, J.; Xiang, X.; Shen, X. Energy Efficient Scheduling and Management for Large-Scale Services Computing Systems. IEEE Trans. Serv. Comput. 2017, 10, 217–230. [Google Scholar] [CrossRef]
  116. Fernandas, Y.E.; Vasanthi, M.S. Energy Efficient Mechanism for Green Computing in Wireless Storage Area Networks. In Proceedings of the 2015 International Conference on Communications and Signal Processing (ICCSP), Melmaruvathur, India, 2–4 April 2015; pp. 1311–1314. [Google Scholar]
  117. Hua, Y.; Feng, D. Needle in a haystack: Cost-Effective data analytics for real-time cloud sharing. In Proceedings of the 2014 IEEE 22nd International Symposium of Quality of Service (IWQoS), Hong Kong, China, 26–27 May 2014; pp. 159–167. [Google Scholar]
  118. Yadav, A.K.; Garg & Ritika, M.L. Docker Containers vs. Virtual machine-based Virtualization. In Emerging Technologies in Data Mining and Information Security; Springer: Berlin/Heidelberg, Germany, 2019; pp. 141–150. [Google Scholar]
  119. Charband, Y.; Navimipour, N.J. Online knowledge sharing mechanisms: A systematic review of the state-of-the-art literature and recommendations for future research. Inf. Syst. Front. 2016, 18, 1131–1151. [Google Scholar] [CrossRef]
  120. Yi, S.; Yuhe, L.; Yu, W. Cloud Computing Architecture Design of Database Resource Pool Based on Cloud Computing. In Proceedings of the 2018 International Conference on Information Systems and Computer Aided Education (ICISCAE), Changchun, China, 6–8 July 2018; pp. 180–183. [Google Scholar]
  121. Zhang, W.-Z.; Elgendy, I.A.; Hammad, M.; Iliyasu, A.M.; Du, X.; Guizani, M.; El-Latif, A.A.A. Secure and Optimized Load Balancing for Multitier IoT and Edge-Cloud Computing Systems. IEEE Internet Things J. 2021, 8, 8119–8132. [Google Scholar] [CrossRef]
  122. Kaur, T.; Chana, I. Energy Efficiency Techniques in Cloud Computing. ACM Comput. Surv. 2015, 48, 1–46. [Google Scholar] [CrossRef]
  123. Cohen, J. A Coefficient of Agreement for Nominal Scales. Educ. Psychol. Meas. 1960, 20, 37–46. [Google Scholar] [CrossRef]
  124. Ramasamy, V.; Alotaibi, Y.; Khalaf, O.I.; Samui, P.; Jayabalan, J. Prediction of groundwater table for Chennai Region using soft computing techniques. Arab. J. Geosci. 2022, 15, 827. [Google Scholar] [CrossRef]
  125. Khalaf, O.I.; Ogudo, K.A.; Sangeetha, S.K.B. Saneetha, Design of Graph-Based Layered Learning-Driven Model for Anomaly Detection in Distributed Cloud IoT Network. Mob. Inf. Syst. 2022, 2022, 6750757. [Google Scholar] [CrossRef]
  126. Banumathy, D.; Khalaf, O.I.; Romero, C.A.T.; Indra, J.; Harma, D.K. Cad of bcd from thermal mammogram images using machine learning. Intell. Autom. Soft Comput. 2022, 34, 667–685. [Google Scholar] [CrossRef]
  127. Sharma, B.; Hashmi, A.; Gupta, C.; Khalaf, O.I.; Abdulsahib, G.M.; Itani, M.M. Hybrid Sparrow Clustered (HSC) Algorithm for Top-N Recommendation System. Symmetry 2022, 14, 793. [Google Scholar] [CrossRef]
  128. Anuradha, D.; Subramani, N.; Khalaf, O.I.; Alotaibi, Y.; Alghamdi, S.; Rajagopal, M. Chaotic Search-and-Rescue-Optimization-Based Multi-Hop Data Transmission Protocol for Underwater Wireless Sensor Networks. Sensors 2022, 22, 2867. [Google Scholar] [CrossRef] [PubMed]
  129. Janniekode, U.M.; Somineni, R.P.; Khalaf, O.I.; Itani, M.M.; Chinna Babu, J.; Abdulsahib, G.M. A Symmetric Novel 8T3R Non-Volatile SRAM Cell for Embedded Applications. Symmetry 2022, 14, 768. [Google Scholar] [CrossRef]
  130. Edeh, M.O.; Khalaf, O.I.; Tavera, C.A.; Tayeb, S.; Ghouali, S.; Abdulsahib, G.M.; Richard-Nnabu, N.E.; Louni, A. A Classification Algorithm-Based Hybrid Diabetes Prediction Model. Front. Public Health 2022, 10, 829519. [Google Scholar] [CrossRef]
  131. Sundas, A.; Badotra, S.; Alotaibi, Y.; Alghamdi, S.; Khalaf, O.I. Modified Bat Algorithm for Op’imal VM’s in Cloud Computing. CMC-Comput. Mater. Contin. 2022, 72, 2877–2894. [Google Scholar] [CrossRef]
  132. Wang, X.; Liu, J.; Liu, X.; Liu, Z.; Khalaf, O.I.; Ji, J.; Ouyang, Q. Ship feature recognition methods for deep learning in complex marine environments. Complex Intell. Syst. 2022, 1–17. [Google Scholar] [CrossRef]
  133. Jayapradha, J.; Prakash, M.; Alotaibi, Y.; Khalaf, O.I.; Alghamdi, S.A. Heap Bucketization Anonymity—An Efficient Privacy-Preserving Data Publishing Model for Multiple Sensitive Attributes. IEEE Access 2022, 10, 28773–28791. [Google Scholar] [CrossRef]
  134. Hemavathi; Akhila, S.R.; Alotaibi, Y.; Khalaf, O.I.; Alghamdi, S. Authentication and Resource Allocation Strategies during Handoff for 5G IoVs Using Deep Learning. Energies 2022, 15, 2006. [Google Scholar] [CrossRef]
  135. Kavitha, C.; Mani, V.; Srividhya, S.R.; Khalaf, O.I.; Tavera Romero, C.A. Early-Stage Alzheimer’s Disease Prediction Using Machine Learning Models. Front. Public Health 2022, 10, 853294. [Google Scholar] [CrossRef]
  136. Puri, T.; Soni, M.; Dhiman, G.; Khalaf, O.I.; Alazzam, M.; Khan, I.R. Detection of Emotion of Speech for RAVDESS Audio Using Hybrid Convolution Neural Network. J. Healthc. Eng. 2022, 2022, 8472947. [Google Scholar] [CrossRef]
  137. Mohan, P.; Subramani, N.; Alotaibi, Y.; Alghamdi, S.; Khalaf, O.I.; Ulaganathan, S. Improved Metaheuristics-Based Clustering with Multihop Routing Protocol for Underwater Wireless Sensor Networks. Sensors 2022, 22, 1618. [Google Scholar] [CrossRef]
  138. Rawat, S.S.; Alghamdi, S.; Kumar, G.; Alotaibi, Y.; Khalaf, O.I.; Verma, L.P. Infrared Small Target Detection Based on Partial Sum Minimization and Total Variation. Mathematics 2022, 10, 671. [Google Scholar] [CrossRef]
  139. Khalaf, O.I.; Romero, C.A.T.; Hassan, S.; Iqbal, M.T. Mitigating Hotspot Issues in Heterogeneous Wireless Sensor Networks. J. Sensors 2022, 2022, 7909472. [Google Scholar] [CrossRef]
  140. Khaparde, A.R.; Alassery, F.; Kumar, A.; Alotaibi, Y.; Khalaf, O.I.; Pillai, S.; Alghamdi, S. Differential Evolution Algorithm with Hierarchical Fair Competition Model. Intell. Autom. Soft Comput. 2022, 33, 1045–1062. [Google Scholar] [CrossRef]
  141. Khan, N.A.; Khalaf, O.I.; Romero, C.A.T.; Sulaiman, M.; Bakar, M.A. Application of Intelligent Paradigm through Neural Networks for Numerical Solution of Multiorder Fractional Differential Equations. Comput. Intell. Neurosci. 2022, 2022, 2710576. [Google Scholar] [CrossRef]
  142. Gill, H.S.; Khalaf, O.I.; Alotaibi, Y.; Alghamdi, S.; Alassery, F. Fruit Image Classification Using Deep Learning. Comput. Mater. Contin. 2022, 71, 5135–5150. [Google Scholar] [CrossRef]
  143. Revathy, A.; Boopathi, C.S.; Khalaf, O.I.; Romero, C.A.T. Investigation of AlGaN Channel HEMTs on β-Ga2O3 Substrate for High-Power Electronics. Electronics 2022, 11, 225. [Google Scholar] [CrossRef]
  144. El-Saleh, A.A.; Alhammadi, A.; Shayea, I.; Alsharif, N.; Alzahrani, N.M.; Khalaf, O.I.; Aldhyani, T.H.H. Measuring and Assessing Performance of Mobile Broadband Networks and Future 5G Trends. Sustainability 2022, 14, 829. [Google Scholar] [CrossRef]
  145. Subramani, N.; Mohan, P.; Alotaibi, Y.; Alghamdi, S.; Khalaf, O.I. An Efficient Metaheuristic-Based Clustering with Routing Protocol for Underwater Wireless Sensor Networks. Sensors 2022, 22, 415. [Google Scholar] [CrossRef]
  146. Qi, J.; Wang, S.; Psaraftis, H. Bi-level optimization model applications in managing air emissions from ships: A review. Commun. Transp. Res. 2021, 1, 100020. [Google Scholar] [CrossRef]
Figure 1. Four conditions by which a system can drop or lose energy.
Figure 1. Four conditions by which a system can drop or lose energy.
Sustainability 14 06256 g001
Figure 2. Evolution of green cloud computing.
Figure 2. Evolution of green cloud computing.
Sustainability 14 06256 g002
Figure 3. Survey methodology.
Figure 3. Survey methodology.
Sustainability 14 06256 g003
Figure 4. Sources of the research papers used in this review.
Figure 4. Sources of the research papers used in this review.
Sustainability 14 06256 g004
Figure 5. Search keywords.
Figure 5. Search keywords.
Sustainability 14 06256 g005
Figure 6. Selection methodology used in this review paper.
Figure 6. Selection methodology used in this review paper.
Sustainability 14 06256 g006
Figure 7. Article review process.
Figure 7. Article review process.
Sustainability 14 06256 g007
Figure 8. Distribution of PSs over publication years.
Figure 8. Distribution of PSs over publication years.
Sustainability 14 06256 g008
Figure 9. Historical energy usage and estimated energy usage for different categories. Source: [70].
Figure 9. Historical energy usage and estimated energy usage for different categories. Source: [70].
Sustainability 14 06256 g009
Figure 10. Percentage of electricity used in 2014 by US data centers.
Figure 10. Percentage of electricity used in 2014 by US data centers.
Sustainability 14 06256 g010
Figure 11. Types of techniques for energy efficiency.
Figure 11. Types of techniques for energy efficiency.
Sustainability 14 06256 g011
Figure 12. Taxonomy of hardware-level energy-efficiency technologies.
Figure 12. Taxonomy of hardware-level energy-efficiency technologies.
Sustainability 14 06256 g012
Figure 13. Distribution of research papers among different types of hardware techniques.
Figure 13. Distribution of research papers among different types of hardware techniques.
Sustainability 14 06256 g013
Figure 14. Taxonomy of software-level energy-efficiency techniques.
Figure 14. Taxonomy of software-level energy-efficiency techniques.
Sustainability 14 06256 g014
Figure 15. Taxonomy of multiple server-level energy-efficiency techniques (continued software level).
Figure 15. Taxonomy of multiple server-level energy-efficiency techniques (continued software level).
Sustainability 14 06256 g015
Figure 16. Distribution of research papers among different types of software techniques.
Figure 16. Distribution of research papers among different types of software techniques.
Sustainability 14 06256 g016
Figure 17. Virtualization (multiple virtual server on one single physical server).
Figure 17. Virtualization (multiple virtual server on one single physical server).
Sustainability 14 06256 g017
Figure 18. Taxonomy of VM consolidation energy-efficiency techniques.
Figure 18. Taxonomy of VM consolidation energy-efficiency techniques.
Sustainability 14 06256 g018
Figure 19. Distribution of research papers among different types of consolidation techniques.
Figure 19. Distribution of research papers among different types of consolidation techniques.
Sustainability 14 06256 g019
Figure 20. Taxonomy of bio-inspired energy-efficiency techniques.
Figure 20. Taxonomy of bio-inspired energy-efficiency techniques.
Sustainability 14 06256 g020
Figure 21. Distribution of research papers among different types of bio-inspired technology.
Figure 21. Distribution of research papers among different types of bio-inspired technology.
Sustainability 14 06256 g021
Figure 22. Taxonomy of thermal-management techniques.
Figure 22. Taxonomy of thermal-management techniques.
Sustainability 14 06256 g022
Figure 23. Distribution of research papers among different types of thermal techniques.
Figure 23. Distribution of research papers among different types of thermal techniques.
Sustainability 14 06256 g023
Figure 24. Classification power-management techniques.
Figure 24. Classification power-management techniques.
Sustainability 14 06256 g024
Figure 25. Taxonomy of power-management techniques.
Figure 25. Taxonomy of power-management techniques.
Sustainability 14 06256 g025
Figure 26. Distribution of research papers among different types of techniques.
Figure 26. Distribution of research papers among different types of techniques.
Sustainability 14 06256 g026
Figure 27. Energy consumption of a data center.
Figure 27. Energy consumption of a data center.
Sustainability 14 06256 g027
Figure 28. Various techniques proposed by year.
Figure 28. Various techniques proposed by year.
Sustainability 14 06256 g028
Table 1. Comparison with other cited surveys.
Table 1. Comparison with other cited surveys.
Survey
Paper
Hardware
Techniques
Software
Techniques
VM
Consolidation
Power Aware TechniquesBio-Inspired
Techniques
Thermal and
Cooling
Techniques
Non-
Technical
Technologies
Issues and
Improvisions
Threats to
Validity
Visual
Representation of
Results
[18]ΧΧΧΧΧΧΧΧ
[100]ΧΧΧΧΧΧΧΧ
[122]ΧΧΧ
[25]ΧΧΧΧΧΧΧΧ
[74]ΧΧΧΧΧΧΧΧ
[51]ΧΧΧΧΧΧΧΧ
[60]ΧΧΧΧΧΧΧΧ
[146]ΧΧΧΧΧΧΧΧΧ
[36]ΧΧΧΧΧΧ
[93]ΧΧΧΧΧΧΧΧ
[7]ΧΧΧΧΧΧΧΧ
Our
Survey
Table 2. Research questions.
Table 2. Research questions.
Research Questions (RQs)
RQ1—Which measuring parameters are considered for power consumption in our selected studies?
RQ2—Describe the impact of high energy consumption by cloud data centers. Explain the estimation on total energy consumption by data centers.
RQ3—What is the current status of cloud energy efficiency?
RQ4—What kind of different techniques used for energy efficiency were proposed by our selected studies?
RQ5—Describe various energy-efficiency techniques employed at the hardware level.
RQ6—Describe the various energy-efficiency techniques employed at the software level.
RQ7—Explain the various techniques for virtual machine consolidation applied at data centers.
RQ8—Describe the different power-aware management techniques.
RQ9—Explain the various bio-inspired techniques for energy efficiency applied at data centers.
RQ10—Describe the different thermal-management and cooling technologies employed at data centers.
RQ11—Explain the key aspects that make the cloud favorable for reducing carbon footprints and the better utilization of energy. Describe non-technical technologies employed at cloud data centers for enhancing energy efficiency.
Table 3. Mapping between the PICOC criteria and SLR RQs.
Table 3. Mapping between the PICOC criteria and SLR RQs.
PICOC CriterionAddressed Research ProblemsResearch Questions
PopulationCloud energy consumption, carbon emissionRQ1, RQ3, RQ4
InterventionEnergy-efficiency techniques considered by the PSsRQ5, RQ6, RQ7, RQ8, RQ9, R10
ComparisonNA
OutcomesVarious energy-related issuesRQ11
ContextAcademic and industryRQ2
Table 4. Search strings determined for digital libraries.
Table 4. Search strings determined for digital libraries.
Digital Library Search String
ScienceDirect((cloud OR power OR computation OR energy OR cloud computing OR power-consumption OR emission OR function OR {power-consumption} OR {energy-consumption}) AND (energy ∗ OR {energy-efficiency} OR {efficiency-techniques} OR {power-optimization} OR {Environment} OR {Data-center} OR {applications}))
IEEE Xplore(cloud OR processors OR power OR saving OR methods OR techniques OR protocols OR scheduling OR “power-consumption” OR “energy efficiency”) AND (energy ∗ OR re-saving∗ Or optimization)
SpringerLink((cloud OR processors OR code OR saving OR method OR techniques OR scheduling OR emission OR “energy-efficiency” OR “power-efficiency”) AND (energy-efficient techniques ∗ OR “power-optimization” OR “energy-mitigation” OR “energy reduction” OR “energy efficient ”))
Scopustitle-abs-key ((cloud OR processor OR saving OR methods OR techniques OR scheduling OR carbon OR emission OR {power-consumption} OR {energy-efficient}) AND (energy ∗ OR {efficiency} OR {energy-efficient} OR {energy optimization} OR {energy-reduction} OR {energy-consumption}))
ACM Digital Library(Title:(cloud OR processors OR saving OR methods OR techniques OR scheduling OR emission OR “energy -efficiency” OR “power-efficiency”) AND (energy-efficient techniques ∗ OR “power-optimization” OR “energy-mitigation” OR “energy reduction” OR “energy efficient)OR Keywords: (cloud OR processors OR saving OR methods OR techniques OR scheduling OR emission OR “energy-efficiency” OR “power-efficiency”)) AND (Title:(energy-efficient techniques∗ OR “power-optimization” OR “energy-mitigation” OR “energy reduction” OR “energy efficient”)) OR Abstract:(“power-efficiency”) AND (energy-efficient techniques∗ OR “power-optimization” OR “energy-mitigation” OR “energy reduction” OR “energy efficient”)) ORKeywords:(“power-efficiency”) AND (energy-efficient techniques ∗ OR “power-optimization” OR “energy-mitigation” OR “energy reduction” OR “energy efficient”)))
Wiley Online Library((“power-efficiency”) AND (energy-efficient techniques∗ OR “power-optimization” OR “energy-mitigation” OR “energy reduction” OR “energy efficient” OR “power-optimization” OR “energy-mitigation” OR “energy reduction”) in Article-Titles OR (cloud OR processors OR saving OR methods OR techniques OR scheduling OR emission OR “energy -efficiency” OR “power-efficiency”) in Abstract OR (cloud OR processors OR saving OR methods OR techniques OR scheduling OR emission OR “energy -efficiency” OR “power-efficiency”) in Keywords) AND ((energy-efficiency ∗ OR energy-consumption ∗) in Full Text)
Table 5. Inclusion and exclusion criteria for papers.
Table 5. Inclusion and exclusion criteria for papers.
Inclusion Criteria
1.Article presenting the importance and use of energy efficiency.
2.Clearly describes the technique from energy efficiency in cloud computing.
3. Peer-reviewed and written by an academic researchers or industry professional.
4. Research paper covering different aspects of energy and the issues generated by it.
5. Published in the field of cloud computing and in reputable journals, conferences, and magazines.
6. Research papers that describe the direct role of energy in cloud data centers as well as in generating carbon footprints.
Exclusion Criteria
1. Research articles that are not in context of energy efficiency in the cloud, rather describing some other theme, such as computer networks.
2. Has common challenges and references.
3. Summary of conversations, workshops, book chapters, or conferences.
4. Research papers that are not written in English.
5. Duplicate research articles, e.g., extended version published in difference venues.
Table 6. Quality assessment results for papers.
Table 6. Quality assessment results for papers.
Q’IdQuestionNumber of PSs (In Percentage)
Research Design and Conduct
YesPartiallyNoNA
QA1 Are the research objectives of the study clear?87.6%12.4%0%0%
QA2Is the analysis approach clearly described to assess the impact?76.6%18.6%4.8%0%
QA3 Are the attributes or parameters on which the impact is assessed clearly stated?89.5%10.6%0%0%
QA4Are the characteristics and merits of energy efficiency in the cloud provided?65.4%22.2%12.4%0%
Data analysis
QA5Is the dataset size defined adequately?72%12.5%14.7%0.08%
QA6Are the research articles’ programming languages mentioned?96.2%1.5%1.5%0.08%
QA7Are the obtained results statistically important?19.5%0%80.5%0%
Conclusion
QA8Are the validity threats, improvements, and limitations discussed?53.8%11.3%34.9%0%
QA9Are all the questions from the study answered?98.9%0.9%0%0%
QA10Does its conclusion accompany the empirical data and results?96.2%3.8%0%0%
Table 7. Power-measuring parameters used to measure the power consumption of the data center.
Table 7. Power-measuring parameters used to measure the power consumption of the data center.
MetricFormulaExplanation
Power usage Effectiveness (PUE)PUE = Total Facility Power/
Total Equipment Power
This is the ratio of the application and infrastructure’s energy consumption to the overhead energy consumption.
DCiEDCiE = 1/PUE DCiE = IT Equipment The inverse of PUE. Designing datacenters mostly relies on PUE and DCiE metrics.
Power/Total Facility Power
Green energy coefficientGreen energy consumed/total energy consumed In the data center, this is a measure of the amount of green energy (i.e., energy derived from renewable sources) being consumed. Used to assess the environmental impact of a data center’s operation.
Compute Power Efficiency(IT Equipment Utilization * it equipment power)/total facility power This is a measure of the computational capacity of the data
center or of the overall power used.
Energy reuse factorERF = Re-sued Energy Used/Total
Energy Consumed
A metric of reusable energy (energy that is consumed outside the data center).
Data center productivityUseful Work-done/Total resource (total
resource taken to produce this useful work)
A measure of the amount of fruitful work yielded by data center.
Thermal power designThe maximum power a computer chip
can consume while a process is in execution
Determines the maximum power needed
by cooling the computer equipment.
SWaP (Space, Wattage and PerformancePerformance space * powerThis is a data center Sun Microsystems metric.
It was developed for computing the resources and capacity needed by a data center.
Table 8. Power-measuring parameters used to measure the carbon consumption of a data center.
Table 8. Power-measuring parameters used to measure the carbon consumption of a data center.
MetricFormulaExplanation
Carbon usage effectivenessTotal CO2 emission from energy used/total energy consumedTo measure greenhouse gases in environment by the data center.
Water usage effectivenessWater used/EITA measure of the water requirements by a data center
Table 9. Power-measuring parameters used to measure the energy consumption efficiency of a data center.
Table 9. Power-measuring parameters used to measure the energy consumption efficiency of a data center.
MetricFormulaExplanation
Energy Star = (EPA)Efficient energy usageFor specific ICT components and products
EUCoC (European Code of Conducts)Energysubmit and compare efficiency data.
LEED (Leadership in Energy and Environmental Design)Characteristic for building for effective energy usageProvide powerful tool for evaluating facility
Table 10. Comparison of various hardware-level energy-efficient techniques.
Table 10. Comparison of various hardware-level energy-efficient techniques.
ReferencesTechnologyDescriptionAdvantagesDrawbackImplementation EnvironmentResults
[S2]Greener PCs for the EnterpriseNetwork Connectivity Proxy technique supports always connected with PCs, which allows them to sleep when inactive—Identified Idle machine problem and reduce energy consumption—Some devices cannot go into idle mode, such as processorsBuilt their own environment to testFor conserving resources and reducing the average cost of ownership, the connectivity proxy network retains network presence in idle situations for PCs, thus allowing them to be asleep.
[S9]Green Cloud architectureAllows better live VM migration and VM deployment optimization and online monitoring—Significantly reduces energy consumption—Not able to meet the criteria for web business models, such as OLTP.Tremulous
online real-
time game as VM
application
—Saves 27% of energy
—Help in consolidate workload
[S1]Energy-Efficient framework dedicated to Cloud architecturesDiscover how much energy virtualized worlds consume and how much does it cost—Saves both money and time
—High load balancing can be achieved
—Cloud pricing and accounting problems are not pursuedModern multicore platform—Save up to 25% of the power usage by cloud nodes
—integrated approaches for load management and migration
[S3]Novel Green Cloud frameworkOffer innovative energy-efficient scheduling, VM machine image, and image processing modules with which we can address new ways of energy conservation.—Reduces energy
Consumption and Enhancesperformance
—Scheduling system does not consider power and thermal issues.Open Nebula project in multi-core cluster—Savings of around 1750 Kwh per year.
[S4]Display power management technologiesEnergy wastage due to monitors, display screens of laptop due to in attentive user, laptop detects when the user is not looking at the screen—Reduces energy
consumption and increase system efficiency
—Ecological objects can interfere with the proposed detection
system, perhaps our procedure might assume that the customer is still active.
Samsung NT-R20 laptop (Intel Core 2 dual processor), webcam (consume 1.5 W)—boost low power
operation up to 50
Percent and increases energy saving
to 5–13%.
—Reduction is 13% more than timeout-based DPM approach
[S20]Scheduling algorithms with
dynamic voltage scaling (DVS)
Power-aware scheduling algorithms that provides a equilibrium between power consumption and job deadline constraints—minimize power consumption
—Reduce maintenance costs and improve system reliability
—Offers no budget-constrained scheduling
—only for static energy consumption
GridSim toolkitSystem load is reduced, energy efficiency is increased with no acceptance ratio loss.
[S5]Energy-Aware Resource-Efficient Workflow Scheduling under Deadline constraint. (EARES-D)EARES-D uses DVFS to schedule the DAG workflow centered on the earliest completion time—High resource utilization rate, reduce the Virtual Machine overhead
—Energy consumptions, energy cost and Co2 emissions are decreased
Complex to buildCloud simenergy efficiency gains of up to 30% on an average but 25% rise in resource utilization.
[S6]Novel Green Cloud Computing FrameworkNew energy-efficient scheduling, Service
Oriented by VM Image
Energy consumption
Improving system
efficiency
Energy-efficient
Scheduling, Service
Oriented VM Image
Open NebulaNew techniques for saving considerable energy while
limited impact on output was found.
[S7]Integrated Green Cloud ArchitectureIt is a customized Green Cloud middleware framework; the Green Cloud Broker offers the user as a platform for efficient cloud task distribution.—Access cloud services in the greenest way
—Reduce totalconsumer power usage
Implanted on the client’s end1 Ethernet Switch, 1 Router and 2800 Mb/s HP DL380 G5 serversThe analytical model is used to demonstrate how by
using the local, private, and public service provider energy efficiency is achieved.
[S8]Cloud-Oriented Green-Computing Architecture.
for eLearning-Applications. (COGALA)
Green aware E-learning
Applications that focused on energy consumption
reduction
—reduce costs and energy consumption,
—Can be deployed to maintain needed software
High complexity of
implementation
and operation as it is a theoretical model
Theoretical modelApplications, such as e-
Learning can use
COGALA
by which can reduce
costs, reduce energy
usage
[S41]Green Cloud
Broker
To deal with resource procurement problem they use mechanism design methods to Dynamically determine whether the work submitted will be distributed and paid.—Reduces cloud
users overhead time
minimize power consumption
—decrease the operational cost
—Inability to implement in a practical frameworkpythonProved that the selection of a
greener cloud service provider is successful done by proposed system.
[S10]UEFI based firmware methodologyfirmware that enhances the DRAM locality and DRAM power limiting accuracyenergy measurement as well as efficient energy limiting—No accurate energy measurement schemeSPEC-Power bench mark sets on a single socket Intel Xeon class ProcessorTotal energy consumption is improved by 15% with containerization memory and increased energy measurement inaccuracy by 14.1%
[S11]Measuring Energy and Power with PAPIExpanded API (PAPI) Efficiency for calculating
and monitoring of energy and resource values
—Support for the most current device efficiency metrics—Measurement overhead
—For total system power alone, knowledge
may not be enough to optimize
PLASMA and MAGMAto within 10% using only four hardware performance counters.
[S12]Measure power and performance at the chip levelReports and study of chip power and output assessed over five generations of process technology—Optimize to manage workloads
—Measure power and performance
The amount of information and results do not lead to a precise conclusion.Pentium.4. Atom.i5energy efficiency is increased
[S13]Energy-Efficient Hardware Accelerators for Machine LearningOffers a hardware accelerator designed specifically for neural networks Special emphasis is placed on the effects of memory on productivity and energy—It can help to retrieve common computational
primitives and locality characteristics behind the techniques.
Does not involves Field Programmable Gate Arraybit-accurate C++ simulator450.65 x speedup over a GPU, with a 150.31 x average energy reduction
[S14]HEARtH:
Heterogeneous
Multicore Platform
Offers a runtime scheduling policy for selecting the best core, and a power-management algorithm for a given application—Efficient use of different cores, achieved more efficiency than DVFS, optimizes CPU platfom energy.Use of heterogeneous CPU is complex than homogeneous CPUIntel.Core.2.Duo.T9900 works on 5 nm and a 32 nm and 22 nm processorproposed algorithm saves by 44% of energy
[S29]G-PMRM: Energy-efficient resource managementa winner determination algorithm used for VM provisioning—satisfies the
Loser
independent
property
—strategy-proof and save energy
No support for dynamic resource managementAmazon EC2The allocation of the G-PMRM is much quicker than the VCG-PMRM
[S16]Green-
Power
Allocation
.in Relay-
Assisted.
Multiuser Networks: A. Pricing-Based. Approach
In order to balance electricity consumption and the manageable sum rate, a cost-based approach was developed.—energy-
efficient
multiuser relay
networks
The convergence of the proposed algorithm had only theoretically demonstrated.MATLABShow the impact on the tradeoff of EE
[S17]Efficient Green Control Algorithm for Cost Optimizationcost optimization by using three N-Control policies SI, ISN, SN policies—significantly improve the response time and cost savings
—Reduce
running
Expenses
down without
compromising SLA
N-Control policy limits more switchingElastic Compute Cloud (EC2)Experimental results show that a system with the SI policy can reach a greater cost-effectiveness than other policies when there has a lower startup cost.
[S21]Near-
Threshold
Voltage
Design
For
Green.
Computing
Circuit method and advantages of NTV architecture, are demonstrated—Enhancing
energy
efficiency significantly of silicon system’s
—designed for achieving robustness
Achieving
reliability is
difficult
Intel’s IA-32 Research. ProcessorBuilt-In.Self-Test work in a continuous loop
[S22]Schedule Real-Time and Non-Real-Time Tasks an energy aware algorithmSystem that manages real-time and non-real-time energy intensive tasks without compromising reliability or performance.Consume less energyNot work effectively for long jobs spanned over a large period of time.MATLABThis saves up to 58% of energy compared with NPM
and 4% of energy in different situations relative to ES.
[S15]Physical
Machine
Resource
Management in
Clouds: G-PMRM
Using auction-based setting and design the Algorithm is designed to determine winner determinations such that each PM is loser-distributed independently.—Strategy-proof greedy mechanism.
—robust strategy
Only suitable for
deployment in real cloud settings
Amazon EC2Shows that the PMRM determines the allocation even more rapidly than VCG-PMRM, achieves better social benefits and achieves higher wages than the VCG-PMRM.
[S19]Dynamic virtual machine migration in peak loadEnergy distribution heuristics method
which provides data center resources to consumers
—Very good
Technique, Based On
Current allocation
of resources, that
improves energy
efficiency of the data
center
Overhead on network may occur to increased migrations of virtual machinesCloud simleads to a large decrease in cloud energy usage Compared with techniques of static allocation of resources in data centers
[S26]Intelligent placement of Virtual Machines in Geo-distributed Cloud for cost efficiencyPropose a two-phase Cut-and-Search meta heuristic cost-aware-algorithm, which approximates the best tradeoff point between the two cost terms—optimizing
the Overall
operating
cost
—significant
performance
improvement is
seen
Achieving accuracy is difficultCloudsimReaches a quite optimized point of trade off in fair computation time and random placement by 50 percent and partial optimization by 10–20 percent.
[S35]Deterministic Algorithms and Adaptive Heuristics algorithm for Dynamic Consolidation of Virtual MachinesTemperature-threshold-based detection of over-utilized Host—Avoids
hardware
wear out. Keeps
host in safe
condition.
—A significant
level of commitment to the service levels agreement while also being significant energy efficient.
Not
implemented
on real-world
Cloud
platform
CloudSimRandom or adaptive algorithms have to be constructed for improving the efficiency of optimal deterministic algorithms.
[S24]DC as Dispatchable Loads to.Harnes- Stranded-PowerIn contrast, adding data centers that are dispatchable by the power grid both reduces stranded power and improves both grid cost and stability, even at high RPS.—reduces stranded power and improves both grid cost and stability
—overall energy
Expenses were
Substantially
reduced
—identification of impacts of load constraints requires for practical
implementation
—Adding data
Centers with
renewable
generation (wind farms) can be harmful to RPS goals
—increases both power and thermal generation
realistic power grid systemDispatch able computing loads achieve duty factors of 60–80%
[S25]modern optimized energy management system architecture for data centers
under GENiC project
The system proposed incorporates energy optimization by monitoring and regulation of IT workloads, refrigeration of data centers, local energy generation and heat recovery waste.More optimized and high utilization of resources Results are produced via simulation
Real time implementation is required
virtual
C130 data centre test-bed
Optimized and upto 40% energy saving
Table 11. Comparison of various software-level based energy-efficiency techniques.
Table 11. Comparison of various software-level based energy-efficiency techniques.
ReferencesTechnologyDescriptionAdvantagesDrawbackImplementation
Environment
Results
[S18]Energy-aware Heuristic algorithmThe placement of applications is designed as a bin-packing problem.Reduces energy consumption dramaticallyOptimization is still absent iVIC framework with Python, Linux 2.6.18 as an OS, Xen as a hypervisor—Saving 10% to 13% of electricity
—Uses-fewer Powerful nodes
As compared to Algorithm, such as
first-Fit and best-fit
[S114]Algorithm for clustering followed by energy aware supply provisioningCentralized online clustering is being useddesired QoS is provided No optimization of costsA simulator with C++ event driven Named
kento-perf
—8% to 25% of energy savings
[S28]for VM selection using Reallocation algorithm and Modified Best-Fit Decreasing (MBFD) algorithmAdaptive consumption criteria are used in dynamic VM consolidation.—Creates energy savings while maintaining SLA violations to as low as 1%
—Attempting to cut down operating costs
Does not involve
Consumption
of electricity by Memory, and Resources for network
CloudSim—Consumes almost the same amount as Energy still nearly guarantees less than 1% SLA Infringements and lower the number of VM migration compared to other algorithms, such as DVFS and non-power-conscious policies
[S58]VM scheduling based on SLA resource constraintsAnalysis of energy used by VM and chooses the shortest response time to fulfil SLA requirementsSupports resource
provisioning and SLA fulfillment
—No live migration support for VM
—Ignores network energy usage, I/O devices and Graphics processing unit
OpenStack, Ubuntu-11.10-server-amd
and a hypervisor
—Requires limited physical equipment in comparison with the level of workload Schedule of the Round-Robin (RR)
—Electricity consumption is lower than RR Scheduling
—Low violations of the SLA
[S30]Multi objective scheduling on
private clouds
Least-load-first algorithm—Saving more resources and time
—Makes a high load-balance feasible
Hard to fulfill and
challenging to achieve various objectives
System consists of four PCs with four processors, 4 GB of Memory, 300 GB of Rom. Ethernet.Eucalyptus is used to build a private cloud environment.—Perform better in terms of average response time than the greedy algorithm; however, it is far behind the RR scheduling algorithm
—It conserves more energy, while RR does not save energy than a greedy algorithm.
—The solution is even perfect than the greedy algorithm and RR in load balancing case
[S31]Energy-efficient scheduling
scheme
Apply VM migrations for allocation of workload and using a rental Program for Management—To limit the number of VMs to minimal
—Spends workloads equally between VMs
Fast response time and limited level of utilization VirtualBox 3.1 Cloud
Environment
hybrid algorithms
—Contains fewer VMs than bull method, hybrid algorithms and RR
—saves about 21% of the energy, more then the amount retained by RR
[S55,S70]Algorithms for single allocation
request and global optimization request
Built on the conventional computation approach to minimize energy consumption and carbon emissions—SLAs are being followed
—greenhouse gases emissions are being decreased
Implementation and operation very complex It uses servers
with single, dual, quad, and six cores.
—up to 31% energy saved
—Reduction in greenhouse
emissions by up to 87%
[S35,S57]Algorithm for VM placement and optimal online deterministic algorithm and
MBFD Algorithm
Using Migratory Minimization (MM) or maximum potential growth strategies,
a random choice to pick VM for migration is there
—Operates well for heterogeneous VMs and infrastructu’e
—Does not depend on the type of workload
Lack of real-world cloud platform implementationCloudSim Toolkit As limits range from 30% to 70%: The minimum migration approach resulted in 66 percent reduction in power, 83 percent, and 23 percent, in comparison to the frequency scaling of dynamic voltage, non-power-awareness and ST schemes.
—When thresholds range from 50% to 90%: MM provides 87%, 74% and 43% of energy savings
[S36]Multistart local search algorithmDependent on the IaaS cloud model and multi-start local search heuristicOffers the desired quality of service and decreases energy consumption—More complicated and cost of procurement does not take GHG emissions into accountXML and OpenNebula cloud—The energy-aware multi-start local search algorithm for an Open-Nebula based
Cloud (EMLS-ONC) is typically having 1.3% higher Vm migration
—conveys 15.6% more enhancement than EMLS-ONC
[S32]Fixate on IaaSEnergy-Efficient Cloud Global Optimization
Algorithm
Enhances performanceNo cost optimization
involved
Operated on a single-, dual-, quad- and six-core server deployed in three types of resource centers: old, normal, and new data centers.44.76, 34.47, and 30.7 MWt of power used in various simulators using the Cloud Global Optimization Algorithm. These values are much lesser than those of the RR and greedy approach
[S37]Energy-aware scheduler for infrastructure clouds called OptSchedHandles IaaS workloadsThe data center and VM heterogeneity were efficiently managedThe total absence of the publicly available workload in implementationPythonIn comparison with RR and first-fit algorithms,
OptSched reduces CMU by up to 60.1% and 16.7%
[S38]Borrowed-virtual time Algorithms, application management algorithm, and Cluster leader
algorithms
Works in a clustered cloud organization within the IaaS framework—Application of SaaS and PaaS, private clouds and hybrids
—Applicable to DVFS-techniques processors
—Great difficulty in communication
—The system is not very capable of handling a rapid rise in system load
Clusters with variable sizes, namely, 20, 40, 60, 80, and 100 cluster nodes —Around 70% of servers operate in the optimal area, with only 5% running in the two undesired regions and 25% in the two sub-optimal areas, in a 20–100 server cluster.
[S39]energy-efficient method of deployment of VMsUsed the MCC approach to balance energy consumption with SLA.Total energy consumption and no SLA violation CloudSim toolkit.NA
[S40]Ant Colony OptimizationUtilizes ACO meta-heuristic for VM placement.Energy consumption reduced and low cost Time consuming
Not for dynamic environment
simulation toolkit based on JAVA.The algorithm proposed to achieve high energy benefits by using the server, including fewer machines, and determining almost optimal solutions that can be applied in an entirely distributed environment.
[S34]Energy-aware VM-Selection
VM-Allocation technique
Uniform Distributed
Discrete Random
Variable is used for
VM-Selection
significant cost savings
Maintain QoS and reduce power consumption.
NACloudSim toolkitThis method in comparison with static resource allocation approaches contributes to significant energy usage reduction in cloud data centers.
[S41]MADLVF: An.Energy Efficient.Resource Utilization ApproachFirst VM is selected with Least CPU utilizationReduce CO2 emission
and energy consumption, resource utilization
Higher SLA violationsCloudSim toolkitAs far as if we talk about energy consumption and migration metrics the proposed algorithm outperforms over MADRS significantly.
[S42]Energy-Efficient Virtual Machine allocation technique EE-ISVM allocation algorithm based on interior searchReduce energy consumption and resource underutilizationNACloudSimEE-IS saves on average 30% of energy
[S43]To optimize
performance and energy efficiency a Novel resource
allocation algorithms are being proposed
Proposed UMC and
VDT algorithms for VM-Allocation
Optimize energy
consumption and
resource utilization is done in efficient way while maintaining QoS
Energy consumption drops as the peak load of VMs was not achieved in same time phaseCloudsimNA
[S44]An energy-efficient power-law oriented VM management framework.Proposed a power
consumption
algorithm for video
streaming datacenters
Energy efficient resource Utilization Simple and not tested with real time environment CloudSim with Pythonhighest average utilization of resources was achieved by VMPL as compared with the other algorithm that offer better SLA performance.
-Reduce energy usage and optimize the average utilization of servers
[S45]An Energy-Saving
Virtual-Machine
Migration
Statistical based
energy consumption
algorithm used for VM-Migration
Power consumption in an optimize way while providing Qos.When the workload became too high or too low, algorithm energy efficiency rates are not that optimalCloudsimEnergy consumption 5% less, and energy efficiency can be improved by 6–10%, although the service quality cannot decrease dramatically.
[S46]Exact allocation and
migration algorithms
Linear integer
programming model and bin packaging
model are used for VM-Migration
Reduction in energy consumption and limit the number of migrations Java and the CPLEX as linear solver The outcomes show the advantages of merging allocation and migration and their potential to achieve substantial energy savings
[S47]Efficient VM-Placement and
VM-Migrationin
servers
Three VM-Placement
policies FDT, DRT,
and DDT are used
Less power
consumption, reduce VM migration
number of
SLA violations is still high
CloudSim SimulatorFDT reduced the of SLA violation 59.66%, Drt 27%
Ddt 71.52%
[S48]Energy-aware resource allocation algorithm for VM-PlacementThe two dimensional heuristic greedy
algorithm is used for
workload prediction
Reduce SLA
Violation, Energy
wastage, Balance
Workload and provide
Scalability
Power-Edge-Blade
servers
Significantly improved energy saving, task balancing and
scalability, relative to CPU-based single objective approaches.
[S49]Dynamic virtual machine
placement algorithm
The statistical
mathematical
framework is used
Workload handling,
reduce power
consumption
Simulation resultsSelf-designed
Simulator
Using the suggested approach, th’ system’s power consumption may be reduced while simultaneously dealing with spikes in workload.
[S50]Energy efficient dynamic
resource management
Logistic regression
and median absolute
derivation models are used for detection
algorithms for, 1 vm placement
Minimize power
consumption of the
data center and avoid
SLA violations.
Not able to deal with dynamic workloadsCloudSimWhile experimentation best value of c to be used is 0.74
[S51]By introducing heterogeneous task
A VM placement algorithm is proposed known as ETVMC
Developed an algorithm for
demand-based VM and machine mapping
Reduce energy use,
make-span and task rejection rates
Results for systems that can change dynamically
during service time in respect of resource demands
CloudsimReduces energy consumption, equipment
and the rate of rejection
[S52]Energy-aware online
VM-Scheduling algorithm
Algorithm for Online VM Scheduling was developedContinue satisfying QoS, reduce energy use and promote hetrogeneity enviroement.No real time implementation Self-designed
Simulator
Minimizing power use and maximization of the provider income
[S53]Energy-aware virtual
machine dynamic
provision and scheduling
For power-aware VM scheduling, the
dynamic round-robin algorithm is proposed
Reduce energy and
physical system use
High complexity Eucalyptus: An
opensource Cloud
middleware
Savings energy 43.7% and physical computer
use 60% less relative to other planning algorithms
[S54]Energy-based accounting
and scheduling of virtual
machines
In-processor tasks
Generation, model is developed for
energy efficiency VM-Scheduling
Reduce energy
consumption, energy credit scheduler
Our model by adding more event counters from other components such as NIC or disksXen HypervisorAddress and measure energy usage errors
within 5% of the overall energy consumption
[S55]EMinTRE-LFT
Energy-efficient
Scheduling of Virtual Machines in IaaS clouds
EMinTRE-LFT algorithm to reduce the amount of the total time spent by all physical devices are equal to a minimum of the total consumption of Energy.minimize system completion time and energy consumptionStill not optimizedOpenStack Nova
scheduler
The total physical server energy consumption of Tian-MFFDE can be reduced by an average of 23.7%
[S99]Energy efficient
scheduling and
management for largescale services computing systems
An online algorithm was provided, which
is based on Lyapunov optimization technology
Reduce energy consumption,
No need of prior information of
user’s request, system performance is also optimized
Near optimal
system profit
Actual trace-based simulation and numerical experimentsNA
[S56]for parallel
task-based applications, dynamic energy-aware scheduling was proposed
Multi heuristic-resource-allocation (MHRA) algorithmReduction in energy consumption and time for executing task.Shows that, depending on the size of the instance and
type of DAG, a large amount of energy can be saved.
Using COMPS programming model to implement a scheduler.It enables the savings to be made in the system configuration by an average of −22.44 percent, −33,17 percent, −27.08% and −31.5% for EP, MT, and WP, SG, respectively.
Table 12. Comparison of consolidation-based techniques.
Table 12. Comparison of consolidation-based techniques.
ReferencesTechnologyDescriptionAdvantagesDrawbackImplementation
Environment
Results
[S76]Energy-efficient.and
QoS.dynamic.
Virtual-machine-consolidation (EQVC) method
Before overloading and migrating the VMs to other Servers, a redundant VMs is chosen from hosts to save energy and ensure QoSEnergy saved and
QoS specifications assured.
Additional factors including memory, disk capacity
and network workload are not included
We experiment with various types
of workload from a real machine
The EQVC solution will significantly improve other conventional approaches in terms of energy usage, QoS assurances and the number of migrations of VMs.
[S59,S69]Energy-efficient
network-aware scheduling for data centers
A hierarchical model that can easily be
adapted to the existing data center topologies
Energy utilization and performance optimization on bases of QoS and traffic requirements —Minimize overheads for storage and processing Inability to incorporate practical mechanisms—DVFS and DPM technique with a Green Cloud simulator
—Three-layer tools
for energy monitoring
—The RR and green scheduler methodology is compared.
—Den’s technique leaves fewer servers idle than the green scheduler.
—RR is the least
efficient in term of energy
—4% more energy consumed by DEN system, slightly more than the green scheduler is used.
[S23]Based on energy policy
for carbon reduction and
meet demands
in performance.
Preemption aware
energy
management
policy (PEMP)
During starvation,
it can perform on-demand switching
Does not recognize the influence on
preemption policy through VM migration and consolidation
Haizea with a Blue Horizon cluster in SDSCPEMP uses less energy than other energy conservation strategies, such as SLA-based Energy Saver Policies (SESP) and generates 18% of energy savings
—Greedy Energy Saver Strategy (GESP), relative to both PEMP and SESP, is the having the least energy consumption
[S60]DVFS-Aware-consolidation for energy-efficient
clouds
Two methodologies were introduced, one based on the DVFS a technique for efficient consumption and the other on VM consolidationEffective energy use, quality of service maintenance and system efficiency improvement. CloudSimSaving of 39.14% for dynamic load conditions
[S61]energy-aware
job consolidation
scheduling
Job consolidation
algorithm by using DVFS technique
Reduce energy
consumption,
physical machine
heterogeneity is being supported
No real-world cloud environment.Cloud sim remains successful in generating effective job replacement policy and ensuring carbon reduction.
[S62]self adaptive approach called SAVEConstructed a probabilistic mechanism for the assignment and
migration of VMs primarily focused on local knowledge
Minimize total energy consumption
Increase efficiency of consolidation of workload,
Optimize memory and network i/o process
Focus on performance of algorithm but not on energy reduction Cloudsim/real testbed30% of reduction of energy consumption is obtained
[S63]topology aware multi objective VM consolidation for cloud data centersImplements a genetic algorithm with multiple objectives
to reduce energy consumption and communication traffic
reduce communication traffic due to topology awareness and
improved energy consumption with preserving SLA violations
NACloudsimNA
[S64]Bayonian-based network (BNEM) modelBayonian-based network (BNEM) model that take 9 factors int consolidation while consolidation in data centersImprove QoS
Energy reduce migration.
NACloudsim demonstrate that the approach will degrade energy usage
substantially, prevent inefficient VM migration and enhance QoS.
[S65]For Aliyun’s Cloud-Scale Computing, a Cost-Efficient Consolidating Service was builtInstance migration cost model by using the worst fir heuristiccost efficiency,
load balance achieved and oscillation free consolidation services
Less optimized Real testbedAttains cost-effective and load- balanced, consolidation service, which is free from oscillation.
[S66]Based on energy
conscious task consolidation
heuristics
ECTC and MaxUtil, Two energy-conscious algorithms for consolidationNo optimization of costOffers no guarantee of high performanceSimulation performed on 50 tasks—The MaxUtil reaches 13% energy savings while ECTC achieves 18% energy savings on average, irrespective of whether if migration is being conducted
[S66]Implies dynamic VM consolidation through the use of adaptive thresholdsFor Modified Best-Fit Decreasing (MBFD) algorithm and VM selection, a reallocation algorithm was proposed—Achieve energy efficiency while ensuring SLA violations
Low at as 1%
—Cost reduction
in operation
It does not involve memory
and network resources power consumption
Cloudsim—When thresholds vary from 50% to 90%: MM policies give 87%, 74%, and 43% energy savings and when a threshold is between 30% and 70%: MM policy provides a 66% savings in energy usage, 83% and 23% as concerning frequency scaling in dynamic voltage.
Table 13. Comparison of bio-inspired techniques.
Table 13. Comparison of bio-inspired techniques.
ReferencesTechnologyDescriptionAdvantagesDrawbackImplementationResults
[S81]Ant Colony OptimizationVirtual machine placement framework using an
ACO in an IaaS environment
Reduces waste of resources and electricity usage and minimizes SLA violationsSometimes it is very complex to do multi-objective optimizationCloudsimDemonstrate that the proposed method can achieve optimum balance in multiple conflict objectives in comparison to the heuristic approach and genetic algorithm
[S83]AVVMC VM consolidationConsolidation technology for VM to use servers across various computing resources in a balanced mannerImproved energy usage as well as drop-in waste of resources No account efficient network
resource utilization.
Java platform is used for implementation on dell i5-2400 CPU with 4 DB ramTackle power usage and waste of
the energy in large virtualized data centers
[S82]VMPACS
algorithm.
To solve the problem of VM placement using ant colony system algorithmTotal resources usage and power usage are greatly reduced effectively and Safely.The algorithm is only suitable for large data centersIntel 2.50 GHz Intel Pentium® dual-core CPU and 3 GB RAMWithin three minutes a complicated placement problem of up to 2000 VM was resolved.
[S84]metaheuristic algorithm based on Ant Colony Optimization (ACO)Algorithm selects a collection of adjacent PMs for data and VMs placementReduce traffic and bandwidth usage across networksOnly homogeneous VMs are Considered,
Multiple requests (batch request) are actually not being considered at the same time.
MapReduce cluster and cloud simPerformance is better as compared to other allocation schemes.
[S85]OEMACSACO-based VMP solution using artificial antsfor reducing the active servers, optimize resource utilization and balancing themnot very sensitive to the parameters,C++ platformEfficient and optimized approach for VMP
[S86]VMPRP algorithm
based on Swarm Intelligence
VM positioning algorithm based on the process of recruitment inside Ant coloniesminimizing both
power consumption and resources wastage, Scalable
In communication between VMs, some additional energy consumption was not consideredCloudSimwastage of resource utilization was less than 10%
[S88]crow search based VM placement algorithm (CSAVMP)Manage and schedule a broad range of VMsMitigate resource wastage and energy consumptionNo real environmentCloudsimPerforms better than GGA and FFD in terms of servers numbers, power consumption and wastage of resources
[S87]TSPCS using meta-heuristic crow search algorithman algorithm to consolidate a massive number of VMs on minimal DCs based on metaheuristic crow searchTime efficiency migration is a costlyJAVA platform with on
i3-380M processor
2.53 GHz and 2 GB RAM.
Proposed methods significantly outperform the classical First Fit (FF)
[S89]Evolutionary Optimal Virtual Machine Placement and
Demand Forecaster
method for developing VMP with demand forecastingminimize the
cost and optimize placement strategy
Only effienct when reservation plan is accurately provisionedMATLAB framework for implementation on a Intel Core 2 Duo 2 GHz processor Mitigate the 2 percent
of the extra costs
[S90]Parallel bio-objective
hybrid genetic
algorithm
Based on
Multi-objective
parallel GA and
ECS heuristics
Proposes deadline-based
optimization
Focus on
Multi-objectives
increases the cost
Fast Fourier Transformation Task Graph establish in real-world —Energy consumption dropped by 47.49% and time reduction by 12.05%.
[S91]Energy-efficient
multitask
scheduling
algorithm
Established on
Google’s
massive data
processing
network
Imposes a load
balancing scheme
—Guarantee of High computing time and d no QoS and cost reduction.performed on data center consisting of 200 servers—optimize the energy
utilization of servers better
from Hadoop MapReduce scheduling
[S92]Modified
genetic
algorithm
Smart and dynamic migration of VMs in combination with Heuristic algorithm (a modified GA)Offers a systematic cost
function for the whole system
—offers an extensive cost function applied to the whole systemCarried out on a flat VPC that enables flawless VM
movements
—The GA-based approach proposed using Cloud net will reduce carbon footprint.
—Carbon footprint decreases by 36.64 percentage but raises energy demand at the rate of 4.96 percent.
—It demonstrates improvement in reducing carbon emission levels for variable loads, when simulated and evaluated under low and high loads.
[S93]Implement an effective and agent-driven solution for energy efficiencyDynamically assigns the CPU and memory bounded tasksSLAs and high throughput and response times—Lacks load
prediction
CloudSim toolkit Yet to be implemented
[S94]algorithm proposed based on ACO Meta-heuristic and Energy-aware workload consolidation algorithm based on ACOBased on ACO metaheuristicNeeds fewer resources and achieves greater
use of energy than other algorithms
—Extreme workload dependence and need resourse details.
—The complex nature of dynamic activities cannot be handled
Java as simulator tool—Based on comparisons, the first fit adaptive ACO algorithm uses fewer host machines and provides better utilization of resources and power savings from the adaptive edition of the first-fit decreasing approach
—4.7% of hosts and 4.1% of power was generally saved
[S95]Hybrid
scheduling
based on ACO
and Cuckoo
search
Based on ACO
and Cuckoo
search, and also
uses a voltage
scaling factor
Efficiently reduces
energy consumption
—Difficult to implementMatlab, IaaS was built via Xen Cloud and PaaS via Windows Azure Microsoft–The hybrid algorithm gains slightly more energy consumption and high performance than ACO according to their numbers of tasks and processors.
[S96]Particle Swarm optimized
Tabu search
mechanism
Energy-efficiency resource consolidation algorithm
for Consolidation of VM
—Maximizes the usage of resources
—Reduces time for execution of tasks
—Boost the acquisition of revenue
—Must increase the profitabilityCloudSim toolkitAs compared to the energy-efficient topology algorithm (ECTC) and the random algo, the proposed approach consumes 67.5 percent less energy.
—Earnings 8.14 times more than random On average
[S97]SSO methodDVS and swarm
intelligence
—Could schedule jobs
for other shared elements
—Infeasible for systems that are not time senstive.—Uses a graph generator to produce large amounts of Directed Acyclic Graphs (DAG) for testing purposes —SSO was compared to PSO, with both achieving
a power reduction of up to 20 percent.
—Particle swarm optimization to a certain
level, more efficient than SSO.
Table 14. Comparison of various thermal and cooling management energy-efficiency techniques.
Table 14. Comparison of various thermal and cooling management energy-efficiency techniques.
ReferenceTechniquesConceptAdvantagesDisadvantagesImplementation EnvironmentResults
[S98]Cooling system for a data center in an emergence situationFormation of an algorithm for emergency decision to be integrated into the new platform for supervisory: GENiCContributed to improve the data center emergency power planMethodology with a three-zone model offers only limited informationtime consumingDynamic building energy simulation methods such as
TRNSYS
Rather than only serving as a design tool, this model might potentially be included into a more comprehensive model-based control system.
[S72]Real-time transient thermal model is developedTemperature control maintained for optimal temperature stability during power failuresManagement the increase
in temperature while power failure
Under standard operating conditions, large units can be costly and inefficient.on its own simulation systemoptimal temperature control during power breaks and show that it would be CRAC fans and cooled water pumps effectively arrange on the UPS
[S100]Free cooling using Direct airside
economizer
Reduce the energy consumption by proper management of indoor temperature and humidity—Low consumption of energy
—The PUE was dramatically reduced
—Humidity needs
To be considered
—Impurities can
Cause malfunctioning
—The air from
the outside must be cold
DeST software49% energy conservation
[S101]Free cooling using Indirect airside economizerEliminating the discharge of excess heat from data center—Reduced chance of air pollution
—reduce the humidification costs
—Accumulation of heat exchanger quantities
—Decreased the heat exchanger
performance and increased rate of maintenance
—only for Large size of data center
DEST (designer’s simulation toolkits) 29% of energy savings
[S102]submerge cooling system123 servers encapsulated in mineral oil used in server blades—No CRACS
or chiller needed
—Increased ability for heat absorption
—No sealed piping and enclosures are required
The need of pumps for liquid renovation
—Dynamic implementation
Its own simulation The cooling of mineral oil has the potential
to minimize cooling costs by up to 50%
[S103]Two-phase cooling systemcooling system for a tele-communications station named
as a micro-channel separate heat pipe
Can eliminate
enormous heat flux
Based on the environment conditionenthalpy difference laboratoryWith a change in indoor and outdoor temperature from
6 °C to 8 °C the cooling power increased by 135%.
[S104]Two-phase cooling by using integrated systemcooling system that is composed of thermosyphon loop, a refrigeration system, and outdoor air convectionhigh energy savings
—Can operate in climate change condition
Presence of active component
(compressor, valves…)
An ISMT, A rotating compressor, an expansion valve and a tube-fin condenser.The energy-saving average for indoor temperatures at
27 °C is 5.4 per cent-47.3 per cent per year
[S105]Two-phase cooling using two phases
immersion
servers were immersed in dielectric fluid the heat was transferred outside the data center with the aid of waterNo need of pumps, fans, economizers, compressors
—reduction of Cost and complexity
Fluid leakage problemCOMSOLPUE is only 1.08
[S106]Cooling system for Container-based data centerWorking on direct-fresh-air-cooled Container-based data center (CDC)By evaporative cooling and circulation of waste heat a combination of direct fresh-air cooling was constructedlack in maintaining the cleanliness of fresh-airConstructed their own simulationReduced electricity by 20.8 percent
[S57]latent heat storage unit by using PCM and
free cooling
PCMs with a natural cold source is introduced to minimize room cooling capacity at base stations—PCMs combined with the cold natural source.Environment dependent Performed based on the operation of the
TBS unit located in five Chinese cities with various climates
50% of energy savings
[S108]Two-phase cooling: heat
pipe
potential use of heat pipe systems for data centers cooling.—No passive componentDepending
on the weather and the wick
structure concerns
A theoretical model 75% of energy savings
[S109]Rack level
liquid cooling is being used
Computational Fluid Dynamics
(CFD) technique is used in cooling system
—The removal of the heat is more
efficient
—Improved working condition
—high maintain
—more costly
Commercial CFD package, ANSYS FLUENT 12.0.1. PUE = 1.3
Table 15. Comparison of various power-aware management scheme-based energy-efficient techniques.
Table 15. Comparison of various power-aware management scheme-based energy-efficient techniques.
ReferencesTechnologyDescriptionAdvantagesDrawbackImplementation
Environment
Results
[S67]Fast array of wimpy nodes (FAWN)Architecture projected
for low-power data intensive computing
High speed flash storage,
and also, for replication.
Low power CPU nodes are more energy effective than standard high-performance CPUs.Intel’s Q6700 quad-core 2 GB DRAM Gpu, Mtron Mobi SSD and Gigabit onboard Graphics and Ethernet Research shows that FAWN clusters can handle around 1000 queries/Joule, suggesting a tremendous potential for a wide I/O workload
[S68]Gordon,
a system architecture for data-centric applications
Processors with limited power consumption,
flash memory and data center Programming systems for Quality improvement and power-consumption reduction.
Energy consumption is optimized Only suitable for low power processors 8 2.4 GHz Core 2 Quad
cluster with 8 GB computers, large hard disk SATA and a 1 Gb Ethernet
Gordon systems will perform better than a disk-based clusters by
1.5 times and offer up to 2.5-times more efficiency per watt.
[S69]Non-dominated Sorting Genetic Algorithm
II (NSGA-II)
An energy-efficient algorithm for resource
allocation for cloud data center based on genetic algorithm.
Optimal and better execution time No account is taken for complex activities, such as task allocation,
internal communications, cost of electricity and load of the data center
1 GHz single-core processors, such as Intel PentiumProposed technique appears capable of performing static distribution to a wide range of individual tasks on Homogeneous single-core servers of the same data center in quadratic time complexity
[S70]genetic algorithm for
power-aware (GAPA) scheduling
For encoding chromosomes, a tree structure is used. Evaluation value of each chromosome
was determined for GA fitness function.
Energy efficient GAPA computational time is high, job deadline time also has not been taken into accountTwo power models of Dell Inc. PowerEdge R620 and the IBM server x3250Baseline scheduling algorithm energy consumption is 130% more than of GAPA
[S71]particle swarm optimization (PSO) based algorithmformulate and explain the heterogeneous
multiprocessor energy-aware real-time scheduling problem
Reduce energy costs and
optimize time for effective solutions.
Not able to solve real-time job scheduling prs1.67 GHz intel core 2
CPU,Cache: 512 KB. Memory: 2074492 KB
Windows 7 OS.
Visual Studio
2003.NET development too.
Meta—heuristic energy awareness based on the PSO consumes 40–50% less energy than the algorithms
based upon GA and SFLA and invests 10% less time in finding solutions than the SFLA algorithm.
[S73]Built on the
DVFS enabled cluster
Power-aware algorithm
for scheduling
—Applicable to clusters with high performance
—Aims to
optimize system performance
Offers no cost
Optimization and no cost
optimization
Open-Nebula and
NBench Linux
DVFS-SIM simulation is used
—Total performance improved
by 34 percent
[S74]intelligent
traffic routing using our Network Traffic Consolidation (NTC)
Describe and evaluate three energy saving solutions in single administrative domain networks without having a major effect on the’network’s ability to achieve the expected levels of efficiency and availabilityEnergy efficient and performance is optimized Not cost-efficient model for users real Web 2.0 workload in a real data center network topologyAchieve 16% power savings with the adjustment of active network elements and additional savings (up to 75%) through the network traffic maintenance and consolidation of server workload
[S75]algorithm to
reduce power consumption that is based on greedy method
Power-aware VM
scheduling algorithm
energy-reduction algorithm based
on greedy algorithm
—Needs better data center architecture for —lacement of server and cooling system Open Nebula —ave 12% of device
power during normal load situations
[S118]Elastic TreeA power manager that adjusts the active network elements set and switches them dynamically to meet changes in traffic loads in the data center.Energy saving No focus performance optimization and fault tolerance, cost efficient Testbed installed from three network vendors by the production of OpenFlow switchesUp to 50% network energy saved from Elastic Tree.
[S33]Based on adaptive
Provisioning schemes, Adaptive-DVFS, and δ-advanced
DVFS
VM Scheduler that shares Power TimeReduces energy consumption
and raises income
for data centers
Works individually without taking
the workload into account
CloudSim and Creat–s four—16-DVFS-enabled
data center
—Adaptive DVFS and Advanced DVFS systems, use less power and generate greater income that is independent of workload.
[S78]Based on CPU and memory intensive
models and also on VM migration
Placement scheduling algorithms, replacement scheduling and
algorithm for energy-conscious migration
Implements workload management and green
sorting algorithms to improve energy efficiency
—requires multiple
VM migrations, which raises
—Large operating
costs and complexity
power meter and Distributed ASCI
Supercomputer 4 (DAS-4)
—Algorithms maximize energy consumption, which require multiple VM
migrations, increasing energy as well as operational costs.
[S79]Task Scheduling energy-efficient algorithm that is DVFS-enabled (DEWTS}based on dynamic
voltage/frequency scaling algorithm.
Substantial energy reduction and maintenance of QoSsystem reliability
and no
real environment
randomly generated
DAG graphs as the tasks set for our experiments
For various parallel applications, DEWTS can reduce overall power consumption by up to 46.5 per cent and balance the scheduling performance.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bharany, S.; Sharma, S.; Khalaf, O.I.; Abdulsahib, G.M.; Al Humaimeedy, A.S.; Aldhyani, T.H.H.; Maashi, M.; Alkahtani, H. A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing. Sustainability 2022, 14, 6256. https://doi.org/10.3390/su14106256

AMA Style

Bharany S, Sharma S, Khalaf OI, Abdulsahib GM, Al Humaimeedy AS, Aldhyani THH, Maashi M, Alkahtani H. A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing. Sustainability. 2022; 14(10):6256. https://doi.org/10.3390/su14106256

Chicago/Turabian Style

Bharany, Salil, Sandeep Sharma, Osamah Ibrahim Khalaf, Ghaida Muttashar Abdulsahib, Abeer S. Al Humaimeedy, Theyazn H. H. Aldhyani, Mashael Maashi, and Hasan Alkahtani. 2022. "A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing" Sustainability 14, no. 10: 6256. https://doi.org/10.3390/su14106256

APA Style

Bharany, S., Sharma, S., Khalaf, O. I., Abdulsahib, G. M., Al Humaimeedy, A. S., Aldhyani, T. H. H., Maashi, M., & Alkahtani, H. (2022). A Systematic Survey on Energy-Efficient Techniques in Sustainable Cloud Computing. Sustainability, 14(10), 6256. https://doi.org/10.3390/su14106256

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop