Energy Sustainability in Cooperating Clouds - Semantic Scholar

3 downloads 105699 Views 2MB Size Report
Cloud Computing, Federation, Virtual Infrastructure Management, Energy Sustainability, Photovoltaic ... investigate the advantages, constrains, and issues for the achievement of such a .... reluctant to absorb energy produced by private re-.
Energy Sustainability in Cooperating Clouds Antonio Celesti, Antonio Puliafito, Francesco Tusa, and Massimo Villari Universit`a degli Studi di Messina, Facolt`a di Ingegneria Contrada di Dio, S. Agata, 98166 Messina, Italy. e-mail: {ftusa,acelesti,apuliafito,mvillari}@unime.it

Keywords:

Cloud Computing, Federation, Virtual Infrastructure Management, Energy Sustainability, Photovoltaic Systems, Renewable Energy, Energy Optimization.

Abstract:

Nowadays, cloud federation is paving the way toward new business scenarios in which it is possible to enforce more flexible energy management strategies than in the past. Considering independent cloud providers, each one is exclusively bounded to the specific energy supplier powering its datacenter. The situation radically change if we consider a federation of cloud providers each one powered by both a conventional energy supplier and a renewable energy generator. In such a context the opportune relocation of computational workload among providers can lead to a global energy sustainability policy for the whole federation. In this work, we investigate the advantages, constrains, and issues for the achievement of such a sustainable environment.

1

Introduction

Federation is the next frontier of cloud computing. Throughout the federation, different small and medium Cloud providers belonging to different organizations, can join each other to achieve a common goal, usually represented by the optimization of their resources. The basic idea is that a Cloud provider has not infinite resources. In order to achieve target business scenarios a Cloud provider may need a flexible infrastructure. Federation allows Cloud providers to achieve such a resilient infrastructure asking additional resources to other federation-enabled Cloud Providers. Cloud federation is much more than the mere use of resources provided by a mega-provider. From a political point of view, the term federation refers to a type of system organization characterized by a joining of partially “self-governing” entities united by a “central government”. In a federation, each self-governing status of the component entities is typically independent and may not be altered by a unilateral decision of the “central government”. Besides cloud mega-providers, also smaller/medium providers are becoming popular even though the virtualization infrastructures they have deployed in their datacenters cannot directly compete with the bigger counterparts. A way to overcome these resource limitation is represented by the promotion of federation mechanisms among small/medium cloud providers. This allows to pick up the advantages of other form of economic model

considering societies, universities, research centers and organizations that commonly do not fully use the resources of their own physical infrastructures. Moreover the traditional market where cloud providers offer cloud-based services to their clients, federation triggers a new market where cloud providers can buy and/or sell computing/storage capabilities and services to other clouds. The advantage of transforming a physical datacenter in a cloud virtualization infrastructure in the perspective of cloud federation is twofold. On the one hand, small/medium cloud providers that rent resources to other providers can optimize the use of their infrastructure, which are often underutilized. On the other hand, external small/medium cloud providers can elastically scale up/down their logical virtualization infrastructure borrowing resources of other providers. Federation enables cloud providers to relocate their services in other ones. In our opinion, this allows to plan more flexible energy sustainability strategies than the past. In this work, we focus on an innovative sustainable federated cloud scenario in which resources are relocated between cloud providers whose datacenters are partially powered by renewable energy generator systems. The federation is seen as a way for reducing energy costs (Energy Cost Saving), but at the same time a possibility to reduce the CO2 emissions (Energy Sustainability). Here we discuss strategies and policies it is possible to apply in federated cloud environments for achieving the just mentioned goals. Specifically our assessment is aimed at the design of an Energy Manager, to be included in our Virtual In-

frastructure Manager, that is the CLoud-Enabled Virtual EnviRonment (CLEVER). Further details on how to achieve such a goal will be provided in Section 4. The manuscript is organized as follows. Section 2 introduce how an energy sustainability strategy can be applied to a federated cloud environment. The energy consumption of a datacenter is affected by different factors including the Power contribution for the Information Technology (IT) equipment (PIT ), the Power contribution for the Electrical (POW) equipments (PPOW ), and the Power contribution for the Cooling (COOL) equipments (PCOOL ). To this regards several energy considerations about cloud datacenters are discussed in Section 3. As said before, a decision algorithm for sustainable federated clouds is presented in Section 4. Section 5 discusses related works. Section 6 summarizes conclusions and lights to the future.

2

Cloud Federation and Energy Sustainability

Federation brings new business opportunities for clouds. In fact, besides the traditional market where cloud providers offer cloud-based services to their clients, federation triggers a new market where cloud providers can buy and/or sell computing/storage capabilities and services from/to other clouds. The advantage of transforming a physical datacenter in a cloud virtualization infrastructure in the perspective of cloud federation is twofold. On the one hand, small/medium cloud providers that rent resources to other providers can optimize the use of their infrastructure, which are often underutilized. On the other hand, small/medium cloud providers can elastically scale up/down their virtualization infrastructure borrowing resources and paying them from other providers. A cloud provider can decide to lend resources to other clouds when it realizes that its datacenter is under-utilized at given times. Typically, datacenters are under-utilized during the night and overutilized during the morning. Therefore, as the datacenter cannot be turned off, the cloud provider may decide to turn the problem into a business opportunity. Instead, a cloud might need to buy resources from other clouds for the following reasons: • The cloud runs out of its storage/computing resources. In order to continue providing cloudbased service to its clients, it decides to buy resources from other clouds. • The cloud needs to deploy a distributed cloudbased service in different geographical locations;

hence, it acquires resources placed in those locations. • The cloud needs to relocate service instances in other clouds. As federation enables cloud providers to relocate their services on other peers belonging to the system, in our opinion, it is possible to carry out more flexible energy-aware scenarios than the past, when we considered independent non-federated clouds. Two possible alternative energy-aware scenarios are: • Energy Cost Saving. Resources are migrated in external cloud providers in order to push down the energy consumption cost of their datacenters. The resources will be migrated in external cloud providers in which the cost of the energy is cheaper. A possible approach could consist into turning off a given datacenter within a site and move the load towards another where the resources renting price is lower than the cost of computing locally. • Energy Sustainability. Resources are migrated in external cloud providers whose datacenters are partially or totally powered by renewable energy just for minimizing the costs due to the environmental impact (e.g. reducing the CO2 emissions). An Energy Sustainability scenario often allows to push down costs, but unfortunately this is not always true. In fact, by now, energy-aware strategies based on sustainable green computing environment have not found so wide diffusion due to the high costs of energy production. In our opinion a sustainable federated cloud environment could push down such costs also allowing to achieve a green computing environment able to reduce the energy costs. Nevertheless, how to achieve such a sustainable federated cloud environment is not clear at all. The main contribution of this work is to propose a possible approach for the achievement of such an environment. Our approach is based on the following idea: “moving the computation toward the more sustainable available cloud datacenter”. This statement is motivated by the following assumptions: 1. Often, the renewable energy generator systems produce more energy than necessary. 2. It is very hard to store the exceeded produced renewable energy (e.g., in batteries). 3. Alternatively, it is becoming very difficult to put the exceeded produced renewable energy in public electric grids. This practice is becoming a problem for Energy suppliers as it implies uncontrolled power surges which are hard to be managed. As this problem is becoming more and more

Figure 1: Example of Sustainable Federated Cloud Environment.

sensitive, the energy suppliers are becoming to be reluctant to absorb energy produced by private renewable energy generator systems. 4. As consequence of 1), 2), and 3) often the exceeded produced renewable energy is wasted. 5. Consequently, it is easier to move the computation toward a datacenter powered by a renewable energy generator system with a high large availability of energy than moving the “green energy” toward the datacenter where the computing has to take place. If we consider a set of datacenters with these features, a sustainable federated cloud environment can allow to save money, maximizing the use of “green energy” and reducing the level of carbon dioxide. An example of such a scenario is depicted in Figure 1. The sustainable federated cloud ecosystem includes four cloud providers: Messina, Sidnay, S˜ao Paulo, and Stuttgard. The electricity suppliers A, B, C, and D are independent companies that provide energy at different costs. In our scenario, the datacenter of each cloud providers is meanly powered by a private renewable energy generator system as primary source of energy. When the primary source of energy is not enough to power the datacenter, the cloud uses the energy of its Electricity Supplier. In addition, each cloud is federated with each other in order to take the advantages of service relocation and resource consolidation. Further details regarding how federate cloud architecture are out of scope of this work. Further details can be found in Section 5. Each cloud provider joins the federation in order to move its computation in other federated clouds

where the production of “green energy” is maximum. In simple terms, we move the computation load towards the more efficient renewable energy generators (in term of produced electricity) maximizing the utilization of the federated clouds in which the workload has been transferred. Considering Figure 1, the four clouds have different latitudes. According to a given period (due to time zone and month), each renewable energy generator system that primarily powers each cloud datacenter can have different energy efficiency compared to each other. These different conditions can depend by different factors according to the adopted source of renewable energy. For example the amount of energy produced by a photovoltaic system depends on the solar radiance which is different hour by hour, day by day, and month by month. In addition, the energy production of a photovoltaic system is also affected by the weather and climate condictions. For simplicity, let us consider the time zone, when the time of Messina is 17:00, the time of S˜ao Paulo is 13:00. In this situation, in S˜ao Paulo the solar radiance of the sun is stronger then the one in Messina. On the other hand, considering both cities in July, the temperature of Messina will be probably higher than the one in S˜ao Paulo. Further similar consideration can be made considering the latitude of the two cities. This scenario implies that if a cloud provider wants to enforce energy sustainability policies on its own datacenter, it relocate its services into other federated cloud providers, chosen according the aforementioned energy consideration. For reasons of Quality of Service (QoS), let us suppose that each cloud provider of the federation has replicated in advance part of its services into other federated clouds.

Considering a federation of Infrastructure as Service (IaaS) clouds, service replication means copying Virtual Machines (VMs) disk-image into other federated providers. In this way the cloud providers that wants to apply energy sustainability policies can turn off the blade center hosting its VMs and turn on the copies of these VMs pre-arranged into other federated cloud datacenters, where the renewable energy production is maximum according to temperature, latitude, and time zone.

• Valves • Unit of Control • ... The entire cooling system of a datacenter can be referred also as HVAC (i.e, Heating, Ventilating, Air-Conditioning) or HVAC(R) (Heating, Ventilating, Air-Conditioning, and Refrigerating). Consequently, the total power consumption of a datacenter can be defined as: PT OT = PIT + PPOW + PCOOL

3

Power Consuption Considerations of a Datacenter

The first step for the achievement of a Sustainable Cloud Federation is to better understand the main factors affecting the total power consumption of a datacenter. As already introduced these factors are PIT , PPOW , and PCOOL . PIT is related to the total power consumption of the IT equipment such as:

(1)

Figure 2 shows the total amount of energy consumption of a datacenter. The percentages of the total power spent in a datacenter can be roughly distributed as follows: PIT = 50%; PPOW = 20%; PCOOL = 30%.

(2)

• CPUs • Storage (i.e., Hard Disk, Tapes, Optical Disks, etc.) • RAM • Switches and Router • Monitors • ... PPOW regards the total power consumption of the Electrical equipment, for example, including: • UPS (Uninterruptible Power Supply). • PSU (Power Supply Unit). • PDU (Power Distribution Unit). • Cable (copper wires characterized by an electrical resistance) • Lights • Batteries • ... PCOOL refers cooling equipment including for example: • Chiller. responsible for making the GAP among the external (outdoor) and internal (indoor) temperatures. • FANs, regarding the Control Room Air Conditioning (CRAC) or to equipment used to discard the heat in the external ambient. • Pumps, responsible for moving the refrigerant substance (or water) inside the distribution pipes.

Figure 2: Typical Consumption of Energy inside a Datacenter

PIT and PPOW are strongly related to the transistors performances. In fact, currently, they have a physical limits that it is not possible to be overcame. However, recent studies are trying to break such limits and the expectation is that future innovation can bring to more performance equipment from the point of view of the energy consumption. In this direction, a recent and interesting dissertation was conducted in The Optimist, the Pessimist, and the Global Race to Exascale in 20 Megawatt (Tolentino and Cameron, 2012). Considering the aforementioned assumption, particular consideration deserves PCOOL . We believe that PCOOL will have a big role in the energy consumption studies in the ICT field. In fact, at present PCOOL is the parameter easier to optimize and how it is described later in this manuscript, cloud federation can help to achieve this goal. In order to model a sustainable federated cloud environment, we consider the datacenter of each cloud provider as a black box which acts an ideal refrigeration machine also known as the Carnot Engine. The

performances of this ideal model is only affected by the temperatures: the black box needs to be cooled as much as depending on the environment where it is placed. Moreover we consider that in such a black box there are both energy coming in, and heat that has to be discarded in the external environment. Figure 3 provides a graphic representation of our datacenter modeling. The energy coming from the left-side, in form of electricity is discarded to the environment in form of heat. This assumption seems to be simplistic, but in the bottom part of the Figure it is possible to understand why this simplification is valid for a datacenter. Considering the First Law of Thermodynamic about the conservation of energy, we have the following equation (see Eq. 3) that states any component/device connected to the Electric Grid transform the energy from one form to other ones (Conservation of Energy). Pin = W + Q (3) Where Pin is the input power over the time (i.e., in hours), W is the energy spent for mechanical works and Q is the energy release as Heat. In a datacenter, Pin is the electric power delivered through copper cables, W is the energy for producing movements (Compressor of a Chiller, FANs, Hard Disk with rotors, Optical Readers, etc.). Finally Q is the Heat produced by components, lights, motors, compressors. In a datacenter, if we assume the PIT and/or PPOW , the contribution to W is negligible. The Compressor in a chiller (PCOOL ) catch a lot of energy, but its work is useful for expelling Heat from inside the datacenter to the outside (environment).

the objective is to identify a datacenter in which relocated services can consume as less as possible, hence in which there is a high availability of sustainable energy that costs as less as possible. The measurement of goodness of a datacenter is given by a number called Power Usage Effectiveness (PUE). It is expressed as the ratio from the total amount of energy consumed as input respect to good part of energy used for IT computations. Values of PUE equals to 1 correspond to an energy efficiency of 100%. PT OT (4) PIT The increasing of the PUE value, corresponds to a greater weight of either PPOW or PCOOL contributions (or both). Typical PUE values for a datacenter are greater than 1 and corresponds to 2-2.5. A similar evaluation index is the DCiE (Datacenter Infrastructure Efficiency). It is expressed in percentage, using the following formula: PUE =

In the perspective of a sustainable federated cloud environment the basic idea is to relocate services and resource between cloud providers selecting the more available sustainable datacenter minimizing the total energy consumption. Furthermore, it is also important looking at reducing the impact of incoming renewable energy both for minimizing costs and maximizing the use of renewable sources. In simple terms,

(5)

1 DCiE

(6)

The PUE is: PUE = 100 ∗

Considering a sustainable federated cloud environment the PUE of two different cloud datacenters with the same equipment, but placed in two different regions, may assume different values. The ambient temperature of each region affects the resulting PUE (either DCiE) depending on the adopted cooling techniques.

4

Figure 3: The first law of thermodynamics; the law of conservation of energy.

PIT PT OT

DCiE = 100 ∗

Decision Algorithm

Considering both the scenario previously described in Section 2 and the considerations already pointed out in this work, in the following we are going to introduce a simple algorithm designed to address the problem of establishing the best site on which a given service has to be deployed for minimizing both energy consumption and costs, on the strength of the actual environmental data collected on each. We assume a given workload has to be executed in our four-sites federated scenario in a given time: we know how many resources will be needed to accomplish that task and how many free computational slots will be available on each site. For each one, we also assume to know the availability of the instantaneous amount of electrical power produced from the photo

Figure 4: The representation of a Datacenter: it is the real installation of the DC in Messina. The APC CUBE uses the HVAC cooling

voltaic equipment, the amount of electrical power absorbed from the HVAC(R) (or free cooling system) and that one used for achieving the computation (there will be also other contribution that will be explained in the following). If we want to find a method for optimizing computation respect to costs, we have to identify an analytic approach able to put together all the parameters characterizing the scenario. Since energy providers apply their fares on the actual amount of consumed kWh, our analytic model will take into account the energy contributions as mean value of the electrical power over a time interval. In order to have a good snapshot of the energy production/absorption when the computing element placement have to be performed, in our case this interval corresponds to one hour. Looking at the scenario depicted in Figure 1, and paying attention to the aforementioned considerations, we might assume the energy consumption of each site is represented from the Eq. 7 EGRID (T, G, s) = [EIT (s) + ECOOL (T ) + EPOW − EPV (T, G)] (7) where each energy term represents the produced/absorbed medium power in a time interval of one hour and is expressed in kWh. EGRID (T, G) is the amount of power grid energy needed from a given site and is related to the other energy contribution terms appearing in the formula. According to this latter, it will depend on the external temperature T where the site operates, the G factor describing the availability of the renewable green energy source (e.g, sun radi-

ation, wind intensity, etc.) and the s parameter associated to the number of computational unit allocated (e.g. virtual machines) on that site. In this specific work, we are assuming each datacenter can rely on a green energy source. For the sake of simplicity, we can restrict our considerations to a scenario where each site takes advantage of just renewable photovoltaic energy. As consequence, we can associate the G factor to the sun radiation factor of the energy plant of a given site. According to the above mentioned assumption, The remaining terms of the expression in order are: EIT , a non-constant factor associated to the energy needed from a data center to perform the computation for a given service that depends on the number s of computational slot to be allocated; ECOOL (T ) an energy factor associated to the HVAC(R) system (or free cooling system) that depends on the external temperature T characterizing the area where the site is working (it is related to the measured PUE for that site); EPOW is the constant factor associated to the energy consumption of the power supply equipment of the data center; finally the last term EPV (T, G) is a function of both external temperature T and sun radiation factor G, related to the mean energy amount made available from the photo voltaic system of that site in the time interval of one hour. If we consider that each site retrieves its electrical power from a different provider, we might estimate the costs needed in terms of electrical power to achieve a workload execution according to the applied fares for kWh. With this assumption and starting from the Eq. 7, you can obtain a new expression 8 related to the energy expenses associated to each site func-

Site

Site 1 Site 2 Site 3 Site 4

TemperatureSun Radi(T [°C]) ation (G [MJ/m2 ]) 35 20 30 18 18 14 23 15

Energy Grid Fare (c [$]) 0.08 0.09 0.07 0.08

Photo Voltaic Energy EPV [kW h] 100 150 80 80

Slots (s)

PUE

Costs [$]

120 90 70 75

3.7 3.2 1.2 2.5

10 13 10 15

Table 1: Data retrieved from 4 different sites that will be given as input for the VIM Energy Manager.

tioning: CGRID (T, G, s, c) = EGRID (T, G, s) · c

(8)

In order to cut costs, the organization of our scenario relying on four different computational sites, can select the one for which the cost values associated to expression 8 is minimum for a given set of parameters (i.e. temperature T , sun radiation G, needed computational slots s and fare f ). To achieve this goal, we can consider data related to those parameters is collected from a sensor network on each site periodically. Depending on these obtained values a table is built considering the mean values of the retrieved parameters. Taking into account such an approach, this particular module of the VIM operating on the datacenters, will take care of computing the costs for each site relying on the physical measurements collected from the available sensors and stored in the table. An example is reported in Table 1. A simple algorithm implemented within the Energy Manager will be able to choose the most convenient site (in terms of energy consumption) where allocate the computation associated to a given service. If two different sites are characterized from the same cost values in a given time, the algorithm implemented within the Energy Manager will evaluate also the amount of photo voltaic energy made available from each site, finally preferring the one where this factor is the greatest. In some situations where energy sustainability is preponderant on cost optimization, the same algorithm could be applied in a different way: the site(s) where the photo voltaic energy production is the greatest is first chosen and, in the case of more sites producing the same photo voltaic energy amount, from the retrieved set, the site where the overall costs are the lowest will be finally selected for allocating the services. The algorithm could be implemented creating a complementary module for the VIM (Energy Manager) that retrieves needed data from the sensor network available in each computational site, computing the associated values of energy contributions (considering the instantaneous electrical power values re-

trieved in that moment) and storing them on the table until the next data refresh (after one hour). Although this approach might be implemented in generic VIM, we designed it having in mind CLEVER (it is an IaaS VIM middleware whose details are provided in (Tusa et al., 2010)). This information will be stored in the (central) database through the database manager plugin (according to the CLEVER terminology). When the VIM has to allocate a new set of s virtual machines for a given service, together with snapshot of physical resource availability of each site (reported in Table 1), it will also invoke the Energy Manager to retrieve information on the most convenient site to which deploy the allocation either in terms of cost minimization or energy sustainability. Since each site can offer a limited number of computational slots, if the virtual machine number needed for a given service are greater than the maximum availability for a site, the load will be split across more sites still considering the satisfaction of the same requirements (cost minimization or energy sustainability). Looking at Table 1, if we suppose executing a service that needs 100 virtual machines, in the first case of cost optimization, the site selected by the Energy Manager will be one between Site 1 or Site 3 (as they guarantee the lowest energy costs: respectively 10 $ and 13 $). The final choice will lead to Site 1 since his photo voltaic energy availability (100 kWh) is greater than the one offered by Site 3 (80 kWh). The EPV availability in this situation is preponderant: although the temperature in Site 3 is 18 °C and allows free cooling as refrigeration methods, the CGRID costs coming from Eq. 8 are still more convenient on Site 1 where the “free cost” energy is offered by the photo voltaic equipment. Furthermore, Site 1 has the availability of s = 120 computational slots and is able to directly satisfy the requested service demand. Otherwise the computation would be split among textitSite 1 and Site 3. Still looking at the same table in the alternative situation we have mentioned before (energy sustainability optimization), the first set of selected sites will be formed by Site 3 and Site 4 (either offering the same amount of EPV = 80 kWh). This time, the final choice

will fall on Site 3 as it is able to offer lower grid energy costs than Site 4 (10 $ against 15 $). Differently from the previous case, the available computational slots of this site is lower than the needed ones. In this case, 70 of the requested VMs will be deployed on Site 3 while the remaining ones on Site 4. As reported in the table, the HVAC(R) system of each data center is characterized in terms of efficiency through the PUE (the values in the table refers to a mean value of the coefficient in the time interval of one hour). High PUE values are associated to better refrigerator systems that allows to use less electrical power to push out heat from the data center. The PUE values are tightly related the ECOOL (T ) values contributing in Eq. 7. The best efficiency in terms of energy spent for cooling is achieved on the site 3, where thanks to the low external environmental temperature, it is possible to use the free cooling technique thus reaching a PUE = 1.2.

5

Related Works

In the section hereby, the early part analyzes works falling into energy saving and green energy topics aimed at datacenter. While in latter part several works dealing with cloud and federation are reported. The section concludes the survey with Green IT solutions aimed at clouds. The work we highlight just below show as the problematic we are trying to address is an hot topic indeed. Many works dealing with datacenters and sustainability exist in the scientific literature, however our contribution tries to give an answer in the are of cooperating clouds. The authors in (Moore et al., 2005) for optimizing the energy usage in a datacenter introduced the concept of temperature aware workload placement. They wrote an interesting dissertation that in-deep analyze the effects of cooling cost against the IT computations. A markable assessment also regards the real COP (Coefficient of Performance) determination. The COP curve (with its relative formalization) they presented in this work represents the real evaluation of a chilled-water CRAC unit at the HP Labs Utility Data Center. The work in (Wang et al., 2011) highlights an innovative cooling strategy that leverages thermal storage to cut the electricity bill for cooling. The authors claimed the system does not cause servers in a datacenter to overheat. They worked on Computational Fluid Dynamics (CFD) to consider the realistic thermal dynamics in a datacenter with 1120 servers. A Workload Distribution for Internet datacenters

is proposed in (Abbasi et al., 2010), where the server provisioning algorithm is aware of the temperature distribution in a DC. The authors try to find a way where the utilization constraints (in term of capacity and performance constrains) are satisfied and energy consumption is minimized. The proposed formula contains the COP of a datacenter, under their assessment. Modeling a thermal behavior of a datacenter is a challenging work due to the high number of physical parameters need to be considered. An interesting model along with a close-loop control system is described in (Zhou and Wang, 2011). The authors assessed a datacenter with many CRACs. The inlet temperature of many racks is investigated for accomplishing the Partition in Zone of a datacenter for an efficient decentralized control.

5.1

Cloud and Federation

Technological solutions for cloud computing infrastructures are increasing day by day, and the vision where companies use computational facilities according to a pay-per-use model similarly to other utilities like electricity, gas and water is becoming true. However, the concept of cloud federation is quite recent. Cloud federation refers to mesh of clouds that are interconnected based on open standards to provide a universal decentralized computing environment where everything is driven by constraints and agreements in a ubiquitous, multi-provider infrastructure. In this paragraph, we provide an overview of currently existing solutions in the field, taking into account initiatives born in academia, industry and major research projects. Most of the work in the field concerns the study of architectural models able to efficiently support the collaboration between different cloud providers focusing on various aspects of the federation. The authors in (Buyya and Ranjan, 2010) propose a decentralized technique using a structured peer-to-peer network supporting discovery, deployment and output data collection of a (PaaS) middleware (Aneka). The system is structured as a set of Aneka coordinator peers deployed in each cloud (or in the extreme case on each node of the cloud) offering discovery and coordination mechanisms useful for federation. A central point is represented by the Distributed Hash Table (DHT) overlay which is adapted in order to take into account multidimensional queries (e.g., “find all nodes with Linux operating system, two cores with 2.0 GHz and 2GB or RAM”). However it is not clear how this approach

deals with dynamism as discovery and matchmaking are carried out by a third part node (imposed by the DHT) which can be subject to failures. In our previous work (Celesti et al., 2010) we describe an architectural solution for federation by means of a Cross-Cloud Federation Manager (CCFM), a software component in charge of executing the three main functionalities required for a federation. In particular, the component explicitly manages: i) the discovery phase in which information about other clouds are received and sent, ii) the match-making phase performing the best choice of the provider according to some utility measure and iii) the authentication phase creating a secure channel between the federated clouds. These concepts can be extended taking into account green policies applied in federated scenarios. In (Buyya et al., 2010b), the authors propose a more articulated model for federation composed of three main components. A Cloud Coordinator manages a specific cloud and acts as interface for the external clouds by exposing well-defined cloud operations. The Cloud Exchange component implements the functionality of a registry by storing all necessary information characterizing cloud providers together with demands and offers for computational resources. Lastly, the Cloud Broker represents the touch point for users to enact the federation process; it interacts with the Cloud Exchange to find appropriate cloud providers and with the Cloud Coordinator to define the resource allocation satisfying the needed QoS. A FP7 European Project focusing on cloud federation is RESERVOIR. In Rochwerger et al (see (Rochwerger et al., 2011)), the authors define a RESERVOIR cloud as decentralized federation of collaborating sites. RESERVOIR introduces an abstraction layer that allows developing a set of high level management components that are not tied to any specific environment. In RESERVOIR, several sites can share physical infrastructure resources on which service applications can be executed. Each site is partitioned by a virtualization layer into Virtual Execution Environments (VEEs). These environments are fully isolated runtime modules that abstract away the physical characteristics of the resource and enable sharing. The virtualized computational resources, alongside with the virtualization layer and all the management enablement components, are referred to as the VEE Host. In RESERVOIR a service application is a set of software components which work to achieve a common goal in which each one can be deployed in the same or in different sites. A RESERVOIR cloud federation is homogeneous (i.e., each site has to run the same middleware) and decen-

tralized. It occurs at the IaaS level and needs a static a-priori configuration in each site. The dissertation in (Kiani et al., 2012) describes the large-scale context provisioning. The authors remarked that the adoption of context-aware applications and services has proved elusive so far, due to multi-faceted challenges in cloud computing area. Indeed existing context aware systems are not ideally placed to meet the domain objectives, and facilitate their use in the emerging cloud computing scenarios. The authors identified what the challenges are in heterogeneous cloud contexts. In particular many works are addressed considering the simplified use of a central context management component e.g. a context (cloud) broker. The use of a predominant focus upon designing for static topologies of the interacting distributed components. Presumptions of a single administrative domain or authority and context provisioning within a single administrative, geographic or network domain. Finally they recognized a lack of standardisation with respect to simple, flexible and extensible context models, for the exchange of contextual and control information between heterogeneous actors. We can admit, in the near future, the federation among clouds still needs extra efforts to be concretized.

5.2

Datacenter and Virtualization

In the recent past even more datacenters are looking to increase their flexibility, in particular exploiting the Virtualization Technology using different type of Virtual Machine Managers (VMMs). The VMM is the layer in which the hardware virtualization is accomplished. It hides the physical characteristics of a computing platform, instead showing another abstract computing platform. The software that controls the virtualization used to be called a “control program” at its origins, but nowadays the term Hypervisor is preferred. A Virtual Machine is a software program that emulates a specific hardware system. Each virtual machine is like a “machine within the machine” and runs like just a real physical computer. This software layer emulates the operating system and allocates hardware resources such as the CPU, disk, network controllers, etc. The main Hypervisors currently used to virtualize hardware resources are: Xen (www.xen.org, 2012), KVM (www.linux kvm.org, 2012), VMware (www.vmware.com, 2012), VirtualBox (www.virtualbox.org, 2012), Microsoft HyperV (www.microsoft.com, 2012), Oracle VM (www.oracle.com, 2010), IBM POWER Hypervisor (PR/SM) (publib.boulder.ibm.com, 2010),

Apple Parallel Server (www.apple.com, 2010), etc. For example one type of a VMM can be a physical machine with the Xen hypervisor para-virtualizer (en.wikipedia.org, 2012b) controlling it (in this case the VM are Xen domains), whereas another type can be a machine with the necessary software to host KVM (full-virtualizer (en.wikipedia.org, 2012a)), and so on. Authors in (Krishnan et al., 2011) have studied the run-time behavior of many Virtual Machines achieving a good reference model useful for describing VM workloads. In particular they introduced several models for characterizing CPUs, RAMs, Disk and I/O within VMs under different working conditions (i.e. percentage of load, power consumption, etc.). This work represents a good staring point for evaluating green metrics along with VMs and Hypervisors. The work in (Mukherjee et al., 2009) uses recent technological advances in data center virtualization and proposes cyber-physical, spatio-temporal , thermal-aware job scheduling algorithms that minimize the energy consumption of the data center under certain performance constraints. Authors remark that savings are possible by being able to temporally spread the workload, assign it to energy-efficient computing equipment. They propose a solution able to reduce the heat recirculation and therefore the load on the cooling systems. Their paper provides three categories of thermal-aware energy-saving scheduling techniques that is: first-come first-serve with back-filling, first scheduling algorithm with thermalaware placement, and offline genetic algorithm for scheduling to minimize thermal cross-interferences. The mathematical models they introduced are well defined with a detailed description and good final analysis. A complex analysis in energy-efficient and Cloud computing has been carried out from Buyya et al. in (Buyya et al., 2010a). The authors in this work tried to define an architectural framework and principles for energy-efficient Cloud computing. They claimed their work is useful for investigating energyaware resource provisioning and allocation algorithms that provision datacenter resources to client applications in a way that improves the energy efficiency of the datacenter. Their solution should introduce autonomic and energy-aware mechanisms that self-manage changes in the state of resources effectively and efficiently to satisfy service obligations and achieve energy efficiency. The investigation involves heterogeneous workloads of various types of Cloud applications and develop algorithms for energy-efficient mixing and mapping of VMs to suitable Cloud resources in addition to dynamic con-

solidation of VM resource partitions. However the work the authors presented does not provide an indepth analytic model of what they assessed. The work in (Moghaddam et al., 2011) falls in server consolidation solution leveraging the deployment of VMs inside a datacenter. This VMs consolidation is aimed to carbon footprint minimization. In particular the authors presented a solution for a Low Carbon VPC (they called it LCVPC), and build a carbon footprint model of a reference Virtual Private Cloud (VPC). They introduced a model proven on a simulation platform network. The approach looks similar to our assessment in considering clouds spread around the world and using more energy providers (a wide portfolio with more sources), but authors do not provide any solution on how to apply this model on different clouds. A similar work to our solution, but a very preliminary work has been presented in (Yamagiwa and Uehara, 2012) The authors try to construct a cloud platform which can operate in an area without the electricity such as a disaster situation and/or in a remote place. In their solution there is a battery attached to a PC. The photovoltaic power generation is used to supply this PC. In the paper is not clear what the meaning of cloud is for authors’ point of view. In reality they used an Ubuntu Linux Machine with an Hypervisor. Several VMs are executed, in our view this testbed is not a cloud at all.

6

Conclusions

Nowadays, a sensitive problem is finding the right combination between high performance datacenter and energy sustainability. In this work, considering a scenario of cloud federation, we proposed a methodology for enabling sustainable cooperating clouds. Considering photovoltaic energy generation systems, our approach is based on an energy and temperaturedriven strategies in which the computation workload of a cloud is moved toward the most efficient sustainable federated cloud. According to such a strategy and considering a federated CLEVER-based scenario, we defined an algorithm for the management of VM allocation according to energy and temperaturedriven policies. In future works, we plan to consider also heterogeneous cooperating clouds.

REFERENCES Abbasi, Z., Varsamopoulos, G., and Gupta, S. K. S. (2010). Thermal aware server provisioning and workload dis-

tribution for internet data centers. In HPDC, pages 130–141. Buyya, R., Beloglazov, A., and Abawajy, J. H. (2010a). Energy-efficient management of data center resources for cloud computing: A vision, architectural elements, and open challenges. CoRR, abs/1006.0308. Buyya, R. and Ranjan, R. (2010). Special section: Federated resource management in grid and cloud computing systems. Future Generation Comp. Syst., 26(8):1189–1191. Buyya, R., Ranjan, R., and Calheiros, R. N. (2010b). Intercloud: Utility-oriented federation of cloud computing environments for scaling of application services. In Proceedings of the 10th International Conference on Algorithms and Architectures for Parallel Processing (ICA3PP 2010, pages 21–23. Springer. Celesti, A., Tusa, F., Villari, M., and Puliafito, A. (2010). Three-phase cross-cloud federation model: The cloud sso authentication. In Proceedings of the 2010 Second International Conference on Advances in Future Internet, AFIN ’10, pages 94–101, Washington, DC, USA. IEEE Computer Society. en.wikipedia.org (2012a). Full Virtualization. The virtualization technique used to provide a certain kind of virtual machine environment, that is a complete simulation of the underlying hardware. http://en.wikipedia.org/wiki/Full-virtualization, Jan 2012. en.wikipedia.org (2012b). Paravirtualization. The virtualization technique that presents a software interface to virtual machines that is similar but not identical to that of the underlying hardware. http://en.wikipedia.org/wiki/Paravirtualization, Jan 2012. Kiani, L., Anjum, A., Bessis, N., and Hill, R. (2012). Largescale context provisioning. In 2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems, CISIS 2012. Krishnan, B., Amur, H., Gavrilovska, A., and Schwan, K. (2011). Vm power metering: feasibility and challenges. SIGMETRICS Perform. Eval. Rev., 38(3):56– 60. Moghaddam, F. F., Cheriet, M., and Nguyen, K. K. (2011). Low carbon virtual private clouds. IEEE 4th International Conference on Cloud Computing. IEEE Computer Society. Moore, J., Chase, J., Ranganathan, P., and Sharma, R. (2005). Making scheduling ”cool”: temperatureaware workload placement in data centers. In ATEC’05: Proceedings of the USENIX Annual Technical Conference 2005 on USENIX Annual Technical Conference, page 5, Berkeley, CA, USA. USENIX Association. Mukherjee, T., Banerjee, A., Varsamopoulos, G., Gupta, S. K. S., and Rungta, S. (2009). Spatio-temporal thermal-aware job scheduling to minimize energy consumption in virtualized heterogeneous data centers. Comput. Netw., 53(17):2888–2904. publib.boulder.ibm.com (2010). The POWER Hypervisor, the abstraction layer between the hardware

and firmware and the operating system instances for GX host channel adapter (HCA) implementations. http://publib.boulder.ibm.com/infocenter/powersys/ v3r1m5/, August 2010. Rochwerger, B., Breitgand, D., Epstein, A., Hadas, D., Loy, I., Nagin, K., Tordsson, J., Ragusa, C., Villari, M., Clayman, S., Levy, E., Maraschini, A., Massonet, P., Munoz, H., and Toffetti, G. (2011). Reservoir - when one cloud is not enough. Computer, 44:44–51. Tolentino, M. E. and Cameron, K. W. (2012). The optimist, the pessimist, and the global race to exascale in 20 megawatts. IEEE Computer, 45(1):95–97. Tusa, F., Paone, M., Villari, M., and Puliafito., A. (2010). CLEVER: A CLoud-Enabled Virtual EnviRonment. In 15th IEEE Symposium on Computers and Communications, 2010. ISCC ’10. Riccione. Wang, Y., Wang, X., and Zhang, Y. (2011). Leveraging thermal storage to cut the electricity bill for datacenter cooling. In Proceedings of the 4th Workshop on Power-Aware Computing and Systems, HotPower ’11, pages 8:1–8:5, New York, NY, USA. ACM. www.apple.com (2010). The Mac: Real versatility through virtualization. http://www.apple.com/business/solutions/it/virtualization.html, August 2010. www.linux kvm.org (2012). KVM (for Kernel-based Virtual Machine) is a full virtualization solution for Linux on x86 hardware. http://www.linux-kvm.org, August 2012. www.microsoft.com (2012). Microsoft Hyper-V Server 2008 R2, the stand-alone product that provides a reliable and optimized virtualization solution. http://www.microsoft.com/hyper-vserver/en/us/default.aspx August 2012. www.oracle.com (2010). Oracle VM, ther virtualization software that fully supports both Oracle and non-Oracle applications, and delivers more efficient performance. http://www.oracle.com/us/technologies/virtualization/oraclevm/, August 2010. www.virtualbox.org (2012). x86 virtualization software package developed by Oracle. http://www.virtualbox.org/, Jan 2012. www.vmware.com (2012). Virtualization - the essential catalyst for enabling the transition to secure cloud computing: http://www.vmware.com/it/, Jan 2012. www.xen.org (2012). Xen Hypervisor - Leading open source hypervisor for servers: http://www.xen.org/, Jan 2012. Yamagiwa, M. and Uehara, M. (2012). A proposal for development of cloud platform using solar power generation. In 2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems, CISIS 2012. Zhou, R. and Wang, Z. (2011). Modeling and control for cooling management of data centers with hot aisle containment. In ASME 2011 International Mechanical Engineering Congress & Exposition.