12
Renewable and Cooling Aware Workload Management for Sustainable Data Centers * Zhenhua Liu 1 , Yuan Chen 2 , Cullen Bash 2 , Adam Wierman 1 , Daniel Gmach 2 , Zhikui Wang 2 , Manish Marwah 2 , and Chris Hyser 2 1 Computing & Mathematical Sciences, California Institute of Technology, Pasadena, CA, USA, E-mail: {zhenhua, adamw}@caltech.edu 2 Sustainable Ecosystem Research Group, HP Labs, Palo Alto, CA, USA, E-mail: fi[email protected] ABSTRACT The demand for data center computing increased signifi- cantly in recent years resulting in huge energy consumption. Data centers typically comprise three main subsystems: IT equipment provides services to customers; power infrastruc- ture supports the IT and cooling equipment; and the cool- ing infrastructure removes the generated heat. This work presents a novel approach to model the energy flows in a data center and optimize its holistic operation. Tradition- ally, supply-side constraints such as energy or cooling avail- ability were largely treated independently from IT workload management. This work reduces cost and environmental im- pact using a holistic approach that integrates energy supply, e.g., renewable supply and dynamic pricing, and cooling sup- ply, e.g., chiller and outside air cooling, with IT workload planning to improve the overall attainability of data cen- ter operations. Specifically, we predict renewable energy as well as IT demand and design an IT workload management plan that schedules IT workload and allocates IT resources within a data center according to time varying power sup- ply and cooling efficiency. We have implemented and eval- uated our approach using traces from real data centers and production systems. The results demonstrate that our ap- proach can reduce the recurring power costs and the use of non-renewable energy by as much as 60% compared to ex- isting, non-integrated techniques, while still meeting opera- tional goals and Service Level Agreements. 1. INTRODUCTION Data centers are emerging as the “factories” of this genera- tion. A single data center requires a considerable amount of electricity and data centers are proliferating worldwide as a result of increased demand for IT applications and services. As a result, concerns about the growth in energy usage and emissions have led to social interest in curbing their energy usage and emissions. These concerns have led to research ef- forts in both industry and academia. An emerging solution is the incorporation of renewable on-site energy supplies and alternative cooling supplies into the design of data centers. * This work is done during Zhenhua Liu’s internship at HP Labs. Zhenhua Liu and Adam Wierman are also partly supported by NSF grant CNS 0846025 and DoE grant DE- EE0002890. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. SIGMETRICS’12, June 11–15, 2012, London, UK. Copyright 2012 ACM 978-1-4503-0262-3/11/06 ...$10.00. The problem addressed by this paper is how to use these re- sources most effectively during the operation of data centers. Most of the efforts toward this goal focus on improving the efficiency in one of the three major data center “silos”: (i) IT, (ii) cooling, and (iii) power. Significant progress has been made in optimizing the energy efficiency of each of the three “silos” enabling sizeable reductions in data center en- ergy use, e.g., [13, 14, 24, 41, 11, 33, 39, 6, 32]; however, the integration of these silos is clearly an important next step. To this end, a second generation of solutions for data cen- ters has begun to emerge. This work focuses on the inte- gration of different “silos”, e.g., [20, 28, 30, 10]. An example of such work is the dynamic thermal management of air- conditioners based on load at the IT rack level [20, 9]. How- ever, to this point, supply-side constraints such as renewable energy and cooling availability are still largely treated inde- pendently from IT workload management, e.g., scheduling. Particularly, current workload management solutions are not designed and operated to take advantage of time variations in renewable energy availability and cooling efficiencies. The potential of integrated, dynamic approaches has been realized in some other domains, e.g., cooling management solutions for buildings that predict weather and power prices to dynamically adapt the cooling control have been proposed [4]. The goal of this paper is to start to realize this potential in data centers. Intuitively, the potential of an integrated approach can be seen from the following three observations. First, most data centers support a range of IT workloads, including both critical interactive applications that run 24x7, e.g., Internet services, and delay tolerant, batch-style ap- plications, e.g., scientific applications, simulations, financial analysis, and image processing, which we refer to as batch workloads. Generally, batch workloads can be scheduled to run anytime as long as they finish before deadlines. This enables significant flexibility for workload management. Second, the availability and cost of the power supply, e.g., renewable energy supply and electricity price, is often dy- namic over time, and so dynamic control of the supply mix can help to lower CO2 emissions and offset costs. Thus, thoughtful workload management can have a great impact on energy consumption and costs, e.g., by scheduling non- critical IT demand in a manner that follows the availability of the renewable generation. Third, many data centers nowadays are cooled by multiple means through a cooling micro grid, e.g., combining tradi- tional mechanical chillers, airside economizers, and waterside economizers. Within a micro grid, each cooling approach has a different efficiency and capacity that depends on IT work- load, cooling generation mechanism and external conditions, e.g., outside air temperature, humidity. This provides op- portunities to optimize cooling cost by “shaping” IT demand according to time varying cooling efficiency and capacity, etc. The three observations above highlight that there is con- siderable potential for integrated management of the IT,

Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

Renewable and Cooling Aware Workload Management forSustainable Data Centers∗

Zhenhua Liu1, Yuan Chen2, Cullen Bash2, Adam Wierman1, Daniel Gmach2, Zhikui Wang2, Manish Marwah2, and Chris Hyser2

1Computing & Mathematical Sciences, California Institute of Technology, Pasadena, CA, USA, E-mail: {zhenhua, adamw}@caltech.edu

2Sustainable Ecosystem Research Group, HP Labs, Palo Alto, CA, USA, E-mail: [email protected]

ABSTRACTThe demand for data center computing increased signifi-cantly in recent years resulting in huge energy consumption.Data centers typically comprise three main subsystems: ITequipment provides services to customers; power infrastruc-ture supports the IT and cooling equipment; and the cool-ing infrastructure removes the generated heat. This workpresents a novel approach to model the energy flows in adata center and optimize its holistic operation. Tradition-ally, supply-side constraints such as energy or cooling avail-ability were largely treated independently from IT workloadmanagement. This work reduces cost and environmental im-pact using a holistic approach that integrates energy supply,e.g., renewable supply and dynamic pricing, and cooling sup-ply, e.g., chiller and outside air cooling, with IT workloadplanning to improve the overall attainability of data cen-ter operations. Specifically, we predict renewable energy aswell as IT demand and design an IT workload managementplan that schedules IT workload and allocates IT resourceswithin a data center according to time varying power sup-ply and cooling efficiency. We have implemented and eval-uated our approach using traces from real data centers andproduction systems. The results demonstrate that our ap-proach can reduce the recurring power costs and the use ofnon-renewable energy by as much as 60% compared to ex-isting, non-integrated techniques, while still meeting opera-tional goals and Service Level Agreements.

1. INTRODUCTIONData centers are emerging as the “factories” of this genera-

tion. A single data center requires a considerable amount ofelectricity and data centers are proliferating worldwide as aresult of increased demand for IT applications and services.As a result, concerns about the growth in energy usage andemissions have led to social interest in curbing their energyusage and emissions. These concerns have led to research ef-forts in both industry and academia. An emerging solutionis the incorporation of renewable on-site energy supplies andalternative cooling supplies into the design of data centers.

∗This work is done during Zhenhua Liu’s internship at HPLabs. Zhenhua Liu and Adam Wierman are also partlysupported by NSF grant CNS 0846025 and DoE grant DE-EE0002890.

Permission to make digital or hard copies of all or part of this work forpersonal or classroom use is granted without fee provided that copies arenot made or distributed for profit or commercial advantage and that copiesbear this notice and the full citation on the first page. To copy otherwise, torepublish, to post on servers or to redistribute to lists, requires prior specificpermission and/or a fee.SIGMETRICS’12, June 11–15, 2012, London, UK.Copyright 2012 ACM 978-1-4503-0262-3/11/06 ...$10.00.

The problem addressed by this paper is how to use these re-sources most effectively during the operation of data centers.

Most of the efforts toward this goal focus on improvingthe efficiency in one of the three major data center “silos”:(i) IT, (ii) cooling, and (iii) power. Significant progress hasbeen made in optimizing the energy efficiency of each of thethree “silos” enabling sizeable reductions in data center en-ergy use, e.g., [13, 14, 24, 41, 11, 33, 39, 6, 32]; however, theintegration of these silos is clearly an important next step.

To this end, a second generation of solutions for data cen-ters has begun to emerge. This work focuses on the inte-gration of different “silos”, e.g., [20, 28, 30, 10]. An exampleof such work is the dynamic thermal management of air-conditioners based on load at the IT rack level [20, 9]. How-ever, to this point, supply-side constraints such as renewableenergy and cooling availability are still largely treated inde-pendently from IT workload management, e.g., scheduling.Particularly, current workload management solutions are notdesigned and operated to take advantage of time variationsin renewable energy availability and cooling efficiencies.

The potential of integrated, dynamic approaches has beenrealized in some other domains, e.g., cooling managementsolutions for buildings that predict weather and power pricesto dynamically adapt the cooling control have been proposed[4]. The goal of this paper is to start to realize this potentialin data centers. Intuitively, the potential of an integratedapproach can be seen from the following three observations.

First, most data centers support a range of IT workloads,including both critical interactive applications that run 24x7,e.g., Internet services, and delay tolerant, batch-style ap-plications, e.g., scientific applications, simulations, financialanalysis, and image processing, which we refer to as batchworkloads. Generally, batch workloads can be scheduled torun anytime as long as they finish before deadlines. Thisenables significant flexibility for workload management.

Second, the availability and cost of the power supply, e.g.,renewable energy supply and electricity price, is often dy-namic over time, and so dynamic control of the supply mixcan help to lower CO2 emissions and offset costs. Thus,thoughtful workload management can have a great impacton energy consumption and costs, e.g., by scheduling non-critical IT demand in a manner that follows the availabilityof the renewable generation.

Third, many data centers nowadays are cooled by multiplemeans through a cooling micro grid, e.g., combining tradi-tional mechanical chillers, airside economizers, and watersideeconomizers. Within a micro grid, each cooling approach hasa different efficiency and capacity that depends on IT work-load, cooling generation mechanism and external conditions,e.g., outside air temperature, humidity. This provides op-portunities to optimize cooling cost by “shaping” IT demandaccording to time varying cooling efficiency and capacity, etc.

The three observations above highlight that there is con-siderable potential for integrated management of the IT,

Page 2: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

IT Power

Cooling Power

IT Equipment

(Interactive Workloads / Batch Jobs)

Outside Air

Economizer

Chiller

Plant

Water

Economizer

Grid Power

Wind Power

PV Power

CRACSCRAC

Power

Energy Storage

Pow

er I

nfr

astr

uct

ure

Figure 1: Sustainable Data Center

cooling, and power subsystems of data centers. Providingsuch an integrated solution is the goal of this work. Specif-ically, we provide a novel workload scheduling and capacitymanagement approach that integrates energy supply, e.g.,renewable supply, dynamic energy pricing, and cooling sup-ply, e.g., chiller cooling, outside air cooling, into IT workloadmanagement to improve the overall energy efficiency and re-duce the carbon footprint of data center operations.

A key component of our approach is demand shifting,which schedules non-critical IT workload and allocates ITresources within a data center according to the availabilityof renewable power supply and the efficiency of cooling. Thisis a complex optimization problem due to the dynamism inthe supply and demand and the interaction between them.For example, on one hand, given the lower electricity priceand cooling cost of outside air cooling at night, batch jobsshould be scheduled to run at night. On the other hand, be-cause more renewable power, e.g., solar, is available aroundnoon, we should do more work during the day to reduce re-curring power cost and environmental impact.

At the core of our design is a model of the costs within thedata center, which is used to formulate a constrained con-vex optimization problem. The workload planner solves thisoptimization in order to determine the optimal demand shift-ing. An optimization-based approach for workload manage-ment has been popular in the research community recently,e.g., [22, 27, 34, 40, 24, 30, 26, 23]. The key contributions ofthe formulation considered here compared to the prior litera-ture are (i) the addition of a detailed cost model for and opti-mization of the cooling component of the data center, whichis typically ignored in the optimization-based designs; (ii)the consideration of both interactive and batch workloads;and (iii) the derivation of important structural properties ofthe solution to the optimization.

In order to validate our integrated design, we have imple-mented a prototype of our approach for a data center thatincludes solar power and outside air cooling. Using our im-plementation, we perform a number of experiments on a realtestbed to highlight the practicality of the approach (Section5). In addition to validating our design, our experiments arecentered on providing insights into the following questions:

(1) How much benefit (operational cost, environmental im-pact) can be obtained from renewable and cooling-awareworkload management planning?

(2) Is net-zero1 grid power consumption [15] achievable?

(3) Which renewable source is more valuable? What is theoptimal renewable portfolio?

1By “net-zero” we mean that the total energy usage over afixed period is less than or equal to the local total renewablegeneration during that period. Note that this does not meanthat no power from the grid is used during this period.

24 48 72 96 120 144 1680

50

100

150

hour

rene

wabl

e ge

nera

tion

(kW

)

solarwind

Figure 2: One week renewable generation

24 48 72 96 120 144 1680

2

4

6

hour

elec

tricit

y pr

ice (c

ents

/kW

h)

Figure 3: One week real-time electricity price

2. SUSTAINABLE DATA CENTEROVERVIEW

Figure 1 depicts an architectural overview of a sustainabledata center design. The IT equipment includes servers, stor-age and networking switches that support applications andservices hosted in the data center. The power infrastructuregenerates and delivers power for the IT equipment and cool-ing facility through a power micro-grid that integrates gridpower, local renewable power supplies such as photovoltaic(PV) and wind and energy storage. The cooling infrastruc-ture provides, delivers and distributes the cooling resourcesto extract the heat from the IT equipment. In this example,the cooling capacity is delivered to the data center throughthe Computer Room Air Conditioning Units (CRACs) fromthe cooling micro grid that consists of air economizer, watereconomizer and traditional chiller plant. We discuss thesethree key subsystems in detail in the following sections.

2.1 Power InfrastructureAlthough renewable energy is in general more sustainable

than grid power, the supply is often time-varying in a mannerthat depends on the source that provides the power, the lo-cation of power generators, and the local weather conditions.Figure 2 shows the power generated from a 130kW PV in-stallation for a data center and nearby 100kW wind turbinein California for one week. The PV generation shows regularvariation while that from the wind is much less predictable.How to manage these supplies is a big challenge for applica-tion of renewable power in a sustainable data center.

Despite the usage of renewable energy, data centers muststill rely on non-renewable energy, including off-site grid powerand on-site energy storage, due to availability concerns. Gridpower can be purchased at either pre-defined fixed rate or on-demand time-varying rate, and Figure 3 shows an example oftime-varying electricity price over 24 hours. Note that theremight also be an additional charge for the peak demand.

Local energy storage technologies can be used to store andsmooth the supply of power for a data center. A varietyof technologies are available [3], including flywheels, batter-ies, and other systems. Each has its costs, advantages anddisadvantages. Energy storage is still quite expensive andthere is power loss associated with energy conversion andcharge/discharge. Hence, it is critical to maximize the useof the renewable energy that is generated on site. An idealscenario is to maximize the use of renewable energy whileminimizing the use of storage.

Page 3: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

2.2 Cooling SupplyDue to the ever-increasing power density of IT equipment

in today’s data centers, a tremendous amount of electricity isused by the cooling infrastructure. According to [32] , about1/3 to 1/2 of the data center total power consumption goesto the cooling system including CRAC units, pumps, chillerplant and cooling towers.

Lots of work has been done to improve the cooling ef-ficiency through, e.g., smart facility design, real-time con-trol and optimization [6, 39, 32]. Traditional data centersuse chillers to cool down the returned hot water from theCRAC (Computer Room Air Conditioning Units) throughthe mechanical refrigeration cycles since they can providehigh cooling capacity continuously. However, compressors ofthe chillers consume a large amount of power [43, 31]. Re-cently, “chiller-less” cooling technologies have been adoptedto remove or reduce the dependency on mechanical chillers.In the case with water-side economizers, the returned hot wa-ter is cooled down by facilities such as dry coolers or evapo-rative cooling towers. The cooling capacity may also be gen-erated from cold water from seas or lakes [43]. In the caseof air economizers, cold outside air may be introduced af-ter pre-processed through filtering and/or humidification/de-humidification to cool down the IT equipment directly whilethe hot air can be rejected back into the environment.

However, these so-called “free” cooling approaches are ac-tually not really free [43]. First, there is still non-negligibleenergy cost associated with these approaches, e.g., blowersdriving outside air through data center need to work againstthe air flow resistance and therefore consume power. Second,the efficiency of these approaches can be greatly affected byenvironmental conditions such as ambient air temperatureand humidity, compared with that of traditional approachesbased on mechanical chillers. The cooling efficiency and ca-pacity of the economizers can vary widely along with time ofthe day, season of the year, and geographical locations of thedata centers. These approaches are usually complementedby more stable cooling resources such as chillers, which pro-vides opportunities to optimize the cooling power usage by“shaping” IT demand according to cooling effectiveness.

2.3 IT WorkloadThere are many different workloads in a data center. Most

of them fit into two categories: interactive or transactionaland non-interactive or batch. The interactive workloads suchas internet services or business transactional applicationstypically run 24x7 and process real time user requests, whichhave to be completed within a certain time (i.e., responsetime). Non-interactive batch jobs such as scientific applica-tions, simulations, financial analysis, and image processingare often delay tolerant and can be scheduled to run anytimeas long as progress can be made and the jobs finish beforethe deadline (i.e., completion time). This provides an opti-mization opportunity for workload management to “reshape”non-critical demand based on the varying renewable energysupply and cooling supply.

Interactive workloads are characterized by different stochas-tic properties for request arrival, service demand, and ServiceLevel Agreements (SLAs, e.g., average response time or per-centile delay). Figure 4 shows a 7-day normalized workloaddemand (CPU usage) trace for a popular photo sharing andstorage web service, which has more than 85 million regis-tered users in 22 countries. demand .We can see that theworkload shows significant variability and exhibits a cleardiurnal pattern, which is typical for most interactive work-loads in data centers.

Batch jobs are defined in terms of total resource demand(e.g, CPU hours), starting time, completion time as well asmaximum resource consumption (e.g., a single thread pro-gram can use up to 1 cpu). Conceptually, a batch job can

24 48 72 96 120 144 1680

0.2

0.4

0.6

0.8

1

hour

norm

alize

d CP

U de

man

d

Figure 4: One week interactive workload

run at anytime with different resources as long as it finishesbefore the specified completion time. Our integrated man-agement approach exploits this flexibility to make use of re-newable energy and efficient cooling when they are available.

3. MODELING AND OPTIMIZATIONAs discussed above, time variations in renewable energy

supply availability and cooling efficiencies provide both op-portunities and challenges for managing IT workloads in datacenters. In this section, we present a novel design for renew-able and cooling aware workload management that exploitsopportunities available to improve the attainability of datacenters. In particular, we formulate an optimization problemfor adapting the workload scheduling and capacity allocationto varying supply from power and cooling infrastructure.

3.1 Optimizing the cooling substructureWe first derive the optimal cooling substructure when mul-

tiple cooling approaches are available in the substructure.We consider two cooling approaches: the outside air cool-ing which supplies most of the cooling capacity, and coolingthrough mechanical chillers which guarantees availability ofcooling capacity. By exploring the heterogeneity of the ef-ficiency and cost of the two approaches, we represent theminimum cooling power of the substructure as a function ofthe IT heat load.

In the following discussion, we define cooling effectiveness2

as the cooling power divided by the IT power to represent thecooling efficiency. By cooling capacity we mean how muchheat the cooling system can extract from the IT equipmentand reject into the environment. In the case of outside aircooling, the cold air from outside is assumed pushed intothe return ends of the CRAC units while the hot air fromthe outlets of the server racks is exhausted to the ambientenvironment through ducts.

Outside Air CoolingThe energy usage of outside air cooling is mainly the powerconsumed by blowers, which can be approximated as a cubicfunction of the blower speed [8, 43]. We assume that capacityof the outside air cooling is under tight control, e.g., throughblower speed tuning, to avoid over-provisioning. Then theoutside air capacity is equal to the IT heat load at the steadystate when the latter does not exceed the total air cooling ca-pacity. Based on basic thermal transfer theory [16], the cool-ing capacity is proportional to the air volume flow rate. Theair volume flow rate is approximately proportional to blowerspeed according to the general fan laws [8, 43]. Therefore,outside air cooling power can be defined as a function of ITdemand d as fa(d) = kd3, 0 ≤ d ≤ d, k > 0, which is a convexfunction. The parameter k depends on the air conditions aswell. It is actually proportional to the temperature differ-ence, (tRA − tOA), where tOA is the outside air temperatureand tRA is the temperature of the (hot) exhausting air fromthe IT racks based again on the thermal transfer theory. Asone example, Figure 5(a) shows the cooling effectiveness for

2Similar to PUE, a larger number here implies low efficiency.

Page 4: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

0 10 20 30 40 500

0.1

0.2

0.3

0.4

0.5

IT power (kW)

cool

ing

effe

ctive

ness

OAT = 30OAT = 25OAT = 20

(a) Outside air cooling

0 10 20 30 40 500

0.1

0.2

0.3

0.4

0.5

IT power (kW)

cool

ing

effe

ctive

ness

OAT = 30OAT = 25OAT = 20

(b) Chiller cooling

Figure 5: Cooling ineffectiveness comparison

an outside air cooling system (assuming the exhausting airtemperature is 35◦C) under different outside air tempera-tures (OAT). The maximum capacity of this cooling system

can be modeled as d = α(tRA − tOA). The parameter α > 0represents the maximal outside air mass flow rate when theblowers run at the highest possible speed.

Chilled Water CoolingFirst-principle models of chilled water cooling systems, in-cluding the chiller plant, cooling towers, pumps and heatexchangers, have been notoriously complicated [16, 9, 31].

In this paper, we consider an empirical chiller efficiencymodel that was built on actual measurement of an oper-ational chiller [31]. Figure 5(b) shows the effectiveness ofthe chiller. Different from the outside air cooling, the chillercooling effectiveness does not change much with temperatureand the variation over different IT load is much smaller thanthat under outside air cooling. In the following analysis, thechiller power consumption is approximated as fc(d) = γd,where d is again the IT demand and γ > 0, a constant thatdepends on the chiller. This assumption is only for simplicityreasons and our analysis actually applies to the case of anarbitrary strictly increasing and convex chiller cost function.

Cooling optimizationAs shown in Figure 5, the efficiency of outside air cooling ismore sensitive to IT load and the OAT than is that of chillercooling. Furthermore, the cost of outside air cooling is higherthan that of the chiller when the IT load exceeds a certainvalue because its power increases very fast (super-linearly)as the IT demand increases, in particular for high ambienttemperatures. The heterogeneous cooling efficiencies of thetwo approaches and the varying properties of along with airtemperature and heat load provide opportunities to optimizethe cooling cost by using proper cooling capacity from eachcooling supply as we discuss below, or by manipulating theheat load as we show in later sections.

For a given IT power demand d and outside air tempera-ture tOA, there exists an optimal cooling capacity allocationbetween outside air cooling and chiller cooling. Assume thecooling capacities provided by the chiller and outside air ared1 and d2 respectively (d1 = d−d2). From the cooling modelsintroduced above, the optimal cooling power consumption is

c(d) = mind2∈[0,d]

γ(d− d2)+ + kd32 (1)

This can be solved analytically, which yields

d∗2 =

{d if d ≤ dsds otherwise

where ds = min{√

γ/3k, d}

, and the optimal outside air

cooling capacity is d∗1 = d− d∗2. So,

c(d) =

{kd3 if d ≤ dskd3

s + γ(d− ds) otherwise(2)

0 10 20 30 40 500

2

4

6

8

10

IT power (kW)

optim

al c

oolin

g po

wer (

kW)

OAT = 30OAT = 25OAT = 20

Figure 6: Optimal cooling power

is the cooling power of the optimal substructure, which isused for the optimization in later sections. Figure 6 illus-trates the relationship between cooling power and IT loadfor different ambient temperatures. We see that the coolingpower is a convex function of IT power, higher with hotteroutside air. We also prove that the optimal c(d) is a convexfunction of d in [25], which is important for using this as acomponent of the overall optimization for workload manage-ment.

3.2 System ModelWe consider a discrete-time model whose timeslot matches

the timescale at which the capacity provisioning and schedul-ing decisions can be updated. There is a (possibly long) timeperiod we are interested in, {1, 2, ..., T}. In practice, T couldbe a day and a timeslot length could be 1 hour. The man-agement period can be either static, e.g., to perform thescheduling every day for the execution of the next day, ordynamic, e.g., to create a new plan if the old scheduling dif-fers too much from the actual supply and demand. The goalof the workload management is at each time t to:

(i) Make the scheduling decision for each batch job.

(ii) Choose the energy storage usage.

(iii) Optimize the cooling infrastructure so as to minimizethe total cost during [1, T ].

We assume the renewable supply at time t is r(t), whichmay be a mix of different renewables, such as wind and PVsolar. We denote the grid power price at time t by p(t)and assume p(t) > 0 without loss of generality3. To modelenergy storage, we denote the energy storage level at timet by es(t) with initial value es(0) and the discharge/chargeat time t by e(t), where positive or negative values meandischarge or charge, respectively. Also, there is a loss rateρ ∈ [0, 1] for energy storage. We therefore have the relationes(t+1) = ρ(es(t)−e(t)) between successive timeslots and werequire 0 ≤ es(t) ≤ ES,∀t, where ES is the energy storagecapacity. Though there are more complex energy storagemodels [37], they are beyond the scope of the paper.

Assume there are I interactive workloads. For interactiveworkload i, the arrival rate at time t is λi(t), the mean servicerate is µi and the target performance metrics (e.g., averagedelay, or 95th percentile delay) is rti. In order to satisfy thesetargets, we need to allocate interactive workload i with ITcapacity ai(t) at time t. Here ai(t) is derived from analyticmodels (e.g., M/GI/1/PS, M/M/k) or system measurementsas a function of λi(t) because performance metrics strictlyimprove as the capacity allocated to the workload increases,hence there is a sharp threshold for ai(t). Note that oursolution is quite general and does not depend on a particularmodel.

Assume there are J classes of batch jobs. Class j batchjobs have total demand Bj , maximum parallelization MPj ,starting time Sj and deadline Ej . Let bj(t) denote theamount of capacity allocated to class j jobs at time t. Wehave 0 ≤ bj(t) ≤ MPj ,∀t and

∑t bj(t) ≤ Bj . Given the

3If at some time t we have negative price, we will use upthe total capacity at this time, then we only need to makedecisions for other timeslots.

Page 5: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

above definitions, the total IT demand at time t is given by

d(t) = Σiai(t) + Σjbj(t). (3)

We assume the total IT capacity is D, so 0 ≤ d(t) ≤ D,∀t.Note here d(t) is not constant, but instead time-varying as aresult of dynamic capacity provisioning.

3.3 Cost and Revenue ModelThe cost of a data center includes both capital and operat-

ing costs. Our model focuses on the operational, energy cost.Meanwhile, by servicing the batch jobs, the data center canobtain revenue. We model the data center cost by combin-ing the energy cost and revenue from batch jobs. Note that,to simplify the model, we do not include the switching costsassociated with cycling servers in and out of power-savingmodes; however, the approach of [24] provides a natural wayto incorporate such costs if desired.

To capture the variation of the energy cost over time, welet g(t, d(t), e(t)) denote the energy cost of the data centerat time t given the IT demand d(t), optimal cooling power,renewable generation, electricity price, and energy storageusage e(t). For any t, we assume that g(t, d(t), e(t)) is non-decreasing in d(t), non-increasing in e(t), and jointly convexin d(t) and e(t).

This formulation is quite general, and captures, for exam-ple, the common charging plan of a fixed price per kWh plusan additional “demand charge” for the peak of the averagepower used over a sliding 15 minute window [29], in whichcase the energy cost function consists of two parts:

p(t) (d(t) + c(d(t))− r(t)− e(t))+

and

ppeak(

maxt

(d(t) + c(d(t))− r(t)− e(t))+),

where p(t) is the fixed/variable electricity price per kWh,and ppeak is the peak demand charging rate. We could alsoinclude a sell-back mechanism and other charging policies.Additionally, this formulation can capture a wide range ofmodels for server power consumption, e.g., energy costs asan affine function of the load, see [13], or as a polynomialfunction of the speed, see [41, 5].

We model only the variable component of the revenue4,which comes from the batch jobs that are chosen to be run.Specifically, the data center gets revenue R(b), where b isthe matrix consisting of bj(t), ∀j, ∀t. In this paper, we focuson the following, simple revenue function

R(b) = ΣjRj(Σt∈[Sj ,Ej ]bj(t)

),

where Rj is the per-job revenue.(Σt∈[Sj ,Ej ]bj(t)

)captures

the amount of class j jobs finished before their deadline.

3.4 Optimization ProblemWe are now ready to formulate the renewable and cooling

aware workload management optimization problem. Our op-timization problem takes as input the renewable supply r(t),electricity price p(t), optimal cooling substructure c(d(t)),and IT workload demand ai(t), Bj and related information(the starting time Sj , deadline Ej ,maximum parallizationMPj), IT capacity D, energy storage capacity ES, and gen-erates an optimal schedule of each timeslot for batch jobsbj(t) and energy storage usage e(t), according to the avail-ability of renewable power and cooling supply such that spec-ified SLAs (e.g., deadlines) and operational goals (e.g., min-imizing operational costs, etc) are satisfied.

4Revenue is also derived from the interactive workload, butfor the purposes of workload management the amount ofrevenue from this workload is fixed.

This is captured by the following optimization problem:

minb,e

Σtg(t, d(t), e(t))− ΣjRj(Σt∈[Sj ,Ej ]bj(t)

)(4a)

s.t.Σtbj(t) ≤ Bj , ∀j (4b)

es(t+ 1) = ρ(es(t)− e(t)), ∀t (4c)

0 ≤ bj(t) ≤MPj , ∀j, ∀t (4d)

0 ≤ d(t) ≤ D, ∀t (4e)

0 ≤ es(t) ≤ ES, ∀t (4f)

Here d(t) is given by (3). (4b) means the amount of servedbatch jobs cannot exceed the total demand, and could be-come Σtbj(t) = Bj if finishing all class j batch job is re-quired. (4c) updates the energy storage level of each times-lot. We also incorporate constraints on maximum paralleliza-tion (4d), IT capacity (4e), and energy storage capacity (4f).We may have other constraints, such as a “net zero” con-straint that the total energy consumed be less than the totalrenewable generation within [1, T ], i.e.

∑t(d(t) + c(d(t))) ≤∑

t r(t). Note that, though highly detailed, this formulationdoes ignore some important concerns of data center design,e.g., reliability and availability. Such issues are beyond thescope of this paper; nevertheless, our designs merge nicelywith proposals such as [36] for these goals.

In this paper, we restrict our focus from optimization (4a)to (5a), but the analysis can be easily extended to otherconvex cost functions.

minb,e

Σtp(t)(d(t) + c(d(t))− r(t)− e(t))+ − ΣjRj

(Σt∈[Sj ,Ej ]bj(t)

)(5a)

s.t.Σtbj(t) ≤ Bj , ∀j (5b)

es(t+ 1) = ρ(es(t)− e(t)), ∀t (5c)

0 ≤ bj(t) ≤MPj , ∀j,∀t (5d)

0 ≤ d(t) ≤ D, ∀t (5e)

0 ≤ es(t) ≤ ES, ∀t (5f)

Note that this optimization problem is jointly convex inbj(t) and e(t) and can be efficiently solved in polynomialtime in the worst-case, and quite quickly in practice.

Given the significant amount of prior work approachingdata center workload management via convex optimization[22, 27, 34, 40, 24, 30, 26], it is important to note the keydifferences between our formulation and prior work–our for-mulation is the first, to our knowledge, to incorporate renew-able generation, storage, an optimized cooling micro-grid,and batch job scheduling with consideration of both pricediversity and temperature diversity. Prior formulations haveincluded only one or two of these features. This “univer-sal” inclusion is what allows us to consider truly integratedworkload management.

3.5 Properties of the optimal workloadmanagement

The usage of the workload management optimization de-scribed above depends on more than just the ability to solvethe optimization quickly. In particular, the solutions foundmust be practical if they are to be adopted in actual datacenters.

In this section, we provide characterizations of the optimalsolutions to the workload management optimization, whichhighlight that the structure of the optimal solutions facil-itates implementation. Specifically, one might worry thatthe optimal solution requires highly complex scheduling ofthe batch jobs, which could be impractical. For example,

Page 6: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

if a plan schedules too many jobs at a time, it may not bepractical because there is often an upper limit on how manyworkloads can be hosted in a physical server. The followingresults here show that such concerns are unwarranted.

Energy usage and costAlthough it is easily seen that the workload managementoptimization problem has at least one optimal solution,5 ingeneral, this solution is not unique. Thus, one may worrythat the optimal solutions might have very different prop-erties with respect to energy usage and cost, which wouldmake capacity planning difficult. However, it turns out thatthe optimal solution, though not unique, has nice propertieswith respect to energy usage and cost.

In particular, we prove that all optimal solutions use thesame amount of power from the grid at all times. Thus,though the scheduling of batch jobs and the usage of energystorage might be very different, the aggregate usage of powerfrom the grid is always the same. This is a nice featurewhen considering capacity planning for the power system.Formally, this is summarized by the following theorem, whichis proved in [25].

Theorem 1. For the simplified energy cost model (5a),suppose the optimal cooling power c(d) is strictly convex ind. Then, the energy usage from the grid, (d(t) + c(d(t)) −r(t) − e(t))+, at each time t is common across all optimalsolutions.

Though Theorem 1 considers a general setting, it is notgeneral enough to include the optimal cooling substructurediscussed in Section 3.1, which includes a strictly convexsection followed by a linear section (while in practice, thechiller power is usually strictly convex in IT power, and sat-isfies the requirement of Theorem 1). However, for this set-ting, there is a similar, slightly weaker result that still holds–the marginal cost of power during each timeslot is commonacross all optimal solutions. This is particularly useful be-cause it then provides the data center operator a benchmarkfor evaluating which batch jobs are worthy of execution, i.e.,provide revenue larger than the marginal cost they wouldincur. Formally, we have the following theorem, which isproved in [25].

Theorem 2. For the simplified energy cost model (5a),suppose c(d) is given by (2). Then, the marginal cost ofpower, ∂

(p(t)(d(t) + c(d(t))− r(t)− e(t))+

)/∂(d(t)), at each

time t is common across all optimal solutions.

Complexity of the schedule for batch jobsA goal of this work is to develop an efficient, implementablesolution. One practical consideration is the complexity ofthe schedule for batch jobs. Specifically, a schedule mustnot be too “fragmented”, i.e., have many batch jobs beingrun at the same time and batch jobs being split across alarge number of time slots. This is particularly importantin virtualized server environments because we often need toallocate a large amount of memory for each virtual machineand the number of virtual machines sharing a server is oftenlimited by the memory available to virtual machines even ifthe CPUs can be well shared. Additionally, there is alwaysoverhead associated with hosting virtual machines. If werun too many virtual machines on the same server at thesame time, the CPU, memory, I/O and performance can beaffected. Finally, with more virtual machines, live migrationsand consolidations during runtime management can affectthe system performance.

5This can be seen by applying Weierstrass’ theorem [7], sincethe objective function is continuous and the feasible set iscompact subset of Rn.

historical power traces, weather, configuration

forecasted PV power

forecasted workload

temperature, electricity price, IT capacity, cooling parameters

historical workload traces

workload schedule, resource provisioning plan

PV Power Forecaster

Capacity and Workload Planner

Runtime Workload Manager

IT Workload Forecaster

SLAs, operational goals

Figure 7: System Architecture

However, it turns out that one need not worry about anoverly “fragmented” schedule, since there always exists a“simple” optimal schedule. Formally, we have the followingtheorem, which is proved in [25].

Theorem 3. There exists an optimal solution to the work-load management problem with at most (T + J − 1) of thebj(t) are neither 0 nor MPj.

Informally, this result says that there is a simple optimalsolution that uses at most (T +J−1) more timeslots, or cor-responding active virtual machines, in total than any othersolutions finishing the same number of jobs. Thus, on aver-age, for each class of batch job per timeslot, we run at most(T+J−1)

TJmore active virtual machines than any other plan

finishing the same amount of batch jobs. If the number ofbatch job classes is not large, on average, we run at mostone more virtual machine. In our experiments in Section 5,the simplest optimal solution only uses 4% more virtual ma-chines. Though Theorem 3 does not guarantee that everyoptimal solution is simple, the proof is constructive. Thus,it provides an approach that allows one to transform an op-timal solution into a simple optimal solution.

In addition to Theorem 3, there are two other properties ofthe optimal solution that highlight its simplicity. We statethese without proof due to space constraints. First, whenmultiple classes of batch jobs are served in the same timeslot,all of them except possibly the one with the lowest revenueare finished. Second, in every timeslot, the lowest marginalrevenue of a batch job that is served is still larger than themarginal cost of energy from Theorem 2.

4. SYSTEM PROTOTYPEWe have designed and implemented a supply-aware work-

load and capacity management prototype in a productiondata center based on the description in the previous section.The data center is equipped with on-site photovoltaic (PV)power generation and outside air cooling. The prototypeincludes workload and capacity planning, runtime workloadscheduling and resource allocation, renewable power and ITworkload demand prediction. Figure 7 depicts the system ar-chitecture. The predictors use the historical traces of supplyand demand information to predict the available power of anon-site PV, and the expected interactive workload demand.The capacity planner takes the predicted energy supply andcooling information as inputs and generates an optimal ca-pacity allocation scheme for each workload. Finally, the run-time workload manager executes the workload plan.

The remainder of this section provides more details oneach component of the system: capacity and workload plan-ning, PV and workload prediction, and runtime workloadmanagement.

4.1 Capacity and Workload Planner

Page 7: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

The data center has mixed energy sources: an on-site pho-tovoltaic (PV) generation tied to grid power. The coolinginfrastructure has a cooling micro-grid, including outside aircooling and chiller cooling. The data center hosts interac-tive applications and batch jobs. There are SLAs associatedwith the workloads. Though multiple IT resources can beused by IT workloads, we focus on CPU resource manage-ment in this implementation. We use a virtualized serverenvironment where different workloads can share resourceson the same physical servers.

The planner takes the following inputs: power supply (timevarying PV power and grid power price data), interactiveworkload demand (time varying user request rates, responsetime target), batch job resource demands (CPU hours, ar-rival time, deadline, revenue of each batch job), IT con-figuration information (number of servers, server idle andpeak power, capacity) and cooling configuration parameters(blower capacity, chiller cooling efficiency) and operationalgoals. We use the optimization (5a) in Section 3.4 with thefollowing additional details.

We first determine the IT resource demand of interac-tive workload i using the M/GI/1/PS model, which gives

1µi−λi(t)/ai(t)

≤ rti. Thus, the minimum CPU capacity

needed is ai(t) = λi(t)µi−1/rti

, which is linear function of the

arrival rate λi(t). We estimate µi through real measure-ments and set the response time requirement rti accordingto the SLAs. While the model is not perfect for real-worlddata center workloads, it provides a good approximation.Although important, performance modeling is not the focusof this paper. The resulting average CPU utilization of in-teractive workload i is 1 − 1

µirti, therefore its actual CPU

usage at time t is ai(t)(

1− 1µirti

), the remaining ai(t)

1µirti

capacity can be shared by batch jobs. For a batch job j,assume at time t it shares nji(t) ≥ 0 CPU resource withinteractive workload i and uses additional nj(t) ≥ 0 CPUresource by itself, then its total CPU usage at time t isbj(t) = Σinji(t) + nj(t), which is used to update Constraint(5b) and (5d). We have an additional constraint on CPU ca-pacity that can be shared Σjnji(t) ≤ ai(t) 1

µirti. Assume the

data center has D CPU capacity in total, so the IT capacityconstraint becomes Σiai(t) + Σjnj(t) ≤ D. Although ouroptimization (5a) in Section 3.4 can be used to handle ITworkload with multi-dimensional demand, e.g., CPU, mem-ory, here we restrict our attention to CPU-bound workloads.

The next step is to estimate the IT power consumption,which can be done based on the average CPU utilization

Pserver(u) = Pi + (Pb − Pi) ∗ u

where u is the average CPU utilization across all servers, Piand Pb are the power consumed by the server at idle andtheir fully utilized state, respectively. This simple model hasproved very useful and accurate in modeling power consump-tion since other components’ activities are either static orcorrelate well with CPU activity [13]. Assuming each serverhas Q CPU capacity, using the above model, we estimate theIT power as follows6:

d(t) =Σiai(t)

Q(Pi + (Pb − Pi) ∗ ui) +

Σjnj(t)

QPb,

where ui =(

1− 1µirti

+Σjnji(t)

ai(t)

).

The cooling power can be derived from the IT power ac-cording to the cooling model (2) described in Section 3.1.

6Since the number of servers used by an interactive workloador a class of batch jobs is usually large in data centers, wetreat it as continuous.

24 48 72 96 120 144 1680

50

100

150

hour

PV g

ener

atio

n (k

W)

actualpredicted

Figure 8: PV prediction

By solving the cost optimization problem (5a), we thenobtain the a detailed capacity plan, including at each time tthe capacity allocated to each class j of batch jobs bj(t) (fromnji(t) and nj(t)), and interactive workload i ai(t), energystorage usage e(t), as well as optimal cooling configuration(i.e., capacity for outside air cooling and chiller cooling).

It follows from Section 3.4 that this problem is a convexoptimization problem and hence there exist efficient algo-rithms to solve this problem. For example, disciplined con-vex programming can be used. Under this approach, convexfunctions and sets are built up from a small set of rules fromconvex analysis, starting from a base library of convex func-tions and sets. Constraints and objectives that are expressedusing these rules are automatically transformed to a canon-ical form and solved. In our prototype, the algorithm isimplemented using Matlab CVX [17], a modeling system forconvex optimization.

We then utilize the Best Fit Decreasing (BFD) method [38]to decide how to place and consolidate the workloads at eachtimeslot. More advanced techniques exist for optimizing theworkload placement [11], but they are out of the scope ofthis paper.

4.2 PV Power ForecasterA variety of methods have been used for fine-grained en-

ergy prediction, mostly using classical auto-regressive tech-niques [12, 21]. However, these works do not explicitly usethe associated weather conditions as a basis for modeling.[35] considered the impact of the weather conditions explic-itly and used an SVM classifier in conjunction with a RBFkernel to predict solar irradiation. We use a similar approachfor PV prediction. In order to predict PV power generationfor the next day, we use a k-nearest neighbor (k-NN) basedalgorithm. The prediction is done at the granularity of one-hour time periods. The basic idea is to search for the most“similar” days in the recent past (using one week workedwell here7) and use the generation during those days to esti-mate the generation for the target hour. The similarity be-tween two days is determined using features such as ambienttemperature, humidity, cloud cover, visibility, sunrise/sunsettimes on those days, etc. In particular, the algorithm usesweighted k-NN, where the PV prediction for hour t on the

next day is given by yt =Σi∈Nk(xt,D)yi/d(xi,xt)

Σi∈Nk(xt,D)1/d(xi,xt), where yt is

the PV predicted output at hour t, xt is the feature vector,e.g., temperature, cloud cover, for the target hour t obtainedfrom the weather forecast, yi is the actual PV output forneighbor i, xi is the corresponding feature vector, d is thedistance metric function, Nk(xi,D) are k-nearest neighborsof xi in D. k is chosen based on cross-validation of historicaldata.

Figure 8 shows the predicted and actual values for thePV supply of the data center for one week in September2011. The average prediction errors vary from 5% to 20%.The prediction accuracy depends on occurrence of similarweather conditions in the recent past and the accuracy of theweather forecast. The results in [25] show that a ballpark

7If available, data from past years could also be used.

Page 8: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

50 100 150 2000

2

4

6

8x 109

period (hour)

powe

r

Period = 24

(a) FFT analysis

6 12 18 240

0.2

0.4

0.6

0.8

1

hour

norm

alize

d cp

u de

man

d

predicted workloadactual workload

(b) Workload prediction

Figure 9: Workload analysis and prediction

approximation is sufficient for planning purposes and oursystem can tolerate prediction errors in this range.

4.3 IT Workload ForecasterIn order to perform the planning, we need knowledge about

the IT demand, both the stochastic properties of the inter-active application and the total resource demand of batchjobs. Though there is large variability in workload demands,workloads often exhibit clear short-term and long-term pat-terns. To predict the resource demand (e.g., CPU resource)for interactive applications, we first perform a periodicityanalysis of the historical workload traces to reveal the lengthof a pattern or a sequence of patterns that appear periodi-cally. Fast Fourier Transform (FFT) can be used to find theperiodogram of the time-series data. Figure 9(a) plots thetime-series and the periodogram for 25 work days of a realCPU demand trace from an SAP application. From this wederive periods of the most prominent patterns or sequencesof patterns. For this example, the peak at 24 hours in theperiodogram indicates that it has a strong daily pattern (pe-riod of 24 hours). Actually, most interactive workloads ex-hibit prominent daily patterns. An auto-regressive modelis then used to capture both the long term and short termpatterns. The model estimates w(d, t), the demand at timet on day d, based on the demand of the previous N daysas w(d, t) = ΣNi=1ai ∗ w(d − i, t) + c. The parameters arecalibrated using historical data.

We evaluate the workload prediction algorithm with sev-eral real demand traces. The results for a Web applicationtrace are shown in Figure 9(b). The average prediction er-rors are around 20%. If we can use the previous M timepoints of the same day for the prediction, we could furtherreduce the error rate.

The total resource demand (e.g., CPU hours) of non-interactiveworkloads or batch jobs can be obtained from users or fromhistorical data or through offline benchmarking [42]. Likesupply prediction, a ballpark approximation is enough, aswe will see in [25].

4.4 Runtime Workload ManagerThe runtime workload manager schedules workloads and

allocates CPU resource according to the plan generated bythe planner. We implement a prototype in a KVM-based vir-tualized server environment [1]. Our current implementationuses a KVM/QEMU hypervisor along with control groups(cgroups), a new Linux feature, to perform resource alloca-tion and workload management [1]. In particular, it executesthe following tasks according to the plan: (1) create and startvirtual machines hosting batch jobs; (2) adjust the resourceallocation (e.g., CPU shares or number of virtual CPUs) toeach virtual machine; (3) migrate and consolidate virtual ma-chines via live migration. The workload manager assigns ahigher priority to virtual machines running interactive work-loads than virtual machines for batch jobs via Cgroups. Thisguarantees that resources are available as needed by interac-tive applications, while excess resources can be used by thebatch jobs, improving server utilization.

5. EVALUATIONTo highlight the benefits of our design for renewable and

cooling aware workload management, we perform a mixtureof numerical simulations and experiments in a real testbed.We first present trace-based simulation results in Section 5.1,and then the experimental results on the real testbed imple-mentation in Section 5.2.

5.1 Case StudiesWe begin by discussing evaluations of our workload and ca-

pacity planning using numerical simulations. We use tracesfrom real data centers. In particular, we obtain PV supply,interactive IT workload, and cooling data from real data cen-ter traces. The renewable energy and cooling data is frommeasurements of a data center in California. The data cen-ter is equipped with 130kW PV panel array and a coolingsystem consisting of outside air cooling and chiller cooling.We use the real-time electricity price of the data center loca-tion obtained from [18]. The total IT capacity is 500 servers(100kW). The interactive workload is a popular web serviceapplication with more than 85 million registered users in 22countries. The trace contains average CPU utilization andmemory usage as recorded every 5 minutes. Additionally,we assume that there are a number of batch jobs. Half ofthem are submitted at midnight and another half are sub-mitted around noon. The total demand ratio between theinteractive workload and batch jobs is 1:1.5. The interactiveworkload is deemed critical and the resource demand must bemet while the batch jobs can be rescheduled as long as theyfinish before their deadlines. The plan period is 24-hours andthe capacity planner creates a plan for the next 24-hours atmidnight based on renewable supply and cooling informationas well as the interactive demand. The plan includes hourlycapacity allocation for each workload. We assume perfectknowledge about the workload demand and renewable sup-ply, and we study the impact of prediction errors and themix of interactive and batch workloads in [25].

We explore the following issues: (i) How valuable is re-newable and cooling aware workload management? (ii) Isnet-zero possible under renewable and cooling aware work-load management? (iii) What portfolio of renewable sourcesis best?

How valuable is renewable and cooling awareworkload management?We start with the key question for this paper: how muchcost/energy/CO2 savings does renewable and cooling awareworkload management provide? In this study, we assumehalf of batch jobs must be finished before noon and anotherhalf must be finished before midnight. We compare the fol-lowing four approaches: (i) Optimal, which integrates supplyand cooling information and uses our optimization algorithmto schedule batch jobs; (ii) Night, which schedules batch jobsat night to avoid interfering with critical workloads and totake advantage of idle machines (this is a widely used so-lution in practice); (iii) Best Effort (BE), which runs batchjobs immediately when they arrive and uses all available ITto finish batch jobs as quickly as possible; (iv) Flat, whichruns batch jobs at a constant rate within the deadline periodfor the jobs.

Figure 10 shows the detailed schedule and power consump-tion for each approach, including IT power (batch and in-teractive workloads), cooling, and supply. As shown in thefigure, compared with other approaches, Optimal reshapesthe batch job demand and fully utilizes the renewable sup-ply, and uses non-renewable energy, if necessary, to completethe batch jobs during this 24-hour period. These additionalbatch jobs are scheduled to run between 3am and 6am or11pm and midnight, when the outside air cooling is most ef-ficient and/or the electricity price is lower. As a result, our

Page 9: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(a) Optimal

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(b) Night

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(c) Best Effort (BE)

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(d) Flat

Optimal Night BE Flat0

500

1000

1500

2000

powe

r (kW

h)

grid powerrenewable

(e) Power usage comparisonsenergy cost CO2 emission

0

0.5

1

1.5

2

norm

alize

d va

lue

OptimalNightBEFLAT

(f) Efficiency comparisons

Figure 10: Power cost minimization while finishingall jobs

solution reduces the grid power consumption by 39%-63%compared to other approaches (Figure 10(e)). Though notclear in the figure, the optimal solution does consume a bitmore total power (up to 2%) because of the low cooling ef-ficiency around noon. Figure 10(f) shows the average recur-ring electricity cost and CO2 emission per job. The energycost per job is reduced by 53%-83% and the CO2 emissionper job is reduced by 39%-63% under the Optimal schedule.

The adaptation of the workload management to the avail-ability of renewable energy is clear in Figure 10. Less clear isthe importance of managing the workload in a manner thatis “cooling aware”.

The importance of cooling aware scheduling: As dis-cussed in Sections 2.2 and 3.1, the cooling efficiency andcapacity of a cooling supply often vary over time. This isparticularly true for outside air cooling. One important com-ponent of our solution is to schedule workloads by taking intoaccount time varying cooling efficiency and capacity. To un-derstand the benefits of cooling integration, we compare ouroptimal solution as shown in Figure 10(a) with two solu-tions that are renewable aware but handle cooling differently:(i)Cooling-oblivious ignores cooling power and considers ITpower only, (ii)Static-cooling uses a static cooling efficiency(assuming the cooling power is 30% of IT power) to esti-mate the cooling power from IT power and incorporates thecooling power into workload scheduling.

Figures 11(a) and 11(b) show Cooling-oblivious and Static-cooling schedules, respectively. As shown in the figures, bothschedules integrate renewable energy into scheduling and runmost batch jobs when renewables are available. However, be-cause they do not capture the cooling power accurately, theycannot fully optimize workload schedule. In particular, byignoring the cooling power, Cooling-oblivious underestimatesthe power demand and runs more jobs than the available PVsupply and hence uses more grid power during the day. Thisis also less cost-efficient because the electricity price peaks atthat time. On the other hand, by overestimating the coolingpower demand, Static-cooling fails to fully utilize the renew-able supply and results in inefficiency, too. Figures 11(c)

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(a) Cooling-oblivious

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(b) Static-cooling

Optimal Cooling−oblivious Static−cooling0

500

1000

1500

2000

powe

r (kW

h)

grid powerrenewable

(c) Power usage comparisonsenergy cost CO2 emission

0

0.5

1

1.5

2

norm

alize

d va

lue

OptimalCooling−obliviousStatic−cooling

(d) Efficiency comparisons

Figure 11: Benefit of cooling integration

and 11(d) compare the total power consumption and the en-ergy efficiency (i.e., normalized grid power usage and carbonemission per job) of these two approaches and Optimal, re-spectively. By accurately modeling the cooling efficiency andadapting to time variations in cooling efficiency, our solutionreduces the energy cost by 20%-38% and CO2 emission by4%-28%.

The importance of optimizing the cooling micro-grid:Optimizing the workload scheduling based on cooling effi-ciency is only one aspect of cooling integration. Anotherimportant aspect is to optimize the cooling micro-grid, i.e,using proper amount of cooling capacity from each cooling re-source. This is important because different cooling supplies(e.g., outside air and chiller cooling) can exhibit differentcooling efficiencies as the IT demand and external condi-tions such as outside air temperature changes. Our solutiontakes this into account and optimizes the cooling capacitymanagement in addition to IT workload management.

We compare the following three approaches: (i) Optimaladjusts the cooling capacity for outside air cooling and chillercooling based on the dynamic cooling efficiency, which is de-termined by IT demand and outside air temperature, (ii) FullOutside Air (FOA) always uses outside air cooling at its fullcapacity, (iii) Chiller only uses the chiller cooling only. Allthree solutions are renewable and cooling aware and sched-ule workload according to the renewable supply and coolingefficiency. They finish the same number of batch jobs. Thedifference is how they manage cooling resources and capacity.

Figure 12 shows the cooling capacity from outside air cool-ing and chiller cooling for these three solutions. As shown inthis figure, Optimal uses outside air only during night whenit is more efficient, and combines outside air cooling andchiller cooling during other times. In particular, our solutionuses less outside air cooling and more chiller cooling between12pm and 3pm as outside air cooling is less efficient due tohigh IT demand and outside air temperature at that time. Incontrast, FOA always runs outside air at full capacity, evenwhen the outside air is less efficient. Figures 12(g) and 12(h)compares the power, cost, and cooling consumption of thethree approaches. By optimizing the cooling substructure,our solution reduces the cooling power by 67% over FOAand 48% over Chiller only.

Is net-zero energy consumption possible with renew-able and cooling aware workload management?Now, we switch our goal from minimizing the cost incurredby the data center to minimizing the environmental impact ofthe data center. Net-zero is often used to describe a buildingwith zero net energy consumption and zero carbon emission

Page 10: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(a) Optimal plan

0 6 12 18 240

50

100

150

200

hour

cool

ing

capa

city

(kW

)

outside airchiller

(b) Optimal cooling capacity

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(c) FOA plan

0 6 12 18 240

50

100

150

200

hour

cool

ing

capa

city

(kW

)

outside airchiller

(d) FOA cooling capacity

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(e) Chiller only plan

0 6 12 18 240

50

100

150

200

hour

cool

ing

capa

city

(kW

)

outside airchiller

(f) Chiller only cooling capac-ity

Optimal Full outside air Chiller only0

100

200

300

400

cool

ing

powe

r (kW

h)

(g) Cooling powerenergy cost CO2 emission

0

0.5

1

1.5

2

norm

alize

d va

lue

OptimalFull outside airChiller only

(h) Efficiency comparisons

Figure 12: Benefit of cooling optimization

annually.Recently, researchers have envisioned how net-zerobuilding concepts can be effectively extended into the datacenter space to create a net-zero data center, whose totalpower consumption is less than or equal to the total powersupply from renewable. We explore if net-zero is possiblewith renewable and cooling aware workload management indata centers and how much it will cost.

By adding a net-zero constraint (i.e., total power consump-tion ≤ total renewable supply) to our optimization problem,our capacity planner can generate a net-zero schedule. Fig-ure 13(a) shows our solution (Net-zero1 ) for achieving a netzero operation goal. Similar to the optimal solution shownin Figure 10(a), Net-zero1 optimally schedules batch jobsto take advantage of the renewable supply; however, batchjobs are only executed when renewable energy is availableand without exceeding the total renewable generation, andthus some are allowed to not finish during this 24 hour pe-riod. In this case, about 40% of the batch jobs are delayedtill a future time when a renewable energy surplus may ex-ist. Additionally, some renewable energy is reserved to offsetnon-renewable energy used at night for interactive workloads.This excess renewable power is reserved in the afternoonwhen the outside air temperature peaks, thus minimizingthe energy required for cooling.

A key to achieving net zero is energy storage. By maximiz-ing the use of the renewable directly, Net-zero1 can reducethe dependency on storage and hence the capital cost. Tounderstand the benefit, we compare Net-zero1 with anotherschedule, Net-zero2, which runs the same number of batch

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(a) Net-zero1

0 6 12 18 240

50

100

150

200

hour

powe

r (kW

)

interactivebatch jobcooling power

IT capacityrenewable

(b) Net-zero2

Figure 13: Net Zero Energy

0 0.2 0.4 0.6 0.8 10

2000

4000

6000

8000

solar ratio

grid

pow

er u

sage

(kW

h)

storage = 0storage = 200kWhstorage = 400kWh

(a) With turing off servers

0 0.2 0.4 0.6 0.8 10

2000

4000

6000

8000

solar ratio

grid

pow

er u

sage

(kW

h)

storage = 0storage = 200kWhstorage = 400kWh

(b) Without turning offservers

Figure 14: Optimal renewable portfolio

jobs but distributes the batch jobs over 24-hours as shownin Figure 13(b). Both approaches achieve this net-zero goal,but Net-zero2 uses 287% more grid power compared to oursolution Net-zero1. As a result, the demand on storage ismuch higher. The energy storage sizes of Net-zero1 andNet-zero2 are 82kWh and 330kWh, respectively. Using anestimated cost of 400$/kWh [3], this difference in energy de-mand results in $99,200 more energy storage expenditure forNet-zero2.

What portfolio of renewable sources is best?To this point, we have focused on PV solar as the sole sourceof renewable energy. Since wind energy is becoming increas-ingly cost-effective, a sustainable data center will likely useboth solar and wind to some degree. The question thatemerges is which one is better source from the perspectiveof optimizing data center energy efficiency. More generally,what is the optimal portfolio of renewable sources?

We conduct a study using the wind and solar traces de-picted in Figure 2. Assuming an average renewable supply of200kW, we vary the mix of solar and wind in the renewablesupply. For each mix, we use our capacity management opti-mization algorithm to generate an optimal workload sched-ule. We compare the non-renewable power consumption fordifferent renewable mixes for two cases (turning off unusedservers and without turning off unused servers). Figure 14shows the results as a function of percentage of solar withdifferent storage capacity. As shown in the figure, the opti-mal portfolio contains more solar than wind because solar isless volatile and the supply aligns better with IT demand.

However, wind energy is still an important component anda small percentage of wind can help improve the efficiency.For example, the optimal portfolio without storage consistsof about 60% solar and 40% wind. As we increase the stor-age capacity, the energy efficiency improves and wind energybecomes less valuable. This is because wind is more volatileat night and so storage will incur heavier losses.

In summary, though solar is a better source for local datacenters, a small addition of wind can help improve energyefficiency. This is particularly true when there is no storageor the storage capacity is small. Additionally, recent workhas shown that the value of wind increases significantly whengeographically diverse data centers are considered [27, 26].

5.2 Experimental Results on a Real TestbedThe case studies described in the previous section high-

Page 11: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

light the theoretical benefits of our approach over existingsolutions. To verify our claims and ensure that we have apractical and robust solution, we experimentally evaluate ourprototype implementation on a real data center testbed andcontrast it with a current workload management approach.

5.2.1 Experiment SetupOur testbed consists of four high end servers (each with

two 12-core 1.8GHz processors and 64 GB memory) and thefollowing workloads: one interactive Web application, and 6batch applications. Each server is running Scientific Linux.Each workload is running inside a KVM virtual machine.The interactive application is a single-tier web applicationrunning multiple Apache web servers and batch jobs aresysbench [2] with different resource demands. httperf [19]is used to replay the workload demand traces in the formof CGI requests, and each request is directed to one of theWeb servers. The PV, cooling data, and interactive workloadtraces used in the case study are scaled to the testbed ca-pacity. We measure the power consumption via the server’sbuilt-in management interfaces, collect CPU consumptionthrough system monitoring and obtain response times ofWeb servers from Apache logs.

5.2.2 Experiment ResultsWe compare two approaches: (i) Optimal is our optimal

design, (ii) Night, which runs batch jobs at night. For eachplan, the runtime workload manager dynamically starts thebatch jobs, allocates resources to both interactive and batchworkloads and turn on/off servers according to the plan. Fig-ures 15(a) and 15(b) shows the CPU allocation of our opti-mal solution and the actual CPU consumption measured inthe experiment, respectively. The results show that actualresource usage closely follows the power generated by the ca-pacity planner. The results for Night are similar. We furthercompare the predicted power usage in the plan and the actualpower consumption in the experiment for both approaches.From Figures 15(c) and 15(d), we see that the actual powerusage is close to the power plan.

We then compare the power consumption and performanceof the two approaches. Figure 16(a) shows the power con-sumption. Optimal does more work during the day when therenewable energy source is available.Night uses additionalservers from midnight to 6am to run batch jobs while oursolution starts batch jobs around noon by taking advantageof renewable energy. Compared with Night, our approachreduces the grid power usage by 48%. One thing worth men-tioning is that the total power is not quite proportional tothe total CPU utilization as a result of the large idle partof the server power. This is most noticeable when the num-ber of servers is small, as we see in the experimental results,Figures 15(b) and 15(c). When the total number of serversincreases, the impact of idle power decreases and Optimalwill save even more grid power.

One reason that batch jobs are scheduled to run at nightis to avoid interfering with interactive workloads. Our ap-proach runs more jobs during the day when the web serverdemand is high. To understand the impact of this on per-formance, we compare the average response time of the webserver. The results show that both approaches are almostidentical: 153.0ms for Optimal, compared to 156.0ms forNight. See [25] for more details. This is because both solu-tions satisfy the demand of web server and Linux KVM/Cgroupsscheduler is preemptive and enables CPU resources to be ef-fectively virtualized and shared among virtual machines [1].In particular, assigning a much higher priority to virtual ma-chines hosting the web servers guarantees that resources areavailable as needed by the web servers.

In summary, the real experiment results demonstrate that(i) the optimization-based workload management scheme trans-

0 6 12 18 240

20

40

60

80

hour

CPU

utiliz

atio

n

interactivebatch job

(a) CPU allocation (Optimal)

0 6 12 18 240

20

40

60

80

hour

CPU

utiliz

atio

n

interactivebatch jobs

(b) CPU consumption (Opti-mal)

0 6 12 18 240

200

400

600

800

1000

hour

powe

r (W

att)

planexperiment

(c) Power (Optimal)

0 6 12 18 240

200

400

600

800

1000

hour

powe

r (W

att)

planexperiment

(d) Power (Night)

Figure 15: Comparison of plan and experimental re-sults

0 6 12 18 240

200

400

600

800

1000

1200

hour

powe

r (W

att)

OptimalNightrenewable supply

(a) Power consumptionOptimal Night

0

5

10

Powe

r (kW

h)

grid powerrenewable

(b) Total power consumption

Figure 16: Comparison of optimal and night

lates effectively into a prototype implementation, (ii) com-pared with traditional workload management solution, Op-timal significantly reduces the use of grid power without de-grading the performance of critical demand.

6. CONCLUDING REMARKSOur goal in this paper is to provide an integrated workload

management system for data centers that takes advantage ofthe efficiency gains possible by shifting demand in a way thatexploits time variations in electricity price, the availabilityof renewable energy, and the efficiency of cooling. There aretwo key points we would like to highlight about our design.

First, a key feature of the design is the integration ofthe three main data center “silos”: cooling, power, and IT.Though a considerable amount of work exists in optimizingefficiencies of these individually, there is little work that pro-vides an integrated solution for all three. Our case studies il-lustrate that the potential gains from an integrated approachare significant. Additionally, our prototype illustrates thatthese gains are attainable. In both cases, we have taken caseto measurements from a real data center and traces of realapplications to ensure that our experiments are meaningful.

Second, it is important to point out that our approachuses a mix of implementation, modeling, and theory. At thecore of our design is a cost optimization that is solved by theworkload manager. Care has been taken in designing andsolving this optimization so that the solution is “practical”(see the characterization theorems in Section 3.5). Buildingan implementation around this optimization requires signifi-cant measurement and modeling of the cooling substructure,and the incorporation of predictors for workload demand andPV supply. We see one role of this paper as a proof of conceptfor the wide-variety of “optimization-based designs” recentlyproposed, e.g., [22, 27, 34, 40, 24, 30, 26].

There are a number of future directions building on this

Page 12: Renewable and Cooling Aware Workload Management for Sustainable Data …users.cms.caltech.edu/~adamw/papers/demandShaping.pdf · 2012-03-12 · Renewable and Cooling Aware Workload

work including integrating reliability and a more detailedstudy of the role of storage. But, perhaps the most excitingdirection is the potential to consider renewable and coolingaware workload management of geographically diverse datacenters as opposed to the local workload management consid-ered here. As illustrated in [27, 26] geographical diversity canbe a significant aid in handling the intermittency of renew-able sources and electricity price fluctuations. For example,the fact that fluctuations in wind energy are nearly uncorre-lated for significantly different locations means that wind canprovide a nearly constant baseline supply of energy if work-load can be adapted geographically. Another future directionwe would like to highlight is the task of understanding theimpact of workload management on the capacity planning ofa data center, e.g. the size of renewable infrastructure, thecapacity of IT and cooling infrastructure, and the capitalexpense and to minimize the Total Cost Ownership (TCO).

AcknowledgementsWe are grateful to many members of Sustainable Ecosys-tem Research Group at HP Labs. Chandrakant Patel hasprovided great support and guidance at various stages ofthis work. Niru Kumari provided valuable information onchiller cooling model. Martin Arlitt, Amip Shah, SergeyBlagodurov and Alan McReynolds offered helpful feedback.We also thank the anonymous reviewers and our shepherd,Christopher Stewart, for their valuable comments and help.

7. REFERENCES[1] Kvm kernel based virtual machine.

http://www.redhat.com/f/pdf/rhev/DOC-KVM.pdf.[2] Sysbench: a system performance benchmark.

http://sysbench.sourceforge.net/.[3] Electricity Energy Storage Technology Options: A White

Paper Primer on Applications, Costs and Benefits. 2010.[4] Clean urban energy: Turn buildings into batteries. 2011.[5] L. L. H. Andrew, M. Lin, and A. Wierman. Optimality,

fairness and robustness in speed scaling designs. In Proc. ofACM Sigmetrics, 2010.

[6] C. E. Bash, C. D. Patel, , and R. K. Sharma. Dynamicthermal management of aircooled data centers. In Proc. ofITHERM, 2006.

[7] D. P. Bertsekas. Nonlinear Programming. Athena Scientific,1999.

[8] F. Bleier. Fan Handbook: Selection, Application andDesign. New York: McGraw-Hill, 1997.

[9] T. Breen, E. Walsh, J. Punch, C. Bash, and A. Shah. Fromchip to cooling tower data center modeling: Influence ofserver inlet temperature and temperature rise acrosscabinet. Journal of Electronic Packaging, 133.

[10] Y. Chen, D. Gmach, C. Hyser, Z. Wang, C. Bash,C. Hoover, , and S. Singhal. Integrated management ofapplication performance, power and cooling in data centers.In Proc. of NOMS 2010, 2010.

[11] J. Choi, S. Govindan, B. Urgaonkar, andA. Sivasubramaniam. Power consumption prediction andpower-aware packing in consolidated environments. IEEETransactions on Computers, 59(12), 2010.

[12] D. R. Cox. Prediction by exponentially weighted movingaverages and related methods, 1961.

[13] X. Fan, W.-D. Weber, and L. A. Barroso. Powerprovisioning for a warehouse-sized computer. In Proc. ofACM ISCA, 2007.

[14] A. Gandhi, M. Harchol-Balter, and C. L. R. Das. Optimalpower allocation in server farms. In Proc. of ACMSigmetrics, 2009.

[15] D. Gmach, Y. Chen, A. Shah, J. Rolia, C. Bash,T. Christian, and R. Sharma. Profiling sustainability ofdata centers. In Proc. of ISSST, 2010.

[16] J. Holman. Heat Transfer. 8th ed. New York: McGraw-Hill,1997.

[17] http://cvxr.com/cvx/.[18] http://www.ferc.gov/market-oversight/mkt electric.

[19] http://www.hpl.hp.com/research/linux/httperf/.[20] W. Huang, M. Allen-Ware, J. Carter, E. Elnozahy,

H. Hamann, T. Keller, C. Lefurgy, J. Li, and J. R.Karthick Rajamani. Tapo: Thermal-aware poweroptimization techniques for servers and data centers. InProc. of IGCC, 2011.

[21] A. Kansal, J. Hsu, S. Zahedi, and M. B. Srivastava. Powermanagement in energy harvesting sensor networks, 2007.

[22] K. Le, O. Bilgir, R. Bianchini, M. Martonosi, and T. D.Nguyen. Capping the brown energy consumption of internetservices at low cost. In Proc. IGCC, 2010.

[23] M. Lin, Z. Liu, A. Wierman, and L. L. Andrew. Onlinealgorithms for geographical load balancing.http://http://www.cs.caltech.edu/ zliu2/onlineGLB.pdf.

[24] M. Lin, A. Wierman, L. L. H. Andrew, and E. Thereska.Dynamic right-sizing for power-proportional data centers. InProc. of INFOCOM, 2011.

[25] Z. Liu, Y. Chen, C. Bash, A. Wierman, D. Gmach,Z. Wang, M. Marwah, and C. Hyser. Renewable and coolingaware workload management for sustainable data centers[extended version]. Technical report.

[26] Z. Liu, M. Lin, A. Wierman, S. H. Low, and L. L. H.Andrew. Geographical load balancing with renewables. InProc. ACM GreenMetrics, 2011.

[27] Z. Liu, M. Lin, A. Wierman, S. H. Low, and L. L. H.Andrew. Greening geographical load balancing. In Proc.ACM Sigmetrics, 2011.

[28] J. Moore, J. Chase, P. Ranganathan, , and R. Sharma.Making scheduling cool: Temperature-aware workloadplacement in data centers. In Proc. of USENIX AnnualTechnical Conference, 2005.

[29] S. Ong, P. Denholm, and E. Doris. The impacts ofcommercial electric utility rate structure elements on theeconomics of photovoltaic systems. Technical ReportNREL/TP-6A2-46782, National Renewable EnergyLaboratory, 2010.

[30] E. Pakbaznia and M. Pedram. Minimizing data centercooling and server power costs. In Proc. of ISLPED, 2009.

[31] C. Patel, R. Sharma, C. Bash, and A. Beitelmal. Energyflow in the information technology stack. In Proc. ofIMECE, 2006.

[32] C. D. Patel, R. K. Sharma, C. E. Bash, and A. Beitelmal.Thermal considerations in cooling large-scale high computedensity data centers. In Proc. of ITHERM, 2002.

[33] R. Raghavendra, P. Ranganathan, V. Talwar, Z. Wang, andX. Zhu. No ”power” struggles: Coordinated multi-levelpower management for the data center. In Proc. ofASPLOS, 2008.

[34] L. Rao, X. Liu, L. Xie, and W. Liu. Minimizing electricitycost: Optimization of distributed internet data centers in amulti-electricity-market environment. In Proc. ofINFOCOM, 2010.

[35] N. Sharma, P. Sharma, D. Irwin, and P. Shenoy. Predictingsolar generation from weather forecasts using machinelearning. In Proc. of Second IEEE International Conferenceon Smart Grid Communications(SmartGridComm), 2011.

[36] E. Thereska, A. Donnelly, and D. Narayanan. Sierra: apower-proportional, distributed storage system. TechnicalReport MSR-TR-2009-153, Microsoft Research, 2009.

[37] R. Urgaonkar, B. Urgaonkar, M. Neely, andA. Sivasubramaniam. Optimal power cost managementusing stored energy in data centers. In Proc. of ACMSigmetrics, 2011.

[38] V. Vazirani. Approximation Algorithms. Springer, 2003.[39] Z. Wang, A. McReynolds, C. Felix, C. Bash, and C. Hoover.

Kratos: Automated management of cooling capacity in datacenters with adaptive vent tiles. In Proc. of IMECE, 2009.

[40] P. Wendell, J. W. Jiang, M. J. Freedman, and J. Rexford.Donar: decentralized server selection for cloud services. InProc. of ACM Sigcomm, 2010.

[41] A. Wierman, L. L. H. Andrew, and A. Tang. Power-awarespeed scaling in processor sharing systems. In Proc. ofINFOCOM, 2009.

[42] E. A. Yolanda Becerra, David Carrera. Batch job profilingand adaptive profile enforcement for virtualizedenvironments. In Proc. of ICPDNP, 2009.

[43] R. Zhou, Z. Wang, A. McReynolds, C. Bash, T. Christian,and R. Shih. Optimization and control of cooling microgridsfor data centers. In Proc. of ITherm, 2012.