You are on page 1of 34

Green Computing

Johan Lilius

January 19, 2012

Johan Lilius

Green Computing

1/67

Introduction

Contents
1 Introduction
2 The economics of Data-Centers: Why Power Matters
3 Energy-proportional computing
4 Servers built on mobile processors
5 Summary
6 Bibliography

Johan Lilius

Green Computing

2/67

Introduction

What is Green-Computing (Green-ICT)?

Green Computing refer to 2 dierent things


1
2

Reducing energy consumption of ICT


Using ICT to reduce energy consumption

Goal: reduce carbon footprint


This presentation : Reducing energy consumption of ICT

Johan Lilius

Green Computing

3/67

Introduction

Green-Computing aspects [11]

Start
1992 EPA Energy star rating

5 issues (2009)
1
2
3
4
5

Johan Lilius

E-Waste
Data-centers and Servers
PCs, Monitors and Workstations
Software
Telecommuting

Green Computing

4/67

Introduction

E-Waste

Recycling issue
estimates that over 25 billion computers, televisions, cell
phones, printers, gaming systems, and other devices
have been sold since 1980,
2 million tons of unwanted electronic devices in 2005
alone,
with only 15 to 20 percent being recycled

Material is transfered to developing countries


Clear environmental hazards

Johan Lilius

Green Computing

5/67

Introduction

Data centers & Servers


Energy efficiency

Power draw of computing clusters is becoming an increasing fraction of


their cost1

Power consumption is an issue

9/3/10&

Johan Lilius

The density of the datacenters that house them is in turn limited by

Driver
money,
not and
environmental
aspects
their ability
to supply
cool 1020 kW
of power:-(
per rack and up to

1020 MW per datacenter


Future datacenters may require as much as 200 MW, and datacenters
are being constructed today with dedicated electrical substations to
feed them.
1Kenneth

G. Brill: The Invisible Crisis in the Data Center: The Economic Meltdown of Moore's Law
Uptime Institute, 2009

Green Computing

2&

6/67

Introduction

PCs, Monitors, Workstations

5th upgarde Energy Star requirements for monitors (1


April 2009)
Requires a 20 percent increase in electrical eciency.
Estimate if all monitors comply,
saving of roughly $1 billion per year in energy expenses
and
avoid GHG emissions equivalent to 1.5 million cars

Johan Lilius

Green Computing

7/67

Introduction

Software

Cost of Spam (McAfee, 2008)


62 trillion spam messages in 2008
0.3 grams of carbon dioxide (CO2) per message
annual spam energy use 33 terawatt hours (tWh)
equivalent to the electricity used in 2.4 million homes
every year, with the
same GHG emissions as 3.1 million automobiles using
two billion US gallons of gasoline

How we use the computing systems influences power


consumption

Johan Lilius

Green Computing

8/67

Introduction

Telecommuting

An ITIF report found that, if only 14 percent of existing


American oce jobs were converted to work-from-home jobs,
the savings would be dramatic: estimated at 136 billion
vehicle travel miles annually in the US by 2020 and 171
billion miles by 2030

Johan Lilius

Green Computing

9/67

Introduction

This presentation

How to reduce energy consumption in datacenters


Central concept: Energy Proportionality
Use only as much computation power that is needed for
the task at hand

Still a lot of R&D to do!

Johan Lilius

Green Computing

10/67

The economics of Data-Centers: Why Power Matters

Contents
1 Introduction
2 The economics of Data-Centers: Why Power Matters
3 Energy-proportional computing
4 Servers built on mobile processors
5 Summary
6 Bibliography

Johan Lilius

Green Computing

11/67

The economics of Data-Centers: Why Power Matters

Energy cost in data-centers I


Energy efficiency

9/3/10&

Johan Lilius

Power draw
computing
clusters
is becoming
an increasing
Power
drawofof
computing
clusters
is becoming
an fraction of
their cost1
increasing fraction of their cost

The density of the datacenters that house them is in turn limited by

The
the and
datacenters
isand
in up to
their density
ability to of
supply
cool 1020that
kW ofhouse
power them
per rack
1020
MW perby
datacenter
turn
limited
their ability to supply and cool 1020 kW
Future datacenters may require as much as 200 MW, and datacenters
of
per rack and
upwith
to dedicated
1020 MW
per substations
datacenterto
arepower
being constructed
today
electrical
feed
them.
Future datacenters require as much as >100 MW, and
Kenneth G. Brill: The
Crisisconstructed
in the Data Center: The
Economic
Meltdown
of Moore's Law
datacenters
areInvisible
being
today
with
dedicated
Uptime Institute, 2009
electrical substations to feed them.
1

Facebook, Lule, budget 120MW


Green Computing

2&

12/67

The economics of Data-Centers: Why Power Matters

Energy cost in data-centers II

Exascale computing roadmap, budgets of >100MW for


supercomputing facilities

Nuclear powerplants:
Loviisa reactors 488MW nominal power
Olkiluoto new reactor 1600 MW nominal power

Johan Lilius

Green Computing

13/67

The economics of Data-Centers: Why Power Matters

Cost breakdown for data-centers I

A $200M facility capable of delivering 15MW of critical


load (server power)
Facility: $200M for 15MW DC (15 yr Amortization)
Servers: $2k/each, roughly 50,000 (3 yr Amortization)
Commercial Power: $0.07/kWh
5% cost of money

Johan Lilius

Green Computing

14/67

period of the equipment. This converts the capital expenses to


effective cost per month. And, by considering amortization
periods, we normalize long lived and short lived capital and
The economics of Data-Centers: Why Power Matters
recognize each appropriately. In this model, land, taxes, security
and administration are not included due to their relatively small
Cost breakdown
data-centers
II
contribution tofor
overall
costs.

defined as IT
tells us what
actually gets d

These terms
1.7 states tha
servers), we d
systems (air c
same as a DC
to the facility
also tells us t
lost in power

PUEs vary g
as low as 2
facilities are
reports, howe

In this explo
generation fa
understood te
available equ
1.7, putting i
But it is not
innovations. A
best and form
where the pow

Figure 1: Monthly Server, Power, and Infrastructure Costs


Figure 1 shows
thatarepower
costs
are related
much lower than
Infrastructure
costs
mostly
power

infrastructure costs, and also much less than the servers


50% of costs are power related
themselves. Servers are the dominant cost, but, before we
conclude that power is only 23% of the total, its worth looking
Looking mor
more
closely.
Infrastructure
includes
the
building,
power
Johan Lilius
Green Computing
15/67 definition of
distribution, and cooling. Power distribution and cooling make up
power to the
82% of the costs of infrastructure [2] with the building itself down
remaining 41
in
the
12-15%
range.
Power
distribution
is
functionally
related
to
The economics of Data-Centers: Why Power Matters
To understan
the power consumed in that sufficient power distribution
going, we loo
equipment is required to distribute the maximum amount of power
both easier to
Measures
(PUE, DCiE) I
consumed. Cooling is also functionally related to power in that
track. Lookin

PUE: Power Usage Eectivenes


PUE =

TotalPower
ITEquip.Power

DCie: Data Center Infrastructure Eectiveness


DCiE =

ITEquip.Power
100%
TotalPower

Google average Q3 2011: PUE 1.16


CSC Kajaani data-center goal: PUE 1.15
PUE is highly dependent on outside temperature

Johan Lilius

Green Computing

16/67

The economics of Data-Centers: Why Power Matters

Measures (PUE, DCiE) II


Nordic countries are interesting
Google: Summa
Facebook: Lule

Added advantage: green power


Facebook datacenter will be powered by hydropower
Diesel only as backup

Other ideas:
Use feed excess heat into city-wide heating system
(Gaudeamus datacenter, Helsinki)

NOTE PUE is a measure on the eectiveness of the


cooling, not of the energy-eciency of the computing!!!
Quite a lot remains to be done to cut down the total
power consumption of the system

efficiency of each conversation from the power delivered by the


utility at 115,000V through to deliver to the servers at 208V.

Johan Lilius
Starting at

pricing out at more than $2M. Most facilities will have at least 1
extra generator (N+1) and many facilities will have 2 spares

Green
Computing
(N+2) allowing one to be in maintenance, one to fail on startup
2,
we see the
utility

the upper left corner of Figure


delivers us 115kV and we first step it down to 13.2kv. The 13.2kv
feed is delivered to the Uninterruptable Power Supply (UPS). In
this case we use a battery-based UPS system, but rotary systems
are also common. This particular battery-based UPS is 94%
efficient, taking all current through rectifiers to direct current and
inverting it all back
to AC. Rotary designs are
usuallyPower
more
Thethen
economics
of Data-Centers:
Why
efficient than the example shown here and bypass designs can
exceed 97% efficiency. In this example, a non-bypass UPS
installation, all power flowing to UPS protected equipment (the
servers and most of the mechanical systems) is first rectified to
DC and then inverted back to AC. All the power destined to the
servers flows through these two conversions steps whether or not
there is a power failure, and these two conversion steps contribute
the bulk of the losses, bringing down the UPS efficiency to 94%.
More efficient bypass UPSs avoid these losses by routing most
power around the UPS in the common, non-power failure case.

and still to be able to run the facility at full load during a power
failure. A 2.5MW generator will burn just under 180 gallons/hour
of diesel so environmentally conscious operators work hard to
minimize their generator time. And the storage of well over
100,000 gallons of diesel at the facility brings additional cost,
storage space, insurance risk, and maintenance issues.

17/67

Matters

After the UPS, we step down the 13.2kV voltage to 480V and
then that is further stepped down to 208V for distribution to the
critical load, the servers. In this facility, we are using very high
quality transformers, so we experience losses of only 0.3% at each
transformer. We estimate that we lose a further 1% in switch gear
and conductor losses throughout the facility.

Energy losses in Power Distribution

We know we deliver 59% of the facility power to the critical load


and, from the electrical distribution system analysis above, we
know we lose 8% of total power to power distribution losses. By
subtraction, we have 33% lost to mechanical systems responsible
for data center cooling.

Figure 2: Power Distribution


For longer term power outages, there are usually generators to
keep the facility operational. The generation system introduces
essentially no additional losses when not being used but they
greatly increase the capital expense with a 2.5MW generator

Johan Lilius

In summary, we have three 99.7% efficient transformers, a 94%


efficient UPS and 1% losses in distribution for an overall power
distribution loss of 8% (0.997^3*0.94*0.99 => 0.922).

Green Computing
3

18/67

The economics of Data-Centers: Why Power Matters

AC vs DC?
Ishikari Datacenter in Ishikari City, Hokkaido Japan
Use High-Voltage (400V) DC power as much as possible

Johan Lilius

Green Computing

19/67

The economics of Data-Centers: Why Power Matters

Energy losses in Mechanical Systems

Figure 3: Mechanical Systems


Several observations emerge from this summary. The first is that
power distribution is already fairly efficient. Taking the 8%
efficiency number down to 4% to 5% by using a 97% efficient
UPS and eliminating 1 layer of power conversion is an easy
improvement. Further reductions in power distribution losses are
possible but the positive impact cant exceed 8% so were better
rewarded looking to improvements in the mechanical systems,
where we are spending 33% of the power, and in the servers,
where we are dissipating 59% of the power.

Johan

4.1 System Balance

Looking back 25 years, we have experienced steady improvement


in CPU performance and, for a given algorithm, increased
performance generally requires increased data rates. In the high
performance computing world, this is reported in bytes/FLOP but
its just as relevant in the commercial processing world. More
CPU performance requires more memory bandwidth to get value
from that increase in performance. Otherwise, the faster processor
just spends more time in memory stalls and doesnt actually get
more work done. For the bulk of the last 25 years, CPU
performance improvements have been driven by design
The CEMS project focused on the latter, increasing the efficiency
improvements and clock frequency increases. Having hit the
Lilius
Green Computing
of the servers themselves.
power wall, were now less reliant on clock frequency
improvements than in the past and more dependent upon increases
4. CEMS Introduction
in core counts. But the net is that processor performance
From the previous section, we understand that 59% of the power
continues to grow unabated and this is expected to continue.
dissipated in a high-scale data center is delivered to the critical
load. Generally that is a good thing in that power delivered to the
Looking at the first row of Table 1, from Dave Pattersons

20/67

The economics of Data-Centers: Why Power Matters

Power losses in Servers

100%

45%

90%

40%

80%

Power per com ponent

50%

35%

70%

30%

60%

25%

50%

20%

40%

15%

30%

10%

20%

5%

10%

0%

C u m u l a t iv e p o w e r

Electricity consumption in a typical data


center

0%
Load

PSU

Chiller

UPS

VRs

Server CRAC fan PDU CW pump Total


fans
baseline

Source: Intel Systems Technology Lab 2008

Johan Lilius

Green Computing

21/67

The economics of Data-Centers: Why Power Matters

Heat density I
Power dissipated / per unit area
Chip-level issue
Building-level issue

Chip level 40W/cm2

Shuttle re-entry 100W/cm2

Building level
Blade server > 5kVA
Finnish sauna 6kVA
Most dense datacom products > 9kVA

One cannot pack servers into a small space


Cooling & Air Handling Gains

Intel

Verari

Tighter control of air-flow increased delta-T


Container takes one step further with very
little air in motion, variable speed fans, &
tight feedback between CRAC and load
Sealed enclosure allows elimination of small,
inefficient (6 to 9W each) server fans
Intel

2009/4/1

Johan Lilius

http://perspectives.mvdirona.com

14

Green Computing

22/67

The economics of Data-Centers: Why Power Matters

Processor

Issues:
1
2

Johan Lilius

Where does processor power go?


What can we do about it?

Green Computing

23/67

The economics of Data-Centers: Why Power Matters

Processor Power I

22% yearly increase

Johan Lilius

Green Computing

24/67

The economics of Data-Centers: Why Power Matters

Processor Power II
Concrete values for AMD and Intel

Johan Lilius

Green Computing

25/67

The economics of Data-Centers: Why Power Matters

Processor Power III


Cooler in 1993 and in 2005

Johan Lilius

Green Computing

26/67

The economics of Data-Centers: Why Power Matters

Processor Power IV

How far can you go?


AMD FX Processor Takes Guiness World Record
http://bit.ly/ux5ORA

Johan Lilius

Green Computing

27/67

The economics of Data-Centers: Why Power Matters

Power budget of processor

Exact numbers dicult if impossible to obtain


Intel Penryn (2007) uses 50% of silicon for L2/L3 caches
Measurement on ARM shows that memory subsystem
uses 50% of total power

Johan Lilius

Green Computing

28/67

The economics of Data-Centers: Why Power Matters

Processor power
Power as function of voltage and frequency
P = c V2 f
c is activity factor
What can you do?
1

Use techniques to shutdown unused parts of the chip


Adds complexity to chip, shutdown circuits add to
energy consumption

Run the chip at lower frequencies and/or lower voltage


Does not scale linearly downwards: running at 20% of
capacity may still use 50% of energy

Dead-end?

Johan Lilius

Green Computing

29/67

Energy-proportional computing

Contents
1 Introduction
2 The economics of Data-Centers: Why Power Matters
3 Energy-proportional computing
4 Servers built on mobile processors
5 Summary
6 Bibliography

Johan Lilius

Green Computing

30/67

Energy-proportional computing

Introduction
Energy-proportional computing introdcued by Barroso
and Hlzle at Google[2]
ments because minor traffic fluctuations or any internal disruption, such as hardware or
software faults, could tip it over
the edge. Moreover, the lack of a
0.025
reasonable amount of slack
makes
regular
operations
exceedingly complex because
0.02
any maintenance task has the
potential to cause serious service
disruptions. Similarly, well-pro0.015
visioned services are unlikely to
spend significant amounts of
time completely idle because
0.01
doing so would represent a substantial waste of capital.
Even during periods of low ser0.005
vice demand, servers are unlikely
to be fully idle. Large-scale services usually require hundreds of
0
servers and distribute the load
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0
over these machines. In some
CPU utilization
cases, it might be possible to
completely idle a subset of servers
Figure 1. Average CPU utilization of more than 5,000 servers during a six-month period.
during low-activity periods by,
Servers are rarely completely idle and seldom operate near their maximum utilization,
for example, shrinking the numinstead operating most of the time at between 10 and 50 percent of their maximum
ber of active front ends. Often,
utilization levels.
though, this is hard to accomplish because data, not just comvoltage-frequency scaling. Mobile devices require high putation, is distributed among machines. For example,
performance for short periods while the user awaits a common practice calls for spreading user data across
response, followed by relatively long idle intervals of many databases to eliminate the bottleneck that a censeconds or minutes. Many embedded computers, such tral database holding all users poses.
as sensor network agents, present a similar bimodal
Spreading data across multiple machines improves
usage model.4
data availability as well because it reduces the likeliThis kind of activity pattern steers designers to empha- hood that a crash will cause data loss. It can also help
size high energy efficiency at peak performance levels hasten recovery from crashes by spreading the recovand in idle mode, supporting inactive low-energy states, ery load across a greater number of nodes, as is done
such as sleep or standby, that consume near-zero energy. in the Google File System.6 As a result, all servers must
However, the usage model for servers, especially those be available, even during low-load periods. In addition,
used in large-scale Internet services, has very different networked servers frequently perform many small backcharacteristics.
ground tasks that make it impossible for them to enter
Figure 1 shows the distribution of CPU utilization lev- a sleep state.
els for thousands of servers during a six-month interWith few windows of complete idleness, servers canval.5 Although the actual shape of the distribution varies not take advantage of the existing inactive energysignificantly across services, two key observations from savings modes that mobile devices otherwise find so
Figure 1 can be generalized: Servers are rarely com- effective. Although developers can sometimes restrucpletely idle and seldom operate near their maximum uti- ture applications to create useful idle intervals during
lization. Instead, servers operate most of the time at periods of reduced load, in practice this is often difficult
between 10 and 50 percent of their maximum utiliza- and even harder to maintain. The Tickless kernel7 exemtion levels. Such behavior is not accidental, but results plifies some of the challenges involved in creating and
from observing sound service provisioning and distrib- maintaining idleness. Moreover, the most attractive inacuted systems design principles.
tive energy-savings modes tend to be those with the highAn Internet service provisioned such that the average est wake-up penalties, such as disk spin-up time, and
load approaches 100 percent will likely have difficulty thus their use complicates application deployment and
meeting throughput and latency service-level agree- greatly reduces their practicality.
Fraction of time

0.03

Observation: Servers are idling a lot of the time


Where does idling come from:
There is not enough computation
Waiting for I/O
Waiting for memory
Johan Lilius

Green Computing

34

31/67

Computer

Energy-proportional computing

Energy-proportional computing I
Barroso and Hlzle suggest dynamic power control of
nodes in the datacenter
The datacenter would only keep a needed number of
nodes online to match the computational requirements
Goal: given a load curve match it with computational
power as closely as possible
Energy Waste
Energy Waste

Computation
Requirement
Energy Waste
Energy Waste
Computation
Requirement

Johan Lilius

Green Computing

32/67

Energy-proportional computing

Energy-proportional computing II

Big servers, large granularity => low


energy-proportionality
Small servers, small granularity => better
energy-proportionality
Small servers will have less computational power, is this
a problem?

Johan Lilius

Green Computing

33/67

Energy-proportional computing

Memory Wall

big discrepancy between processor speed and memory


access speeds
processor speed has been growing with 50-100%
annually
memory speed has been growing with 7% annually

speed gap that is growing all the time


to avoid wait times, and keep deep pipelines working,
large caches are required, which use a lot of power
complex memory subsystems DDR1-3 are developed
which also consume power

Johan Lilius

Green Computing

34/67

Energy-proportional computing

Amdahl blades[12] I

A balanced computer system needs:


sequential I/O per sec - Amdahl number
2 memory with Mbyte/MIPS ratio of close to 1 - Amdahl
memory ratio
3 performs on I/O operation per 50k instructions Amdahl IOPS ratio
1

Graywulf system (2008, state of the art architecture)


Amdahl number 0.56, Amdahl ratio 1.12, Amdahl IOPS
ratio 0.014
Data cannot be fed fast enough into the system

Johan Lilius

Green Computing

35/67

Energy-proportional computing

Amdahl blades[12] II

Compare Graywulf with COTS systems


Table 2: Performance, power, and cost characteristics of various data-intensive architectures.

GrayWulf
ASUS
Intel
Zotac
AxiomTek
Alix 3C2

CPU
[GHz]

Mem
[GB]

SeqIO
[GB/s]

RandIO
[kIOPS]

Disk
[TB]

Power
[W]

Cost
[$]

21.3
1.6
3.2
3.2
1.6
0.5

24
2
2
4
2
0.5

1.500
0.124
0.500
0.500
0.120
0.025

6.0
4.6
10.4
10.4
4.0
N/A

22.5
0.25
0.50
0.50
0.25
0.008

1,150
19
28
30
15
4

19,253
820
1,177
1,189
995
225

Relative
Power
1.000
0.017
0.024
0.026
0.013
0.003

Amdahl numbers
Seq Mem Rand
0.56
0.62
1.25
1.25
0.60
0.40

1.13
1.25
0.63
1.25
1.25
1.00

0.014
0.144
0.156
0.163
0.125

(essentially the same today). We note that for the


three times more power than the equivalent Intel
It2008
would
seem that running atmately
lower
frequencies gives
SSD-based systems the cost and disk size columns in
system (240 W vs. 84 W).
Table 2 represent projections for a 250 GB drive with
better
powerand eciency?
the same performance
a projected cost of $400 at

Johan Lilius

the end of 2009, in line with historic SSD price trends.


Power consumption varies between 15W-30W depending on the chipset used (945GSE, USW15, ION) and
generally agrees with the values reported in the motherboards specifications. The current university rate for
electric power at JHU is $0.15/kWh. The total cost of
power should include the cost for cold water and air conditioning, thus we multiply the electricity cost by 1.6
[7]. Table 2 presents these cumulative costs.
Lastly, we present the different Amdahl numbers and
ratios for the various node types. It is clear that, compared to the GrayWulf and Alix, the Atom systems, esGreen Computing
pecially with dual cores, are better balanced across all
three dimensions.
Scaling Properties. Table 3 illustrates what happens
when we scale the other systems to match the Gray-

5. DISCUSSION

The nature of scientific computing is changing it is


becoming more and more data-centric while at the same
time datasets continue to double every year, surpassing petabytes. As a result, the computer architectures
currently used in scientific applications are becoming
increasingly energy inefficient as they try to maintain
sequential read I/O performance with growing dataset
sizes. The scientific community therefore faces the following dilemma: find a low-power alternative to existing systems or stop growing computations on par with
the size of the data. We thus argue that it is unavoidable
36/67
to build scaled-down and scaled-out systems comprising large numbers of compute nodes each with a much
lower relative power consumption at a given sequential
read I/O throughput.

Servers built on mobile processors

Contents
1 Introduction
2 The economics of Data-Centers: Why Power Matters
3 Energy-proportional computing
4 Servers built on mobile processors
5 Summary
6 Bibliography

Johan Lilius

Green Computing

37/67

Servers built on mobile processors

Introduction

Mobile processors are very energy ecient


Could we used mobile processors (e.g ARM) to
implement energy proportional data centers?
Research problems:
Are mobile processors more energy ecient compared to
IA processors?
How would one implement energy proportionality?

Our goal: build an energy proportional cluster based on


ARM processors

Johan Lilius

Green Computing

38/67

Servers built on mobile processors

Commercial drivers

Money
Mobile processors are cheap

Optimized for low energyconsumption


Architecture
Low-power states

Heat-density
It will be possible to place servers into spaces where it
currently is not possible

Johan Lilius

Green Computing

39/67

Servers built on mobile processors

Energy eciency of ARM processors


Comparison of Energy per Instruction (i.e. what could
Challenges
be ideally
achieved?)
Energy efficiency: ARM vs Pentium
Processor

EPI(nJ)

NominalCPI

Power consumption

ARM720T

0.22

2.2

65mW

ARM926EJ-S

0.46

1.6

95mW

ARM1136J-S

0.63

1.4

115mW

Processor

EPI (nJ)

Nominal CPI

Power consumption

Pentium 4

48

2.59

5060mW

Pentium M

15

2320mW

Core Duo

11

1190mW

ARM720T is 220 times more energy efficient than a Pentium 4

Do 220 ARM720Ts have more computational power


than a Pentium 4? For which workloads?
9/3/10&

Johan Lilius

5&

Green Computing

40/67

Servers built on mobile processors

Hardware
Versatile Express Quad-core Cortex A9, 1GB DDR2,

400Mhz
Tegra 250 Dual-core Cortex-A9, 1GB DDR2, 1G

Johan Lilius

Green Computing

41/67

Servers built on mobile processors

Benchmarks

Autobench and Apache 2 HTTP server


static web pages

SPECweb2005
more demanding web services

Erlang
micro benchmarks
real world SIP proxy

Johan Lilius

Green Computing

42/67

Servers built on mobile processors

Autobench and Apache

Johan Lilius

Green Computing

43/67

Servers built on mobile processors

SPECWeb2005

Johan Lilius

Green Computing

44/67

Servers built on mobile processors

Erlang

Johan Lilius

Green Computing

45/67

Servers built on mobile processors

Results

The performance of 2 ARMv7 based ARM cortex-A9


was measured and evaluated and compared to Xeon
processors
Measurements show that the Cortex A9 can be up to
11 times more ecient with the Apache server
3.6 times more ecient with Erlang base SIP proxy
2.9 times more ecient with the SPECweb2005

Johan Lilius

Green Computing

46/67

Servers built on mobile processors

Implementation of energy proportianality

Since the computational capacatiy of an ARM processor


is lower that an IA processer we need more of them
This enables a more fine-grained control of the
computational power

Johan Lilius

Green Computing

47/67

Servers built on mobile processors

Evaluation Platform

Few fast CPUs dissipate much power and have rough


power granularity
Instead use slow but many mobile CPUs to increase the
power granularity
Evaluation done on a cluster using ARM Cortex-A8
(Beagleboard)
Low energy consumption and low price tag
Running Ubuntu 11.04 with Linux 2.6.34

Johan Lilius

Green Computing

48/67

Servers built on mobile processors

The problem with DVFS


The CPU is the main energy consumer in the server
Power managers have been used to scale the
performance of the CPU according to the demand
DVFS scales the voltage and frequency
DVFS does not scale the power dissipation linearly !

Johan Lilius

Green Computing

49/67

Servers built on mobile processors

Power Manager using Sleep States


The power manager is used to dynamically adjust the
system capacity to the workload
The manager wakes up cores when the capacity is too
low and shuts down cores when capacity is unnecessary
high
Goals:
Good power-to-workload proportionality resulting in
little energy waste
Show a minimal performance degradation to users
Scale in large clusters

Johan Lilius

Green Computing

50/67

Servers built on mobile processors

System level power manager


Switches on/o CPU cores in the cluster
Operates on systems level, meaning that it controls the
whole cluster as one entity
Master core controlling the other workers
Workers taking orders: Sleep/Wake and workload

Johan Lilius

Green Computing

51/67

Servers built on mobile processors

Simulation framework

Johan Lilius

Green Computing

52/67

Servers built on mobile processors

Simulation framework
Quality of Service (QoS) is our current measurement of
performance
QoS shows how many % of the incoming requests are
handled in a time frame
The unhandled requests are moved to the next time
frame and result in a QoS drop
QoS drop is a result of latency that triggers a deadline
miss
Simulation framework can select static cores that are
not altered by the power manager, i.e. always running

Johan Lilius

Green Computing

53/67

Servers built on mobile processors

CPU characteristics

Data based on ARM Cortex-A8 benchmarks


The power dissipation of the CPU is 1.4W at full speed
Waking up a core takes about 650 ms
The load capacity of a CPU was benchmarked with the
tool Autobench resulting in 5 requests/second for a 248
KB file
An arbitrary amount of CPUs can be simulated with the
framework assumed that the CPU specs are given

Johan Lilius

Green Computing

54/67

Servers built on mobile processors

Comparison with DVFS

Johan Lilius

Green Computing

55/67

Servers built on mobile processors

Power savings vs QoS

Johan Lilius

DVFS has the highest QoS but wastes much energy


Up to 60% in energy can be saved with only 4%
degradation in QoS
20% energy can still be saved if less than 1% QoS
degradation isGreen
requested
Computing

56/67

Servers built on mobile processors

Other results and future work


Implementation of a real-life demo based on cluster
consisting of 8 XM BeagleBoards
This is done and we have validated the results

Implementation of PID power manager into Linux


Scheduler
Validated the approach using LinSched
Currently implementing the framework into Linux kernel

Video transcoding over the cloud


Evaluation of dierent control mechanisms
State based control theory vs. PID

Evaluation of more complex settings:


Is the approach applicable with VMs running on the
cluster?

Johan Lilius

Green Computing

57/67

Summary

Contents
1 Introduction
2 The economics of Data-Centers: Why Power Matters
3 Energy-proportional computing
4 Servers built on mobile processors
5 Summary
6 Bibliography

Johan Lilius

Green Computing

58/67

Summary

Summary
1

PUE measures the overall power loss in the datacenter,


it does not measure the total power consumption

ARM based processors are more energy ecient that IA


based processors for typical datacenter loads

IA based processors are more energy ecient in


computations with large computational kernels

Energy Proportionality is the key to lower total power


consumption in datacenters

It is feasible to implement Energy Proportional Power


Managers into modern OS kernels

Many interesting research challenges remain!

Johan Lilius

Green Computing

59/67

Summary

Acknowledgements
The Lean Server Team
Prof. Johan Lilius
Dr. Sebastien Lafond
M.Sc. Simon Holmbacka
M.Sc. Fareed Johkio
M.Sc. Tewodros Deneke
M.Sc Fredric Hllis
Alumni
M.Sc.
M.Sc.
M.Sc.
M.Sc.

Johan Lilius

Jens Smeds
Olle Swanfeldt-Winter
Joachim Sjlund
Joakim Nylund

Green Computing

60/67

Summary

Our Papers I
Simon Holmbacka, Sbastien Lafond, Johan Lilius, A
PID-Controlled Power Manager for Energy Ecient Web
Clusters. In: International Conference on Cloud and
Green Computing (CGC2011), 721-728, IEEE Computer
Society, 2011.
Simon Holmbacka, Jens Smeds, Sbastien Lafond,
Johan Lilius, System Level Power Management for
Many-Core Systems. In: (Ed.), Workshop on Micro
Power Management for Macro Systems on Chip, 2011.
Sbastien Lafond, Simon Holmbacka, Johan Lilius, A
System Level Power Management for Web Clusters. In:
COST Action IC0804 on Energy Eciency in Large Scale
Distributed Systems, 2nd Year, 127-131, IRIT, 2011.
Johan Lilius

Green Computing

61/67

Summary

Our Papers II
Sbastien Lafond Johan Lilius Simon Holmbacka, Power
Proportional Characteristics of an Energy Manager for
Web Clusters. In: Embedded Computer Systems:
Architecture, Modeling and Simulation (SAMOS) 2011,
8, IEEE pres, 2011.
Olle Svanfeldt-winter, Sbastien Lafond, Johan Lilius,
Cost and Energy Reduction Evaluation for ARM Based
Web Servers. In: International Conference on Cloud and
Green Computing (CGC2011), 480-487, IEEE Computer
Society, 2011.

Johan Lilius

Green Computing

62/67

Bibliography

Contents
1 Introduction
2 The economics of Data-Centers: Why Power Matters
3 Energy-proportional computing
4 Servers built on mobile processors
5 Summary
6 Bibliography

Johan Lilius

Green Computing

63/67

Bibliography

References I
David Andersen, Jason Franklin, Michael Kaminsky,
Amar Phanishayee, Lawrence Tan, and Vijay Vasudevan.
FAWN: a fast array of wimpy nodes.
SOSP 09: Proceedings of the ACM SIGOPS 22nd
symposium on Operating systems principles, October
2009.
LA Barroso and U Holzle.
The case for energy-proportional computing.
Computer, 40(12):3337, 2007.
BG Chun, G Iannaccone, G Iannaccone, R Katz, G Lee,
and L Niccolini.
An energy case for hybrid datacenters.
ACM SIGOPS Operating Systems Review, 44(1):7680,
2010.
Johan Lilius

Green Computing

64/67

Bibliography

References II
James Hamilton.
Cooperative Expendable Micro-Slice Servers (CEMS):
Low Cost, Low Power Servers for Internet-Scale Services.
pages 18, December 2008.
James Hamilton.
Data Center Eciency Best Practices.
pages 129, April 2009.
Urs Holzle.
Brawny cores still beat wimpy cores, most of the time.
IEEE Micro, pages 12, June 2010.

Johan Lilius

Green Computing

65/67

Bibliography

References III
W Lang, JM Patel, and S Shankar.
Wimpy Node Clusters: What About Non-Wimpy
Workloads?
2010.
Jacob Leverich and Christos Kozyrakis.
On the Energy (In)eciency of Hadoop Clusters.
pages 15, July 2009.
Reijo Maihaniemi.
Energy Ecient ICT.
Presentation, pages 127, September 2009.
J Manko, R Kravets, and E Blevis.
Some computer science issues in creating a sustainable
world.
Computer, 41(8):102105, 2008.
Johan Lilius

Green Computing

66/67

Bibliography

References IV

S Ruth.
Green IT More Than a Three Percent Solution?
IEEE Internet Computing, 2009.
Alexander Szalay, Gordon Bell, H Huang, Andreas Terzis,
and Alainna White.
Low-power amdahl-balanced blades for data intensive
computing.
SIGOPS Operating Systems Review, 44(1), March 2010.

Johan Lilius

Green Computing

67/67

You might also like