You are on page 1of 35

Datacenter Basics

Fred Chong
290N Green Computing
Storage Hierarchy

Figure : Storage hierarchy of a Warehouse-scale computer


Performance Variations

Figure : Latency, bandwidth and capacity of a Warehouse-scale computer


Server Comparison
HP Integrity Superdome – HP ProLiant ML350 G5
Itanium2
Processor 64 sockets, 128 cores (dual- 1 socket, quad-core, 2.66GHz
threaded), 1.6GHz Itanium2, X5355 CPU, 8MB last-level cache
12MB last-level cache
Memory 2048GB 24GB
Disk 320,974GB, 7056 drives 3961GB, 105 drives
storage
TPC-C $2.93/tpmC $0.73/tpmC
price/perfor
mance
price/perfor $1.28/transactions per minute $0.10/transactions per minute
mance
(server HW
only)
price/perfor $2.39/transactions per minute $0.12/transactions per minute
mance
(server HW
only) (no
discounts)
Cost proportional to Power
• Cost proportional to power delivered
• Typically $10-20/W
• Power delivery
– 60-400kV transmission lines
– 10-20kV medium voltage
– 110-480V low voltage
UPS
• Uninterruptible Power Supply
• Batteries or flywheel
• AC-DC-AC conversion
• Conditions the power feed
– Removes spikes or sags
– Removes harmonic distortions
• Housed in a separate UPS room
• Sizes range from hundreds of kW to 2MW
PDUs
• Power Distribution Units
• Breaker panels
– Input 200-480V
– Output many 110V or 220V
– 75-225kW in 20-30A circuits (max 6 kW)
• Redundancy from two independent power
sources
Paralleling
• Multiple generators or UPSs
– Feed a shared bus
• N+1 (one failure)
• N+2 (one maintenance, one failure)
• 2N (redundant pairs)
Cooling
Cooling Steps
• 12-14 C coolant
• 16-20 C air at CRAC (Computer Room AC)
• 18-22 C at server intake
• Then back to chiller
“Free Cooling”
• Pre-cool coolant before chiller
• Water-based cooling towers use evaporation
– Works in moderate climate – freezes if too cold
• Glycol-based radiator outside the building
– Works in cold climates
Cooling is Critical
• Datacenter would fail in minutes without
cooling
– Cooling backed up by generators and UPSs
• Adds > 40% critical electrical load
Airflow
• 100 cfm (cubic feet per minute) per server
• 10 servers would require 1000 cfm from
perforated tiles
• Typically no more than 150-200W / sq ft
power density
• Recirculation from one server’s hot air into the
intake of a neighbor
– Some avoid with overhead ducts
Variations
• In-rack cooling
– Water cooled coils next on the server
– Cost of plumbing
– Damage from leaks (earthquake zones!)
• Container-based datacenters
– Shipping container 8’ x 8.5’ x 40’
– Similar to in-rack cooling but for the whole
container
– Higher power densities
Power Efficiency
• PUE – power usage efficiency
– Datacenter power infrastructure
Poor PUEs
• 85% of datacenters PUE > 3
• Only 5% PUE = 2.0
• Chillers take 30-50% overhead
• CRAC 10-30% overhead
• UPS 7-12% overhead (AC-DC-AC)
• Humidifiers, PDUs, lighting
• EPA “achievable” PUE of 1.4 by 2011
Improvements
• Evaporative cooling
• Efficient air movement
• Eliminate power conversion losses
• Google PUE = 1.21
• Several companies PUE = 1.3
A more comprehensive metric
Computation  1   1   Computation 
Efficiency       
Total Energy  PUE   SPUE   Total Energy to Electronic Components 
(a) (b) (c)

• (b) SPUE – server power usage efficiency


• (c) computation energy efficiency
SPUE
• Power delivered to components directly
involved in computation:
– Motherboad, disks, CPUs, DRAM, I/O cards
• Losses due to power supplies, fans, voltage
regulators
• SPUE of 1.6-1.8 common
– Power supplies less than 80% efficient
– Voltage regulators less than 70% efficient
• EPA feasible SPUE < 1.2 in 2011
TPUE
• Total PUE = TPUE = PUE * SPUE
• Average of 3.2 today (2.2 Watts wasted for
every Watt in computation)
• PUE 1.2 and SPUE 1.2 would give 2X benefit
• TPUE of 1.25 probably the limit of what is
economically feasible
Computing Efficiency
• Area of greatest potential
• Hardest to measure
• SPECpower
• Joulesort
• Storage Network Industry Association
SPECPower Example
Server Load
Load vs Efficiency
Pwr50 Pwr10 Pwr10sub Eff50 Eff10 Eff10sub

180

160

140

120

100
(%)

80

60

40

20

0
0 5 10 15 20 25 30 35 40 45 50 55 60 65 70 75 80 85 90 95 100
Load level (% of peak)
Sleep

Knitting

Playing cards

Typing

Bowling

Cooking

Ballroom

Laundry

Walking

Soccer

Tennis

Lambada

Cycling (race)

Jumping rope

Running

Ax chopping
Human Dynamic Range

0
500
1000
1500
2000

Watts
Component Efficiency
CPU DRAM Disk Other

100.00

90.00

80.00

70.00
Power (% of peak)

60.00

50.00

40.00

30.00

20.00

10.00

0.00
Idle 7 14 21 29 36 43 50 57 64 71 79 86 93 100
Compute load (% )
CPU Voltage Scaling
100
2.4 GHz
90

1.8 GHz
80

70
Power (% of peak)

1 GHz
60

50

40

30

20

10
DVS savings (%)
0
Idle 7 14 21 29 36 43 50 57 64 71 79 86 93 100
Compute load (%)
Disks
• As much as 70% power to keep drives spinning
• 1000X penalty to spin up and access
• Multiple head, low RPM drives [Gurumurthi]
Server Power Supplies
Power Provisioning
• $10-22 per deployed IT Watt
• Given 10 year depreciation cycle
– $1-2.20 per Watt per year
• Assume $0.07 per kilowatt-hr and PUE 2.0
– 8766 hours in a year
– (8766 / 1000) * $0.07 * 2.0 = $1.22724
• Up to 2X cost in provisioning
• eg. 50% full datacenter = 2X provisioning cost
Time at Power Level

80 servers
800 servers
8000 servers
Oversubscription Opportunity
• 7% for racks (80)
• 22% for PDUs (800)
• 28% for clusters (8000)
– Could have hosted almost 40% more machines
Underdeployment
• New facilities plan for growth
• Also discretization of capacity
– Eg 2.5kW circuit may have four 520W servers
• 17% underutilized, but can’t have one more

You might also like