You are on page 1of 108

Data Center

1.

INTRODUCTION

2.

OVERVIEW AND ROLE OF DATA CENTER

3.

DATA CENTER STANDARD TIA-942

4.

DATA CENTER HISTORY FROM 60S TO THE


CLOUD

5.

GREEN DATA CENTER

DCDA Consulting Services

DCDA Consulting Services is


what it is abbreviated from, Data
Center Design and Assessment.
We concentrate our service in
Designing Data Centers; from
scratch, or motivated by
thorough assessments of your
existing Data Centers.

DCDA Consulting is
proud to be one of the
very few local companies
who specializes in Data
Center Design and
Assessment .

DCDA Consulting Services

HIGH
AVAILABILITY
NETWORK

EXPERT
Trained Employees

PROCESSES
Standardization, Simplicity, Documentation

TECHNOLOGY
Data Processing, Communication, Data Storage

Reliability
Basis

ESSENTIAL NETWORK PHYSICAL INFRASTRUCTURE


Power Supply, Rack, Cooling, Maintenance & Management

DCDA Consulting Services


Core Services
Data Center Infrastructure System

Design and Assessment

Services
that meets TIA-942 standards and best practices

Extended Services
To provide you with comprehensive assistance, we follow
through our design and aid you with the implementation
of your Data Center project.

Data Center Project Management


Control and manage the construction and integration of
all aspects of your data center.

Structured Cabling System Design


For Data Centers, Offices, Plants and Industries.

Training on Data Center Planning and


Design Customized training on Data Center Planning
and Design.

Design & Assessment

Scope of Services
1. Floor Plan (Layout) of a Data Center (Architecture &
Interior)

2. Structured Cabling System / Cable Management /


Telecommunication

3. Electrical Design & Power Management


4. Cooling
5. Physical Security System (Access Control System
and CCTV System)

6. Fire Protection System (Detection and


Suppression)

7. Environmental Monitoring System (EMS)


 (optional
optional) Structural Evaluation

The final deliverable is in the form of a Tender Document


that cover client requirements, design and material
specification, inclusive of shop drawings, and complete
BoQ for item 1 to 7.

DCDA Consulting Services


List of Experiences

No

Client

Data Center Project

Dimension

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18

ICON+ PLN
Sigma
Bentoel Malang
HMSampoerna
Hortus Securitas
Lintas Arta JTL
BII Data Center
Depdiknas Puspitek
Medco HP,
Teradata, Bandung
Bukopin DC Renovation
ICON+ PLN
Barclay Bank
Bakri Bumi Resources
ICON+ PLN
PT. Dizamatra
Bank Indonesia
PT. POS Indonesia (Persero)

Design
Design
Design
Design
Server Room Design
Design
Assessment Services (AUDIT)
Design
Design
Design
Project Management
Tender Doc
DC & SCS Project Management
Design
Project Management
Business Plan & DC Design
Design & Modelling
Design, Assessment

800 m2
800 m2
100 m2
100 m2
30 m2
100 m2
500 m2
120 m2
135 m2
100 m2
200 m2
800 m2
150 m2 &100 m2
80 m2
800 m2
500 m2
1451 m2
300 m2 & 200 m2

19
20
21
22
23

Bakri Bumi Resources


Niko Resources
PLN Pusat
Kementerian Keuangan
Garuda Indonesia

Project Management
Design Server Room
Project Management
Data Center Design
Data Center Assesment and Design

80 m2
15 m2
576 m2
1000 m2
1070m2

Tier
Tier3 to 4
Tier3
Tier2
Tier2
Tier2
Tier2
Tier1
Tier2
Tier2
Tier2
Tier2
Tier3 to 4
Tier2
Tier2
Tier3 to 4
Tier2
Tier3
Tier2
Tier2
Tier1
Tier2
Tier2
Tier3

7
I.

Our
Approach

TIA-942 Telecommunications
Infrastructure Standards for Data Centers

II. Best Practices and Green DC


III. Real Hands-on Experiences of 20++ Data Center Projects
IV. Keep up with the latest development of DC

Data Center
Design &
Assessment
Reference

Data Center

1.

INTRODUCTION

2.

OVERVIEW AND ROLE OF DATA CENTER

3.

DATA CENTER STANDARD TIA-942

4.

DATA CENTER HISTORY FROM 60S TO THE


CLOUD

5.

GREEN DATA CENTER

What is a Data Center


A DATACENTER IS A FACILITY THAT IS DESIGNED
AND BUILT SPECIFICALLY TO HOUSE COMPUTERS,
STORAGE ARRAYS, AND TELECOMMUNICATIONS
(NETWORKING) SYSTEMS.

Data Centers, are specialized building or rooms,


hosting ICT equipments such as computing and networking
systems to keep the companys critical applications and
websites running smoothly and properly (24/7 x 365 days and
ongoing)

Sample Medium to Large Data Center

10

Small Data Center & Server Room

11

Movie of a Data Center


12

SECURE-24
OR

PARIS TELECITY

Important role of data centers


Data centers play an important role in today's
business & personal computing, our daily life and
are integral to cloud computing.

In most of the cities, our life relies on the functioning and


availability of one or multiple data centers. It is not an
overstatement.
Most of the things in every segment of human activity such as
energy, lighting, telecommunications, internet, transport, urban
traffic, banks, security systems, public health, and entertainment
are controlled by data centers.

13

Today everyone use Data


Center services

14

15

Most of Mobile-Phone
Facilities, Features and
Applications use Data
Center Services
(from SMS, Chat, Social
Networkingetc)

16

Most Trafficked Sites Online April 2012

17

We live in an Exponential World

Time to reach target market of 50


million people
40

Radio
38 years

35

Years

30
25
20
15
10
5
0

18

Television
13 years
Internet
4 years

iPod
3 years

Facebook
2 years

We live in an Exponential World

With 900 million registered users,


Facebook would be the 3rd largest
country in the world

Facebook

19

users

2009

We live in an Exponential World

First commercial text message was


sent in December of 1992

20

Today, the number of text


messages sent and received
everyday exceeds the total
population on the planet.

So, why Data Center is


Important?
The millions of web
pages are provided
by servers and their
supporting
infrastructure.

The millions of
Facebook users
are supported by
servers and their
supporting
infrastructure.

Every one of
The 30+ billion
Google searches
performed every
month are run on
servers and their
supporting
infrastructure

21

The billions of
text messages
sent every day
are supported
by servers and
their
supporting
infrastructure.

Data Center Today

Data Center has evolved from


the domain of the IT
Department to being a key topic
in a boardroom

22

Data Center

23

1.

INTRODUCTION

2.

OVERVIEW AND ROLE OF DATA CENTER

3.

DATA CENTER STANDARD TIA-942

4.

DATA CENTER HISTORY FROM 60S TO THE


CLOUD

5.

GREEN DATA CENTER

Data Center Standard TIA-942


TELECOMMUNICATIONS INFRASTRUCTURE
FOR DATA CENTERS

24

What is a Data Center


25

Called or written as:


 Computer Room
 Data Center
 Data Centre
 Server Room
 Network Room
 Network Closer
 Telecommunication Room

What is a Data Center




Data Center is the description given when the entire site and building are utilized
exclusively as a data center site.

 Data Center is a larger space composed of:

Computer Room(s)
Telecommunication Room(s)
Entrance Room(s)
Mechanical Room(s)
Electrical Room(s)
Network Operations Center
Staging Area, Storage, Loading Dock
Common Areas
General Office Space
26

Data Center Function


Provide a safe and secure place to
locate mission critical equipment
Provide sufficient power protection
to maintain the critical load
Provide adequate environmental
controls to operate the space within
the parameter of the critical load
Provide communication
connectivity both inside and outside
of the space

27

OVERVIEW OF DATA CENTER


Based on

TIA--942 & Relevant Update


TIA

TIA-942 Telecommunications
TIAInfrastructure Standards for Data Centers
 Intended for use by data center designers early in the building development process,

TIA--942 covers the following


TIA

Tiered reliability
Site space and layout
Cabling management & infrastructure
Environmental considerations
29

1. TIA
TIA--942 Data Center Tiers classifications
What does they mean?
A simple explanation is:
Guidelines to build a DC based on the level of performance as desired/dictated by the business
requirement

Tier 1

Tier I, single path for power and cooling distribution, NO redundant components;
99.671% availability (max cumulative annual downtime is 346 hrs or 14.5 days).

Tier II, single path for power and cooling distribution, WITH Redundant
components; 99.741% availability (max cumulative annual downtime is 272 hrs or
11.5 days).

Tier 2

Tier III, is composed of multiple active power and cooling distribution paths, but
onlyIII,
one
active,of
redundant
components,
andcooling
is concurrently
maintainable,
Tier
is path
composed
multiple active
power and
distribution
paths, but only
(max cumulative
annual
downtime
is 19 hrsmaintainable,
or < 1 day).
99.982%
availability
one
path active,
has redundant
components,
and
is concurrently

Tier 3

99.982% availability (max cumulative annual downtime is 19 hrs or < 1 day).




Tier 4

Tier IV, is composed of multiple active power and cooling distribution paths, has

Tier
IV, is composed
of multiple
active
power and
coolingavailability
distribution(max
paths,
has
redundant
components,
and is fault
tolerant;
99.995%
cumulative
redundant
components,
and is fault tolerant; 99.995% availability (max cumulative
annual downtime
is 5hrs)
annual downtime is 5hrs)
 Source: Uptime Institutes

30

DC Tiers: Similarities and Differences


TIER I

TIER II

TIER III

TIER IV

Building Type

Tenant

Tenant

Stand-alone

Stand-alone

Staffing

None

1 shift

1 + shitfs

24 by forever

Number of delivery paths

Only 1

Only 1

1 active
1 passive

2 active

Redundant Components

N+1

N+1 or N + 2

2 (N + 1) or S + S

Support space to raised floor ratio

20%

30%

80% - 90%

100%

Initial watts/ft2

20-30

40-50

40-60

50-80

Ultimate watts/ft2

20-30

40-50

100-150

150+

Raised floor height

12" / 30cm

18/45cm

30"
30"--36" /
75
75--90cm

30"-36" /
75
75--90cm

150
85 (=734)
(=415)

175
100(=857)
=488)

250
150 (=1225)
(=732)

250 (=1225)
150+
(GE 732)

Utility voltage

208, 480

208, 480

12-15kV

12-15kV

Months to implement

3 to 6

15 to 20

15 to 20

Year first deployed

1965

1970

1985

1995

Construction $/ft2 raised floor

$450

$600

$900

$1,100+

Annual IT downtime due to site

28.8 hrs

22.0 hrs

1.6 hrs

0.4 hrs

Site Availability

99.671%

99.749%

99.982%

99.995%

Floor loading pounds/ft2 (kg/sqm)

Source: Uptime Institute

TIA-942 higher requirements

31

Tiers other Highlights


Tier-I

Tier-II

Tier-III

Tier-IV

ARCHITECTURAL
Multi-tenant occupancy within
building

no restriction

Allowed only if
occupancies are nonhazardous

Allowed if all tenants


are data centers or
telco companies

Allowed if all tenants


are data centers or
telco companies

Ceiling Height

2.6 m (8.5 ft)


minimum

2.7 m (9.0 ft)


minimum

3 m (10 ft) minimum

3 m (10 ft) minimum

Operations Center

no requirement

no requirement

yes

Yes

Floor loading capacity


superimposed live load

7.2 kPa (150 lbf/sq ft)


= 734 kg/m2

8.4 kPa (175 lbf/sq ft)


= 857 kg/m2

12 kPa (250 lbf/sq


ft) = 1225 kg/m2

12 kPa (250 lbf/sq


ft) = 1225 kg/m2

Floor hanging capacity for


ancillary loads
suspended from below

1.2 kPa (25 lbf/sq ft)


= 122 kg/m2

1.2 kPa (25 lbf/sq ft)


= 122 kg/m2

2.4 kPa (50 lbf/sq ft)


= 245 kg/m2

2.4 kPa (50 lbf/sq ft)


= 245 kg/m2

150 (=734)

175 (=857)

250 (=1225)

250 (=1225)

Structural

Floor loading
pounds/ft2 (kg/sqm)

TIA-942 Telecommunications
TIAInfrastructure Standards for Data Centers
Intended for use by data center designers early in the building development process,

TIA--942 covers the following


TIA

Tiered reliability
Site space and layout
Cabling management & infrastructure
Environmental considerations
33

TIA-942 Telecommunications Infrastructure Standards for Data Centers

2. Site Space and Layout


Data Center should be designed with plenty of flexible empty
space that can accommodate future racks or cabinets.

The space surrounding the DC must also be considered for


future growth and planned for easy annexation.

The standard

recommends specific functional areas,


which helps to define equipment placement based on the
standard hierarchical star topology design for regular commercial
spaces.

34

Functional Areas in a Data Center


(area-area fungsional dari Data Center)
HDA (horiz. distr. area)
serve as the distribution points for
horizontal cabling and house
crossconnects and equipment for
distributing cable to the EDA.

Backbone cabling
provides connections
between ER, MDA, and
HDAs

ER (entrance room)
is the location for access
provider equipment and
demarcation points.

MDA (main distr. area)

EDA (equipment distr. area)


contains equipment cabinets and racks.
Horizontal cables are terminated with
patch panels in the EDA.

Horizontal cabling
provides connections
between the HDA, ZDA,
and EDA.

is a centrally located area that houses


the Main Crossconnect as well as core
routers and switches for LAN and
SAN infrastructures

(Site Space and Layout) DC TOPOLOGY

TIA-942 Compliant Data Center


Access Provider

Offices, Operation
Center, Supports
Room

Entrance Room
(Career Equipment
& Demarcation)

Access Provider

Backbone
Cabling

Horizontal Cabling
Telecommunication
Room
(Office & Operation
Center LAN Switches)

(ISP/Co-Lo)

MAIN DISTRIBUTION
AREA
(Router, Backbone, LAN/SAN
switches, PBX, M13 Muxes

COMPUTER
ROOM
Backbone Cabling
Horizontal Cabling

Horizontal
Distribution Area

Horizontal
Distribution Area

Horizontal
Distribution Area

Horizontal
Distribution Area

(LAN, SAN, KVM,


Switches)

(LAN, SAN, KVM,


Switches)

(LAN, SAN, KVM,


Switches)

(LAN, SAN, KVM,


Switches)

Equipment Distribution
Area
(Rack/Cabinets Servers etc)

Equipment Distribution
Area
(Rack/Cabinets Servers etc)

Equipment Distribution
Area
(Rack/Cabinets Servers etc)

Zone Distribution Area

Equipment Distribution
Area
(Rack/Cabinets Servers etc)

36

Enterprise Data Center Design


(Corporate DC or CDC)
Access Provider

Entrance Room
(Career Equipment
& Demarcation)

Access Provider

Most common
DC Topology
 Consolidate MC

and HC (the main


cross-connect and
horizontal crossconnect) in a single
main distribution
area

 Horizontal cable

(90 m copper, 300


m optical)

Offices, Operation
Center, Supports
Room

MAIN DISTRIBUTION
AREA
(Router, Backbone,
LAN/SAN switches, PBX,
M13 Muxes)

COMPUTER
ROOM

Backbone Cabling
Horizontal Cabling

Zone Distribution
Area

Equipment Distribution
Area

Equipment
Distribution Area

Equipment
Distribution Area

(Rack/Cabinets Servers etc)

(Rack/Cabinets Servers etc)

(Rack/Cabinets Servers etc)

37

Hot & Cold Aisle approach  Dictate Rack Arrangement

Most of the Design we did is using (1)


Overhead for Telco & (2) Power under
RF

Telecommunication
Cable tray

Power Cable tray

38

TIA-942 Telecommunications
TIAInfrastructure Standards for Data Centers
Intended for use by data center designers early in the building development process,

TIA--942 covers the following


TIA

Tiered reliability
Site space and layout
Cabling management & infrastructure
Environmental considerations
39

3. Cabling Management & Infrastructure


Sample end result of direct/point-to-point connection
and example of Poor Use of Rack Space, No
Migration Planning and worse Cable Management.

40

Telecommunication  Cabling Management &


Infrastructure
1. Direct Connect
Three cabling
connection
method

2. Interconnection
3. CrossCross-connection

1. Direct Connect or point to


point is:
unstructured
Definitely will lead to common familiar
mess / spaghettis in the Data/IT room.
*) Tolerable for limited usage in the same rack.
But unacceptable for the growth expected in
Data Centers

Switches
Rack

Servers
Rack

41

Cabling Management & Infrastructure


2. InterInter-Connection
in the MDA or HDA
48 or xx ports
Patch Panel
permanent
Horizontal Cabling

Patch Cord
into Server

Servers

Patch Cord
into Switch

Switches

HDA or MDA

EDA

 Prevalent architecture in LAN, works well in smaller Data Centers


 For larger, Tier III and IV Data Centers design limits growth
42

Cabling Infrastructure

3. CrossCross-Connection
Permanent cable to EDA or SAN storage
(Backbone and Horizontal cabling)

PATCH CORD CROSS


CONNECTIONs (easier MAC
MAC))

CORE
SWITCHES

Core
Switches

EDA
(Server Room)

MDA or HDA (Network Room)

Permanent single-ended cable


to the back of wiring center

43

TIA-942 Data Center Topology and Structure


Cabling in the DC
Access Provider

Offices, Operation
Center, Supports
Room

Entrance Room
(Career Equipment
& Demarcation)

MAIN DISTRIBUTION
AREA

Access Provider

COMPUTER
ROOM

(Router, Backbone, LAN/SAN


switches, PBX, M13 Muxes)
Backbone Cabling
Horizontal Cabling

Zone Distribution
Area

Equipment Distribution
Area

Equipment Distribution
Area

Equipment
Distribution Area

(Rack/Cabinets Servers etc)

(Rack/Cabinets Servers etc)

(Rack/Cabinets Servers etc)

Logical Topologies for Data Center




While TIA-942 standards help


guide the data center physical
infrastructure, the logical data
center infrastructure does not have
a standards body helping with
design.
Logical architectures as shown
in the left Figure vary based on
customer preference and are
also guided by the electronics
manufactures.

 Most logical architectures

can be broken into four


layers:
1.
2.
3.
4.

Core
Aggregation
Access
Storage

three different types of switches




Access Switch


Core Switch


46

A core switch is located in the core of the network and serves to interconnect edge switches. The core layer routes
traffic from the outside world to the distribution layer and vice versa. Data in the form of ATM, SONET and/or
DS1/DS3 will be converted into Ethernet in order to enter the Data Center network. Data will be converted from
Ethernet to the carrier protocol before leaving the data center.

Distribution Switch


An access switch or an edge switch is the first point of user access (and the final point of exit) for a network. An edge
switch will allow the servers to connect to the network. Multimode optical fiber is the typical media that connects
the edge devices to the servers within the data center. Edge switches are interconnected by core switches.

Distribution switches are placed between the core and edge devices. Adding a third layer of switching adds
flexibility to the solution. Firewalls, load balancing and content switching, and subnet monitoring take place,
aggregating the VLANs below them. Multimode optical fiber will be the typical media running from the distribution
layer to the core and edge devices.

Not every data center will have all three layers of switching. In smaller Data Centers the core and
distribution layer are likely to be one and the same.

MAPPING LOGICAL ARCHITECTURES TO TIA-942


TIA-942 Physical
Architecture

Maps to

Logical Architecture Area

MDA=Main Distribution Area

Core and Aggregation

HDA=Horizontal Distr. Area

Aggregation

ZDA=Zone Distribution Area


EDA=Equipment Distr. Area

Access and Storage

MAPPING LOGICAL ARCHITECTURES TO TIA-942


In implementing this structured cabling
design, the datacenter will be segmented based
on the logical topology shown in left Figure.
The segmentation will be as follows:
1. Collapse the core switching LAN, SAN

and aggregation switching in the MDA


2. Segment the access layer into three zones:

front-end, application and back-end


3. Segment the storage into a separate zone.
4. Each zone will use a middle-of-the-rack

interconnect solution for the cabling.


5. Within each zone, the EDAs will utilize

a top-of-the-rack interconnect. The


EDAs will serve the electronics in each cabinet
and the ZDAs will serve the EDAs.
6. The ZDAs will homerun back to the MDA,

where they will terminate in a main crossconnect (MC).

Data Center Topologies

A good data center layout adapts flexibly to new needs and enables a high degree of
documentation and manageability at all times.

Customers can choose a variety of cabling structures from:




end of row or dual end of row

middle of row

top of rack, or

two row switching.

Patching: Interconnect & Crossconnect

In two Racks

In One Rack

End of Row

From each dualdual-end


end--ofofrow racks LAN & SAN
Switch back to MDA core
switches

Dual end of Row

Middle of Row

Sample!!! Implementation of
LOGICAL ARCHITECTURES TO TIA-942
using ZDA in Middle of Row

Top of Rack Data


(LAN switches @each-rack)

From each server rack


LAN Switch back to
MDA core switches

LAN Switch

Cisco Nexus Data Center Portfolio

Nexus 7000
Data Center Core/Aggregation

Nexus 5000
Unified Server Access

Nexus 2000
Remote Module

Nexus 1000V
VM-Aware Switching

Top of Rack Cisco


-cisco ucs Nexus 7000
From server rack (nexus 2000 or 1000V to end of row (nexus 5000)
then back to MDA (nexus 7000)

LAN Fabric
Extender

Nexus 2000 /1000V

Nexus 5000

Top of Rack Full

Each EDA Rack has


its own SAN, LAN &
KVM switch
LAN & SAN
 Back to MDA

Two Row
(traditional implementation early TIATIA-942 implementation)
implementation)

No Active (network) Devices in EDA only


cable management  cablings back to Access
& Core Switches in MDA

Sample  50
50%
% Filled Under Raised floor Tray

60

Pathways and Spaces - Overhead

Benefits
 Alleviates congestion beneath access floor
 Creation of segregated pathways
 Minimize obstruction to cold air

Concerns

Requires adequate space above racks

Infrastructure provisions to support the pathway

Cabling may be exposed

61

Pathways and Spaces Access Floor

Benefits
Pedestals create infrastructure
pathways
Utilization of real estate
Cabling is hidden

Concerns

Could restrict cold airflow

Creating segregated pathways

Accessibility to cables

62

Proper cabling Installation & labeling in the pathway, work


area, equipment area and wiring center is a must

63

Quality and Documented end result of


cabling management is highly important

64

Sample manageable cabling center

Typical large fiber patch frame

UTP Main/
Horizontal
crossconnect

65

TIA-942 Telecommunications
Infrastructure Standards for Data Centers
Intended for use by data center designers early in the building development process,

TIA--942 covers the following


TIA

Tiered reliability
Site space and layout
Cabling management & infrastructure
Environmental considerations
66

4. Environmental considerations Green Data Center

Include: Fire suppresion, humidity levels, operating temperatures,


architectural, electrical (power), and mechanical system specifications

Power requirements:
is based on the desired reliability tier and may include two or more power feeds from the utility,
UPS, multiple circuits, and on site generator.
Power needs is what is required for all existing devices and anticipated in the future. Power
requirements must also include all support equipment such as UPS, generators, conditioning
electronics, HVA, lighting, etc.

Cooling:
The standard recommends the use of adequate cooling equipment as well as a raised floor systems for
more flexible cooling.
Additionally standard states that cabinets and racks should be arranged in alternating patterns to
create hot and cold aisle.

67

Development of IT Equipment

TODAY

1964 to early 1990s

1-RU Blade Servers 


42 server per-rack
LAN 84 per-rack
SAN: 84 per cabinet

68

Blade Servers
in the 42U

Racks

7U Blade each
having 14 Servers

1-RU Blade Servers


42 server perper-rack
LAN 84 perper-rack
SAN: 84 percabinet

Blade Servers cluster


84 server perper-rack
LAN 168 perper-rack
SAN: 168 percabinet

69

Hot & Cold Aisle approach


Dictate Rack/Cabinet Arrangement

70

Cold Aisle Containment

Hot Aisle Containment

71

Data Center

72

1.

INTRODUCTION

2.

OVERVIEW AND ROLE OF DATA CENTER

3.

DATA CENTER STANDARD TIA-942

4.

DATA CENTER HISTORY FROM 60S TO THE


CLOUD

5.

GREEN DATA CENTER

Data Center
1960s
1960s

Prior to 1960 (1945), the Army developed a huge
machine called ENIAC (Electronic Numerator,
Integrator, Analyzer, and Computer):
Weighed 30 tons
Took up 1,800 sq ft of floor space
Required 6 full-time technicians to keep it running.

73

Up until the early 1960s, computers were primarily


used by government agencies. They were large
mainframes stored in rooms what we call a
datacenter today. In the early 1960s many
computers cost about $5 million each and time on
one of these computers could be rented for $17,000
per month.

By the mid 1960s, computer use developed


commercially and was shared by multiple parties.

American Airlines and IBM teamed up to develop a


reservation program termed the Sabre system. It
was installed on 2 IBM 7090 computers, located in
a specially designed computer center in Briarcliff
Manor, New York. The system processed 84,000
telephone calls per day.

Punch cards
Early computers often used punch cards for input both of
programs and data. Punch cards were in common use until the
mid-1970s. It should be noted that the use of punch cards predates
computers. They were used as early as 1725 in the textile industry
(for controlling mechanized textile looms).

Above: Card from a Fortran program: Z(1) = Y + W(1)

Above left: Punch card reader. Above right: Punch card


writer.

From PINGDON

Magnetic drum memory


Invented all the way back in 1932 (in Austria), it was widely used
in the 1950s and 60s as the main working memory of computers.
In the mid-1950s, magnetic drum memory had a capacity of
around 10 kB.

Above left: The magnetic Drum Memory of the UNIVAC computer. Above
right: A 16-inch-long drum from the IBM 650 computer. It had 40 tracks, 10 kB
of storage space, and spun at 12,500 revolutions per minute.

The hard disk drive


The first hard disk drive was the IBM Model 350 Disk File that
came with the IBM 305 RAMAC computer in 1956. It had 50 24inch discs with a total storage capacity of 5 million characters
(just under 5 MB).

Above: IBM Model 350, the first-ever hard disk drive.

Data Center
1970s
1970s


Datacenters in the US began documenting


formal disaster recovery plans in 1973. These
functions were batch operations and not
complex in nature.

Mainframes required special cooling and in


the late 1970s, air-cooled computers moved
into offices.

In 1973, the minicomputer Xerox Alto was a


landmark step in the development of personal
computers because of its graphical user
interface, bit-mapped high-resolution screen,
large internal and external memory storage,
mouse, and special software.

In 1977, the worlds first commercially


available local area network, ARCnet was first
put into service at Chase Manhattan Bank,
New York, as a beta-site. It was local area
network using token-ring architecture,
supporting data rates of 2.5 Mbps, and
connecting up to 255 computers.

78

Data Center 1980s


1980s The Rise of the PC


During the 1980s, the computer industry


experienced the boom of the microcomputer
era thanks to the birth of the IBM Personal
Computer (PC).

Mainframes were extremely expensive to use


and required enormous resources in space,
operation and cooling.

Starting in 1985, IBM provided more than $30


million in products and support over the course
of 5 years to a supercomputer facility
established at Cornell University in Ithaca,
New York.

In 1988, IBM introduces the IBM Application


System/400 (AS/400), and quickly becomes
one of the worlds most popular business
computing systems.

As information technology operations started


to grow in complexity, companies grew aware
of the need to control IT resources.

79

The first hard drive to have more than 1 GB in capacity was the
IBM 3380 in 1980 (it could store 2.52 GB). It was the size of a
refrigerator, weighed 550 pounds (250 kg), and the price when it
was introduced ranged from $81,000 to $142,400.

Above left: A 250 MB hard disk drive from 1979. Above


right: The IBM 3380 from 1980, the first gigabyte-capacity
hard disk drive.

The floppy disc


The diskette, or floppy disk (named so because they were flexible), was
invented by IBM and in common use from the mid-1970s to the late
1990s. The first floppy disks were 8 inches, and later in came 5.25 and
3.5-inch formats. The first floppy disk, introduced in 1971, had a capacity
of 79.7 kB, and was read-only. A read-write version came a year later.

Above left: An 8-inch floppy and floppy drive next to a regular


3.5-inch floppy disk.Above right: The convenience of easily
removable storage media.

Data Center
1990s
1990s


Client-Server Computing - Data


Center History: Microcomputers (now
called servers) started to find their
places in the old computer rooms and
were being called data centers.

Companies were putting up server rooms


inside their company walls with the
availability of inexpensive networking
equipment.

Late 1990s The Internet & Non-Stop


Computing : The boom of data centers
came during the dot-com bubble.
Companies needed fast Internet
connectivity and nonstop operation to
deploy systems and establish a presence
on the Internet.

Many companies started building very


large facilities to provide businesses with a
range of solutions for systems deployment
and operation.

82

As companies began demanding a permanent presence on the


Internet, network connectivity and colocation services grew
into a business requirement, Internet providers and hosting
companies and quickly growing Internet Data Centers
consisting of hundreds and often thousands of servers were
created. The data center as a service model became common
for most companies.

83

84

Data Center 2000s Energy Efficiency


concerns and Green Data Center
2000s





85

As of 2007, the average datacenter


consumes as much energy as 25,000 homes.
There are 5.75 million new servers deployed
every year.
The number of government data centers has
gone from 432 in 1999 to 1,100+ today.
Data centers account for 1.5% of US energy
consumption and demand is growing 10%
per year.
Facebook launched the OpenCompute
Project, providing specifications to their
Prineville, Oregon data center that uses 38%
less energy to do the same work as their
other facilities, while costing 24% less.
As the growth of online data grows
exponentially, there is opportunity (and a
need) to run more efficient data centers.

GOOGLE SECURITY & GREEN DC AND

86

Cloud Computing Definition

Cloud computing is a general term for anything that involves


delivering hosted services over the Internet. These services are
broadly divided into three categories: Infrastructure-as-a-Service
(IaaS), Platform-as-a-Service (PaaS) and Software-as-a-Service
(SaaS). The name cloud computing was inspired by the cloud
symbol that's often used to represent the Internet in flowcharts
and diagrams.

88

89

Cloud Provider

90

Cloud Computing Services

91

92

EXPLAINING THE
CLOUD COMPUTING
CLOUD COMPUTING EXPLAINED BY
CHRISTOPHER BARRAT

93

94

Data Center

95

1.

INTRODUCTION

2.

OVERVIEW AND ROLE OF DATA CENTER

3.

DATA CENTER STANDARD TIA-942

4.

DATA CENTER HISTORY FROM 60S TO THE


CLOUD

5.

GREEN DATA CENTER

Green Data Center

Background
 Data center spaces can consume up to

100 to 200 times as much electricity


as standard office spaces  they are
prime targets for energy-efficient
design measures that can save money
and reduce electricity use.
 However, the critical nature of data

center loads elevates many design


criteriachiefly reliability and high
power density capacityfar above
energy efficiency.

98

What does Green Data Center mean?


A green data center is an enterprise class computing facility that is entirely built,
managed and operated on green computing principles. It provides the same features
and capabilities of a typical data center but uses less energy and space, and its
design and operation are environmentally friendly.

Holistic approach of Green Data Center shall cover

Green Design
Green
Procurement
& clean

Green
Operation &
sustainability

Green
disposal

Whats more and Why


 A green data center is one that can operate with maximum

energy efficiency and minimum environmental impact. This


includes the mechanical, lighting, electrical and IT equipment
(servers, storage, network, etc.).

Why Green Data Center matters?


Within corporations, the focus on green data centers is
driven primarily by a desire to reduce the tremendous
electricity costs associated with operating a data
center.
Going green is a way to reduce operating expense
significantly for the IT infrastructure.

What does it mean to be Green


 Sustainability


Process Efficiency

Energy Usage

Material Usage

LEED (Leadership in Energy and


Environmental Design)

 Measure Lifecycle Cost




(The concept of including acquisition, operating, and disposal costs when


evaluating various alternatives.)

DC Energy Consumption & PUE

PUE  Power Usage Effectiveness

one very
important
aspect for
measurement
of Green DC

PUE = Total Facility Power / IT Equipment Power

Detailed Calculation
Total IT Load

94kW

Cooling Infrastructure

80kW

Power System Load

24kW

Lighting Load

2kW

Total Facility Load

200kW

PUE

2.13

PUE

Level of Efficiency

3.0

Very Inefficient

2.5

Inefficient

2.0

Average

1.5

Efficient

105

Best Practices for Green Data


Center
 The best practices for energy-efficient data center design spans

the categories of:

1.

Information Technology (IT) systems and their environmental


conditions,

2.

Data center air management, cooling and electrical systems, on-site


generation, and heat recovery.

IT system energy efficiency and environmental conditions are presented first because
measures taken in these areas have a cascading effect of secondary energy savings for
the mechanical and electrical systems.

 Metrics and benchmarking values to evaluate data center

systems energy efficiency.

Summary for Best Practice of


Green Data Center

Operation &
management

design
Selection and
procurement

Key elements in a
green data center

E N D - DCDA Consulting Services


108

You might also like