You are on page 1of 46

Avoiding the $25 Million Server

Data Center Power, Cooling, and Growth

Prepared for:
IBM Corporation

By:
Sine Nomine Associates
43596 Blacksmith Square May 1, 2007
Ashburn, VA 20147
Avoiding the $25 Million Server

_________________________________________________________________________

Copyright © 2007 Sine Nomine Associates

All rights reserved. Trademarks and brand names mentioned in this document are used only for
identification purposes. Sine Nomine Associates acknowledges the ownership of all product
trademarks by their respective manufacturers.

This report was developed by Sine Nomine Associates with IBM cooperation and funding. This
report utilizes publicly available information provided by various companies and sources,
including IBM. The opinions are those of the report's author, and do not necessarily represent
IBM's position on these issues.

© 2007 Sine Nomine Associates ii


Avoiding the $25 Million Server

Document History
Revision/
Date Nature of Change
Reissue

May 1, 2007 Version 01 New Release

© 2007 Sine Nomine Associates iii


Avoiding the $25 Million Server

Table of Contents

1 INTRODUCTION............................................................................................1
1.1 Constraints on Data Center Growth ......................................................................................... 1

1.2 The $25 Million Server ............................................................................................................... 2

1.3 Discrete Servers, Blade Servers, and System z ......................................................................... 2

1.4 Power and Cooling Impact of Virtualization............................................................................ 3

1.5 About Sine Nomine Associates................................................................................................... 4

2 DATA CENTER SCENARIO MODEL............................................................5


2.1 Fundamental Assumptions......................................................................................................... 5

2.2 Sources of Input Data ................................................................................................................. 5

2.3 Structure of the Model................................................................................................................ 6


2.3.1 Globals .......................................................................................................................................... 6
2.3.2 Platforms ....................................................................................................................................... 7
2.3.3 Workloads ..................................................................................................................................... 9
2.3.4 Scenarios..................................................................................................................................... 11

3 RESULTS IN DETAIL ..................................................................................13


3.1 Workload Aggregation ............................................................................................................. 13

3.2 Constrained Server Density ..................................................................................................... 14

3.3 Total Cost of Ownership........................................................................................................... 16


3.3.1 Energy Cost................................................................................................................................. 18

4 CONCLUSIONS...........................................................................................21
4.1 Power and Cooling .................................................................................................................... 21

4.2 Total Cost of Ownership........................................................................................................... 21

4.3 Sine Nomine Associates' Recommendations ........................................................................... 22

4.4 Avoiding the $25 Million Server .............................................................................................. 23

APPENDIX A: POWER/COOLING MODEL.......................................................24

© 2007 Sine Nomine Associates iv


Avoiding the $25 Million Server

APPENDIX B: EXAMPLE CALCULATION........................................................39

BIBLIOGRAPHY ................................................................................................41

List of Tables

Table 1: Aggregate workload per server used in model ................................................................ 14

Table 2: Constrained logical server density per enclosure ............................................................ 15

Table 3: Total logical servers in entire data center........................................................................ 20

List of Figures

Figure 1: Relative frequency of occurrence of server types in the model ..................................... 13

Figure 2: Logical server density per rack location unit ................................................................. 16

Figure 3: TCO per logical server for one, three, and five years .................................................... 17

Figure 4: Fifth-year incremental TCO for each scenario .............................................................. 18

Figure 5: Power utilization per rack location unit ......................................................................... 19

Figure 6: Annual power cost per logical server............................................................................. 20

© 2007 Sine Nomine Associates v


Avoiding the $25 Million Server

1 Introduction
In this paper Sine Nomine Associates has analyzed the problem of functional density in a modern
data center, considering traditional rack-mount discrete servers, new-generation blade servers,
and virtual servers on both Intel and System z mainframe platforms.

This paper presents the results of deploying each of these designs in a hypothetical urban-area
data center, owned by a single large enterprise or shared by multiple medium-sized enterprises (as
would be the case in a hosting provider). This scenario was selected to illuminate an emerging
constraint on Fortune 500 IT: the fundamental operational limits of existing high-value data
center space.

1.1 Constraints on Data Center Growth

Modern data centers have benefited from, and have been impeded by, recent technological
changes in computer architecture. Faster processors, denser mass storage, and inexpensive
memory chips allow more and more work to be delivered from a single system. On the other
hand, as applications grow more complex, more resources are demanded from each application.
Also, modern semiconductors1 consume power in direct proportion to their clock speed and to
their gate count. As both circuit complexity and clock speed increase, the result is a greater-than-
linear increase in power consumption per server.

Additionally, rack density has increased dramatically in recent years. Servers that formerly would
have occupied two or even three rack units (2U or 3U) now fit into a single 1.75 inch rack unit
(1U). Blade servers increase the density beyond 1 server per unit by orienting the circuit boards
vertically and cooling the system with carefully designed, highly efficient airflow components. A
typical blade server can pack 14 blades into 9 rack units, while another packs 16 blades into 10
rack units. The density increases, but the power per server is largely dependent on processor and
memory, not board configuration, and therefore the required power per rack increases
significantly in today's environment.

Although some data centers are located in rural areas where real estate is relatively inexpensive,
many are found in densely populated urban areas where real estate is extremely expensive or
unavailable at any practical price. Nevertheless, urban data centers remain vital for organizations,
because they already are in service and therefore incur only recurring costs rather than initial
acquisition costs, and because they are close in space and time to other operational centers of the
company, its customers and partners. In both urban and rural settings, there is usually a limit to
how much electrical power can be delivered to a given facility. In the rural area, it might (or
might not) be possible to "throw money" at the utility company to obtain a larger power feed, but
in urban areas there is often simply no more capacity to be had at any price.

This is important because data centers are ultimately constrained not just by space, but also by
power. This leads to the problem of the "$25 Million Server."

1
This statement refers to metal-oxide semiconductor (MOS) chips, rather than bipolar junction chips. MOS
(and CMOS) chips are used for almost all microprocessor and memory circuits, with bipolar chips now
being used only for ultra-high-speed functions such as multi-gigabit-per-second network transceivers.

© 2007 Sine Nomine Associates 1


Avoiding the $25 Million Server

1.2 The $25 Million Server

Data processing needs of companies typically grow over time, as the company grows in size and
as the demands of the marketplace lead to increasingly complex and numerous systems. Given a
data center of a certain size, with a limited amount of power and cooling available, the density of
servers steadily increases. Eventually, it simply becomes impossible to deploy one more server in
that data center. For a well-managed data center, under disciplined administration, this point
occurs when the nominal power capacity (less a prudent safety factor) is reached. For companies
that fall short of perfection, this point occurs when the already-overbooked power and cooling
systems finally fail, leading to overheated servers, tripped breakers or shutdowns, or brownouts.

Whether the breakpoint is a prudently chosen safety stop, or an empirically determined failure
point, it is inescapable, and the company's only choice is to build a new data center. Thus, in
effect, that one additional server resulted in an expense of tens of millions of dollars, and has
become known in data center design parlance as "the $25 million server." The purpose of this
paper is to identify ways in which careful selection of server platform can push the $25 million
server as far into the future as possible.

1.3 Discrete Servers, Blade Servers, and System z

This paper examines multiple configurations for each of three different platform types: discrete
servers, blade servers, and mainframes, comparing their strengths and weaknesses.

As microcomputers2 began to appear in large-scale data centers, the natural tendency was to
reshape a traditional PC mechanically so that it would fit into a standard 19 inch equipment rack.
Early PC servers occupied 4 standard 1.75 inch rack units (vertically) each, or perhaps even more,
but advancing technology has reduced typical servers to 2U or 1U sizes. Currently, a discrete
microcomputer server packs two or more processors into a 1U chassis.

To push for even higher density, server manufacturers took a cue from telecommunications
equipment makers and began offering blade servers. These use vertically oriented plug-in boards,
with one server each, in an enclosure that is typically 7U to 10U in height. Sixteen compute nodes
in a 10U chassis, for example, results in a density of 1.60 servers per rack unit, a 60%
improvement over standard 1U discrete servers. Blades also offer the advantage of internal
network switches and (frequently) storage area network (SAN) switches, so fewer external switch
ports are needed for a given number of servers. Typically, a blade deployment needs only two
network and two SAN ports per blade chassis, not per blade, whereas discrete servers need two of
each port per server.3

Both discrete and blade servers can be operated in either physical or virtual mode. Physical mode
dedicates a compute node to each logical server, and provides maximum isolation at the expense
of a great deal of underutilized resources, since a typical modern CPU is much faster than all but
the most intensively compute-bound applications require. Virtualized microcomputer servers
offer strong isolation in software, with a hardware assist, and allow resource sharing and leveling

2
In this paper, the term "microcomputer" refers to any computer in which the central processing unit is a
microprocessor. All x86-based servers are considered microcomputers in this context.
3
This assumes N+N redundancy, the most common configuration.

© 2007 Sine Nomine Associates 2


Avoiding the $25 Million Server

between logical servers on a single physical platform. Virtualization also offers advantages in
manageability and in business continuity planning, because virtualization tools such as VMware
allow management of the virtual system images in a robust way.

Mainframes such as the IBM System z take consolidation with virtualization to its logical limit,
concentrating dozens, hundreds, or even thousands of servers into a single physical machine.
Although it is theoretically possible to run a single system image on a mainframe, practical
experience shows that this is the optimal solution in only a few very specialized situations.
Mainframes offer full virtualization at the hardware level, and they are optimized to have load
leveling and resource sharing managed by the hypervisor, not the guest operating system or
application.

Whereas a typical microcomputer has an extremely fast CPU and often a relatively modest I/O
capability, the mainframe has processors that are only modestly faster than a microcomputer, but
I/O capabilities that far outstrip those of an ordinary microcomputer. The fabrication budget and
larger physical size budget of a mainframe allow ultra-wide internal data paths with sophisticated
"channel" controllers that offload almost all of the device management and data transfer work
from the main CPU to a dedicated I/O processor. Mainframe CPUs spend much less of their time
servicing low-level device interrupts. Even so, the mainframe is not suited for extremely
compute-intensive tasks such as ray-trace rendering of 3D graphics.

Each of the three platforms has its benefits, and its drawbacks, and this paper looks at all three in
a hypothetical enterprise data center with an aggregate workload that closely simulates workloads
Sine Nomine Associates has observed in actual customer sites.

1.4 Power and Cooling Impact of Virtualization

System administrators often separate servers for reasons other than a need for additional hardware
resources. Organizational boundaries may require (or encourage) segregation of servers along
departmental or project lines, even though a single server could have been shared from a purely
technical standpoint. Servers may be dedicated to major customers or business partners due to
contractual constraints, or because these external organizations need administrative-level access
to certain applications. Some applications require separate servers because they are incompatible
with other software that might logically coexist on the same server if not for that fact. Finally, for
applications with low average utilization and high peak utilization for brief intervals, it is
necessary to provision dedicated servers to meet the peak demand, rather than the average
demand that dominates the server's workload.

The result of any of these situations is a data center full of dedicated servers that are
underutilized, albeit for very good business reasons, and this leads to low productivity of the
energy used by the systems in aggregate.

Virtualization allows underutilized physical servers to be consolidated in a way that maintains the
isolation of their operating systems, application runtime environment, security context, network
broadcast domain, and administrative authority. Load leveling in the virtual data center enables
planners to share the peak-handling overhead among applications whose instantaneous workloads
are unlikely to peak at the same time. Having fewer servers in the data center proportionally
reduces the power consumption of the electronics and of the associated cooling machinery,
resulting in substantial savings of recurring costs.

© 2007 Sine Nomine Associates 3


Avoiding the $25 Million Server

1.5 About Sine Nomine Associates

Sine Nomine Associates is an IT engineering consultancy experienced in multiple hardware and


software platforms, from microprocessor-based systems through UNIX workstations and
minicomputers, and into large-scale mainframe systems. Sine Nomine Associates engineers work
with UNIX, Linux, Windows, Macintosh, and mainframe operating systems, taking a vendor-
neutral approach to solving customer problems.

© 2007 Sine Nomine Associates 4


Avoiding the $25 Million Server

2 Data Center Scenario Model


To analyze the power and cooling impact of virtualization and the various platforms on which
virtual servers can be run, Sine Nomine Associates has created a detailed mathematical model of
a hypothetical data center. The model is presented in Appendix A of this paper. This section
explains that model in detail.

The basic question this model answers is, "Given that the data center cannot expand in space or
power, how can the enterprise expand to meet future needs, and what is the best platform to allow
maximum growth potential at the least cost?" The intent of the model is to determine how best to
avoid the "$25 million server" problem.

The model will show how to avoid this scenario through server consolidation onto virtualized
hardware, including both microcomputers and System z mainframes.

2.1 Fundamental Assumptions

The model simulates an urban-area data center of a given size, with the presumption of no room
available onsite for further expansion and little or no additional electrical power available at that
site. This may sound artificially constrained, but Sine Nomine Associates has worked with many
customers in New York City, Los Angeles, Chicago, and Washington DC areas who face this
very situation.

A single large enterprise with a company-owned data center is the focus of this study, though
most of the conclusions are applicable to a collocation facility shared by several medium-sized
enterprises. In all scenarios considered here, the model assumes that the company has broadly
deployed storage area network (SAN) technologies with efficient aggregation and management
capabilities. The space, cost, and power required by the SAN arrays and management, and data
backup systems, are not included in this model, on the principle that the quantity of enterprise
data storage is not significantly affected by the choice of processing platform. This is the one
assumption that may not directly translate between a single large enterprise and multiple medium-
sized enterprises, although one could presume multiple smaller SANs as well.

All of the servers in this model run Debian Linux, so that the cost of the operating system is
equalized (at zero, in fact) between all platforms.4 Hypervisors (VMware for the Intel platforms,
and z/VM for the mainframe) are included in the pricing calculations of this model, however.

2.2 Sources of Input Data

The model accepts as its inputs a number of data sources, including:

4
Since Windows can only run on Intel servers, there is no reason to compare Windows across Intel and
mainframe platforms. If an enterprise runs a mixed Windows and Linux environment, this model can still
be useful -- just assume that the Windows servers exist in a separate section of the data center whose square
footage is not included in these calculations, and that the Windows environment is not subject to migration
from one platform to another.

© 2007 Sine Nomine Associates 5


Avoiding the $25 Million Server

• Constants that represent the hardware characteristics of the various platforms and
configurations being compared.

• Constants that represent the capabilities and limitations of the hypothetical data center,
such as its physical size, available power per unit area, and other parameters.

• Constants for occupancy factors defining the efficiency with which the "white tile" space
in the data center can be utilized, amount of cooling energy required per watt of
computing energy, and similar factors.

• Costing figures that capture the national average cost per kilowatt-hour of electricity, the
costs of hardware and software for the various platforms and configurations, and the cost
of data center construction per unit area and unit of available power.

• Performance constants that allow qualitative differences between platforms to be


normalized so that, at least at an estimate level, the capacity of different architectures can
be compared.

• Constants that represent typical loading factors for various components of a typical server
under various common application loads, and the relative frequency with which each of
these server types is likely to occur in an enterprise data center.

Wherever possible, the input constants have been obtained from specific manufacturer
specification datasheets, well-respected published reports, or internal findings developed by Sine
Nomine Associates in capacity planning and performance measurement projects. In a few cases,
such as the total size of the hypothetical data center, an arbitrary constant must of necessity be
selected, and these were chosen based on Sine Nomine Associates' extensive experience with
multiple large- and medium-sized customers ranging from Fortune 500 financial companies to
mid-sized insurance and manufacturing enterprises.

2.3 Structure of the Model

The model was deliberately broken down into sections that are decoupled, as much as possible,
from one another. This makes debugging easier, makes the model easier to understand, and keeps
it as configuration-neutral as possible. The sections of the model include the following:

2.3.1 Globals

The Globals page provides constants, and wherever possible the numbers were generated from
objectively verifiable external sources. Constants on this page include standard units per rack, the
arbitrarily chosen size of the hypothetical data center, and electricity cost data (from the United
States Department of Energy).

There are occupancy constants that reflect the desired number of empty units (for future
expansion) per rack, and the "IT yield" ratio which defines how much of the data center's "white
tile" area can actually be occupied by racks. This figure was obtained from an Uptime Institute
report based on statistics gathered from numerous actual data centers. The IT yield compensates
for service aisles, building support columns, electrical and network wiring chases, cooling ducts,
and in-room configuration workspace.

© 2007 Sine Nomine Associates 6


Avoiding the $25 Million Server

The Globals page also contains parameters for the non-platform infrastructure components, such
as network switches, storage area network (SAN) switches, and keyboard-video-mouse (KVM)
switches. These items are presumed to be selected independently of the processing platform. The
choice of platform will cause the number of needed Ethernet, SAN, and KVM ports to vary, but
in this model it does not change the type of device selected for each of these roles. The
infrastructure parameters also include a multiplier that determines the level of redundancy of each
infrastructure component (2X for Ethernet and SAN, 1X for KVM).

The Globals page is specifically constrained so that its data, whether calculated or constant,
depends on no field from any other page. The intent is that the global constants are independent
of platform and workload. The pricing data that appears on this page is per-unit pricing for
specific components, with no extended totals based on quantity anywhere on this page.

2.3.2 Platforms

The second page contains the hardware and costing details for each platform, without regard to
the workload. A few fields are calculated based on the global constants, but no fields on this page
are in any way dependent on the workload or scenario pages.

2.3.2.1 Intel Configurations

The Intel systems were all configured and priced to use Intel Xeon Woodcrest 5140 processors
running at 2.33 GHz clock speed. Other configurations of the same server brands can use faster or
slower processors, but this particular configuration was selected because it was available across
all of the servers being compared, and is a very common implementation for corporate data
centers. The basic configuration of each Intel system type is considered a typical build-out for a
standalone Internet or intranet server, whereas the expanded system contains an additional
processor and much more RAM, so that it is more suitable for a virtualized platform.

The blade servers considered in this model consolidate up to 10 processor blades into a 7U
chassis, for a density of about 1.43 servers per rack unit. The non-blade Intel servers, of course,
offer a density of exactly 1.00 servers per rack unit.

The pricing for the Intel systems was obtained from the manufacturer's online catalog and from
quotes obtained from their presales team.

2.3.2.2 Mainframe Configurations

In the case of the IBM System z mainframes, three configurations are considered, but in all cases
the assumption is that z/VM will be used as a hypervisor to virtualize the platform. It is almost
universally agreed that the System z is not an economically viable platform for running a single
Linux instance, nor is it viable for most real-world applications running one server per logical
partition (LPAR).5 Thus, this model does not consider those configurations.

5
Logical Partitions (LPARs) are a way to divide a single physical system into a relatively small number of
isolated sub-units that can be viewed by the operating system as if they were standalone systems. LPARs in
the mainframe environment are typically used for gross system partitioning (e.g., test vs. development vs.
production) rather than fine-grained partitioning at the individual application level.

© 2007 Sine Nomine Associates 7


Avoiding the $25 Million Server

IBM does not publish list prices for mainframes, preferring to allow the customer to receive a
client-specific quotation from the IBM sales representatives. The prices reported here are a
composite of actual prices that have been paid by known customers, and published IBM prices for
components that do have a list price.

2.3.2.3 Infrastructure Differences

The quantities of network switch ports and SAN ports are differently calculated for the System z
than they are for the Intel systems. The Intel systems use a calculated value based on the number
of ports per infrastructure component, the redundancy multiplier, and the chassis size of the
components. This calculation, plus the mandated number of open slots per rack, iterates with the
size of the server enclosure to determine how many rack units are available for actual server
enclosures per rack. This, in turn, is used to calculate the number of servers that can be installed
in a rack.

For the System z mainframe, most of those calculations are not applicable. The System z has all
the SAN and Ethernet ports it needs, and a single locally attached console replaces KVM. There
is no need for a KVM-like switch function because there is only one physical machine in the
"rack". Also, rack-level calculations do not apply because the mainframe occupies one full rack,
period.

Note that neither the Intel systems nor the mainframes include upstream SAN and Ethernet
switching or routing. This has the net effect of giving a slight model bias to the Intel platform, but
the error is negligible because of the use of rack-local switches to fan-in the connections.

2.3.2.4 Normalizing Platforms to Common Specifications

Comparison of mainframe processors to microprocessors is difficult and widely debated among


computing experts, because the processors and their support components are radically different.
The differences are qualitative, not just quantitative. Microprocessors are typically categorized
into CISC (complex instruction set computer) or RISC (reduced instruction set computer), with
RISC systems having simpler binary operations that complete at blinding speed, whereas CISC
machines provide complex binary operations that take longer but do more. There is, of course
much overlap between these categories. Mainframe processors are, in a sense, an ultra-CISC
design, with machine instructions carrying out very complex operations in microcode or
nanocode.

A particular distinguishing feature of the mainframe environment is that virtualization of


hardware is much more developed and mature than even the best microprocessor-based systems.
A larger portion of the mainframe instruction set is virtualized than the Intel instruction set, which
means virtual environments require less software emulation of hardware. Mainframe
administrators routinely run nested virtualized environments for testing and development
situations, something that is almost totally impractical on microcomputers.

Mainframes also possess ultra-high-speed interconnection mechanisms and massive-width


internal data buses, allowing extremely fast transfers over virtual LANs. In fact, some transfers
are actually memory-mapped so that the transfer never actually happens, and the data is simply
handed off between sender and receiver by pointer manipulation.

© 2007 Sine Nomine Associates 8


Avoiding the $25 Million Server

As a result of these architectural differences, it is impossible to provide a single metric that


directly converts Intel gigahertz to mainframe MIPS specifications.6 However, based on extensive
field experience Sine Nomine Associates typically uses a figure of 6.0 MHz per MIPS for the
latest generation System z processors from IBM. Though this number is a simplification, it is
necessary in any kind of a capacity-planning model to provide some sort of normalization
mechanism. Sine Nomine Associates has assisted many customers with system integration in a
mixed mainframe/microcomputer environment, and this ratio has proven to be a reasonable
approximation for typical workloads.

In the Platforms page, the processing capacity of each system is converted to mainframe MIPS
using the above figure, and these MIPS numbers are carried forward to the scenario calculations.

2.3.3 Workloads

The effectiveness of processing platforms depends heavily on the nature of the workload, and any
responsible expert will advise clients to carefully consider their specific workload when choosing
a mix of platforms.

For the sake of this model, rather than choosing a single arbitrary workload "per server", the
Workloads page catalogs a total of eight unique workloads and their typical resource
requirements, based on published benchmarks and monitoring data collected by Sine Nomine
Associates on various projects. This allows the model to flexibly simulate detailed scenarios for
comparison, allowing much more precise analysis of relative configuration strengths. The
workload types include:

• Static web server -- This server type assumes a traditional publishing-type web site, with
static and lightly dynamic content. Lightly dynamic content would consist of server-side
includes (SSI), blogs, or content management systems written in Perl or PHP.

• Database server -- The database server is presumed to run an SQL-compliant relational


database manager as its primary function. The model assumes that the database server
most frequently processes simple or medium-complexity queries, and that large and
complex queries are less common.

• Domain name server (DNS) -- The DNS server is the quintessential lightly loaded
machine. DNS queries typically return only a small amount of data, and the data is read-
only except for occasional changes. This is almost perfectly optimized for cache
performance within the DNS server and within the operating system.

• Ecommerce web server -- In a multi-tier application, this server would represent the
customer-facing outer tier that processes HTTP queries from client browsers and

6
Originally, the acronym MIPS stood for Millions of Instructions per Second, representing the number of
binary-level operations that a processor could do per second. This, however, becomes a rather amorphous
concept on modern high-complexity processors because, among other things, instructions take vastly
different amounts of time to complete. A full thread-context switch obviously takes much longer than a
simple arithmetic operation, for example. In the mainframe context, then, MIPS has become a shorthand
terminology that represents a relative, rather than absolute, processor speed. In essence, MIPS is a way to
compare one mainframe model to another when predicting capacity of a proposed upgrade.

© 2007 Sine Nomine Associates 9


Avoiding the $25 Million Server

provides the presentation-layer logic for the application. This may include portal
components being merged into the overall page, AJAX processing of dynamic-content
form fields, and adaptation to browser differences. This layer specifically excludes the
core business logic of the application as well as database services.

The relative frequency of this profile is higher than that of the business logic server (see
below) because in many intermediate-complexity situations this tier may function without
a separate business logic server, connecting directly to the database server.

• File server -- A typical departmental file server performs very little processing, but a
great deal of network and disk I/O. The load imposed upon the file server is largely
independent of the protocols in use, so this workload profile could represent a server
offering NFS, CIFS (SMB), AFS, or some combination of these.

• Business logic server -- This represents the second tier of a three-tier client/server
application, residing between the ecommerce web server and the database server. The
business logic server in this profile is assumed to run a Java application in an enterprise-
grade container such as IBM WebSphere, JBoss, or BEA WebLogic.

• Management and/or monitoring server -- Servers that collect and analyze log
messages from other servers, and which are used as a host for monitoring and system
management applications, perform a great deal of network I/O and a moderately-high
amount of disk I/O. Typically, log aggregation uses efficient buffering algorithms to
reduce the amount of time spent on disk writes, but they cannot control the characteristics
of the inbound network traffic as closely.

• Email server -- The model’s final common server combination is the email server. In a
typical enterprise, a small number of email relays receive mail from the Internet, and a
small number transmit outbound mail to the Internet. Within the company, however,
there are usually a larger number of mailbox servers that hold user inboxes and archived
messages. The mailbox servers perform almost no processing, though the relay servers
spend much of their time scanning for spam, malware, and other undesired content. In all
cases, however, the I/O load of an email server is large. The profile shown in this model
is an aggregate assuming a few relay servers and a larger number of mailbox servers.

The individual workloads are aggregated into a weighted average workload, based on the relative
frequency of these servers in enterprises where Sine Nomine Associates has done other projects.
For example, static web servers and file servers are quite common, often occurring at a rate of
one per department, whereas domain name servers (DNS) are relatively rare because one server
can reasonably support thousands of users.

The load factor for each workload profile is segregated according to specific resources, including
the main processor, main memory (RAM or working storage), disk subsystem, and network
interface. The processor workload is further divided into application, kernel, and cryptographic
loads, with the kernel load representing the time that the operating system spends managing
physical devices such as disk drives and network interfaces.

The workloads are normalized to a discrete Intel server with one processor running at 2.33 GHz,
locally-controlled SATA disk storage, no cryptographic coprocessing, and ordinary Ethernet
adapters.

© 2007 Sine Nomine Associates 10


Avoiding the $25 Million Server

Finally, each workload profile includes a ratio estimating how much of the network traffic will be
server-to-server rather than server-to-client. This is important in the virtualized environment,
because much of the server-to-server traffic will be routed on virtual LANs that have extremely
low overhead. These ratios are based on Sine Nomine Associates' direct experience working with
Fortune 500 data center implementations around the world.

2.3.4 Scenarios

The final page of the model brings together a matrix of platforms and workload to produce the
model's scenario results. This page is the logical opposite of the Globals page, in that it consumes
data from all the other pages but does not feed back to any of them.

2.3.4.1 Primary Calculations

The scenario page initially calculates a "nominal" result for each configuration, based on its
theoretical density and capacity. The theoretical density is adjusted for infrastructure components
(network, SAN, and KVM switches) but not for power limitations of the data center facility.

This calculation provides a number of servers that could fit into the data center based on their
own architectural limits. In the case of the virtualized environments, this is also limited by the
basic factors of the aggregate per-server workload.

2.3.4.2 Power Limitations

Real-world data centers do not have unlimited power, and it is quite common, especially in urban
areas, for data center density to be constrained not by rack capacity, but rather by power and
cooling limitations.

This section of the page takes this into account. From the Globals page, the page obtains a figure
for watts per square foot, and then derates the average rack density based on the total amount of
power available to the entire facility.

2.3.4.3 Platform Adjustments

The platform adjustments section encapsulates the qualitative differences among platforms,
including:

• On both Intel and mainframe platforms in this scenario, some of the usual low-level disk
processing is offloaded because of the presence of the storage area network. The
offloading factor is higher in the mainframe because of an additional level of
coprocessing at the "channel" level, but both platforms receive a favorable adjustment
here, varying in degree only.

• On the mainframe, built-in cryptographic coprocessors offload the vast majority of SSL
encryption workload. The platform adjustments remove this from the overall CPU load
on the mainframe systems but not on the Intel systems, which lack dedicated
cryptographic coprocessors.

© 2007 Sine Nomine Associates 11


Avoiding the $25 Million Server

• On a microcomputer, swapping to disk (due to overcommitted RAM resources) is a very


expensive operation, but the mainframe's VDISK capability and high-level I/O channel
processing make swapping much less expensive, proportionally. Thus, on a mainframe it
is reasonable to overcommit memory at a much higher factor than on a microcomputer.

• Mainframes use dedicated coprocessors for network and disk I/O (in fact, there is no such
thing as a "dumb" I/O adapter in a mainframe). The platform adjustments account for this
offloading in the network and disk workload factors.

• The microcomputers rely on frequent interrupts to the main processor to handle I/O
devices, whereas the mainframe uses a channel-oriented I/O interface that accepts high-
level commands from the guest operating system. As a result, mainframes spend much
less of their time executing kernel-level code under Linux, and more of their time is
available for application-level processing.

• Finally, on both mainframe and microcomputer systems, virtualization improves the


speed of server-to-server network connections by allowing much of this traffic to be
routed over virtual networks that operate (basically) at memory-bus speeds. The platform
adjustment examines the aggregate workload's ratio of server-to-server versus server-to-
client traffic, and factors this against the number of logical servers per physical enclosure
for each platform, arriving at a figure that estimates the percentage of overall network
traffic that would be virtualized within the specific platforms.

2.3.4.4 Workload Calculations

The workload calculations are performed to determine, for the nominal cases, which of the load,
power, and architectural constraints will limit each configuration's total capacity per unit. The
result of this calculation is a limit of logical servers per chassis constrained by CPU, RAM, disk,
cryptography, network, or architectural limits. The final result is the worst-case (minimum) of
these values for each platform.

2.3.4.5 Adjusted Workload Calculations

The adjusted workload calculations work just like their unadjusted counterparts, except that the
loading constraints take into account the platform adjustments explained previously. These are
the results, in the opinion of Sine Nomine Associates, which most accurately represent the real-
world capacity of the data center, accounting for all the factors that are considered in this model.

It is the adjusted workload calculations which feed into the TCO calculations that follow.

2.3.4.6 Total Cost of Ownership (TCO) Calculations

The TCO calculations, in contrast to the complexity of the other parts of this model, are very
straightforward, tallying fixed costs of acquisition for the first year (only), and then adding costs
of software support/maintenance and electricity for each year the systems remain in use. TCO is
computed for one, three, and five year intervals. No cost-of-capital (net present value) or
projected growth rates are factored into these calculations, because these are numbers determined
largely by corporate policy and accounting rather than by engineering factors.

© 2007 Sine Nomine Associates 12


Avoiding the $25 Million Server

3 Results in Detail
The calculations in the Sine Nomine Associates model yielded numbers which came as no
surprise to the SNA engineering team, who have worked extensively with both microcomputer
and mainframe platforms in customer data centers. As expected, processor resources constrain the
mainframe in all its configurations. This, combined with the notional limit of 8 mainframes per
data center, results in a total count of logical servers per data center that is much lower than if
microcomputers were used throughout. The lesson to be drawn from this is, quite simply, that one
would not want to replace the entire data center with mainframes for a typical aggregated
workload. Rather, the mainframe would be applied selectively to replace microcomputer servers
in those situations where its strengths are greatest. All of Sine Nomine Associates' customers who
use mainframe Linux are also using microcomputers where appropriate, deploying the proverbial
"right tool for every job."

3.1 Workload Aggregation

The aggregated workload depends on the identified characteristics of each server category (web
server, database server, etc.) and on the frequency of occurrence of each type of server in a
typical database. The following figure shows the frequency of occurrence used in this model.

Workload Frequency of Occurrence

Static w eb server

18% Database server


25%
DNS or LDAP server
4%
e-Commerce w eb server
4%

File server
4%
Business logic server
7%
Management and monitoring
4%
34% server
E-mail server

Figure 1: Relative frequency of occurrence of server types in the model

The model is structured so that the notional "count" of servers in the aggregation process is not
significant; only the ratio factors into the result. Note that this in itself is not a model output, but
rather an input assumption presented here for reference only.

When the workload characteristics are aggregated, the result is as follows:

© 2007 Sine Nomine Associates 13


Avoiding the $25 Million Server

Aggregated Workload Aggregated Workload in Notional Units


Characteristics

Weighted application
5% 20.23 MIPS
CPU load
Weighted working
282.95 282.95 MB RAM
storage set
Weighted disk
66% 258.21 unitless
bandwidth
Weighted crypto load 6% 23.71 unitless
Weighted network
86% 334.51 unitless
bandwidth
Weighted kernel CPU
8% 29.64 unitless
load
Weighted server-to-
13%
server net ratio

Table 1: Aggregate workload per server used in model

The aggregated workload represents the amount of work assigned to each logical server in the
data center, for the model. The percentages represent a ratio to the nominal full-load capacity of a
typical microcomputer with one Intel Xeon 5140 processor at 2.33 GHz and local disk storage.
The memory is a working storage requirement in megabytes, and the processor utilization is
converted to equivalent mainframe MIPS.7 All of the workload characteristics are adjusted within
the scenarios to account for multiple processors and architectural differences of the hardware.

The disk and network utilizations are considerably higher than processor utilization. Sine Nomine
Associates has found this to be the case in almost every site where we have worked. There are, of
course, exceptions with compute-intensive applications such as finite element modeling, financial
market models, and graphics rendering, but the quantity of such servers in a typical business is
small compared to the count of web, file, email, and database servers.

3.2 Constrained Server Density

The maximum server densities are calculated based on workload, physical space, infrastructure
requirements, and power and cooling load limits of the data center in which they reside. In the
scenarios, the architecturally-adjusted workload factors are applied to these densities, resulting in
a set of constrained server densities based on each load factor. The worst case for each platform
thus establishes the final density limit for that platform in this data center, and those densities are
fed into the TCO models.

The table below shows the results of the constraint modeling, with the worst case for each
platform scenario highlighted. Note that the discrete microcomputer servers tend to be
constrained by architectural limitations rather than other factors; this is reflective of their under-

7
See explanation in footnote 6 on page 9.

© 2007 Sine Nomine Associates 14


Avoiding the $25 Million Server

utilization in the typical environment. The virtualized microcomputers in this model were
constrained by network bandwidth rather than by processor or memory limits.

IBM IBM
Dell Dell System z System z IBM
Dell PowerEdge Dell PowerEdge Business Business System z
PowerEdge 1950 PowerEdge 1955 Class S07 Class S07 Enterprise
1950 virtualized 1955 virtualized Small Expanded Class S18

Logical servers
per chassis
5.27 10.55 52.73 105.46 58.24 101.91 262.07
based on CPU
load

Logical servers
per chassis
14.48 115.81 14.48 115.81 115.81 231.62 694.86
based on RAM
capacity

Logical servers
per chassis
3.01 6.01 30.05 60.11 109.99 192.48 494.95
based on disk
bandwidth

Logical servers
per chassis
16.37 32.73 163.65 327.31 1197.87 2096.27 5390.41
based on crypto
load

Logical servers
per chassis
based on
network 1.16 2.64 13.11 26.59 216.46 378.81 974.09
bandwidth,
excluding intra-
chassis traffic

Logical servers
per chassis
1 16 10 160 36000 36000 36000
based on
architecture

Table 2: Constrained logical server density per enclosure

The above numbers are servers per chassis, not per rack. The table illustrates the constraint
behavior within each platform, and is not intended for direct comparison between platforms.

When examining server density per rack, the chart below shows again that the virtualized
environments have a substantial advantage over the discrete servers. The largest System z density
calculated by the model was 262 logical servers per mainframe. A recent published conference
presentation8 by a company using System z and Linux for server consolidation reports that the

8
The presentations were made at SHARE 107, Baltimore, Maryland, August 14-18, 2006, by a company
located in the midwestern United States. The presentations are available from http://www.linuxvm.org/.

© 2007 Sine Nomine Associates 15


Avoiding the $25 Million Server

company is running over 300 servers on one mainframe, and Sine Nomine Associates has
customers running systems of this scale and larger.

Number of Servers per Rack

0 50 100 150 200 250 300

Dell PowerEdge 1950 28

Dell PowerEdge 1950


73
virtualized

Dell PowerEdge 1955 40

Dell PowerEdge 1955


106
virtualized
IBM System z BC S07
58
Small
IBM System z BC S07
101
Expanded

IBM System z EC S18 262

Figure 2: Logical server density per rack location unit

The non-virtualized servers, of course, achieve a density that exactly equals the number of
physical units that will fit into each rack, allowing for infrastructure components and independent
of the load profile. The density of the virtualized platforms depends on the aggregate load that
limits how many logical servers will fit on each physical server.

3.3 Total Cost of Ownership

The initial cost of acquisition of the System z mainframe is somewhat higher per logical server
than the cost of discrete or blade servers, providing a total cost of ownership (TCO) that is higher
in the first year but which increases more slowly than that of the other alternatives over time. For
each additional year, discrete and blade servers consume more power and cooling cost, which
drives up their TCO until it eventually passes that of the mainframe. The following table shows
the TCO for each of the platform scenarios.

The figures cited here are for reference only and do not imply endorsement of this paper by SHARE or by
the company providing the presentations.

© 2007 Sine Nomine Associates 16


Avoiding the $25 Million Server

TCO per Logical Server

Thousands
$0 $2 $4 $6 $8 $10 $12 $14

Dell PowerEdge 1950

Dell PowerEdge 1950


virtualized

Dell PowerEdge 1955


1 year TCO
Dell PowerEdge 1955
3 year TCO
virtualized
5 year TCO
IBM System z BC S07
Small

IBM System z BC S07


Expanded

IBM System z EC S18

Figure 3: TCO per logical server for one, three, and five years

Notice in Figure 3 that the System z TCO for one, three, and five years appears almost invariant,
whereas there is a steady increase for all of the microcomputer systems. Only the fully virtualized
PowerEdge 1955 blade system approaches the low incremental TCO of the mainframe.

One interesting observation from the model is that the TCO per logical server on the System z is
not significantly affected by the size of the mainframe, provided the mainframe is appropriately
sized to the number of virtual servers being deployed. Customers planning System z
consolidation projects should, therefore, consider acquiring only the needed resources initially,
and leveraging the System z's ability to be nondisruptively field-upgraded with additional
capacity to meet growth requirements, ensuring that the available and needed resource levels are
well matched and minimizing long-term TCO.

The System z is highly optimized for virtualized operation, and therefore is not analyzed as a
discrete server. However, the analysis of virtualized versus non-virtualized operation on the
discrete and blade servers actually shows a slight increase in TCO for the virtualized
environments, due to hypervisor software costs that more than offset the power and cooling
savings. On the other hand, virtualization with appropriate management tools can significantly
reduce administrative costs, which is a factor not reflected in this model.

It is instructive to break out the fifth-year incremental TCO per logical server, as shown in the
following chart, to illustrate the annual cost savings of the System z once its acquisition cost has
been absorbed.

© 2007 Sine Nomine Associates 17


Avoiding the $25 Million Server

Fifth-Year Incremental Cost per Logical Server

$0 $500 $1,000 $1,500 $2,000

Dell PowerEdge 1950 $1,841

Dell PowerEdge 1950


$713
virtualized

Dell PowerEdge 1955 $1,100

Dell PowerEdge 1955


$419
virtualized
IBM System z BC S07
$304
Small
IBM System z BC S07
$175
Expanded

IBM System z EC S18 $77

Figure 4: Fifth-year incremental TCO for each scenario

Here the System z has a clear edge, because most of the annual TCO is power and cooling
expenses. The System z also benefits from the fact that z/VM support is a fixed annual price
regardless of the number of systems being supported, so its unit cost decreases as the number of
supported servers increases.

Within the three System z scenarios, there is a clear trend showing the lowest incremental TCO
for the largest system with the most logical servers. The System z has a power consumption that
is not strongly dependent on the number of processor engines enabled, and a more powerful
configuration consumes only modestly more power than an entry-level configuration for the same
system.

This situation explains, to a large extent, why the model imposes a "notional limit" of eight
mainframes in the entire data center. In a practical situation, customers will deploy a few larger
mainframes rather than proliferating small- to mid-sized configurations in quantity, in order to
maximize the savings reflected in Figure 4 as well as saving labor costs and physical space.

3.3.1 Energy Cost

Power cost per logical server is another area of considerable difference among platforms, and the
difference in power consumption accounts for much of the difference in incremental TCO.

Recall that the energy utilization (and therefore cost) in this model includes not only the server
itself, but also any associated infrastructure components and the air conditioning equipment to
remove all of the heat. Typically, every watt of power consumed by electronic equipment requires
an additional 0.7 watts of air conditioning power, according to Rob Snevely's book, Enterprise
Data Center Design and Methodology.

© 2007 Sine Nomine Associates 18


Avoiding the $25 Million Server

Figure 5 compares the total power consumption per rack location unit, accounting for air
conditioning energy as well as energy used by the electronics. With the data center's limitation of
125 watts per square foot,9 the data center is nominally able to supply just under 4200 watts per
rack location unit.10 Thus, the data center is not capable of being fully loaded with servers even
after factoring in lost rack space due to IT yield. This requires a further derating of the data
center, which is accounted for in the model as a limit on the data center's maximum total server
count.

Power per Rack Location by Platform Infrastructure power per rack


Server power per rack
Power (Kilowatts) Total power per rack
0 5 10 15 20 25 30 35 40 45

Dell Pow erEdge 1950

Dell Pow erEdge 1950


virtualized

Dell Pow erEdge 1955


Platform

Dell Pow erEdge 1955


virtualized

IBM System z BC S07


Small

IBM System z BC S07


Expanded

IBM System z EC S18

Figure 5: Power utilization per rack location unit

After that derating is applied, the model calculates the maximum total logical servers for the
entire data center; the results are presented in Table 1. Note that the System z values are limited
by the "notional" constraint of eight mainframes per data center, as described in Section 3.3 and
Appendix B. It would actually be possible to deploy several times this number of servers, but
most companies doing so would purchase a larger System z than the ones considered in this
model, rather than replicating more than eight of these units.

9
Published sources often cite 100 W/sq-ft, but this is bumped up here to give the “benefit of the doubt” and
assumes a high-quality physical facility.
10
Data center designers derate rack density from the theoretical concept of racks butted side-to-side, to
account for air ducts, power distribution and UPS units, operator consoles, building support columns, etc.
The resulting "IT yield" varies depending on data center design, but a recent article by the Uptime Institute
cites a figure of 18% yield, which is used in this model.

© 2007 Sine Nomine Associates 19


Avoiding the $25 Million Server

Dell Dell IBM System z IBM System z


Dell PowerEdge Dell PowerEdge Business Business IBM System
PowerEdge 1950 PowerEdge 1955 Class S07 Class S07 z Enterprise
1950 virtualized 1955 virtualized Small Expanded Class S18

1372 3577 2320 6148 464 808 2096


Table 3: Total logical servers in entire data center

As shown in Figure 6, the annual power cost per logical server varies by orders of magnitude,
with the traditional rack servers in non-virtualized mode faring the worst. For rack servers,
virtualization reduces power per logical server by about 61%, and on blade servers virtualization
reduces power per logical server by about 62%. The blade servers benefit slightly from the
improved performance of virtualized networks within a physical server versus even high-speed
physical networks between servers.

Annual Power Cost per Logical Server

$0 $100 $200 $300 $400 $500 $600 $700 $800 $900 $1,000

Dell Pow erEdge 1950 $920.70

Dell Pow erEdge 1950


$353.14
virtualized

Dell Pow erEdge 1955 $550.06

Dell Pow erEdge 1955


$207.57
virtualized

IBM System z BC S07


$127.23
Small

IBM System z BC S07


$73.06
Expanded

IBM System z EC S18 $32.80

Figure 6: Annual power cost per logical server

Because the System z mainframe's power consumption is largely independent of its processor
count, the high-end mainframe's power consumption per logical server is almost 75% lower than
that of the entry-level mainframe. Note, however, that all of the System z results are substantially
lower (at least 38%) than the best of the microcomputers.

© 2007 Sine Nomine Associates 20


Avoiding the $25 Million Server

4 Conclusions
4.1 Power and Cooling

Two conclusions leap from the analysis in this model, with regard to power and cooling:

• Virtualization of underutilized servers, on any platform, will yield power and cooling cost
savings that are substantial.

• Among virtualization solutions, the System z offers greater potential power and cooling
benefits than any other available platform.

It is important to consider these conclusions together, rather than in isolation from one another.
This model shows that mainframe consolidation tends to be constrained by available CPU
resources, a conclusion that mirrors the practical experience of Sine Nomine Associates and its
customers. However, in applications that are well suited to its architectural features, the System z
offers spectacular power savings, reducing power and cooling costs to a negligible portion of total
and recurring costs.

4.2 Total Cost of Ownership

The model shows that the total cost of ownership (TCO) per logical server is very competitive
between the mainframe and the microcomputer configurations. While in the model the
mainframes do not have a TCO advantage, it should be noted that this model intentionally omits
(as out-of-scope) a number of "soft" factors that will affect the actual TCO at a customer site,
specifically:

• Managing virtual servers typically requires significantly less labor than managing an
equivalent number of physical servers. This is true on both microcomputer and
mainframe platforms, but the management tools on z/VM are more mature and plentiful
than the corresponding tools on the microcomputer platform. This is intended as no
disrespect to VMware, but simply reflects the fact that the mainframe environment has
had several decades to perfect virtualized management tools.

• Power distribution units (PDUs), computer room air conditioners (CRACs), and the
network and SAN infrastructure components all require periodic maintenance and
occasional repairs, none of which are free. The mainframe requires significantly less of
each of these resources, and thus one expects that the maintenance and repair costs would
decrease proportionally. Even in an existing facility with pre-existing power and cooling
systems, one could argue that running these systems at a lower-than-maximum load level
would prolong their useful life. It is certain that doing so provides greater capacity
headroom to handle temporary overloads from other sources.

© 2007 Sine Nomine Associates 21


Avoiding the $25 Million Server

4.3 Sine Nomine Associates' Recommendations

Based on the results of this model, and combined with practical experience at dozens of customer
installations, Sine Nomine Associates offers the following recommendations for customers
seeking to maximize the capability of existing data centers that are constrained by power, cooling,
or space limitations:

1. Identify servers that are unitized for functional partitioning, security, or political
reasons, rather than because of their resource requirements, and consolidate them
onto virtualized platforms. Virtualization is absolutely the best way to recapture
underutilized hardware resources, and to reduce the number of physical servers that must
be purchased, installed, powered, cooled, and maintained. In addition, a properly
managed virtual data center reduces labor costs for deploying new servers and upgrading
existing servers, because of the ability to "clone" prebuilt virtual servers rapidly.

2. Use the System z mainframe as a part of the total virtualization environment. Many
applications are constrained by I/O rather than CPU resources, and many commodity
services such as DNS, directory services, file and print sharing, and email can be
consolidated onto a single mainframe running virtualized Linux with little impact on
clients, even if the clients are using Windows as their operating system. In addition,
business applications written in Java exhibit excellent portability to both Linux and
mainframe Linux platforms, and are excellent candidates for consolidation onto the
System z.

3. On an attrition basis, begin a strategic migration to blades rather than discrete


servers for applications where the mainframe is not the best alternative. The TCO
cost differences between blades and discrete servers are not overwhelming, but neither
are they insignificant. When combining even a modest TCO advantage with the blades'
improvements in manageability, the increased throughput of their internal SAN and
Ethernet switches, and the simplification of external cabling and infrastructure, it is
worthwhile to establish blades as the preferred choice for new hardware.

4. As a strategic direction, move toward storage area network (SAN) virtualized


storage rather than local storage for applications and data. Virtualized storage is an
important foundation for gaining the most benefit from virtualized servers later, and its
availability significantly improves the manageability of virtualized data centers. Note,
however, that a virtualization initiative does not need to wait for full SAN penetration, as
virtualized servers can be migrated from local storage onto the SAN over time.

5. Leverage existing mainframe-native infrastructure such as databases rather than


migrating these to microcomputer equivalents. The internal virtual networking
capability of the System z is an enormous asset for making Oracle or DB2 databases,
CICS applications and application services, and Tivoli or equivalent backup/restore
services available to consolidated applications running under mainframe Linux. If a
mainframe-native service is already available, the Linux guests can transparently utilize
that service in parallel with traditional mainframe applications.

6. Manage and monitor virtualized resources at the hypervisor level rather than
within the guest operating system. An operating system designed to run on discrete
systems will not be aware of the capabilities of an underlying virtualization platform, and
will not allocate, schedule, and prioritize physical resources as well as the hypervisor can.

© 2007 Sine Nomine Associates 22


Avoiding the $25 Million Server

This argues for horizontal scaling, in which application functionality is divided among
virtual servers of a smaller size with the hypervisor sharing resources between them.

4.4 Avoiding the $25 Million Server

The model demonstrates how power and cooling, even over physical space, limit the capacity of
data centers with modern high-performance systems. Virtualization, regardless of the underlying
physical platform, provides significant reduction in power per server, allowing many more
servers to be deployed.

Among virtualized environments, the IBM System z offers by far the best power performance.
When deployed with workloads that are well suited to its architecture -- which includes many of
the most common types of servers found in enterprise data centers -- the System z reduces the
impact of energy costs.

Ultimately, a data center's best configuration will involve selective deployment of multiple
platforms, utilizing microcomputers for the most compute-intensive loads and mainframes for
loads that are network and disk constrained. This synergy of technical strengths can extend the
life and capabilities of legacy data centers, overcoming power, space, and cooling constraints to
support growth of the enterprise.

© 2007 Sine Nomine Associates 23


Avoiding the $25 Million Server

Appendix A: Power/Cooling Model


This Appendix contains the data from the model. Certain spreadsheet cells containing only
intermediate calculations rather than input constants or output results are omitted here for the sake
of brevity, as they do not in themselves provide useful information.

Global Constants

Item Description Value Units Source Notes

Physical dimensions

Number of standard U per


rack 42 rack units

Size of standard Fibre Assumes QLogic SANBox 5602 or


Channel switch 1 rack units equivalent

Size of standard Ethernet Assumes Cisco Catalyst 4948 or


switch 1 rack units equivalent

Assumes Raritan Dominion KX464


Size of KVM switch 2 rack units or equivalent

Typical value for an enterprise data


center in an urban area is 7000-
White tile area of existing data 10000 sq ft based on SNA
center 8000 sq ft experience

Infrastructure assumptions

Power consumption per Fibre Assumes QLogic SANBox 5602 or


Channel switch 120 W equivalent

Power consumption per Assumes Cisco Catalyst 4948 or


Ethernet switch 300 W equivalent

Power consumption per KVM Assumes Raritan Dominion KX464


switch 180 W or equivalent

Ports per Fibre Channel Assumes QLogic SANBox 5602 or


switch 16 count equivalent

Assumes Cisco Catalyst 4948 or


equivalent, allows one port for
Ports per Ethernet switch 47 count uplink

© 2007 Sine Nomine Associates 24


Avoiding the $25 Million Server

Assumes Raritan Dominion KX464


Servers per KVM switch 64 count or equivalent

Performance comparison
assumptions

Published limit for VMware ESX is


128, but real-world installations in
SNA's experience limit to 16 or
less on a single physical server.
Maximum virtual servers per Some of SNA's customers limit to 4
physical server on x86 per physical server as an
platforms 16 count administrative limit.

Theoretical limit in z/VM is 32,000


Maximum virtual servers per in each of 3 scheduling queues,
physical server on System z but a practical limit is 12,000 per
platforms 36000 count queue

An estimated number based on


Estimated x86 to System z published benchmarks for the z9
performance conversion systems by Linux users (not
factor 6.0 MHz per MIPS endorsed officially by IBM)

Minimum memory for a reasonable


boot environment, suitable for
servers like DNS that use very few
memory resources. This
Minimum RAM per Linux establishes a limit on virtual
instance 64.0 megabytes servers on a physical platform.

MIPS in this context are used in


the traditional S/390 sense, not the
Mainframe MIPS per IFL literal million-instructions-per-
engine 355.0 MIPS second sense

Occupancy assumptions

Fibre Channel redundancy


factor 2 multiplier

Ethernet redundancy factor 2 multiplier

KVM redundancy factor 1 multiplier

Arbitrary assumption based on


Empty units per rack for experience in enterprise data
growth allowance 4 rack units centers of customers

Turner and Seader, "Dollars per


kW plus Dollars per Square Foot
Are a Better Data Center Cost
Ratio of infrastructure to Model than Dollars per Square
white-tile area 0.3 multiplier Foot Alone"

© 2007 Sine Nomine Associates 25


Avoiding the $25 Million Server

Kenneth Brill, "2005-2010 Heat


Density Trends in Data Processing,
Rack area as fraction of white Computer Systems, and
tile area ("IT yield") 0.18 multiplier Telecommunications Equipment"

Based on a rack footprint size of 24


Square feet per rack (white- by 36 inches and the area
tile area only), rounded up to percentage ("IT yield") from the cell
0.5 sq ft multiple 33.5 sq ft/rack above

Published sources often cite 100


W/sq-ft, but this is bumped up here
Data center power/cooling to give the “benefit of the doubt”
design limit 125 W/sq ft and assuming a high-quality site

Nominal available power per


rack limit 4187.5 W/rack

Costing assumptions

U.S. Dept. of Energy, national


average for commercial electricity
for 12 months ending December
Electricity cost $0.0936 US$/kWH 2006

Turner and Seader, "Dollars per


kW plus Dollars per Square Foot
Are a Better Data Center Cost
Building cost per sq ft (space Model than Dollars per Square
only) $220 Foot Alone"

Building cost per kVA of


usable UPS output, including Turner, Seader, and Brill, "Tier
switchgear, batteries, diesel Classifications Define Site
gen., HVAC $22,000 US$/kW capacity Infrastructure Performance"

Base cost of System z BC SNA case study, not an official


with 1 IFL and 64GB RAM $585,000 US$ price quote from IBM

Cost of System z additional


IFL $95,000 US$ Published price

Base cost of z/VM per IFL $22,500 US$ Published price

Annual support cost of z/VM


(flat fee) $11,500 US$ Published price

Pricing based on VM ware


Infrastructure Standard Edition,
published at $3750 per 2
processors (normalized to per-
Base cost of VMware ESX per processor here for calculation
processor $1,875 simplicity)

© 2007 Sine Nomine Associates 26


Avoiding the $25 Million Server

Pricing not published by VMware.


Support cost for VMware per Estimated at same as z/VM
year (flat fee) $11,500 US$ support cost

SNA case study, not an official


Cost of System z RAM $8,000 US$/GB price quote from IBM

Rob Snevely, "Enterprise Data


Ratio of HVAC to computer Center Design and Methodology"
kW 0.7 multiplier (Sun Microsystems)

Online price from Provantage.com


Cost of Fibre Channel switch $7,875 as of 2007-04-04

Online price from Buy.com as of


Cost of Ethernet switch $7,652 2007-04-04

Online price from


SuperWarehouse.com as of 2007-
Cost of KVM switch $5,702 04-04

Computed values from


constants

Electricity per kWH per month Computed from the US DoE data
(30 days) $67.39 US$/kW/month above

Maximum racks in existing


data center 238

© 2007 Sine Nomine Associates 27


Avoiding the $25 Million Server

Platform Parameters

IBM System z
Dell PowerEdge Dell PowerEdge Dell PowerEdge Dell PowerEdge IBM System z BC S07 IBM System z
Parameters 1950 1950 virtualized 1955 1955 virtualized BC S07 Small Expanded EC S18
1U rack server Blade server Mainframe Mainframe Mainframe
optimized for optimized for optimized for optimized for optimized for
Category 1U rack server virtual servers Blade server virtual servers virtual servers virtual servers virtual servers
Intel Xeon 5140 Intel Xeon 5140 Intel Xeon 5140 Intel Xeon 5140 IBM Multi-Chip IBM Multi-Chip IBM Multi-Chip
dual-core 2.33 dual-core 2.33 dual-core 2.33 dual-core 2.33 Module (MCM) Module (MCM) Module (MCM)
Processor GHz GHz GHz GHz “IFL” engine “IFL” engine “IFL” engine
Configured
processors per
server 1 2 1 2 4 7 18
Configured memory
per server 4 32 4 32 16 32 96
Size of chassis (rack
units) 1 1 7 7 42 42 42
Maximum servers
per chassis 1 16 10 160 36000 36000 36000
Network ports per
chassis (not allowing
for redundancy) 1 1 2 2 1 1 1
Fibre Channel ports
per chassis (not
allowing for
redundancy) 1 1 2 2 1 1 1
Power consumption
at specified config 670 670 4000 4000 5400 5400 6300

© 2007 Sine Nomine Associates 28


Avoiding the $25 Million Server

IBM System z
Dell PowerEdge Dell PowerEdge Dell PowerEdge Dell PowerEdge IBM System z BC S07 IBM System z
Parameters 1950 1950 virtualized 1955 1955 virtualized BC S07 Small Expanded EC S18
Est. hardware cost
per chassis at
specified config $6,082 $20,421 $53,150 $196,340 $486,000 $899,000 $2,456,000
Est. hypervisor cost
per chassis at
specified config $0 $3,750 $0 $37,500 $90,000 $157,500 $405,000

Computed
Parameters
Maximum chassis
per rack (theoretical,
allowing only
minimum possible
overhead) 34 34 4 4 1 1 1
Theoretical servers
per rack (not
allowing for
overhead) 34 544 40 640 36000 36000 36000
Fibre Channel
switches per rack
(redundancy
included) 6 6 2 2 0 0 0
Ethernet switches
per rack (redundancy
included) 2 2 2 2 0 0 0
KVM switches per
rack 1 1 1 1 0 0 0

© 2007 Sine Nomine Associates 29


Avoiding the $25 Million Server

IBM System z
Dell PowerEdge Dell PowerEdge Dell PowerEdge Dell PowerEdge IBM System z BC S07 IBM System z
Parameters 1950 1950 virtualized 1955 1955 virtualized BC S07 Small Expanded EC S18
Total overhead units
per rack 14 14 10 10 0 0 0
Actual net chassis
per rack 28 28 4 4 1 1 1
Max servers per rack
at best-case virtual
density (where
applicable) 28 448 40 640 36000 36000 36000

Est. infrastructure
cost per rack for
SAN, network, and
KVM $68,256 $68,256 $36,756 $36,756 $0 $0 $0

Estimated equivalent
z9 MIPS per physical
processor 388 388 388 388 355 355 355
Estimated equivalent
z9 MIPS per physical
server 388 776 388 776 1420 2485 6390
Estimated equivalent
z9 MIPS per chassis 388 776 3880 7760 1420 2485 6390
Estimated equivalent
z9 MIPS per rack 10864 21728 15520 31040 1420 2485 6390

© 2007 Sine Nomine Associates 30


Avoiding the $25 Million Server

IBM System z
Dell PowerEdge Dell PowerEdge Dell PowerEdge Dell PowerEdge IBM System z BC S07 IBM System z
Parameters 1950 1950 virtualized 1955 1955 virtualized BC S07 Small Expanded EC S18
Infrastructure power
per rack 1500 1500 1020 1020 0 0 0
Server power per
rack 18760 18760 16000 16000 5400 5400 6300
Total power per rack 20260 20260 17020 17020 5400 5400 6300
Demanded power vs.
available limit 484% 484% 406% 406% 129% 129% 150%

Max number of Max number of Max number of Max number of Max number of
servers servers servers servers servers
represents a represents a represents a represents a represents a
reasonable reasonable reasonable reasonable reasonable
theoretical value, theoretical value, theoretical value, theoretical value, theoretical value,
which will be which will be which will be which will be which will be
further limited by further limited by further limited by further limited by further limited by
other calculations other calculations other calculations other calculations other calculations
Configuration Notes in the model in the model in the model in the model in the model
Processor Processor Processor Processor
selected for selected for selected for selected for
across-the-board across-the-board across-the-board across-the-board
availability, for availability, for availability, for availability, for
easier easier easier easier
comparison comparison comparison comparison

© 2007 Sine Nomine Associates 31


Avoiding the $25 Million Server

IBM System z
Dell PowerEdge Dell PowerEdge Dell PowerEdge Dell PowerEdge IBM System z BC S07 IBM System z
Parameters 1950 1950 virtualized 1955 1955 virtualized BC S07 Small Expanded EC S18
IBM does not IBM does not IBM does not
publish prices on publish prices on publish prices on
mainframes; the mainframes; the mainframes; the
pricing is based pricing is based pricing is based
Priced at Priced at on an estimate- on an estimate- on an estimate-
dell.com and by dell.com and by driven model driven model driven model
Priced online at Priced online at phone to Dell phone to Dell (see Globals for (see Globals for (see Globals for
Pricing Notes dell.com dell.com pre-sales pre-sales constants) constants) constants)
Some values Some values Some values
obtained from obtained from obtained from
QLogic 2462 QLogic 2462 QLogic 2462 QLogic 2462 SNA case SNA case SNA case
dual-channel FC dual-channel FC dual-channel FC dual-channel FC studies; details studies; details studies; details
HBA for SAN HBA for SAN HBA for SAN HBA for SAN not publishable not publishable not publishable
access access access access due to NDA. due to NDA. due to NDA.
Dual-channel Dual-channel Dual-channel Dual-channel
Ethernet included Ethernet included Ethernet included Ethernet included
in base config in base config in base config in base config
Doubly- Doubly-
redundant (2+2) redundant (2+2)
Added redundant Added redundant power supplies power supplies
P/S P/S standard standard
Minimum Minimum Minimum Minimum
onboard DASD onboard DASD onboard DASD onboard DASD
(36 GB, no RAID) (36 GB, no RAID) (36 GB, no RAID) (36 GB, no RAID)
selected selected selected selected
Removed Dell Removed Dell Dell remote Dell remote
remote remote management management
management management interface not interface not
interface interface used used

© 2007 Sine Nomine Associates 32


Avoiding the $25 Million Server

IBM System z
Dell PowerEdge Dell PowerEdge Dell PowerEdge Dell PowerEdge IBM System z BC S07 IBM System z
Parameters 1950 1950 virtualized 1955 1955 virtualized BC S07 Small Expanded EC S18
Enclosure price Enclosure price
estimated at estimated at
$5000, would be $5000, would be
somewhat lower somewhat lower
if purchased in if purchased in
quantity quantity
Price above Price above
reflects 10 reflects 10
identical blades, identical blades,
but no quantity but no quantity
discount is discount is
reflected here reflected here

Workload Characteristics

Management
Workload scenario Static web Database DNS or LDAP e-Commerce Business and monitoring E-mail
assumptions server server server web server File server logic server server server
CPU Network and
Net andDASD I/O and Net and DASDNet andintensive, DASD I/O withNet and
Characteristics DASD I/O moderate CPU Low utilization I/O, crypto DASD I/O moderate I/O moderate CPU DASD I/O
Frequency of
occurrence (unitless) 30 4 8 4 40 4 4 20
Application CPU load
factor (unitless) 0.05 0.25 0.03 0.1 0.02 0.3 0.05 0.03
Working storage set
(megabytes) 256 1536 128 512 128 1024 256 256

© 2007 Sine Nomine Associates 33


Avoiding the $25 Million Server

Management
Workload scenario Static web Database DNS or LDAP e-Commerce Business and monitoring E-mail
assumptions server server server web server File server logic server server server
Relative DASD load
factor (unitless) 0.5 0.9 0.05 0.2 0.9 0.4 0.6 0.8
Relative crypto load
factor (unitless) 0.02 0.02 0 0.2 0.1 0.02 0.1 0.05
Relative bandwidth
factor (unitless) 1 0.75 0.1 0.5 1 0.6 0.75 0.85
Kernel CPU load factor
for I/O (unitless) 0.075 0.0825 0.0075 0.035 0.095 0.05 0.0675 0.0825
Ratio of net I/O that is
server-to-server 0 0.9 0.1 0.5 0 0.98 0.98 0.02

The aggregated workload values are reported in Table 1on page 14.

Scenario Results

Dell Dell
Dell PowerEdge Dell PowerEdge IBM System z
PowerEdge 1950 PowerEdge 1955 IBM System z BC S07 IBM System z
1950 virtualized 1955 virtualized BC S07 Small Expanded EC S18
Platform Summary
Actual net chassis per rack 28 28 4 4 1 1 1
Max servers per rack at best-case
virtual density (where applicable) 28 448 40 640 36000 36000 36000

© 2007 Sine Nomine Associates 34


Avoiding the $25 Million Server

Dell Dell
Dell PowerEdge Dell PowerEdge IBM System z
PowerEdge 1950 PowerEdge 1955 IBM System z BC S07 IBM System z
1950 virtualized 1955 virtualized BC S07 Small Expanded EC S18
Estimated equivalent z9 MIPS per
chassis 388 776 3880 7760 1420 2485 6390
Total power per rack 20260 20260 17020 17020 5400 5400 6300
Demanded power vs. available
limit 484% 484% 406% 406% 129% 129% 150%
Allowable racks in data center
based on design limit 49 49 58 58 184 184 158
Allowable racks in data center,
limited by “notional limit” 49 49 58 58 8 8 8

Platform Adjustments
RAM overcommit factor 1 1 1 1 2 2 2
Offloaded DASD support factor 0.5 0.5 0.5 0.5 0.95 0.95 0.95
Offloaded crypto support factor 0 0 0 0 0.95 0.95 0.95
Offloaded net support factor 0 0 0 0 0.85 0.85 0.85
CPU I/O handling offload factor 0 0 0 0 0.9 0.9 0.9
% of net I/O that is inside chassis 0% 12% 12% 13% 13% 13% 13%

Workload Calculations
Logical servers per chassis based
on CPU load 5.273247496 10.54649499 52.73247496 105.4649499 19.29899857 33.7732475 86.84549356
Logical servers per chassis based
on RAM capacity 14.47619048 115.8095238 14.47619048 115.8095238 57.9047619 115.8095238 347.4285714

© 2007 Sine Nomine Associates 35


Avoiding the $25 Million Server

Dell Dell
Dell PowerEdge Dell PowerEdge IBM System z
PowerEdge 1950 PowerEdge 1955 IBM System z BC S07 IBM System z
1950 virtualized 1955 virtualized BC S07 Small Expanded EC S18
Logical servers per chassis based
on DASD bandwidth 1.502666833 3.005333666 15.02666833 30.05333666 5.49945078 9.624038864 24.74752851
Logical servers per chassis based
on crypto load 16.36525085 32.7305017 163.6525085 327.305017 59.89344384 104.8135267 269.5204973
Logical servers per chassis based
on network bandwidth 1.159899653 2.319799306 11.59899653 23.19799306 28.29995717 49.52492505 127.3498073

Adjusted Workload Calculations


Logical servers per chassis based
on CPU load 5.27 10.55 52.73 105.46 58.24 101.91 262.07
Logical servers per chassis based
on RAM capacity 14.48 115.81 14.48 115.81 115.81 231.62 694.86
Logical servers per chassis based
on DASD bandwidth 3.01 6.01 30.05 60.11 109.99 192.48 494.95
Logical servers per chassis based
on crypto load 16.37 32.73 163.65 327.31 1197.87 2096.27 5390.41
Logical servers per chassis based
on network bandwidth, excluding
intra-chassis traffic 1.16 2.64 13.11 26.59 216.46 378.81 974.09
Logical servers per chassis based
on architecture 1 16 10 160 36000 36000 36000

Logical servers per chassis based


on worst-case for platform 1.00 2.64 10.00 26.59 58.24 101.91 262.07

© 2007 Sine Nomine Associates 36


Avoiding the $25 Million Server

Dell Dell
Dell PowerEdge Dell PowerEdge IBM System z
PowerEdge 1950 PowerEdge 1955 IBM System z BC S07 IBM System z
1950 virtualized 1955 virtualized BC S07 Small Expanded EC S18
Logical servers per rack,
accounting for all limitations 28 73 40 106 58 101 262
Logical servers in data center,
accounting for all limitations 1372 3577 2320 6148 464 808 2096

Cost Calculations
Est. hardware cost per chassis at
specified config $6,082 $20,421 $53,150 $196,340 $486,000 $899,000 $2,456,000
Est. hypervisor cost per chassis at
specified config $0 $3,750 $0 $37,500 $90,000 $157,500 $405,000
Est. infrastructure cost per rack $68,256 $68,256 $36,756 $36,756 $0 $0 $0
Acquisition cost per rack $238,552 $745,044 $249,356 $972,116 $576,000 $1,056,500 $2,861,000
Total acquisition cost $11,689,048 $36,507,156 $14,462,648 $56,382,728 $4,608,000 $8,452,000 $22,888,000

Power utilization for servers (kW) 919 919 928 928 43 43 50


Power utilization for HVAC (kW) 643 643 650 650 30 30 35
Total power utilization (kW) 1562 1562 1578 1578 73 73 85
Power cost per month (US$) $105,266 $105,266 $106,345 $106,345 $4,920 $4,920 $5,728
Power cost per year (US$) $1,263,196 $1,263,196 $1,276,135 $1,276,135 $59,035 $59,035 $68,740
Power cost per year per logical
server $920.70 $353.14 $550.06 $207.57 $127.23 $73.06 $32.80
Hypervisor support cost per year
(US$) $0 $11,500 $0 $11,500 $11,500 $11,500 $11,500

© 2007 Sine Nomine Associates 37


Avoiding the $25 Million Server

Dell Dell
Dell PowerEdge Dell PowerEdge IBM System z
PowerEdge 1950 PowerEdge 1955 IBM System z BC S07 IBM System z
1950 virtualized 1955 virtualized BC S07 Small Expanded EC S18

First year TCO $12,952,244 $37,781,852 $15,738,783 $57,670,363 $4,678,535 $8,522,535 $22,968,240
Three-year TCO $15,478,635 $40,331,243 $18,291,053 $60,245,633 $4,819,606 $8,663,606 $23,128,720
Five-year TCO $18,005,026 $42,880,634 $20,843,323 $62,820,903 $4,960,677 $8,804,677 $23,289,199

TCO Costs PER LOGICAL


SERVER
First year TCO $9,440 $10,562 $6,784 $9,380 $10,083 $10,548 $10,958
Three-year TCO $11,282 $11,275 $7,884 $9,799 $10,387 $10,722 $11,035
Five-year TCO $13,123 $11,988 $8,984 $10,218 $10,691 $10,897 $11,111
Fifth-year incremental cost per
logical server $1,841 $713 $1,100 $419 $304 $175 $77

© 2007 Sine Nomine Associates 38


Avoiding the $25 Million Server

Appendix B: Example Calculation


As an example of the model's operation, let us consider the case of the System z Model S07, in
the small configuration. This section will trace its calculations end-to-end to illustrate how the
results are derived.

Consider first the workload calculation for CPU utilization, which applies to all the server types.
The utilization number for each workload type is multiplied by the frequency of occurrence of
that workload, yielding a number that is in itself not meaningful for comparison but rather is an
intermediate result. These intermediate values are summed and divided by the sum of the relative
occurrence values, which normalizes the CPU utilization back to a unit-weighted average
representing one unit of the aggregated workload. Thus, it does not matter what absolute numbers
are used for the frequency of occurrence; only their proportion to one another is significant. This
process is repeated for each of the system resource calculations (RAM, I/O bandwidth, network
bandwidth, etc.).

The Platforms page shows the System z Model S07 server with 4 IFL processors and 16
gigabytes of main memory (RAM). These are constants arising from the choice of configuration.
This yields a total of 1420 estimated MIPS for the mainframe as a whole. Since the mainframes
considered here occupy essentially one rack location each, the per-chassis and per-rack numbers
are equivalent for this platform.

Note that IBM does not disclose list prices for mainframes, so the acquisition costs are partially
based on amounts paid by Sine Nomine Associates customers. Some mainframe hardware,
notably the IFL engines, is priced publicly by IBM, and in these cases those public list prices are
used.

The Platforms page also computes a ratio of power demanded per chassis versus nominal power
available per chassis, the latter being computed from the data center's watts-per-square-foot value
from the Globals page. In this model, it happens that all of the platforms exceed the nominal
power per rack location, but the amount varies considerably. The mainframe's ratio, expressed as
a percentage, is 129%, meaning that if mainframe racks were to fill the data center at maximum
physical density, the data center would be overcommitted on power by 29%. The ratio computed
here is revisited on the Scenarios page, where it is applied as a limiting factor to limit the
maximum number of racks (for each platform) that can be installed in the data center.

In the case of the mainframe, Sine Nomine Associates has applied an arbitrarily chosen "notional
limit" of 8 mainframes total per data center, based on our experience with actual customers. Thus,
it is the notional limit that constrains the number of mainframes per data center, not the power
limit. For other platforms where no notional limit is imposed, the power allocation eventually
constrains the number of racks. For the configurations considered here, there are no cases where
physical space constrains the number of servers or racks in the entire data center.

Moving on to the Scenarios page, the Platform Summary numbers are simply mirrored from the
Globals and Platforms pages, for convenience. Then a set of constraint-based computations
determines how many logical servers per physical chassis could be deployed, based on each of
the resource types (CPU, RAM, I/O, etc.) defined in the aggregated workload. The constraints are
separately calculated, and it is understood that the final constraint is based on whichever of these
resources is the worst case for any particular platform or configuration. As expected from

© 2007 Sine Nomine Associates 39


Avoiding the $25 Million Server

customer experiences known to Sine Nomine Associates, the mainframe platforms are typically
constrained by CPU resources rather than I/O, memory, or architectural limits.

Rather than using these numbers directly, it is more accurate to apply a set of platform-specific
adjustments to each constraint value, to reflect the qualitative differences between architectures.
For instance, the microcomputers benefit somewhat from offloading of disk management to the
storage area network (SAN), because the main CPU does not have to manage low-level device
interfaces for the drive but rather can treat the SAN as a virtualized resource. The mainframe,
using the same SAN technology, gains this benefit also, but in addition benefits from the fact that
the SAN host bus adapter is connected to an intelligent channel controller, providing an even
higher level of I/O offloading. Linux device drivers in the mainframe environment are built to
leverage this capability.

The disk I/O offloading also is reflected in an adjustment to memory overcommit limits, because
the mainframe offers a feature called VDISK that allows Linux to see "swap space" that operates
at nearly RAM speed, whereas on the microcomputer platform any swapping operation actually
involves disk access. Both Sine Nomine Associates experience and published sources show that
RAM on the mainframe can be overcommitted at a higher level than on a microcomputer. As it
turns out, however, the CPU constraint on the mainframe still is the limiting factor, even with this
adjustment applied.

Similar adjustments are applied to network I/O based on the intelligence of the subsystem on the
different platforms. This is calculated as a reduction in kernel processing time based on the fact
that the Linux operating system delegates much of what would be low-level device handling to
the mainframe's intelligent channel controllers.

For cryptographic processing, the platform-specific adjustment applies all of the cryptographic
workload to the main CPU on the microcomputers, but removes a fixed percentage of this
workload from the main CPU on the mainframe, because the mainframe offers multiple
cryptographic coprocessors as a standard feature. Some of the crypto workload remains on the
main CPU, however, reflecting the small but finite amount of CPU required to supervise the
cryptographic engines.

The end result of the platform-specific adjustments is to accurately model the fact that while the
mainframe has a smaller total CPU cycle count than a rack full of microcomputers, the
mainframe's CPU is utilized more efficiently and is more available (proportionally) to
application-level code rather than being consumed by low-level operating system kernel
functions.

Based on the adjusted constraints for each platform, the constraint calculations are re-executed to
yield the final list of constrained maximum logical server counts. It is no surprise that even here,
the mainframe's CPU is its worst-case constraint and limits the final server count for that
platform. For the various microcomputer platforms, other constraints dominate in some cases.
Note that the constraint selection is applied neutrally to all platforms; that is, it is not required to
be the same on all configurations, but rather is taken as the worst case for each individual
configuration.

Once the actual number of logical servers is known per chassis (and thereby per rack, since the
number of chassis per rack is already known), a final TCO calculation can be performed to
determine the total TCO for the data center, and the total TCO per logical server.

© 2007 Sine Nomine Associates 40


Avoiding the $25 Million Server

Bibliography
Listing of a source in this Bibliography does not imply that the author of that document endorses
or supports the conclusions of this study.

Sine Nomine Associates http://www.sinenomine.net/

IBM System z Mainframes http://www.ibm.com/systems/z/

Dell Rack Servers


http://www.dell.com/content/products/compare.aspx/rack_optimized

Dell Blade Servers


http://www.dell.com/content/products/compare.aspx/blade

The Uptime Institute http://www.uptime.org/

Debian Linux http://www.debian.org/

VMware ESX http://www.vmware.com/products/vi/esx/

Cisco network switches http://www.cisco.com/en/US/products/hw/switches/

QLogic SAN switches http://www.qlogic.com/products/san_pro.asp

Raritan KVM switches


http://www.raritan.com/products/kvmswitches/dominionkx

Kenneth Brill, "2005-2010 Heat Density Trends in Data Processing, Computer Systems, and
Telecommunications Equipment"

Turner and Seader, "Dollars per kW plus Dollars per Square Foot Are a Better Data Center Cost
Model than Dollars per Square Foot Alone"

Turner, Seader, and Brill, "Tier Classifications Define Site Infrastructure Performance"

Rob Snevely, "Enterprise Data Center Design and Methodology" (Sun Microsystems)

U.S. Dept. of Energy, national average for commercial electricity for 12 months ending October
2006, http://www.eia.doe.gov/cneaf/electricity/epm/table5_3.html (data retrieved April 4, 2007)

© 2007 Sine Nomine Associates 41

You might also like