You are on page 1of 34

Capacity Planning

DISCIPLINE FOR DATA CENTER DECISIONS


Table of Contents
1 Introduction ...........................................................................1-1
2 The Value of Capacity Planning................................................2-1
Introduction ............................................................................. 2-1
The Value Proposition for Capacity Planning .............................. 2-2
3 Capacity Planning Processes...................................................3-1
4 How to Do Capacity Planning...................................................4-1
Three Steps for Capacity Planning ............................................. 4-1
Determine Service Level Requirements ...................................... 4-1
Analyze Current Capacity .......................................................... 4-5
Plan for the Future .................................................................... 4-9
Capacity Planning Process ...................................................... 4-13
5 Using Capacity Planning for Server Consolidation ....................5-1
Server Consolidation Defined.................................................... 5-1
Finding the Answers Can Be Difficult ......................................... 5-2
Predict the Future ..................................................................... 5-2
Case Studies ............................................................................ 5-2
Company A Finding Underutilized Resources for Use
as Disaster Backup ................................................................... 5-2
Company B Avoiding Unnecessary Expenditures ..................... 5-4
Conclusion ............................................................................... 5-4
6 Capacity Planning and Service Level Management ...................6-1
Service Level Agreements (SLAs) ............................................... 6-1
Proactive Management of Resources to Maintain
Service Levels........................................................................... 6-2
Service Level Management Scenarios........................................ 6-2
7 Product Focus: TeamQuest Model............................................7-1
TQ-EB01 Rev. A Case Studies ............................................................................ 7-3
Copyright 2004 TeamQuest Corporation EMA Perspective ....................................................................... 7-4
All Rights Reserved 8 Bibliography...........................................................................8-1
Like what you see? Subscribe. See the final page for important legal notices.
Introduction

1 Introduction
Both server capacity planning and network capacity planning are important disciplines for managing
the efficiency of any data center. The focus of this eBook is on server capacity planning.
Traditionally, server capacity planning is defined as the process by which an IT department
determines the amount of server hardware resources required to provide the desired levels of service
for a given workload mix for the least cost. The capacity planning discipline grew up in the mainframe
environment where resources were costly and it took a considerable amount of time to upgrade. As
the data center transitioned to a distributed environment using UNIX, Linux, and Windows servers,
over-provisioning and the introduction of cheap new boxes served as a replacement for capacity
planning. However, as the distributed data center matures we have found these practices to be less
than optimal.
The average utilization on many servers in the distributed data center is way below acceptable
maximums, wasting costly resources. Further, as the economy evolves, IT is expected to do its share
to become more efficient, both in terms of hardware and human resources. And finally, IT is now
becoming an accepted partner in the organizations drive to grow revenue while increasing profits
and remaining competitive for the long term. All of these factors require more mature IT processes,
including the use of capacity planning, a core discipline every company should adopt.
There are a number of types of capacity planning including the following:
1. Capacity benchmarking
2. Capacity trending
3. Capacity modeling
Benchmarking, or load testing, is perhaps the most common, but also the most expensive. The idea
is, you set up a configuration and then throw traffic at it to see how it performs. To do this right, you
need access to a fully-configured version of the target system, which oftentimes makes
benchmarking or load testing impractical, to say the least.
Linear trend analysis and statistical approaches to trending can provide quick and dirty ways to
predict when you will need to do something about performance, but they dont tell you what you
should do to optimally respond. Trending does not provide a way to evaluate alternative solutions to
an impending problem, nor to understand problems unrelated to trending, e.g. server consolidation
or adding an application.
That leaves modeling, which comes in a couple flavors: simulation and analytic modeling. Simulation
modeling can be very versatile and accurate, but requires a great deal of set up effort and time.
Analytic modeling is fast and is potentially very accurate as well. The beauty of modeling is that you
can test various proposed solutions to a problem without actually implementing them. This can
save a lot of time and money.
This eBook explores the modeling approach to capacity planning. However, regardless of the toolset
you use, if any, the key is to adopt mature IT processes, including the strategic use of capacity
planning.
This eBook is organized into a series of chapters, with Chapter 2 following this introduction:
2. The Value of Capacity Planning
The second chapter is excerpted from a paper1 written by Enterprise Management Associates,
an analyst organization focused on the management software and services market. It
explains the benefits of capacity planning.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 1-1
Introduction

3. Capacity Planning Processes


In this chapter we discuss capacity planning in context with other generally accepted IT
processes, specifically using the ITIL model.
4. How to Do Capacity Planning
This section illustrates a methodology for capacity planning with examples using the
TeamQuest product line.
5. Using Capacity Planning for Server Consolidation
Server consolidation is a popular capacity planning technique for making IT operations more
efficient. This chapter explains server consolidation and provides two real-world examples
where capacity planning tools were used in consolidation projects.
6. Capacity Planning and Service Level Management
This chapter discusses key aspects of service level management and the relationship
between service level management and capacity planning.
7. Product Focus: TeamQuest Model
We conclude with another excerpt from EMAs paper1 that highlights TeamQuests capacity
planning solution, TeamQuest Model.
8. Bibliography
This eBook includes material from several resources listed here.
Whether you are learning about capacity planning for the first time, or you are an old hand looking for
some new points of view on the subject, we hope this eBook will prove useful to you.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 1-2
The Value of Capacity Planning

2 The Value of Capacity Planning This chapter is excerpted from


a white paper1 by Enterprise
Management Associates (EMA),
Introduction an analyst organization
specializing in the
The boom-and-bust economy of the past five years has had an management software and
unprecedented effect on IT organizations. In the late 1990s and services market.
2000, IT infrastructures grew at a frantic rate, creating a level of
capacity and staffing that had never been seen before. Yet, with the economic tailspin of the
subsequent three years, most corporations are now seeking new ways to leverage IT to improve
business efficiency and automationwhile severely restricting ITs budgets for personnel, training,
bandwidth, and equipment.
This roller coaster of supply and demand has left many IT managers looking for new, better ways to
manage their resources. While previous IT strategies focused primarily on managing growth, todays
IT organizations now are seeking methods to expand the number of applications and services
available to support the business while cutting costs at the same time. In the current economy IT is
increasingly being asked to do more with less.
From a server perspective, the rise and fall of IT has instigated
In a recent EMA survey,
several shifts in resource management. First, most IT organizations
reclaiming and/or re-
are looking to make better use of existing assets by consolidating
purposing hardware and
underutilized server resources and re-purposing capacity that was
software that is underutilized
purchased during the boom years, but has now been freed up by
was cited as a top priority by
down-sizing and/or reorganization of corporate business units. In a
57% of IT executives
recent EMA survey, reclaiming and/or re-purposing hardware and
responding.
software that is underutilized was cited as a top priority by 57% of
IT executives responding.
Second, IT organizations are re-thinking their previous brute force approach to provisioning server
capacity. During the boom years, when growth was the top priority, many IT organizations responded
to performance problems simply by purchasing additional servers to handle the load. Today, when
every IT dollar is critical, this over-provisioning approach is viewed as wasteful and inefficient,
particularly since downsizing, not growth, is the order of the day. In difficult economic times, IT
organizations must be able to measure server utilization in sharp detail, and make the best possible
use of every processor cycle.
Lastly, many IT organizations have grown weary of traditional server-by-server resource management
efforts and are now evaluating next-generation technologies that would enable servers to pool and
dynamically allocate their capacity to fit the specific needs of applications and services. This dynamic
allocation of capacity, sometimes called virtualization, enables applications to draw capacity as
needed from a collective infrastructure, rather than relying on dedicated servers. This virtualization
technology could eventually pave the way for a new paradigm known as utility computing, in which
users and/or applications draw only the capacity they need from a common infrastructure, just as
homes or businesses draw only the water or electricity they need each month from their local utility.
How can enterprises re-purpose underutilized capacity, implement structured methods for
provisioning servers, and prepare their IT environments for the coming wave of virtualization? The
answer to all three questions is the same: capacity planning. By using capacity planning tools and
processes, enterprises can measure server utilization trends, analyze future capacity needs, and
predict the exact server requirement for a given application, service, or general time period. Some
capacity planning toolsets also offer the ability to model the server environment, testing a variety of
what if scenarios against proposed server configurations to determine the most efficient use of
capacity and server resources. In the following section, we examine the capabilities of capacity
planning and modeling tools, and evaluate the value that they can deliver to the enterprise.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 2-1
The Value of Capacity Planning

The Value Proposition for Capacity Planning


In a difficult economy that has imposed severe restrictions on IT budgets, it may seem
counterintuitive to invest in new technology as a means of cost control. Yet, in the case of capacity
planning technology, a relatively small investment in software can lead to a very large savings in
server hardware and associated costs. While there may be some costs associated with deploying
capacity planning tools and processes, the value of that investment is typically extremely high.
How is that value achieved? One way is through the reductionor,
in some cases, complete elimination of server over-provisioning. Key Capacity Planning
Capacity planning tools enable IT organizations to benchmark the Value Points
exact workload that a server handles over time, then project trends
Reduce or eliminate server
in traffic that the server might need to handle in the future. This
over-provisioning
data, in turn, enables the IT organization to determine the exact
server requirements for a particular situation, ensuring that the Identify and repurpose
server environment will have enough capacity at all times. This underutilized servers
process contrasts sharply with the estimated capacity allocation
Reduce IT operational
method in which the IT organization simply guesses at the capacity
expenditures
requirement, then overbuys servers to ensure that there will always
be sufficient processor power. In many cases, capacity planning Reduce server downtime
may enable the enterprise to slash its server acquisition budget
Improve server
significantly.
performance and
The flip side of this server utilization data is that it helps availability
enterprises to identify underutilized servers and capacity so that
Improve IT budget accuracy
they can be consolidated or re-purposed for other applications and
services. Capacity planning tools help organizations to identify Leverage future
server resources that might be idle, enabling IT departments to technologies
consolidate several applications or services on a single server, or
reallocate server processor power to other services that may need the capacity. This is an essential
capability in todays IT organization, where capacity purchased during the boom years may be
sitting idly while budgets for new server capacity have often been slashed to the bone.
Capacity planning can also help preserve another crucial enterprise resource: IT staffing. Over the
course of a servers life, most enterprises spend more money on maintenance, configuration, and
upgrade of servers than they spent to purchase the server in the first place. For every underutilized
server, there is a group of IT staff who have been tasked with installing, configuring, and maintaining
that server. If server resources can be used more efficientlyeither by eliminating servers or by
enabling a single server to handle multiple applicationsthen the cost of IT staffing drops
proportionately.
A related benefit of capacity planning is the reduction of server downtime. Under the estimated
method of capacity allocation, enterprises typically purchase more servers than they need, then
simply re-provision their servers when there is a capacity overload. As a result, most IT operations
centers are tasked with maintaining and repairing more servers than they can use, yet they still
experience outages or system performance problems when an overload forces the addition or hot-
swapping of a new server in the environment. By contrast, enterprises that use capacity planning
tools generally can anticipate overloads before they occur, thus reducing or eliminating workload-
related outages while enabling the IT operations center to reduce the number of servers it must
maintain.
Capacity planning can also improve the overall performance and availability of the server
environment. While experienced IT staffers may be able to make an educated guess as to their
capacity needs at any given time, such guesses are fallible and may not take all contingencies into
account. As a result, many IT organizations that do not use capacity planning tools occasionally find
their servers oversubscribed, delivering slow response times or timeouts that may slow business

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 2-2
The Value of Capacity Planning

processes or cause customer dissatisfaction. With capacity planning tools, however, the enterprise
can track trendsor, in some cases, model new applications or servicesto anticipate potential
performance and availability problems and eliminate them before they occur.
Capacity planning tools can also help deliver another key benefit in the current economybudget
predictability. During the boom years, enterprises could afford to provision their servers on the fly,
over-provisioning new services at the outset and then purchasing new capacity when existing servers
were oversubscribed. Under todays tight budgets, however, the ability to predict future server needs
is crucial to making intelligent buying decisions. Capacity planning enables the enterprise to develop
an educated server purchasing strategy, and helps to reduce the need for panic purchasing
situations that a server vendor might exploit to charge premium prices for its equipment.
Finally, capacity planning technology is an essential element in the emerging paradigms of system
virtualization and utility computing. In order to create an effective pool of capacity to support all
necessary applications and services, enterprises must have a way to track utilization and predict
future capacity needs. This process of analyzing and projecting capacity requirements will become
even more complex when enterprises begin to employ computing grids that are designed to
support the broad diversity of applications and services used in the enterprise. In the utility
computing model, capacity planning will become as important to the enterprise IT organizations as it
is to power, telephone, or water utilities today.
The next chapter places capacity planning in the overall context of IT processes.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 2-3
Capacity Planning Processes

3 Capacity Planning Processes


The IT Infrastructure Library (ITIL)2 describes a standard set of established processes by which an IT
organization can achieve operational efficiencies and thereby deliver enhanced value to an
organization. ITIL is most commonly used in Europe, originally having been created for the U.K.
government. Because it provides a useful best-practice framework that can be put to use relatively
inexpensively, ITIL is growing in popularity as a means to manage IT services within both commercial
and governmental organizations.
As shown in the following diagram, Capacity Management is part of the Service Delivery area of ITIL.
Capacity Planning is part of the ITIL Capacity Management process.

Figure 3-1
ITIL Service Delivery Processes

Further, ITIL defines three sub-processes in the Capacity Management process:


Business Capacity Management: This sub-process is responsible for ensuring that the
future business requirements for IT Services are considered, planned and implemented
in a timely fashion.
Service Capacity Management: The focus of this sub-process is management of the
performance of live, operational IT Services.
Resource Capacity Management: The focus in this sub-process is the management of the
individual components of the IT infrastructure.
The types of capacity planning that we will cover in this eBook focus on the last two sub-processes,
service and resource capacity management. Specifically, when we build a single system model, we
are usually performing aspects of resource capacity management rather than service capacity
management. This is because services typically span multiple systems. When we combine models
from multiple systems, then we are performing aspects of service capacity management.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 3-1
Capacity Planning Processes

The capacity planning function within capacity management interfaces with many other processes as
indicated in the following diagram:

Figure 3-2
Workloads and Service
As you can see, service level management is where all good IT processes begin, since this is the crux
of the problem. At a minimum, service level management provides service definitions, service level
targets, business priorities, and growth plans. Within given architectural constraints, capacity
planning determines the cheapest optimal hardware and software configuration that will achieve
required service levels, both now and in the future. Capacity planning then feeds configuration and
asset management to make this all a reality so that service delivery can begin. On another process
branch, performance management verifies capacity plans and service levels and provides revised
input into capacity planning when necessary.
The next chapter provides a high-level approach capacity planning best practices.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 3-2
How to Do Capacity Planning

3
4 How to Do Capacity Planning
It is very common for an IT organization to manage system performance in a reactionary fashion,
analyzing and correcting performance problems as users report them. When problems occur,
hopefully system administrators have tools necessary to quickly analyze and remedy the situation. In
a perfect world, administrators prepare in advance in order to avoid performance bottlenecks
altogether, using capacity planning tools to predict in advance how servers should be configured to
adequately handle future workloads.
The goal of capacity planning is to provide satisfactory service levels to users in a cost-effective
manner. This chapter describes the fundamental steps for performing capacity planning. Real life
examples are provided using TeamQuest Performance Software.

Three Steps for Capacity Planning


In this chapter we will illustrate three basic steps for capacity planning:
1. Determine Service Level Requirements
The first step in the capacity planning process is to categorize the work done by systems and to
quantify users expectations for how that work gets done.
2. Analyze Current Capacity
Next, the current capacity of systems must be analyzed to determine how they are meeting the
needs of the users.
3. Planning for the Future
Finally, using forecasts of future business activity, future system requirements are determined.
Implementing the required changes in system configuration will ensure that sufficient capacity
will be available to maintain service levels, even as circumstances change in the future.

Determine Service Level Requirements


We have organized this section as follows:
a. The overall process of establishing service level requirements first demands an understanding of
workloads. We will explain how you can view system performance in business terms rather than
technical ones, using workloads.
b. Next, we begin an example, showing workloads on a system running a back-end Oracle database.
c. Before setting service levels, you need to determine what unit you will use to measure the
incoming work.
d. Finally, you establish service level requirements, the promised level of performance that will be
provided by the IT organization.

Workloads Explained
From a capacity planning perspective, a computer system processes workloads (which supply the
demand) and delivers service to users. During the first step in the capacity planning process, these
workloads must be identified and a definition of satisfactory service must be created.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-1
How to Do Capacity Planning

A workload is a logical classification of work performed on a computer system. If you consider all the
work performed on your systems as a pie, a workload can be thought of as some piece of that pie.
Workloads can be classified by a wide variety of criteria:
who is doing the work Figure 4-1
(particular user or department) Workloads By Department
what type of work is being done Engineering
(order entry, financial reporting) Sales

how the work is being done


(online inquiries, batch database
backups) Marketing

It is useful to analyze the work done on systems in


terms that make sense from a business perspective, using business-relevant workload definitions.
For example, if you analyze performance based on workloads corresponding to business
departments, then you can establish service level requirements for each of those departments.
Business-relevant workloads are also useful when it comes time to plan for the future. It is much
easier to project future work when it is expressed in terms that make business sense. For example, it
is much easier to separately predict the future demands of the human resources department and the
accounts payable department on a consolidated server than it is to predict the overall increase in
transactions for that server.

An Example Using Workloads


The TeamQuest View chart below shows 24 hours (cropped to fit the page) of CPU utilization on an
IBM F50 PowerPC system. The chart is useful, but it provides a birds eye view of performance, at
best.

Figure 4-2
CPU Utilization
In Figure 4-3, below, the Process Table chart of TeamQuest View reveals that during the same 24
hour period, 10,642 individual processes ran on this system. All of the utilization information for all
of those processes was displayed together in our CPU utilization chart. Wouldnt it be nice if we could
show a similar chart, but display utilization based on the major functions being performed on this

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-2
How to Do Capacity Planning

system? Using TeamQuest Performance Software, we can do just that, by going through a process
called workload characterization.

Figure 4-3
Process Table
We will leave the detailed instructions for performing workload characterization to another eBook, or
you can refer to the TeamQuest Performance Software documentation. In a nutshell, workload
characterization requires you to tell TeamQuest Performance Framework how to determine what
resource utilization goes with which workload. This is done on a per process level, using selection
criteria to tell the TeamQuest Performance Framework how to determine which processes belong to
which workloads.
Figure 4-4, below, shows a list of workloads that have been characterized so that the work of each of
the 10,642 processes is attributed to one of seven workloads. These workloads are defined according
to the type of work being done on the system.

Figure 4-4
Workload Definitions

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-3
How to Do Capacity Planning

If you look carefully you will see six explicitly defined workloads in Figure 4-4, but we said there were
seven. The reason is that there is always an OTHER workload in addition to the explicitly defined
workloads. Any resource utilization that does not match the characterization for any of the explicitly
defined workloads becomes associated with OTHER. This ensures that no performance data falls
through the cracks simply because it didnt match any of the defined workloads.
In Figure 4-5, below, pink bars show utilization that did not match the characterization criteria for
any of our define workloads. There is little or no pink in the chart, demonstrating that we have done a
good job of explicitly characterizing most of the work done on this server.

Figure 4-5
CPU Utilization by Workload
All we did was define some workloads based on the type of work being performed on this server, but
notice how much more useful the information is that is provided in our new chart. Workloads can be
very powerful.

Determine the Unit of Work


For capacity planning purposes it is useful to associate a unit of work with a workload. This is a
measurable quantity of work done, as opposed to the amount of system resources required to
accomplish that work.
To understand the difference, consider measuring the work done at a fast food restaurant. When
deciding on the unit of work, you might consider counting the number of customers served, the
weight of the food served, the number of sandwiches served, or the money taken in for the food
served. This is as opposed to the resources used to accomplish the work, i.e. the amount of French
fries, raw hamburgers or pickle slices used to produce the food served to customers.
When talking about IT performance, instead of French fries, raw hamburger or pickle slices, we
accomplish work using resources such as disk, I/O channels, CPUs and network connections.
Measuring the utilization of these resources is important for capacity planning, but not relevant for
determining the amount of work done or the unit of work. Instead, for an online workload, the unit of
work may be a transaction. For an interactive or batch workload, the unit of work may be a process.
The examples given in this chapter use a server running an appointment scheduling application
process, so it seems logical to use a calendar request as the unit of work. A calendar request
results in an instance of an appointment process being executed.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-4
How to Do Capacity Planning

Establish Service Levels


The next step now is to establish a service level agreement. A service level agreement is an
agreement between the service provider and service consumer that defines acceptable service. The
service level agreement is often defined from the user's perspective, typically in terms of response
time or throughput. Using workloads often aids in the process of developing service level
agreements, because workloads can be used to measure system performance in ways that makes
sense to clients/users.
In the case of our appointment scheduling application, we might establish service level requirements
regarding the number of requests that should be processed within a given period of time, or we might
require that each request be processed within a certain time limit. These possibilities are analogous
to a fast food restaurant requiring that a certain number of customers should be serviced per hour
during the lunch rush, or that each customer should have to wait no longer than three minutes to
have his or her order filled.
Ideally, service level requirements are ultimately determined by business requirements. Frequently,
however, they are based on past experience. Its better to set service level requirements to ensure
that you will accomplish your business objectives, but not surprisingly people frequently resort to
setting service level requirements like, provide a response time at least as good as is currently
experienced, even after we ramp up our business. As long as you know how much the business will
ramp up, this sort of service level requirement can work.
If you want to base your service level requirements on present actual service levels, then you may
want to analyze your current capacity before setting your service levels.

Analyze Current Capacity


There are several steps that should be performed during the analysis of capacity measurement data.
a. First, compare the measurements of any items referenced in service level agreements with their
objectives. This provides the basic indication of whether the system has adequate capacity.
b. Next, check the usage of the various resources of the system (CPU, memory, and I/O devices).
This analysis identifies highly used resources that may prove problematic now or in the future.
c. Look at the resource utilization for each workload. Ascertain which workloads are the major users
of each resource. This helps narrow your attention to only the workloads that are making the
greatest demands on system resources.
d. Determine where each workload is spending its time by analyzing the components of response
time, allowing you to determine which system resources are responsible for the greatest portion
of the response time for each workload.

Measure Service Levels and Compare to Objectives


TeamQuest Model is a tool that can help us check measured service levels against objectives. For
example, after building a model of our example system for a three-hour window, 7:00 AM 10:00 AM,
the display below (Figure 4-6) shows the response time and throughput of the seven workloads that
were active during this time.
By looking at the top line of the table, you can tell that the model has been successfully calibrated for
our example system, because the total Measured AR% and Modeled AR% are equal. AR stands for
Active Resource. An active resource is a resource that is made 100% available once it has been
allocated to a waiting process. In this case, the active resource is CPU.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-5
How to Do Capacity Planning

Figure 4-6
Response Time and Throughput
Because no changes have been made in the system configuration, modeled response time and
throughput for each workload should closely match reality. In our example, the response time means
the amount of time required to process a unit of work, which in the case of our application, is an
appointment request process. So this report provides us with an appointment request average
response time for each of our workloads that we could compare with desired service levels.

Measure Overall Resource Usage


It is also important to take a look at each resource within your systems to see if any of them are
saturated. If you find a resource that is running at 100% utilization, then any workloads using that
resource are likely to have poor response time. If your goal is throughput rather than response time,
utilization is still very important. If you have two disk controllers, for example, and one is 50% utilized
and the other is swamped, then you have an opportunity to improve throughput by spreading the
work more evenly between the controllers.

Figure 4-7
Overall Resource Usage
The table above (Figure 4-7) shows the various resources comprising our example server. The table
shows the overall utilization for each resource. Utilization for the four CPUs are shown together
treated as one resource, otherwise each resource is shown separately.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-6
How to Do Capacity Planning

Notice that CPU utilization is about 64% over this period of time (7:00 AM 10:00 AM on January 02).
This corresponds with the burst of CPU utilization that was shown earlier in Figures 4-2 and 4-5.
No resource in the report seems to be saturated at this point, though hdisk2 is getting a lot more of
the I/O than either hdisk0 or hdisk1. This might be worthy of attention; future increases in workloads
might make evening out the disparity in disk usage worthwhile.

Measure Resource Usage by Workload


Figure 4-8, shows the same period again, only now resource utilization is displayed for the
APPOINTMENTS workload. Note that this particular workload is using 59% of the CPU resource, nearly
all of the 64% utilization that the previous table showed as the total utilization by all workloads.
Clearly, the APPOINTMENTS workload is where a capacity planner would want to focus his or her
attention, unless it is known that future business needs will increase the amount of work to be done
by other workloads on this system. In our example, that is not the case. Ramp-ups in work are
expected mainly for the APPOINTMENTS workload.

Figure 4-8
Resource Utilization by Workload
The previous charts and tables have been useful for determining that CPU Utilization is likely to be a
determining factor if the amount of work that our system is expected to perform increases in the
future. Furthermore, we were able to tell that the APPOINTMENTS workload is the primary user of the
CPU resources on this system.
This same sort of analysis can work no matter how you choose to set up your workloads. In our
example, we chose to treat appointment processes as a workload. Your needs may cause you to set
up your workloads to correspond to different business activities, such as a Wholesale Lumber unit vs.
Real Estate Development, thus allowing you to analyze performance based on the different
requirements of your various business units.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-7
How to Do Capacity Planning

Identify Components of Response Time


Next we will show how to determine what system resources are responsible for the amount of time
that is required to process a unit of work. The resources that are responsible for the greatest share of
the response time are indicators for where you should concentrate your efforts to optimize
performance. Using TeamQuest Model we can determine the components of response time on a
workload by workload basis, and you can predict what the components will be after a ramp-up in
business or a change in system configuration.
A components of response time analysis shows the average resource or component usage time for a
unit of work. It shows the contribution of each component to the total time required to complete a unit
of work.

0.3

0.25
All Other AR Queues Queue Delay
All Other AR Queues Service
20-58-00(aixdemo) Queue Delay
0.2
20-58-00(aixdemo) Service
hdisk0(aixdemo) Queue Delay
Seconds

hdisk0(aixdemo) Service
0.15
hdisk1(aixdemo) Queue Delay
hdisk1(aixdemo) Service
hdisk2(aixdemo) Queue Delay
0.1
hdisk2(aixdemo) Service
CPU(aixdemo) Queue Delay
0.05
CPU(aixdemo) Service

0
Step: 1

Figure 4-9
APPOINTMENTS Components of Response Time
Figure 4-9, above, shows the components of response time for the APPOINTMENTS workload. Note
that CPU service time comprises the vast majority of the time required to process an appointment.
Queuing delay, time spent waiting for a CPU, is responsible for the rest. I/O resources made only a
negligible contribution to the total amount of time needed to process each user call.
The ASMAIN workload, shown in Figure 4-10, is more balanced. There is no single resource that is the
obvious winner in the contest for the capacity planners attention (however, make note of the queue
delay for hdisk2.)

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-8
How to Do Capacity Planning

All Other AR Queues Queue Delay


7
All Other AR Queues Service
20-58-00(aixdemo) Queue Delay
6
20-58-00(aixdemo) Service
hdisk0(aixdemo) Queue Delay
5
Seconds

hdisk0(aixdemo) Service
hdisk1(aixdemo) Queue Delay
4
hdisk1(aixdemo) Service

3
hdisk2(aixdemo) Queue Delay
hdisk2(aixdemo) Service
2
CPU(aixdemo) Queue Delay
CPU(aixdemo) Service
1

0
Step: 1

Figure 4-10
ASMAIN Components of Response Time

Plan for the Future


How do you make sure that a year from now your systems wont be overwhelmed and your IT budget
over-extended? Your best weapon is a capacity plan based on forecasted processing requirements.
You need to know the expected amount of incoming work, by workload, then you can calculate the
optimal system configuration for satisfying service levels.
Follow these steps:
a. First, you need to forecast what your organization will require of your IT systems in the future.
b. Once you know what to expect in terms of incoming work, you can use TeamQuest Model to
determine the optimal system configuration for meeting service levels on into the future.

Determine Future Processing Requirements


Systems may be satisfying service levels now, but will they be able to do that while at the same time
meeting future organizational needs?
Besides service level requirements, the other key input into the capacity planning process is a
forecast or plan for the organizations future. Capacity planning is really just a process for
determining the optimal way to satisfy business requirements such as forecasted increases in the
amount of work to be done, while at the same time meeting service level requirements.
Future processing requirements can come from a variety of sources. Input from management may
include:
Expected growth in the business
Requirements for implementing new applications
Planned acquisitions or divestitures
IT budget limitations
Requests for consolidation of IT resources

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-9
How to Do Capacity Planning

Additionally, future processing requirements may be identified from trends in historical


measurements of incoming work such as orders or transactions.

Plan Future System Configuration


After system capacity requirements for the future are identified, a capacity plan should be developed
to prepare for it. The first step in doing this is to create a model of the current configuration. From this
starting point, the model can be modified to reflect the future capacity requirements. If the results of
the model indicate that the current configuration does not provide sufficient capacity for the future
requirements, then the model can be used to evaluate configuration alternatives to find the optimal
way to provide sufficient capacity.

Figure 4-11
Setting Up TeamQuest Model
Our base model in Figure 4-11 was representative of 300 users generating appointment/calendar
requests on our system. From there we have added steps in increments of 30 users until the
incoming work has increased by one-half again as much. This ramp-up in work has been added only
for the APPOINTMENTS workload, because no such increase was indicated for other workloads when
we did our analysis of future processing requirements in the previous section.
TeamQuest Model will predict performance of the current system configuration for each of the steps
we have set up. Figure 4-12 is a chart generated using TeamQuest Model showing the predicted
response time for each workload.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-10
How to Do Capacity Planning

20
18
16
14 A PPO INTMENTS
12 A S MA IN
Seconds

10
B O URNE
8
G PMS
6
4 O THER
2 S TA FF
0 390 S Y S TEM
SYSTEM

STAFF

OTHER

GPMS

BOURNE

300

APPOINTMENTS
ASMAIN

Figure 4-12
Predicted Response Time
As we can see, the response time starts to elongate after 390 users.
Figure 4-13 is a chart showing predicted response time using a stack bar chart that also shows the
components of response time. Notice the substantial increase in CPU wait time after the number of
uses reaches 360. It seems that the performance bottleneck is CPU resource.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-11
How to Do Capacity Planning

1.4

1.2
All Other AR Queues Queue Delay
All Other AR Queues Service
1
20-58-00(aixdemo) Queue Delay
20-58-00(aixdemo) Service
0.8 hdisk0(aixdemo) Queue Delay
Seconds

hdisk0(aixdemo) Service
hdisk1(aixdemo) Queue Delay
0.6
hdisk1(aixdemo) Service
hdisk2(aixdemo) Queue Delay
0.4 hdisk2(aixdemo) Service
CPU(aixdemo) Queue Delay
CPU(aixdemo) Service
0.2

0
300 330 360 390 420 450

Figure 4-13
Predicted Response Time
Figure 4-14 shows the same stack bar chart, but this time TeamQuest Model was told to predict
performance if the system involved was changed to a p670 1100Mhz 4-CPU system. Clearly, the
newer, faster architecture not only allows us substantial growth, but reduces our overall response
time to a more realistic level and still allows us the headroom to experience additional growth if
needed.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-12
How to Do Capacity Planning

1.4

1.2
All Other AR Queues Queue Delay
All Other AR Queues Service
1
20-58-00(aixdemo) Queue Delay
20-58-00(aixdemo) Service
0.8 hdisk0(aixdemo) Queue Delay
Seconds

hdisk0(aixdemo) Service
hdisk1(aixdemo) Queue Delay
0.6
hdisk1(aixdemo) Service
hdisk2(aixdemo) Queue Delay
0.4 hdisk2(aixdemo) Service
CPU(aixdemo) Queue Delay
CPU(aixdemo) Service
0.2

0
300 320 360 390 420 450

Figure 4-14
Predicted Response Time
After Upgrade

Capacity Planning Process


In summary, we have shown these basic steps toward developing a capacity plan:
1. Determine service level requirements
a. Define workloads
b. Determine the unit of work
c. Identify service levels for each workload
2. Analyze current system capacity
a. Measure service levels and compare to objectives
b. Measure overall resource usage
c. Measure resource usage by workload
d. Identify components of response time
3. Plan for the future
a. Determine future processing requirements
b. Plan future system configuration
By following these steps, you can help to ensure that your organization will be prepared for the
future, ensuring that service level requirements will be met using an optimal configuration. You will
have the information necessary to purchase only what you need, avoiding over-provisioning while at
the same time assuring adequate service.
The next chapter provides example applications of capacity planning for todays data center issues.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 4-13
Using Capacity Planning for Server Consolidation

4
5 Using Capacity Planning for Server Consolidation
With the right tools, you can locate underutilized capacity and move applications and subsystems to
take advantage of that capacity, oftentimes delaying the purchase of additional servers. This is just
one of many IT optimization strategies that falls under the definition of server consolidation.
This chapter explains server consolidation and provides two real-world examples where performance
management and capacity planning tools were used in consolidation projects. These projects saved
the organizations involved thousands of dollars in one case, and millions in another, mainly by
helping them to use resources they already had on hand.

Server Consolidation Defined


Consolidation by definition is the act that brings together separate parts into a
single whole. Typically, server consolidation refers to moving work or applications
from multiple servers to a single server. Server consolidation is just one aspect of
capacity planning.

Why, Oh Why?
There are many reasons to consolidate servers, but the underlying reason is
ultimately, To save money. The goal is to lower the Total Cost of Ownership (TCO)
for information technology and increase Return on Investment (ROI). By increasing
efficiency, IT departments can do more with less and make optimal use of what they already have.
In a typical server consolidation effort, you can increase efficiency by:
Reducing complexity
Reducing staff requirements
Increasing manageability
Reducing costs for physical space, hardware systems and software
Increasing stability/availability
This chapter will provide examples where server consolidation was used to increase efficiency by:
Delaying new purchases
Extending the life of existing servers
Optimally utilizing existing servers
When mainframes ruled the earth, everything ran on one or a few big boxes. This simplified
management. Now even where there are mainframes in an organization there are frequently a great
many distributed servers as well. Running many distributed boxes offers a great deal of flexibility,
and when business departments own those boxes rather than IT departments, they get to have
their own playground. Applications can be rolled out without going through a bureaucratic quagmire,
and expensive mainframe expertise is not required. This flexibility allows for rapid deployment, but
the cost is a management nightmare.
Decentralized management and control means a hodge-podge of servers and techniques used to
manage them. Management is primarily reactive, oftentimes with data center managers being
consulted only after disaster strikes, when problems are far out of hand. Because of this, it is not
unusual for the IT department to provide the impetus to consolidate servers.
Today there is a general trend toward server consolidation. The idea is to introduce management
where chaos previously ruled, reducing complexity and optimizing existing IT resources.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 5-1
Using Capacity Planning for Server Consolidation

Finding the Answers Can Be Difficult


Server consolidation sounds simple enough, but it is not. How much capacity is
available on each server? What resources are underutilized? Is the CPU? What
about memory?
What happens if work is moved in from another server? Will there be enough
resources for both applications? Will I/O be a bottleneck after the move? What
will response time be like?
Can more than one application be moved to that server? What are the overall effects of increasing the
work on that server? Is there any space for future growth? How long will it be before business
workloads increase to the point that the CPU or other resources are over-utilized?
These are all true concerns and legitimate questions when talking about server consolidation, and
without the proper tools, finding answers can be difficult.

Predict the Future


The tool of choice for many server consolidation experts is TeamQuest Model. TeamQuest Model uses
analytic modeling based on queuing theory to rapidly evaluate various hypothetical system
configurations. It also predicts performance based on projected business growth. Most importantly
for this chapter, it can predict how applications will perform after servers are consolidated.
With TeamQuest Model you can:
1. Measure work being done on each individual server.
2. Model effects of consolidating work from multiple servers onto a single
server.
3. Quickly identify any performance bottlenecks that are likely to be
caused by the additional work being done on the server.
4. Predict effects of future growth in your business using what if
scenarios. What if the work grows by X percent over X amount of
time?
5. Predict effect on performance of changing hardware configurations.
What if another CPU is added? What if disk drives are added, or another
controller? What if the make or model of server is changed?
The most important benefit of using TeamQuest Model is that you can see the effects of changes
before any work or data are actually moved. You can then be more confident regarding the effects of
the server consolidation before you make any purchases and before data or applications are moved.

Case Studies
Below are two examples of how server consolidation was used in different situations, saving both
companies money. The examples are real, but we have changed the names to protect our customers
interests. Company A used capacity planning and server consolidation to accomplish big cost
savings in their disaster recovery/management plans. Company B discovered they could make do
with existing capacity, and canceled an order for a large server.

Company A Finding Underutilized Resources for Use as Disaster Backup


A disaster management consultant was hired for Company A. Company A had three locations, one in
California, the second in New York, and the third in Kansas City. All three sites used a large SAN
environment so that each site had access to the same data. There were a total of some 900 servers.
The consultant was to come up with a disaster management plan that would be able to recover 200 of
the most critical servers in case of a disaster. 100 of these mission critical servers were located in the

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 5-2
Using Capacity Planning for Server Consolidation

California office, 50 in New York and 50 in Kansas City. In case of a disaster, the mission critical
servers had to be recovered and operational within a 24-hour time period.
This consultant started looking at the workloads and utilizations of each server, and found that there
was one business application per server. Upon interviewing system administrators (SA), the
consultant learned that Company A had at one point tried running multiple applications per server.
This resulted in unacceptably poor performance and the people responsible for the applications
pointing fingers at each other.
The company thought they could resolve the problem quickly by putting each individual business
application on its own server. This turned out to be a very difficult, time consuming and expensive
task. To avoid having to separate applications again in the future, a company-wide policy was
instituted requiring each business application be hosted on its own dedicated server.
Hindsight being 20-20, it would have been prudent to have TeamQuest View installed on
the servers, helping to analyze performance bottlenecks. It might have been possible to
eliminate problems with just a few carefully chosen changes or upgrades. Even better,
TeamQuest Model could have been used to explore alternative configurations to meet
capacity requirements.
Instead, the company chose what appeared to be the most
expedient solution to their performance finger-pointing Underutilization Means
problem; they hosted just one application per server. This is Opportunity
not an uncommon solution, but it usually leads to inefficient
use of IT resources.
Using TeamQuest View the disaster management consultant
analyzed a few of the larger critical servers and found they were
less than 3% to 6% utilized at any given time of the day or
night. There were vast amounts of underutilized capacity that
could potentially be used as backup in the event of a disaster.
The consultant then used TeamQuest Model to evaluate a
potential 10-to-1 server consolidation. He modeled 10 of the
mission critical servers from California, and moved the work to a mission critical server in Kansas
City. The model showed that even with the consolidation of these servers, the server in Kansas City
would be less than 71% utilized.
The consultant then did an analysis of the historical
Consolidation Saves the Day
growth of the workloads on servers that resided in
California for the preceding year. He used what-if
analysis capabilities of TeamQuest Model to predict the
utilization for the next 12 months based on the growth
patterns of the previous year. This revealed that after the
consolidation and the projected growth for the next 12
months, the server would still be less than 80% utilized.
The consultant convinced management to let him use the
server in Kansas City as a backup site for the 10 mission
critical servers. He mirrored the individual server from
California to the one server in Kansas City, and found
performance to be within a small percentage of what had
been projected with TeamQuest Model.
The consultant is currently modeling the entire enterprise, including all three sites in the analysis. It
is quite likely that no new computer resources will be required to provide disaster back-up, resulting
in a very significant savings, a very pleasant surprise for upper management.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 5-3
Using Capacity Planning for Server Consolidation

Company B Avoiding Unnecessary Expenditures


Company B was a small company with a System Administrator (SA) who was in charge of six servers
within his department. One of those servers was leased, and the lease was expiring at the end of the
month. The hardware vendor convinced Company B to order the newest, latest, greatest (and most
expensive) server on the market. If the new server arrived as scheduled, service levels could be
maintained, but the SA was concerned. What would he do if the new server did not arrive before the
old leased equipment had to be returned?
The work being done on the leased server included an instance of Oracle and two custom
applications. Could that processing be temporarily moved to any of the other five servers?
The SA used TeamQuest Model to analyze all six servers. He found that no single server would be able
to take the entire load of the leased server, but he did find a way to distribute the load over two
existing machines. He could move the Oracle instance to a machine that had another small instance
of Oracle running on it. If he added two more CPUs to that server, there would be sufficient capacity
for the next 12 months with the current growth patterns. He found
Eliminated Need for
another server that had enough capacity to host the custom
Expensive Replacement
applications without any modifications for the next 12 months.
With sufficient capacity to cover growth over the next 12 months,
there was no need for the leased server or the replacement server
that had been ordered. The SA quickly took these findings to his
management, and the changes were successfully implemented over
the next weekend. The order for the new server was cancelled, and
the SA received a healthy bonus for saving the company a
substantial amount of money.

Conclusion
These are just two examples of how using a good capacity planning tool to perform server
consolidation can save money. Company B saved thousands of dollars, and Company A literally saved
millions, dramatically demonstrating that performance management and capacity planning is best
done with the proper tools. To do otherwise creates risk instead of minimizing it.
To make optimal use of your organizations IT resources, you need to know exactly the amount of
capacity you have available and how that capacity can be utilized to its fullest.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 5-4
Capacity Planning and Service Level Management

6 Capacity Planning and Service Level Management


In this chapter we discuss service level management and its relationship to capacity planning.
We hear a lot about service level management today, but what do we really mean when by it? In the
boxes below we show side-by-side definitions of service level management and capacity planning.
The service level definition is from Sturm, et al, Foundations of Service Level Management.5

Service Level Management (SLM) Capacity planning


is the disciplined, proactive is the process by which an IT
methodology and procedures used department determines the amount of
to ensure that adequate levels of server hardware resources required to
service are delivered to all IT users in provide the desired levels of service
accordance with business priorities for a given workload mix for the least
and at an acceptable cost.5 cost.

You can clearly see that these two critical IT processes really share common goalsprovide adequate
service for the least cost. So how do you really do this? You must employ an integrated approach to
service level management and capacity planning, as this section describes. Stated slightly
differently:
Capacity planning requires a firm understanding of required service levels
Service level management is successful when the necessary IT architecture and assets
are available
Operational processes should incorporate both capacity planning and SLM to maximize
success
We will now examine two key aspects of service level management, setting service level agreements
and maintaining service levels over time as things change.

Service Level Agreements (SLAs)


Many of the IT organizations we talk to do not have any type of SLA mechanism in place and this is a
concern. How can you adequately provision a service if you do not have goals to shoot for? And how
can you plan upgrades if you do not know what the business expects for this service over time? We
recommend at least recording the following information for each service:
The business processes supported by the service
The priority to the business of these processes
The expected demand for this service and the seasonality, if any, of that demand
The expected growth in demand for this service over the next three years
The worst response time or throughput that is acceptable for this service
You do not have to make this process formal with signed agreements, but it is a very good idea for
both IT and the affected business units to have a copy of the above information to resolve conflicts
should they arise.
With this minimal information, or a more complete and formal SLA if you have it, you are now ready to
use capacity planning to ensure performance that is adequate at a minimum cost.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 6-1
Capacity Planning and Service Level Management

The key SLA questions to answer are:


What hardware configuration is required for the performance and workload level?
Can the service be provisioned on existing systems and meet the requirements?
What hardware upgrade plan must be in place for anticipated business growth?
There are many ways to determine how much hardware capacity you need to deliver a service level.
You can perform load testing, but this assumes you can afford exact duplicates in the test lab of the
production floor, and load testing fails in a consolidated environment anyway. As an alternative, you
can use capacity modeling to determine how much hardware you will need.
About the only way to find out if a service can be provisioned on an existing system with the specified
service level is to use capacity modeling. And what about after it is up and running? How do you
maintain the service level as workloads grow? When do you need to upgrade, if at all? Capacity
modeling is often the best way to know for sure. We will present best practices for doing this in a
moment.

Proactive Management of Resources to Maintain Service Levels


Essentially you must answer the following questions:
Do predicted service levels match real service levels?
Can we consolidate applications or servers to save money and still provide the required
service?
If growth plans change, what does this mean to our upgrade plan?
How do you manage resources to maintain service levels and minimize costs? First you need to
measure service levels and compare to specified and predicted. This requires a performance
management system with workload measurement. Are you providing adequate service levels but at
the expense of over-provisioning? You can find this out by (a) looking at historical resource utilization
for the service and the system and looking for headroom, or (b) by building a model and stressing it to
see where it breaks.
If you have a service over-provisioned for now, is it also for the foreseeable future? If yes, you should
definitely consolidate or provision a new service here. It is always a good idea to check with business
units or other client organizations periodically to verify growth plans in the SLA. What if they recently
changed? The best thing to do is get out your model of this service and plug in the revised growth
projections and see how your upgrade plan looks now.

Service Level Management Scenarios


The following scenarios illustrate capacity planning applications for service level management.

Provisioning a new service to meet an SLA


1. Perform load planning and testing.
2. Baseline the test runs.
3. Build a model of the test environment.
4. Change the model parameters to the production environment.
5. Solve the model for the projected workloads.
6. Compare predicted response time/throughput to service level
parameters.
7. Make necessary changes to planned provisioning.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 6-2
Capacity Planning and Service Level Management

Adding an SLA to an existing service


1. Build a model of the production environment.
2. Solve the model for the projected workloads.
3. Compare predicted response time/throughput to service level
parameters.
4. Make necessary changes to planned provisioning.

Developing an upgrade plan under an SLA


1. Gather projected growth estimates from business units.
2. Build a model of the existing service based on current
configuration and workload.
3. Solve the model for one, two and three year growth projections.
4. Compare predicted response times and throughputs to the SLA.
5. Use modeling what-if scenarios to find least cost, just-in-time
upgrade path.

To summarize, the goals of service level management and capacity planning are tightly interwoven.
Both strive to provide adequate levels of service at an acceptable cost. And both are much more
successful when implemented together as part of a disciplined and unified process.
The final chapter introduces the TeamQuest solution for capacity planning.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 6-3
Product Focus: TeamQuest Model

Most of this chapter is


7 Product Focus: TeamQuest Model excerpted from a white paper1
by Enterprise Management
One of the leading vendors in the capacity planning space is Associates (EMA), an analyst
TeamQuest Corp., which makes tools both for performance analysis organization specializing in the
and capacity planning. Founded in 1991 as a spin-off of Unisys management software and
Corp., TeamQuests software has become one of the industrys services market.
most established capacity planning solutions, and is used by such
major corporations as 20th Century Fox, AT&T, Nike and 3M Corp. While TeamQuest offers a variety of
products designed to help enterprises optimize performance, its primary capacity planning tool is
TeamQuest Model (Model), which offers full planning and modeling capabilities. The full TeamQuest
product line is depicted in Figure 7-1.
One of the first criterion that distinguishes
Model from its competitors is its support
for multiple hardware platforms. While
there are many vendor- or platform-specific
tools for capacity planning, the number of
multi-platform solutions is much smaller.
Many vendors have found it difficult to
instrument tools that can collect and
extrapolate data coming from platforms as
different as Microsoft Windows and Unix,
but that capability is essential for most
large enterprises, which usually support a
variety of different servers.
Another key differentiator is Models
multiple system modeling capability that
allows management of the capacity of a
multi-tiered application from a holistic Figure 7-1
point of view. Models ability to specify and TeamQuest Performance Framework
play what-if with the horizontal scaling of a tier is unique. The impact of changing or adding CPUs may
be easily modeled, showing the impact of adding servers to the tier. Additionally, Model supports
split workloads, showing application flows from one server to the next, recognizing that todays
applications do not move in a single path through the tiers.
From a functional perspective, Models chief benefit is its ability to provide predictive analysis
required for long-term capacity planning. It is a PC-based application that automatically retrieves data
from Windows and Unix servers and then builds models of systems and applications that enable the
user to analyze what if scenarios and predict the effects of proposed changes on systems and
applications before they are made. This capability is essential to IT organizations not only for
planning system purchase and configuration, but to predict the impact of IT changes and escalating
business workloads on the performance of business services.
Two example TeamQuest Model screens are shown below; the first depicts predicted performance as
the number of users increases over time, and the second depicts the projected performance if a
proposed system upgrade is performed. Clearly the proposed upgrade will adequately prepare the
system for future workloads. In fact, it may be overkill. TeamQuest Model would allow the easy
exploration of additional, perhaps less expensive, alternatives.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 7-1
Product Focus: TeamQuest Model

1.4

1.2
All Other AR Queues Queue Delay
All Other AR Queues Service
1
20-58-00(aixdemo) Queue Delay
20-58-00(aixdemo) Service
0.8 hdisk0(aixdemo) Queue Delay
Seconds

hdisk0(aixdemo) Service
hdisk1(aixdemo) Queue Delay
0.6
hdisk1(aixdemo) Service
hdisk2(aixdemo) Queue Delay
0.4 hdisk2(aixdemo) Service
CPU(aixdemo) Queue Delay
CPU(aixdemo) Service
0.2

0
300 330 360 390 420 450

Figure 7-2
TeamQuest Model Shows Predicted Response Time

1.4

1.2
All Other AR Queues Queue Delay
All Other AR Queues Service
1
20-58-00(aixdemo) Queue Delay
20-58-00(aixdemo) Service
0.8 hdisk0(aixdemo) Queue Delay
Seconds

hdisk0(aixdemo) Service
hdisk1(aixdemo) Queue Delay
0.6
hdisk1(aixdemo) Service
hdisk2(aixdemo) Queue Delay
0.4 hdisk2(aixdemo) Service
CPU(aixdemo) Queue Delay
CPU(aixdemo) Service
0.2

0
300 320 360 390 420 450

Figure 7-3
Predicted Response Time After Upgrade
Models most obvious benefits are in planning the purchase and deployment of new hardware. The
software offers an array of pre-built capacity profiles, enabling IT organizations to use real-world data
to perform what if queries on the infrastructure impact or performance of planned hardware
installations. Using Model, enterprises can get a real-life picture of how a new server will affect
capacity and performance, even before that server is purchased or installed.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 7-2
Product Focus: TeamQuest Model

Model also enables enterprises to test the effects of various server consolidation scenarios, making it
easier to find the most efficient configuration of hardware and expose potential consolidation
conflicts before they occur. Model offers an advantage over other capacity planning products in this
area, because it can integrate, analyze, and compare data from multiple servers to aid in server
consolidation efforts.
Aside from server acquisition and consolidation, Model offers a number of other potential benefits for
IT planning. For example, Model can be used to measure the effectiveness of proposed workload
balancing strategies, such as examining the impact of increased activity on a particular server or of
moving some types of processing to a different part of the day. In addition, Model has been used to
model capacity for newly introduced applications, or even the impact of infrastructure consolidation
that may take place as a result of mergers or acquisitions.
While Model can make a significant difference in efforts to evaluate and measure the impact of future
IT initiatives, it can be equally effective in facilitating day-to-day activities, such as the ongoing
monitoring and modeling of server capacity. For example, Model can be used to predict when a
particular server will run out of capacity, given the projected business workload growth rate. This data
can be used to help IT organizations precisely plan their strategies for system upgrade, as well as
providing data on exactly what the upgrade requirements might be.
Model can also be used to support the development, analysis and enforcement of established service
levels, whether they are delivered by internal IT organizations or external service providers. Model
can help organizations model the time, costs, and resources required to maintain different levels of
performance, enabling service providers and their end-user customers to reach mutually-satisfactory
service level agreements that are based on realistic expectations and scientific analysis, rather than
wishful thinking. With Model, IT organizations can benchmark a variety of service levels to determine
their costs and then set realistic customer expectations for maintaining them.
In fact, the sophistication of Models ability to predict performance is one of the elements that sets it
apart from other capacity planning tools on the market. IT organizations can use Model to predict a
variety of performance metrics, including response time, throughput, and queue length as well as
resource utilization. Model is the only capacity planning tool on the market that offers a choice
between analytic modeling and discrete event simulation, which means that users can employ Model
to gain insights on overall performance or on the impact of specific IT events in the enterprise, such
as infrastructure upgrades.

Case Studies
Clearly, Model offers a range of advantages on paper, but how does it work in practice? To answer this
question, EMA interviewed two Model users: the U.S. Patent Office and a US Government
Subcontractor that maintains records and data about U.S. military personnel and their dependents.
The following is a synopsis of those interviews.

United States Patent and Trademark Office


Three years ago, the U.S. Patent Office (www.uspto.gov) had a major performance
problem in a critical application. A key search system was suffering major performance
problems, yet IT technicians needed a method to model and design specific scenarios
to optimize the proposed solution. The technicians contacted Hewlett-Packard, maker
of the Patent Offices HP MeasureWare performance management software, and an HP
contact suggested that the TeamQuest Model performance-modeling tool could save
valuable time in optimizing the proposed solution.
By employing analytic modeling and validating its search system application using Model, the Patent
Office found the I/O bottleneck immediately. After locating the problem, technicians used Model to
help test a variety of channel striping sizes and find the appropriate Fibre Channel configuration.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 7-3
Product Focus: TeamQuest Model

Model enabled the Patent Office to aggregate statistics for each storage drive across the Fibre
Channel, making it possible to identify bottlenecks and then determine the best method for dividing
I/O across the channel to eliminate those bottlenecks.
Today, the Patent Office is using Model regularly to import, analyze, and enhance data from
MeasureWare. Model captures workload data from existing systems, and applies that data to the
Patent Offices new system model, projecting workload growth over the lifespan of the server. This
approach enables the Patent Office to avoid over-provisioning over the life of the system, because the
IT organization can accurately predict the need for future capacity.
And what of the Patent Offices search system? In early 2000, the Patent Office was awarded a
Department of Commerce Silver Medal Award for the Patent Search Performance Teams improvement
of its multimillion-dollar search engine project. IT executives at the Patent Office say that TeamQuest
Model has saved them millions of dollars in over-provisioning costs, and helped make the search
system project a roaring success.

US Government Contractor
A contractor to the US Government that provides services to many government agencies (and whom
requested anonymity) has a database engine that contains information on hundreds of thousands of
military personnel and their dependents. Just a few years old, the system has become enormously
popular, growing from two Unix servers to nearly 200 since it began operations.
Such rapid growth was good news for the company, but bad news for its IT organization, which was
scrambling to keep up with the skyrocketing demand for server capacity. To have any chance to catch
up, the company needed automation and planning technology to facilitate growth. The company also
needed performance reporting and tracking capabilities to ensure good service and debug
performance problems.
In 2001, the companys sole server provider, Sun Microsystems, recommended Model to the
companys IT department. TeamQuest technicians came to the company and installed Model on six
Web servers to collect data and create some sample models for IT staff. Once company officials gave
the go-ahead, the full Model package was installed and operational within a few days.
Since its installation, TeamQuest Model has enabled the company to save millions of dollars in over-
provisioning costs, according to officials associated with the Model deployment. In one instance,
Model modeling helped the company find a configuration that spread the workload across existing
Sun servers, reducing the need for new servers from 30 to 20 and saving the organization millions of
dollars.
Today, Model is helping the company maintain consistent levels of service across its government
customer base, which is a crucial element to its business. Model enables the company to model
server performance and establish service level agreements with realistic thresholds, making it easier
for the organization to meet or beat customer expectations, even as its server infrastructure
continues to evolve.

EMA Perspective
In a difficult economy, every IT dollar matters. The days of solving performance problems by
indiscriminately buying more servers are over, and there is an excellent window of opportunity for
deploying management technology that can optimize server utilization and performance. IT
organizations are looking for ways to make better use of existing server capacity while reducing the
need to buy more processor power.
At the same time, it is important to note that IT budgets are tight, and that most IT organizations are
not prepared to purchase new management or planning software unless they are guaranteed a clear

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 7-4
Product Focus: TeamQuest Model

and fast return on their investment (ROI). Todays management tools must be affordable, easy to
deploy, and show strong results in a very short period of time.
Capacity planning tools meet all of these criteria. Through capacity planning, IT organizations can
realize major cost savings immediately by reducing or eliminating the need to over-provision their
server environments. Capacity planning and modeling technology also can guide IT through the
server consolidation process, and can help set realistic goals for service levels that can be
maintained over time.
EMA has established that TeamQuest is a superior partner for capacity planning. TeamQuest has
been in the performance assurance business since 1991, and has a legion of satisfied customers. The
company has an excellent reputation in the industry, and all of the customers interviewed for this
report offered high praise for the TeamQuest organization, from pre-sales to post-sales. The Model
product has several features that separate it from its competitors, including multi-platform support
and the ability to do broad or event-driven capacity analysis and modeling.
EMA recommends TeamQuest Model to IT executives that are concerned with doing more with less
in todays difficult economy. Model clearly demonstrates the ability to recognize major cost savings
and increased efficiencies in a short amount of time, generating excellent value and a strong ROI for
the IT investment dollar.

Like what you see? Subcribe.


Capacity Planning: Discipline for Data Center Decisions Page 7-5
Bibliography

8 Bibliography
1. The Value Proposition for Capacity Planning, Enterprise Management Associates, 2003.
2. The Office of Government Commerce web site, http://www.ogc.gov.uk.
3. How to Do Capacity Planning, TeamQuest Corporation, TQ-WP23 Rev. B.
4. Consolidating Applications onto Underutilized Servers, TeamQuest Corporation, TQWP20
Rev. A.
5. Sturm, Rick & Morris, Wayne, Foundations of Service Level Management, SAMS, 2000.
Like what you see? Subscribe.

TeamQuest Corporation

Americas Legal Notices


One TeamQuest Way TeamQuest and the TeamQuest logo are registered trademarks in the
Clear Lake, Iowa 50428 US, EU, and elsewhere. All other trademarks and service marks are the
USA property of their respective owners. No use of a third-party mark is to
+1 641 357-2700 be construed to mean such marks owner endorses TeamQuest
products or services.
+1 800 551-8326
info@teamquest.com The names, places and/or events used in this publication are purely
fictitious and are not intended to correspond to any real individual,
group, company or event. Any similarity or likeness to any real
Europe, Middle East and Africa individual, company or event is purely coincidental and unintentional.
Box 1125
NO WARRANTIES OF ANY NATURE ARE EXTENDED BY THE DOCUMENT.
405 23 Gteborg Any product and related material disclosed herein are only furnished
Sweden pursuant and subject to the terms and conditions of a license
+46 (0)31 80 95 00 agreement. The only warranties made, remedies given, and liability
accepted by TeamQuest, if any, with respect to the products described
United Kingdom in this document are set forth in such license agreement. TeamQuest
38 The Old Woodyard cannot accept any financial or other responsibility that may be the
Hagley Hall result of your use of the information in this document or software
material, including direct, indirect, special, or consequential
Hagley damages.
Worcestershire DY9 9LQ
You should be very careful to ensure that the use of this information
+44 (0)1562 881889 and/or software material complies with the laws, rules, and
emea@teamquest.com regulations of the jurisdictions with respect to which it is used.
The information contained herein is subject to change without notice.
Asia Pacific Revisions may be issued to advise of such changes and/or additions.
Level 6, 170 Queen Street
U.S. Government Rights. All documents, product and related material
Melbourne, VIC 3000 provided to the U.S. Government are provided and delivered subject to
Australia the commercial license rights and restrictions described in the
+61 3 9641 2288 governing license agreement. All rights not expressly granted therein
asiapacific@teamquest.com are reserved.

You might also like