You are on page 1of 245

tm

The Definitive Guide To

Converged
Network
Management
Ken Camp
Introduction

Introduction to Realtimepublishers
by Don Jones, Series Editor

For several years, now, Realtime has produced dozens and dozens of high-quality books that just
happen to be delivered in electronic format—at no cost to you, the reader. We’ve made this
unique publishing model work through the generous support and cooperation of our sponsors,
who agree to bear each book’s production expenses for the benefit of our readers.
Although we’ve always offered our publications to you for free, don’t think for a moment that
quality is anything less than our top priority. My job is to make sure that our books are as good
as—and in most cases better than—any printed book that would cost you $40 or more. Our
electronic publishing model offers several advantages over printed books: You receive chapters
literally as fast as our authors produce them (hence the “realtime” aspect of our model), and we
can update chapters to reflect the latest changes in technology.
I want to point out that our books are by no means paid advertisements or white papers. We’re an
independent publishing company, and an important aspect of my job is to make sure that our
authors are free to voice their expertise and opinions without reservation or restriction. We
maintain complete editorial control of our publications, and I’m proud that we’ve produced so
many quality books over the past years.
I want to extend an invitation to visit us at http://nexus.realtimepublishers.com, especially if
you’ve received this publication from a friend or colleague. We have a wide variety of additional
books on a range of topics, and you’re sure to find something that’s of interest to you—and it
won’t cost you a thing. We hope you’ll continue to come to Realtime for your educational needs
far into the future.
Until then, enjoy.
Don Jones

i
Table of Contents

Introduction to Realtimepublishers.................................................................................................. i
Chapter 1: Introduction to Unifying Network Management and Converged IP Communications .1
Convergence Covers Many Areas ...................................................................................................2
Infrastructure Convergence in the Network—the Wiring and Circuit Convergence ..........2
Service Convergence over a Common IP Networking Environment ..................................3
Device Convergence—the Convergence of Desktops, Laptops, Tablets, and PDAs ..........4
Application Convergence—Integration of Enterprise Business Applications with
Network Services .................................................................................................................5
Fixed Mobile Convergence..................................................................................................6
VoIP as a Converged Service ..............................................................................................7
Voice vs. Data..........................................................................................................8
The Cost of Doing Business Drives VoIP ...............................................................8
Technical Definition vs. Market Definition...........................................................10
VoIP or IPT? ..................................................................................................................................12
A Telephone Call Simplified .............................................................................................13
Converting the Analog Signal to Digital ...........................................................................13
When Does VoIP Become IPT? ........................................................................................14
Intended for PC-to-PC Voice Messaging ..........................................................................15
Never Intended to Replace Global Telephony.......................................................15
“Free Voice” Was Compelling and Simple ...........................................................15
Pushes the Edge Over Convergence ..................................................................................15
Multiple Play..........................................................................................................16
New Devices ..........................................................................................................16
VoIP Can Include Advanced Applications ........................................................................17
Converged Communications Leads to Converged Applications ...................................................18
FMC ...................................................................................................................................18
Summary ........................................................................................................................................19
Chapter 2: Key Considerations in Effective Voice and Data Integration for a Changing IT/IP
Landscape ......................................................................................................................................20
Quantifiable Business Processes....................................................................................................20
Web-Centric Businesses ....................................................................................................20
Call Centers........................................................................................................................22
IVR Systems ......................................................................................................................24
CTI .....................................................................................................................................24

ii
Table of Contents

CTI History ............................................................................................................24


Application Integration with CRM and Enterprise Resource Planning Systems ..............27
Customer Relations and CRM ...............................................................................28
Delivering Call Quality with VoIP ................................................................................................30
Traditional Voice Characteristics—the PSTN...................................................................30
IP Traffic Characteristics ...................................................................................................31
Design Considerations and Class of Service .................................................................................33
QoS ................................................................................................................................................36
QoS Approaches in IP Networks .......................................................................................37
QoS—the Signaled Approach................................................................................40
QoS—the Provisioned Approach...........................................................................43
QoS—the Bypass/Shim Approach.........................................................................45
Summary ........................................................................................................................................49
Chapter 3: Business Drivers and Justification ...............................................................................50
Vertical Market Business Drivers for Change ...............................................................................50
Business Sales....................................................................................................................56
Web-Enabled Business ..........................................................................................59
Product Sales..........................................................................................................61
Service Sales ..........................................................................................................61
Financial Services ..............................................................................................................63
Health Care ........................................................................................................................64
Manufacturing....................................................................................................................64
Financial Cost Reduction Drivers for Change ...............................................................................65
Local Telephone Expenses ................................................................................................65
Long-Distance Expenses....................................................................................................65
Inside the Enterprise ..............................................................................................66
Outside the Enterprise............................................................................................66
Support Costs .....................................................................................................................66
Help Desk Support.................................................................................................67
Adds, Moves, and Changes....................................................................................68
Remote and Mobile Workers .................................................................................68
Strategic Drivers for Change .........................................................................................................69
New Applications...............................................................................................................69

iii
Table of Contents

Obsolescence of Legacy Systems ......................................................................................70


Stop Investing in Legacy Technologies.................................................................70
Manufacturer Discontinuation of Product and/or Support.....................................70
Rising Support and Maintenance Costs .................................................................71
Inability of Legacy Solutions to Support Business Needs.....................................71
Summary ........................................................................................................................................71
Chapter 4: Productivity Advantages of Unified Communications ................................................73
How Service Convergence Drives Productivity and Enables New Business Operational Processes74
Management and Utilization of Property...........................................................................76
Cost Reduction and Management ......................................................................................76
Enterprise Agility...............................................................................................................76
Staff Productivity ...............................................................................................................77
The Non-PC Workstation Environment.................................................................77
Evaluating the Call Center Strategy.......................................................................78
Business Process Optimization ..............................................................................78
Productivity in Resource Management..................................................................78
Focusing on Employee Productivity..................................................................................79
Unified Messaging .................................................................................................79
Personal Communications Assistants ....................................................................79
IP Video Solutions .................................................................................................80
Web-Centric Businesses ....................................................................................................80
Integrating Voice with Sales, Service, and Support ..............................................81
Call Centers—Localized, Distributed, Offshore................................................................82
IVR Systems ......................................................................................................................82
Computer Telephony Integration.......................................................................................83
Application Integration with CRM and ERP Systems.......................................................83
Customer Relations and CRM ...............................................................................84
Vertical Market Business Drivers for Change ...............................................................................85
Business Sales and the Web, or Net-Enabled Business.....................................................87
Product and Services Sales ....................................................................................87
Financial Services ..............................................................................................................88
Health Care ........................................................................................................................89
Manufacturing....................................................................................................................89

iv
Table of Contents

Summary ........................................................................................................................................89
Chapter 5: Key Steps in VoIP Deployment and Management for the Enterprise .........................92
Network Readiness Assessment ....................................................................................................92
Ensuring Network Readiness for Converged Services ......................................................92
Plan to Succeed—Don’t Fail to Plan .................................................................................93
Analysis..................................................................................................................94
Planning .................................................................................................................94
Testing....................................................................................................................94
Acquire the Right Resources .................................................................................94
Look to the Future—Think Long Term .................................................................95
Understanding Existing Voice Needs ............................................................................................95
One Voice Factor—Compression Methods .......................................................................96
Voice Call Quality—The PSTN vs. VoIP .........................................................................98
Network Design Considerations ..................................................................................................101
The Performance Envelope..........................................................................................................102
Choosing Performance Envelope Characteristics to Measure .....................................................108
Throughput.......................................................................................................................109
Bandwidth ............................................................................................................109
Response Time.....................................................................................................109
CPU Utilization....................................................................................................110
Network Segment Utilization ..............................................................................110
Integrity/Reliability..........................................................................................................110
Packet Loss ..........................................................................................................110
Jitter......................................................................................................................110
Delay/Latency ......................................................................................................110
Cost ..................................................................................................................................111
Availability ......................................................................................................................111
Uptime—99.999% ...............................................................................................111
Security ............................................................................................................................111
Manageability ..................................................................................................................112
Scalability ........................................................................................................................112
Summary ......................................................................................................................................112
Chapter 6: Impact Analysis, Root Cause, and Event Correlation................................................114

v
Table of Contents

SNMP...........................................................................................................................................114
What Is a MIB?................................................................................................................115
The Architecture of SNMP ..............................................................................................116
Using SNMP ....................................................................................................................118
snmpwalk .............................................................................................................118
Factors to Consider with SNMP ......................................................................................119
Autodiscovery ......................................................................................................120
Negative Implications ..........................................................................................120
ICMP............................................................................................................................................121
ICMP Message Format ....................................................................................................123
Reachability Testing ........................................................................................................124
ping ......................................................................................................................124
Traceroute ............................................................................................................125
Syslog, Data Logging, and the Console.......................................................................................127
Integrating Tools for Event Correlation.......................................................................................128
Network Monitoring ........................................................................................................129
VoIP Service Elements to Monitor ......................................................................129
Monitoring Bandwidth and QoS..........................................................................131
Measurements and Metrics for Voice Quality .....................................................132
The Do-It-Yourself Approach..........................................................................................133
The SLA...............................................................................................................133
Pros and Cons of Rolling Out Your Own Management Platform .......................135
Freeware and Open Source vs. Commercial Products.........................................136
Integrated Commercial Solutions ....................................................................................136
Pros and Cons of a Packaged Management Platform ..........................................138
Summary ......................................................................................................................................138
Chapter 7: Effective Service Availability Management and Capacity Planning.........................139
Introducing FCAPS: A Sustainable Model for Balanced Network Management .......................140
Fault Management ...........................................................................................................142
Configuration Management .............................................................................................143
Accounting/Administration/Asset Management..............................................................144
Performance Management ...............................................................................................145
Security Management ......................................................................................................145

vi
Table of Contents

FCAPS Simplified ...........................................................................................................146


Optimization for Service Availability and Capacity....................................................................148
Trending and Capacity Planning......................................................................................149
Bandwidth ............................................................................................................149
Ports/Lines ...........................................................................................................152
Codec Planning ....................................................................................................152
Optimizing the Infrastructure Optimization.....................................................................152
Optimizing Hardware...........................................................................................153
Optimizing Software............................................................................................153
SLA Optimization............................................................................................................154
Delay ....................................................................................................................154
Delay Variation....................................................................................................155
Packet Loss ..........................................................................................................156
Summary ......................................................................................................................................159
Chapter 8: Effective Network Configuration, Network Fault, and Network Performance
Management.................................................................................................................................161
Fault Management .......................................................................................................................161
Network and Fault Management—An Integration Strategy............................................164
Caveats of Implementation Strategies for NMSs.............................................................164
Recognize, Isolate, and Detect Faults ..............................................................................165
The Role of the NMS.......................................................................................................166
SNMP and Fault Management.............................................................................167
Syslog and Fault Management.............................................................................169
Fault Management and ROI.............................................................................................169
Configuration Management .........................................................................................................170
Collecting and Storing Configuration Data .....................................................................171
Configuration and Change Management .........................................................................171
Performance Management ...........................................................................................................173
QoS and Bandwidth Monitoring..........................................................................178
Collecting and Analyzing the Data ..................................................................................180
Monitoring the Health of the Network.................................................................180
Performance and Utilization Trends ....................................................................180
Administration Management for Performance and Planning ..........................................181
Gathering Usage Statistics ...................................................................................181

vii
Table of Contents

Managing Backups and Synchronization for Performance .................................181


Summary ......................................................................................................................................182
Chapter 9: Effective Security Management.................................................................................184
FCAPS and Security Management ..............................................................................................185
Identifying Risks..............................................................................................................186
Inconsequential Risk Classes...............................................................................187
Significant Risk Classes.......................................................................................187
Introducing Incident Response Planning: CSIRT Basics ................................................192
Building Processes for Defense in Depth ....................................................................................193
User Authentication and the Gold Standard ....................................................................195
Authentication......................................................................................................195
Authorization ...................................................................................................................197
Auditing ...........................................................................................................................197
Prepare to Respond ..........................................................................................................198
Policies, Procedures, and Awareness...............................................................................198
Physical Access and Environment Controls ....................................................................199
Perimeter: First Line of Defense Between the Internet and Internal Networks...............199
Network: Internal Network Layer....................................................................................200
Systems: Server and Client OS Hardening Practices.......................................................202
Application: Application Hardening Practices ................................................................203
Data: Protection of Customer and Private Information ...................................................203
Summarizing Defense in Depth ...................................................................................................203
Prevention ....................................................................................................................................204
Protect the Protection.......................................................................................................205
Detection ......................................................................................................................................205
Reaction .......................................................................................................................................205
Proactive and Reactive Strategies....................................................................................206
Reactive Strategies.......................................................................................................................208
Proactive Strategies......................................................................................................................208
Testing the Strategy .....................................................................................................................210
Summary ......................................................................................................................................210
Chapter 10: Asset Reporting, Audit Compliance, and IT Documentation ..................................211
FCAPS and Asset/Administration/Accounting Management......................................................211

viii
Table of Contents

Accounting/Administration/Asset Management..............................................................211
Managing Billing .................................................................................................212
User Accounting ..................................................................................................213
The Asset Management View..............................................................................213
Why Documentation? Why Process?...........................................................................................214
Creating Repeatable Processes ........................................................................................214
Revisiting the Importance of Knowing Your Environment.............................................214
Be Prepared: Protecting Against Future Business Issues ................................................215
Legal and Regulatory Issues to Consider.....................................................................................216
SOX..................................................................................................................................216
Risk Assessment ..................................................................................................217
Control Environment ...........................................................................................217
Control Activities.................................................................................................217
Monitoring ...........................................................................................................218
Information and Communication.........................................................................218
GLBA...............................................................................................................................218
HIPAA .............................................................................................................................219
Methodologies in Best Practices for Management and Oversight...............................................222
ITIL ..................................................................................................................................222
ISO 17799 ........................................................................................................................226
Business Continuity Planning ..............................................................................226
System Access Control ........................................................................................229
Physical and Environmental Security ..................................................................230
Compliance ..........................................................................................................230
Asset Classifications and Control ........................................................................230
Security Policy .....................................................................................................230
Managing and Protecting the Network ........................................................................................231
Inconsequential Risk Classes...........................................................................................231
Significant Risk Classes...................................................................................................231
Reducing Risk for Single Occurrence Losses..................................................................232
Addressing the Risks........................................................................................................233
Risk Management Life Cycle ..........................................................................................233
Summary ......................................................................................................................................234

ix
Copyright Statement

Copyright Statement
© 2007 Realtimepublishers.com, Inc. All rights reserved. This site contains materials that
have been created, developed, or commissioned by, and published with the permission
of, Realtimepublishers.com, Inc. (the “Materials”) and this site and any such Materials are
protected by international copyright and trademark laws.
THE MATERIALS ARE PROVIDED “AS IS” WITHOUT WARRANTY OF ANY KIND,
EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE,
TITLE AND NON-INFRINGEMENT. The Materials are subject to change without notice
and do not represent a commitment on the part of Realtimepublishers.com, Inc or its web
site sponsors. In no event shall Realtimepublishers.com, Inc. or its web site sponsors be
held liable for technical or editorial errors or omissions contained in the Materials,
including without limitation, for any direct, indirect, incidental, special, exemplary or
consequential damages whatsoever resulting from the use of any information contained
in the Materials.
The Materials (including but not limited to the text, images, audio, and/or video) may not
be copied, reproduced, republished, uploaded, posted, transmitted, or distributed in any
way, in whole or in part, except that one copy may be downloaded for your personal, non-
commercial use on a single computer. In connection with such use, you may not modify
or obscure any copyright or other proprietary notice.
The Materials may contain trademarks, services marks and logos that are the property of
third parties. You are not permitted to use these trademarks, services marks or logos
without prior written consent of such third parties.
Realtimepublishers.com and the Realtimepublishers logo are registered in the US Patent
& Trademark Office. All other product or service names are the property of their
respective owners.
If you have any questions about these terms, or if you would like information about
licensing materials from Realtimepublishers.com, please contact us via e-mail at
info@realtimepublishers.com.

x
Chapter 1

[Editor's Note: This eBook was downloaded from Realtime Nexus—The Digital Library. All
leading technology guides from Realtimepublishers can be found at
http://nexus.realtimepublishers.com.]

Chapter 1: Introduction to Unifying Network Management and


Converged IP Communications
The convergence of voice and data networks has been evolving and gaining momentum for
several years. Although networks today have not converged, they are moving toward
convergence in several ways. Many organizations are implementing Voice over Internet Protocol
(VoIP) in an effort to cut communications costs or leverage the competitive advantage of
integrated services, and VoIP implementers often focus on voice quality and interoperability—
important factors in the delivery of Quality of Service (QoS). However, convergence really
means much more than that.
Today, converging networks and net-centric applications are changing everything about network
management. Network administrators need to manage and monitor a wide variety of network
elements. They have to understand more complex network events and respond more quickly than
ever. Effective management for these integrated network technologies—data, voice, video,
wireless, and so forth—is crucial to network operations. This guide will highlight the overall
service management challenges facing enterprise business and identify common industry best
practices for effectively managing an integrated, unified communications environment using
VoIP. It will offer a series of systematic and holistic techniques for managing the total integrated
network to ensure consistent service delivery and support for ongoing business operations.
The term convergence is generally used in reference to the integration of telephony with data
services and applications as well as video onto a single network. This single network is
frequently assumed to be the Internet, but the convergence of services is bringing voice and data
networks closer together in many ways. These technologies all used dedicated, separate
resources in the past but can now share resources and interact with each other, creating new
efficiencies for business.
The IP data network is evolving much further than just VoIP. Video technologies are blending
and overlapping VoIP. Video VoIP (VVoIP) is becoming an accepted business service. With
new operating system (OS) evolutions ahead and increased difficulty in air travel, many
businesses are seriously exploring video collaboration as an alternative approach to traditional
travel.
Beyond video, mobility is a prime business consideration in today’s business environment. The
evolution of wireless technologies to broadband services increased productivity for mobile
workers. Another looming aspect of convergence is the convergence of the wired enterprise data
network with the wireless cellular networks. This fixed mobile convergence (FMC) will surely
add momentum as the technologies and handsets mature.

1
Chapter 1

A trend is underway in which voice and data communications are merging. The irresistible logic
is that digitized voice is just another kind of data, so why not carry it on the same data links that
handle all your other ones and zeros? The economies of convergence can be considerable—there
is no need to build and support separate voice and data infrastructures when you can have just
one. That combination of infrastructures presents the problem that convergence aims to solve:
data people, who haven’t worried about voice in the past, have to worry about it now. Similarly,
people who used to specialize exclusively in switched voice circuits must adapt to the new
environment.
From enterprise business to small business to consumer, the end user doesn’t care what network
delivers services. The ability to work from any single device, anywhere, any time is more in
demand today than ever in history. This chapter will review the broad aspects and implications of
convergence in several forms.

Convergence Covers Many Areas


Five years ago, when people spoke of convergence, they really meant the evolution to VoIP or IP
Telephony (IPT). Over those years, convergence has emerged with many different shapes and
serves multiple purposes. It means different things to different people. Let’s briefly explore the
different nuances of convergence to fully understand what is happening in the network
technology evolution.

Infrastructure Convergence in the Network—the Wiring and Circuit Convergence


Infrastructure convergence has taken place over nearly 20 years in different forms. The earliest
examples of infrastructure convergence were brought on within the corporate building, driven by
advances in LAN technology and improvements in digital PBXs. This initial convergence was
the shift to a single Category-5 unshielded twisted-pair (UTP) wiring run to every work cubicle.
Wiring plans simplified to bring voice and data services on a single wired infrastructure in the
office. This premise-based convergence was driven by cost and convenience, but it set the stage
for further network infrastructure integration.
Large enterprise business struggled for years with the separation of telephony and data services.
Organizations bought circuit connections from carriers in staggering numbers:
• Primary Rate T-1s (PRIs)
• Basic rate ISDN lines for telecommuters and small offices
• POTS lines for home offices
• Point-to-point circuits to connect PBXs
• Frame Relay T-1s, both full and fractional
• Point-to-point circuits for SNA and other data needs
• POTS lines for dial-up to Remote Access Service (RAS)
Billing and administrative tracking of all this circuit connectivity placed a huge burden on
enterprise business. The validation of circuit billing and Centrex service billing as well as, in
some cases, trying to validate billed minutes of use, is a huge, labor-intensive effort.

2
Chapter 1

Reducing the number of circuits became an obvious solution. Virtual circuit technologies such as
Asynchronous Transfer Mode (ATM) and Frame Relay demonstrated the practical potential for
the large enterprise business to receive all voice and data services converged onto a single “fat
pipe” circuit. Fiber optic cable began replacing copper wire in many areas as Synchronous
Optical Networking (SONET), using lasers to transmit information, provided advances in
carrying capacity of the physical media.

References for ATM and SONET and other technologies: Wikipedia provides an excellent basic
framework for high-level explanation and history of many of the technologies described in this guide.
There are hundreds of excellent technical books on the protocols and technologies mentioned.

Service Convergence over a Common IP Networking Environment


For telephony carriers, infrastructure service convergence was also driven by Internet dial-up
access. The Public Switched Telephone Network (PSTN) has been finely tuned and optimized to
support voice traffic. An average telephone call lasts 3 to 4 minutes in duration. Dialup access to
the Internet presented a new traffic engineering challenge to the telcos. The PSTN needed to
support Internet sessions that might last for hours or days.
The PSTN is designed using blocking switches. What this means is that if every customer
connected to a local central office (CO) picks up the phone at the same time, some will be
blocked. It isn’t cost effective to over-engineer a central office to support 100,000 concurrent
users when traffic studies have indicated that only 30 percent might be using the telephone at any
given point in time.
Dialup access to the Internet forced a change in thinking about how the Internet and PSTN
interact. As Figure 1.1 illustrates, they began as separate, disconnected networks, but dialup
access introduced touch points between the two. As technologies evolved, a converging phase
was reached with numerous high-capacity connections between the two. That is the current state
of network technology.
The fully converged infrastructure network of tomorrow will blend the PSTN and Internet so
closely that many users won’t even detect a distinction. Voice calls might route off the PSTN, a
wired LAN connection using a VoIP phone on the desktop, or a wireless handheld smartphone
that uses WiFi in the office then hands a call off to the cellular network when the user moves out
of WiFi range.

3
Chapter 1

Figure 1.1: Network convergence evolution.

Device Convergence—the Convergence of Desktops, Laptops, Tablets, and PDAs


In business, the way people work has changed dramatically in the past 20 years. The text-based
green screen used to work on a mainframe application isn’t entirely gone. Now that function is
often performed using emulation software on a PC workstation. The PC has evolved as well.
Today, businesses and workers customize and optimize their workstations to provide the most
effective, productive working environment.
Business has become finely attuned to the cost of real estate. This has driven many businesses to
implement telecommuting programs. Reduction in computer size has brought small notebook
computers into the workplace. Today, tablet PCs are becoming quite common.
This shrinkage in PC size has extended beyond the computer into the worker’s purse of pocket.
Personal Digital Assistants (PDAs) became very popular, and today are blending into the
smartphone. Mobile phones today have more computational power and memory than earlier PCs.

4
Chapter 1

Figure 1.2: Device convergence driven by shrinking desktop “real estate” and increased mobility.

It’s also important to recognize convergence on the desktop. Real estate, even desktop real
estate, is a costly commodity. Integrating voice services into the desktop workstation, laptop, or
even the PDA/smartphone is another facet of convergence.

Application Convergence—Integration of Enterprise Business Applications with


Network Services
Perhaps one of the most interesting and potentially productive areas of convergence is the
integration of applications and services. The Web no long relies solely on HTML. Today XML-
based Web services provide communications between front-end applications and back-end
servers. Software development tools such as AJAX and Ruby on Rails as well as plug-in tools
such as Flash provide an increasingly rich end user experience in using Web services.
Combining these advancements in development tools with enterprise applications provides a new
frontier of service convergence that might lead to process re-engineering efforts in workflow.
Major business applications include Customer Relationship Management (CRM) systems that
can integrate all aspects of the relationship life cycle between businesses and customers.
Enterprise Resource Planning (ERP) systems aid in supply-chain management for manufacturing
organizations. They help monitor inventory control, quality assurance processes, and product life
cycle management.

5
Chapter 1

Fixed Mobile Convergence


A more recent development in convergence is the idea of embedding WiFi technology into
mobile handsets. Business users all carry a mobile telephone. There is a clear trend emerging in
the form of a fixed and mobile telephony convergence. The goal is to provide both services on
one handset—it is a critical strategic issue across the telecommunications industry for the fixed
wireline carriers, the mobile wireless carriers, and IP network providers. When this convergence
is accomplished, these operators, regardless of service, will be able to provide voice access to the
fixed line infrastructure of the PSTN. Beyond voice conversations, there is a real independence
in the end user’s device. Location, access technology, and terminal all become irrelevant because
they all work.
As Figure 1.3 shows, using FMC features, a phone call might originate via a softphone on the PC
in a worker’s home. When the worker leaves for the office, the call could be transferred to the
mobile handset, perhaps using WiFi over a home wireless network. While driving to the office,
the call could be continued over the cellular connection to the PSTN. Upon arriving at the office,
this call might automatically be passed from the cellular network onto a corporate WiFi
connection to minimize cellular airtime. Finally, the worker might switch the call over to a rich
media desktop client in the office to collaborate over an application, shift to video conferencing,
or share some other network resources.

Figure 1.3: Example of FMC.

6
Chapter 1

Key drivers for traffic migration between fixed and mobile networks include the fact that
customers gain the ability to choose which network carries voice calls. This choice can be made
based on both cost and convenience. Users making calls from home or office locations (where
both are available) can select the transport network that suits their needs.
Will increasing dependence cause consumers to use mobile devices at home? Absolutely—
business users already do a tremendous volume of mobile calling. The continual downward
spiral in per-minute costs can drive consumers to rely more and more on mobile services. The
rising availability of broadband data services delivered to these handsets assures the spread of
new uses. According to one survey conducted in Britain, the penetration levels for mobile
telephony are as high as 70 percent in several developed countries. Mobile telephony may
present a viable alternative to traditional fixed telephony for many users.
From the mobile operator’s perspective, increased indoor usage of the mobile phone will drive
acceptance. Coverage, capacity, and spectrum constraints might limit their QoS. Using
convergence of technologies to embrace the unlicensed, short-range wireless protocols today and
the longer range WiMax technologies on the horizon, provide an inexpensive means for them to
increase capacity and provide service moving ahead. It’s in their best interest to actively embrace
the convergence of fixed and mobile services.
FMC depends on the handsets to make a wireless connection to an access point and to both make
and receive calls via the fixed line infrastructure. Short-range wireless technologies using the
unlicensed frequency bands give fixed line service providers the opportunity to create new
services without the investment in expensive spectrum licensing or complicated frequency
planning.
The two most likely technologies for wireless interface in the 2.4GHz unlicensed band are
Bluetooth and WiFi. For FMC, there is a key limiting factor—or success factor: the handset.
Compatible handsets are appearing in growing numbers. Bluetooth is available in more than 25
percent of new models. Bluetooth is suitable as a wireless media for FMC and was designed to
handle voice service, but it’s not the sweet spot. WiFi integration into the handset is quickly
becoming a key differentiator for manufacturers. This is a trend that will rapidly expand to the
majority of new handsets over the next year or two.

VoIP as a Converged Service


IPT is one of the most visible, talked about technologies in the marketplace today. Improvements
in performance and cost reduction issues are huge drivers for enterprise business adoption. There
are several factors that make IPT a suitable technology evolution for business
telecommunications. Because there are so many different business needs ranging from small
companies with home offices to very large enterprises spread around the globe, there is no single
“one size fits all” solution for telephony requirements. This should not come as a surprise,
because there isn’t a single solution for using existing telephony services and equipment.

7
Chapter 1

Voice vs. Data


Voice telephony services require a long holding time to support call durations averaging about 4
minutes per phone call. The signal is sensitive to delay and jitter because it is a real-time
interactive communication. In the PSTN, you deliver this service using circuit switched
technology. In some cases, large businesses have implemented private telephone links, such as
T-1 tie lines, between branch offices to provide internal telecommunications from site to site.
Data communications don’t have long holding times or durations. They’re described as bursty
and unpredictable in nature. The durations of data transfers may be quite short, especially with
increases in available bandwidth. Data communications are frequently not a real-time, two-way,
interactive session like a phone call. Email and Web browsing, for example, are not as time-
sensitive as voice. Data services have generally used all the available bandwidth for a very short
duration. That’s beginning to change as QoS becomes an integral part of network operations.
In the past, it wasn’t uncommon for enterprise businesses to have two or three completely
separate networks for conducting business. Voice traffic was often handled through a PBX
connected to the PSTN. Real-time interactive data was transmitted via some type of packet
network, either IP or Frame Relay in many cases. Large file transfer between mainframe systems
or server farms might have been carried out across a dedicated point-to-point connection.

The Cost of Doing Business Drives VoIP


One of the major early drivers behind convergence and VoIP was cost, the job of calculating the
true total cost of separate voice and data networks is a complex process. The network
infrastructure cost is one factor. Telephone services have historically been billed based on
minutes of use. A phone call requires that a circuit be established for the duration of the call, so
this billing system made sense in the past. Data networks are different, and most commonly bill
for either the bandwidth provided or some guaranteed carrying capacity (referred to as
Committed Information Rate—CIR).
Equipment cost can usually be treated as one-time capitol expenditures (CAPEX).This cost is
tied to buying required equipment, routers, computers, telephone systems, wiring and cabling,
and so forth. CAPEX costs are closely scrutinized when businesses invest in new network
solutions, but they only represent one facet of the total cost of ownership (TCO).
Operational expense (OPEX) is another factor that is too easily underestimated. Given that the
topic of this guide is converged network management, it will be digging into a variety of holistic
management issues and techniques. The labor effort required to provide operational support for
additions, moves, and changes to the network are difficult for many organizations to quantify.
Reorganization is for many enterprises, business as usual, with constant upheaval. The costs to
support this continual churn over the lifetime of the network often far outweigh the one-time
equipment costs. For the enterprise managing multiple networks, telephone and voice, this
OPEX may be duplicated for each network. It is very common for an organization to have a
support person or staff for the telephone system and another for the data network. In the past,
these support personnel required different skill sets, but convergence, even in the state
achievable today, can change that.

8
Chapter 1

Administrative costs comprise still another facet of networking that may be underestimated in
some enterprises. There are the basic costs of processing monthly invoices for payment. Beyond
simple accounts payable, there is administrative effort involved in the ongoing monitoring of
network performance compared with monthly billing. Billed minutes of use in the voice network
and committed information rate delivered have to be compared with Service Level Agreements
(SLAs) and validated against traffic reports. The same staff that manages day-to-day operations
might perform this analysis. Although often overlooked and perhaps not high in terms of actual
cost, this function can potentially yield high return, particularly in networks that are volatile and
changing frequently. Holding providers accountable for meeting the terms of their contracts is a
labor-intensive but crucial component of both voice and data network services.
The telephone and data network(s) in an enterprise need holistic treatment, like a living
organism. The health and well-being of these vital business resources must be constantly
monitored. Enterprise business constantly evaluates and assesses employee performance.
Network performance must be proactively managed in a similar fashion. That means vigilant
monitoring throughout the life cycle of the network to ensure maximum performance.
Holistic management of the business network includes working closely with voice and data
service providers. It’s far too easy for changes in a dynamic enterprise to happen without any
correlation to network services. The result is that the network reality diverges from the
requirements. They move in different directions driven by changes in the business.

I’ve helped many clients conduct network analysis and assessment focused on identifying potential
OPEX reduction. In one such case recently, we quickly discovered that the company had downsized
operations by half in the preceding year. What was troubling was that the network billing didn’t reflect
any reduction. The circuit cost component of their OPEX hadn’t changed. When a remote branch
office closed, this company had completely failed to have the network provider disconnect circuits and
stop billing. Several thousand dollars were paid to a provider over the course of a year for idle circuits
that carried no traffic. Although this might seem a laughable example of really poor management, that
isn’t the case. This type of problem is far too common.

Given the technologies generally available, is it practical to build one single network
infrastructure that will support all traffic and service types required to do business? A few years
ago, the answer to that question was probably no. Four or five years ago, only the earliest
adopters of leading-edge technologies were able to take advantage of integrating multiple
services into a single environment. Today, networks are actively converging. Right now, for
many companies, unifying communications onto a single integrated infrastructure, converging
services into single workstations, or integrating applications and services makes sound business
sense. As technologies collide, keep in mind that what didn’t work yesterday may be viable
today, and what doesn’t work today, could well be the de facto standard of tomorrow.

9
Chapter 1

Technical Definition vs. Market Definition


It’s perhaps an important distinction to identify the differences between VoIP and IPT and what
they might mean in the marketplace. From a purist perspective, VoIP is simply the process of
digitizing and packetizing voice. Figure 4.4 shows a common arrangement for making a VoIP
call. Although IP telephone calls can take on many forms, this example represents an IP phone
call in a form that doesn’t directly involve the subscriber in any way. The telephone set in this
example is a traditional telephone connected to a traditional PBX.

Figure 4.4: A variation on IPT—the gateway between business and the PSTN.

The phone call begins when the subscriber lifts the receiver and a dial tone is supplied from the
local PBX. In this case, when the caller keys in the telephone number, that information is
forwarded to an IPT gateway. The gateway provides a conversion point for traditional voice
traffic to be converted to IP and vice versa.

10
Chapter 1

The originating gateway has to convert standard PSTN signals and correlate the dialed telephone
number to the IP address of the terminating gateway that serves the called party. This signaling
information is packetized and sent to the IPT gateway at the far end. The receiving gateway must
decode the IP packets and convert the signals back into traditional voice format for the PBX on
the remote end. When the called party answers the telephone, a complete two-way conversation
can occur.
Some benefits to this approach are obvious. Others are more subtle. This method of VoIP
implementation has been employed by both corporate enterprises and commercial VoIP service
providers (also called an Internet Telephony Service Provider—ITSP):
• Using a packet network to transmit allows the sharing of resources rather than the
dedicated resources for voice and separate resources for data. Although this visual shows
the Internet, many enterprises use their internal wide area packet network to deliver this
service.
• If the Internet is the packet network used, the service provider can now provide two
distinctly different but necessary services from one consolidated backbone infrastructure.
This approach may drive the need for QoS in the packet network.
• Because the ISP or ITSP is not a regulated telephone company, the requirement for
payment of access charges doesn’t exist. Thus, the ITSP can deliver calls cheaper.
However, it also means that the local exchange company doesn’t receive remuneration
while its circuits are being used in many cases. This is good for one part of the industry,
but bad for another part. Telephony traffic is shifting off the PSTN and on to the Internet.
• In this particular implementation of IPT, the end user doesn’t have to convert to a VoIP
phone. The end user doesn’t have to do anything. They might not even know the
telephone call is being carried over a VoIP service. The Internet is a large and growing
network, but the user experience for a voice call may best be served using a telephone set.
There is no added complexity of soft phones or specialized instruments for end users in
this model. This could prove a viable approach for common calling services such as pay
phones, hotel phones, and public phones where other, more advanced services aren’t
required.
• The initial startup cost, or barrier to entry, for a VoIP service provider is very low in
comparison with the startup costs for a new telephone company using traditional PSTN
technologies. IP packet switching technologies provide for greater efficiencies at a lower
cost.

11
Chapter 1

Variations on implementing the converged services network range from straightforward to very
complex. When evaluating how services converge within an enterprise, it’s vital to do so with an
eye to not just the business needs but also the ease of support and network management. Some
other variations on implementations might be:
• IP Centrex services from a managed service provider—This service could be delivered to
the corporate network over one single network connection, with voice and data services
broken out inside the company network.
• A managed IP-PBX inside the corporate network—The key difference between this
approach and that shown in Figure 4.4 is the potential for VoIP telephones or integrated
softphones within the corporate network.
• Internet users might utilize some form of commercial gateway—Notable early examples
of this were http://www.dialpad.com and http://www.net2phone.com. You can expect to
see an increase in Web services providing this sort of service for Web only, or thin client,
connections. In particular, as developers find new ways to embed VoIP services within
the browser, you can expect to see more of this functionality.
• Some companies are using PC-based software to provide internal telephone calling and
collaboration services between employees on the corporate network—This LAN-based
VoIP is efficient within the network, and essentially free of charges. This may be some of
the early use you see for the new Microsoft Live Communications Server. Calls to the
PSTN might require employees to have a traditional telephone using the traditional
network.
• A large multi-location enterprise deploying unified communications might implement
multiple gateways at geographically dispersed office sites. Calls might be directed to the
nearest gateway to the called party, minimizing toll and long-distance charges. This “least
cost routing” approach has long been used in large enterprise voice networks.

VoIP or IPT?
Voice digitization is not a new technology. Digitization began inside the telcos in the 1960s
utilizing what is called T-1 service today. During this time, T-Carrier, as it was called then, was
deployed to provide trunking capacity between telco central offices. Analog transmission
technologies were in widespread use at the time, and digital technologies enabled network
performance improvements in the PSTN that benefited both customers and the phone companies.
The following section explores the basics of a telephone call to set the stage for a discussion of
network requirements and call quality issues in later chapters.

12
Chapter 1

A Telephone Call Simplified


When a person makes a telephone call, there are several basic steps that must occur whether the
call is a videoconference, a fax transmission, or a voice conversation. These tasks are performed
by the service network in each case.
When you speak, your vocal chords vibrate, generating sound waves. These sound waves are an
analog signal. These sound waves travel through the air, and your ears convert them back into
signals your brain can understand. The world around you is very much an analog place.
The telephone converts the sound waves from your spoken conversation into electrical signals
that can be transmitted over copper wires. The telephone set at the receiving end converts that
electrical signal back into analog sounds waves you can hear.
In the telephone network, that electrical signal is converted from analog into a digital signal, a
stream of zeros and ones defined by changes in the electrical state of the circuit. Thus, to carry
on a telephone conversation, your analog voice must be changed from sound waves made up of
vibration of air molecules into an electrical representation of the analog wave. This analog wave
then must be digitized and transmitted across the network. At the other end, the signal is
converted back to analog, then back into sound waves at the earpiece of the telephone handset.

Converting the Analog Signal to Digital


The telephone network today is made of digital central office switches connected by digital trunk
circuits. It makes sense to transmit digital signals. The local loops (also called the last mile) that
connect subscribers and most phones are analog. Analog-to-digital conversion must be
performed somewhere in the network. This conversion is typically accomplished using a
technique known as pulse code modulation (PCM). In order to perform PCM, a coder and
decoder (more commonly called a codec) are needed. PCM takes samples of the voice
conversation 8000 times per second. These samples are then converted into an 8-bit word,
resulting in a 64Kbps sample (8 bits × 8000 samples per second = 64kbps). Each of these 8 bit
samples can be coded into one of 255 different possible combinations. The 8 bits of binary data
can represent values from 0 to 255, but all zeroes cannot be used in this coding scheme.
This guide will touch on the 64Kbps line rate more than once. This line speed is what a standard
voice channel in the PSTN provides. In the United States, a common time-division multiplexing
transmission scheme is used to transmit 24 voice channels as a digital stream of data over a
single circuit. This is the basic voice service provided over a T-1 circuit today, with 24 voice
channels delivered over 1.544Mpbs (due to some overhead). This design of digital facilities and
voice circuits is part of the Synchronous Digital Hierarchy throughout the world. Other parts of
the world don’t use T-1 circuits, but their approach is similar.
By today’s standards, 64Kbps really doesn’t seem like a lot of bandwidth, but the technical
reality is that, depending on which codec is used, it’s far more than necessary to carry a voice
conversation.

A future chapter will touch on coding schemes used to sample and compress audio traffic streams.

13
Chapter 1

Network economy of scale and simple economics drive the industry to work toward reducing the
bandwidth required for a phone call. The less bandwidth required per conversation, the more
conversations the network can carry at one time. Adding carrying capacity through codecs and
compression minimized capital investment in network equipment. This can potentially drive
down the cost of carrying a phone call.

When Does VoIP Become IPT?


The PSTN is a complex and mature implementation of technology that has evolved over more
than a hundred years of use. Though technologies are changing and advancing quickly, the
Internet is quite immature by comparison.
The PSTN has what is referred to as the Advanced Intelligent Network (AIN). 800 number
services provide access to databases of caller information. Caller ID, call waiting, call hold, and
conference calling are common features available to almost all users. E-911 services are
constantly evolving with technology in order to identify the caller’s location to within a very
small geographic radius. This mature network has been optimized over time to provide the best
quality voice and a variety of services that are now taken for granted. The PSTN does far more
than transport voice traffic. It provides a comprehensive and robust suite of telephony services.
The IP networks, or Internet, certainly can’t provide that rich set of services by itself. VoIP has
been achievable for a number of years. Simply packetizing a voice signal is easy. There is far
more to IPT than just carrying voice traffic in IP packets. The full, rich telephony feature set has
to be incorporated before users could give serious consideration to IPT as a viable service in the
production network for enterprise business.
IPT solutions can now bridge the gaps between the PSTN and the IP network. Using the TCP/IP
protocol suite, a network such as the Internet can now provide many aspects of the traditional
telephone network. The IPT market has seen a substantial growth rate over the past several years.
Vendor products have improved, but so has basic network technology.
The issue of VoIP vs. IPT may be one of semantics, but it represents a problem for business
people trying to make a decision. One vendor will talk about VoIP, another about IPT, and a
third will now refer to unified communications. All the scenarios described in this guide are IPT
in some form.
It’s clear that if the end user connects to the PSTN, they’ll require some kind of a telephone. If
they connect to the IP network, they’ll need an IP device of some kind. As business managers,
it’s important not to get hung up in the semantics vendors or service providers use. It’s prudent to
focus on the services being provided to ensure they support business needs.

14
Chapter 1

Intended for PC-to-PC Voice Messaging


Pure IPT between computer users over the Internet was a popular, early hobbyist’s
communication technique. This form of VoIP has been accomplished by users meeting in a chat
room. Today, it’s performed by a staggering array of software solutions including voice-oriented
solutions such as Skype, Gizmo Project, and GoogleTalk. There is also a whole family of instant
messaging (IM) programs including ICQ, AOL Instant Messenger (AIM), MSN Live Messenger,
Yahoo, and others. IM users can easily activate a voice session over the Internet in most popular
client solutions today. Many of these VoIP solutions that began as PC-to-PC calling tools now
have links in and out from the PSTN as well.
Early prophecies for success in IPT viewed the technology as a tool for consumers to combat the
high price of long-distance telephone service. Those arguments seemed to hold water at that
point in time, but they overlooked the technical advances and growth of mobile telephony.
Today, many wireless providers offer nationwide calling at rates cheaper than imaginable 10
years ago. VoIP still aids in reducing the cost of international long distance, but the current
primary driver is more likely to be service integration than cost reduction.

Never Intended to Replace Global Telephony


VoIP wasn’t developed with the ultimate goal of replacing the PSTN—not initially. But VoIP
disruption has occurred and is still gaining momentum. Delivery of voice service is no longer
tied to the vertically integrated Class-5 telco switch. This disruption has been accepted as the
trend to the future of telecommunications.
IP evolution is replacing monolithic, proprietary solutions and broadening into Web and
application services. Innovation is increasing at a faster rate in a dynamic and highly competitive
environment.

“Free Voice” Was Compelling and Simple


The idea of free voice calls spurred hobbyists and early adopters to experiment with an array of
new ideas. The drawback is that this PC-centric group of innovators and early adopters provides
only a limited market. They represent an application and integration proving ground, but not a
mass market.

Pushes the Edge Over Convergence


What will be the accelerant for VoIP and unified communications? The ongoing debate between
cost reduction and enhanced utility will continue into the foreseeable future. Today, the market
sees far too many “more of the same” or “me too” solutions. This limits the advance of both user
experience and penetration. The VoIP developer community has often been providing equivalent
features to the telco environment. VoIP as a replacement for PSTN dial tone is neither useful nor
well received in today’s market.
If applications will fuel the future growth and demand, one option might be to pursue integration
inside what have been termed the walled gardens. An integration layer between proprietary
applications and content providers will expand the reach of converged services to a broader
market.

15
Chapter 1

Vertical communities of interest are early adopters, but early adopters really only provide fuel
for developers. Markets emerge over time through an established pipeline of applications and
services. You never know at the beginning of the development cycle what the “killer app” will
be. Communities are the early adopters and innovators that help create the future. A rich multi-
media, real-time telephony experience exists within the technology, but it’s just taking root and
beginning to expand. One thing is clear: Voice will no longer be driven by the telephone. Voice
services of the future will be driven by the context of the business application.

Multiple Play
Multi-play is a popular marketing term describing the delivery of multiple communications
services by service providers that traditionally only offered one or two of those services. Triple
play and quadruple play are commonly used to describe the combined delivery of high-speed
Internet, television, telephone, and mobile phone services, respectively:
• Dual play—The dual-play service is a marketing term for the provisioning of the two
services: high-speed Internet and telephone service over a single broadband connection.
This has most frequently been used with cable providers.
• Triple play—Three-way convergence is inextricably linked to the underlying
communication infrastructure. A prime example of this would be packaging
communication services in a form so that customers can purchase television, Internet, and
telephony in a single, bundled service subscription.
• Quadruple play—So far, the quadruple play service remains elusive. It’s the triple play
service of broadband Internet access, television, and telephone with the addition of
wireless service provisions. Advancements in WiMax and other leading-edge
technologies are rapidly improving. Transmission over a wireless connection link at
combinations of speeds, distances, and non-line-of- sight conditions may soon make it
possible to never connect to voice or data services by a wire to anything, even while at
home.

New Devices
New devices are driven by technology advances, user expectations, and how people use their
VoIP and mobile phones. There is some relevance evolving in the market surrounding how often
people change phones. Many consumers move to the next new smartphone every 6 months in the
mobile environment, but in business, many people have been using the same telephone set with
the same features for many years.
You get new cell phones regularly because they enhance your productivity in some way. This
leads providers to think about how people interact with their phone set. The comparison of how
you change cell phones sets the stage for other unified communications options.
Although vendors continue to see rising shipments of dedicated IP phones quarterly, there are
other business drivers to consider. Three words describe this new driver—relevance, context, and
presence. The type of telephone set you need varies depending on your context. In some cases,
you’ll be in a meeting or unable to talk and an instant message might be the most relevant
communication method. At other times, you can’t be interrupted at all. Whether it’s a PC-based
softphone, a traditional desktop telephone, or a mobile handset, each offers strengths depending
on the context being used.

16
Chapter 1

Today, people work in distributed, virtual, mobile teams. The work day doesn’t start when you
enter the office, and it doesn’t end when you leave. You need to manage your communications
flow throughout the day. There are hundreds of features and services in existing systems, but
users generally don’t know how to access them. Most people use a very limited subset of the
available features and functionality. Features will become more intuitive. The context of your
work day and how you’re using other business applications will be more central to the
development of converged communications devices as things evolve.
The key questions for device manufactures are:
• What do users expect from their phone? Users interact with many different devices all
day long. Those interactions drive expectations.
• How can we add efficiency via the telephone device?
• How can phones increase productivity?
Phones are moving from being feature focused to user focused. It’s a paradigm shift on the desk
and in your purse or pocket. Can a phone really improve or impact efficiency? Yes, but if it fails
to provide access to features and functions, it can be a frustration and lead to lost productivity.

VoIP Can Include Advanced Applications


The convergence of real-time media carried over IP along with data is where we’ve arrived. That
is the state of technology today. Now we’re looking at how to converge these services onto
application engines. Convergence with enterprise business applications presents another whole
convergence layer. CRM, human resources (HR), and ERP systems all introduce interesting
ideas when blended with telephony services.
Converged communications is the capability to integrate at different layers than just
telecommunications. These other layers are still fairly separated into silos, or vertical groups.
Workers today have multiple communications devices and soft phones. New capabilities such as
real-time directories that provide presence and availability information, conference call setup,
and rich media integration empower the end user.
Convergence integrates the power of multiple platforms that offer service capability in a unified
environment rather than discrete applications or services. The Services Oriented Architecture
(SOA) is moving toward opening up the services of one application for sharing with another
application altogether. This isn’t just VoIP. It isn’t just integrating communications. It’s service
integration through software across multiple business applications. Telephony software in the
converged environment can now act like a business application and interact with other business
applications.
Many of the functions that the hardware enterprise PBX provided can now be performed in
software. Integration at the software level, at the application level, is the key to opening up
integrated services within IP. You can do new things when you integrate business applications
rather than focus on technology.

17
Chapter 1

Converged Communications Leads to Converged Applications


Specialized applications are becoming standard applications. Using the SOA information
resources available to all participants (users, services, and applications) in the network as
independent services may now be accessed in a standardized, converged way. Users don’t need
to change context from one application to another. The end user shouldn’t have to shift the
context of the core application in which he or she is working to make a telephone call. The
developer communities are working to open up integration services to allow all that context
switching to happen behind the scenes. The following list highlights real application examples:
• A Web browser-based interface is the simplest example. It can display state information
to users. The browser can easily show present state in an address book or telephone
directory, with click-to-dial capability.
• A mobile device that can run thin client or browser-based applications presents another
example. This device makes it easy to authenticate as a mobile WiFi user and then
display state information based on that IP connectivity. The mobile device simply
becomes a thin client using Web services on the network. Again, directories and presence
state are prime examples of integrating productivity tools.
• For large enterprises, think about the whole cumbersome process of voice and data adds,
moves, and changes for new employees and people shuffling from cube to cube. That’s
historically a huge labor effort. Why not enable an HR system that interacts with the
voice and data services network to automatically configure telephony services for new
employees?
Converging business applications is absolutely where the entire business community is headed.

FMC
Mobility is a huge issue, and the advances ahead with FMC make this a subject of interest across
many business sectors. The environment is ripe for mobile VoIP. The delivery of enterprise
applications, information, and services to a mobile worker is vital today. One factor that has
made mobile VoIP viable today is near-universal standardization on Session Initiation Protocol
(SIP).
At a recent developer conference, informal polling of attendees pointed out that email usage on
mobile phones was at about a 50 percent usage rate. We are becoming mobile data users more
and more. Enterprise users are a huge piece of the broad mobile services market and have been
since 1973. Although mobile usage is on the rise and ripe for convergence of VoIP and mobile
services, user interactions and experiences are distracting and degrade from the overall the user
experience.
Radio characteristics present another challenge. Current WiFi technology is very chatty in
standby mode, draining mobile device batteries and shortening usable lifetime. Because of this,
dual-mode devices tend to suffer from measurably greatly shortened battery life. There is
industry work needed to make the WiFi device more effective as a phone.

18
Chapter 1

A recent survey of mobile professionals produced some interesting results. The most active
mobile users in business are either customer-facing workers or management. More than 75
percent were happy with coverage. About half felt that dual mode could improve coverage for
better service. Churn in the mobile industry is driven by coverage issues more than any other
factor. Most mobile carriers are now down to about a 2 percent churn rate.
Business today is very mobile. 70 million Americans use their cell phones for work. 70 percent
of all cell calls start in WiFi-enabled areas (home, office, hotel, and so on). By 2009, the industry
expects that 25 percent of new mobiles will be smartphones. One really interesting note from the
survey about business and mobility: 67 percent of mobile professionals receive or make more
than 25 percent of their business calls on mobile phones. Mobiles are very important devices to
enterprise business. (These survey numbers come from a collection of surveys conducted by
FirstHand, Nortel, RHK, and Gartner.) Mobile VoIP is much more than just terminating a SIP
session on a small handheld device. It’s about enabling a productive personal or business
experience.

Summary
You can wait for evolution or you can begin using the technologies that are available today and
ride the wave toward the fully converged networks of tomorrow. Even as developers are defining
the architecture, boundaries, requirements, and interfaces for a shared unified communications
network of tomorrow, the current solutions provide a wide open space to grow and evolve
existing business services.
The chapters ahead will review key considerations for integration, business drivers that make
sense, and productivity advantages. This guide will dig further into the key success factors for
deployment and management of converged networks. It will investigate event correlation across
multi-service networks, and dig into availability management and capacity planning in the world
of unified communications. This guide will explore fault, configuration, performance, and
security management in this exciting new world. Finally, it will take a look at asset and
compliance management.
The next chapter will hone in on business processes most often impacted by convergence. In
addition, it will examine the issue of call quality and how to deliver total quality network
services, including VoIP in the converged network.

19
Chapter 2

Chapter 2: Key Considerations in Effective Voice and Data


Integration for a Changing IT/IP Landscape
The integration or convergence of voice, video, and data can provide a business with a
competitive edge when effectively implemented. There are several business models and
operating environments that present opportunities for strategic consideration when planning for
this change. The key factors for success involve leveraging the integrated features to provide the
greatest support for existing processes.
This chapter will look first at high-level business models and processes that are often impacted
by service convergence. Later, the chapter will delve into the issues of call quality that affect
every organization implementing an integrated service solution. Quality is often the single
biggest factor in a successful implementation, so this chapter will explore a variety of approaches
for delivering total quality network services, with a focus on integrated VoIP.

Quantifiable Business Processes


Business crosses a wide array of sectors, each having unique business requirements to support
aspects of the core business. Call centers may play a central role for many businesses,
particularly those in financial services, insurance, or travel. They also play a key role in many
other sectors as smaller customer support teams. Interactive voice response (IVR) systems are
frequently automated to reduce the requirement for staffing and provide information to
customers. Computer telephony integration (CTI) isn’t a new concept with the deployment of
VoIP, but in many cases, it becomes easier. CTI may provide levels of service and application
integration previously outside the financial grasp of some organizations.

Web-Centric Businesses
In addition to the traditional sectors of business, the integration of the Internet has heightened
awareness of four distinct business models in the Web-centric world of e-business. Some of
these models fit with large enterprise; others are more amenable to small business and have been
used by many an e-business startup company.
In the open market model, anyone can be a buyer, and anyone can be a seller. There’s no
centralized control, and minimal trust involved. There isn’t particularly high value to integration
of enterprise systems because relationships may be ephemeral. Market leadership for these e-
businesses requires being in the right place at the right time, with the right solution at the right
price. OASIS and eBay are good examples of the open market business model.

20
Chapter 2

The alliance model is more common in larger businesses. It embraces a distributed corporate
environment with multiple leaders of the pack. The goal of these alliances is frequently
optimization of specific solutions to solve identified customer problems. Alliances are often
formed among the best and brightest in their respective fields. High levels of integration in
services and applications between partners bring tremendous value to an alliance. Sun, IBM,
Oracle, and Netscape demonstrate this model via the Java Alliance.
The aggregation model is typically adopted by the leader in a business sector. The aggregator
positions itself between producers and consumers, providing access to products. Integration with
consumers may be low, but integration with the producers and internally across the aggregator
enterprise can add very high value. Wal-Mart represents a perfect example of this model.
The value chain model is adopted by most businesses. Every business is, in some facet, the
leader of the pack in its’ particular sector. Process optimization within the enterprise is crucial to
business success. The leader focuses on optimizing the value chain through service and
application integration rather than aggregating buyers and sellers. Cisco Systems, Dell, and
Amazon represent value chain leaders in the world of e-business.
There is also an emerging competitive model, driven by competition between clusters, not
individual companies. This model is often a matter of survival for businesses that need
partnerships to deliver a complete solution set to market. In this model, companies maximize the
user of distributors, resellers, and retail sales chains. Integration of services and applications may
focus on facilitative technologies such as Web-based Electronic Data Interchange (EDI) to drive
e-commerce online.
Business in the new economy of the Internet requires instantaneous reaction to customer
behavior. Being nimble and responsive can provide a competitive edge in having the right
solution for customers at the right time.

Examples of near-instantaneous change include Wal-Mart and K-Mart. Wal-Mart monitors computer
inventories closely in stores. If a customer buys a package of tennis balls, the inventory and shipping
systems are automatically updated to ship a replacement set to a specific store. If a store
experiences a run on tennis balls, or an unusual trend exceeding normal thresholds for a product
occurs, stores all across a geographic area may find they receive increased delivery of tennis balls to
support the trend.
K-Mart takes a different approach, using outside data sources to drive product stocking. One often-
used example is their use of the National Weather Service (NWS). If K-Mart notes inclement weather
headed for a particular part of the country, their stores may receive umbrellas and raincoats for
stocking, in support of anticipated demand.

21
Chapter 2

Call Centers
Businesses implement call centers to effectively administer incoming product support or provide
information to customers. Call centers most often interact directly with consumers and are often
the lifeline of the customer relationship. Outbound call centers are used for telemarketing,
clientele management, and debt collection. The call center might also be a broader contact
center, handling postal mail, fax communications, and email for an enterprise or business unit.
Call centers are generally built using large, open work spaces with workstations including
computers, telephones with headsets, and supervisory stations. Some businesses build centralized
call centers; others distribute call centers at diverse locations, connecting them with voice and
data technology. Global companies may have call centers around the world, with each being the
primary center as the time of day changes. Location, coupled with time, can provide the
advantage of using resources during the local business day.
Call centers are linked to the corporate computer network, including mainframes,
microcomputers, and LANs. Increasingly, the voice and data services are linked through CTI.
Most large enterprise businesses use call centers to interact with their customers. What has
changed with the widespread deployment of VoIP is the barrier to entry. Now a midsized
company can easily deploy a centralized or distributed call center, leveraging integrated
technologies and bringing new services to customers that were previously out of reach.
Call center technology essentially provides a queuing network coupled with workforce planning
and management to achieve desired service levels. A common example of call center service
levels might require that at least 80 percent of the callers are answered within 20 seconds or no
more that 3 percent of the customers hang up, due to their impatience, before being served. Call
centers provide information about traffic patterns and business patterns as well. Statistics
gathered can help determine whether a single large call center is more effective at providing
customer service (answering calls) than several distributed, smaller ones might be.
Centralizing call management into a call center aims at improving business operations and
reducing costs while providing a standardized, streamlined, uniform service for consumers. The
efficiency of a repeatable process makes this approach ideal for large companies with extensive
customer support needs.
Call centers use an array of voice and data technologies and provide a prime candidate for
integrated services and applications. These technologies can ensure that customer service agents
are kept as productive as possible and that calls are queued and processed as quickly as possible,
producing the desired service levels for customers. Some of these technologies include:
• Automatic call distribution (ACD) groups
• Analysis tools to review agent activity
• Optimization analysis for outbound call centers, referred to as Best Time to Call (BTTC)
• Interactive voice response systems (IVR) to improve efficiency by reducing agent time
on the phone
• CTI
• Predictive dialing for outbound calling
• Integration with Customer Relationship Management (CRM) systems
• Web collaboration and online chat tools for customer support

22
Chapter 2

Beyond the features and technologies, there is a standard suite of typical performance metrics
used in the call center management methodology. When a company invests in a call center, it’s
crucial to monitor performance levels and ensure return on the investment. The most common
metrics include:
• Measuring the average delay callers wait in a queue for an agent to come on the line
• Call duration times, generally referred to as Average Talk Time (ATT)
• The time an agent spends on the total call, including preparation, conversation, and after-
call work or wrap-up; this is called Average Handling Time (AHT)
• The percentage of calls that get answered within an established call pickup time; this is
typically called the service level (for example, 90 percent of calls are to be answered
within 30 seconds)
• The raw number of telephone calls each agent handles per hour; this is used as a measure
of agent productivity
• How much time an agent spends after getting off the phone in closing out the customer
transaction; this is generally called either Wrap-Up or After Call Work (ACW)
• Calls that completely resolve the customer’s question or problem on the first call may be
tagged with an indicator of First Call Resolution (FCR); the FCR rate is often used as a
broad, overall measure of call center performance
• Either the number or percentage of calls that are abandoned by the customer; the higher
the number or percentage, the more indicative this is of long holding times, driving
customers to hang up
• Idle time is the percentage of time agents spend either not on the phone with customers or
in ACW; high idle time may be an indication that the call center is overstaffed during a
particular time period; it’s used for trend analysis to balance staffing of work shifts in
large call centers
• Quality assurance (QA) monitoring may be performed by either a QA team or supervisor;
customers hear a standard “Your call may be monitored for quality purposes.” message
on a routine basis
The acceptance and widespread deployment of VoIP has enabled the staffing of call centers with
remote agents. These agents work from home, often on flexible part-time schedules. In the past,
they would use a basic-rate ISDN line, but widespread broadband deployments have made high-
speed Internet access easy to couple with VPN and VoIP solutions to provide a fully integrated
solution for teleworkers.
Clearly, the call center has been a key resource in large customer service organizations for many
years. Today, the reduced barrier to entry in call centers has made the technology more available
to a wide set of business environments.

23
Chapter 2

IVR Systems
IVR is a computerized system that allows a caller to choose options from a voice menu and
otherwise interface with a computer system. Typically, the IVR system plays prerecorded voice
prompts, and the caller responds by pressing numbers on the telephone keypad to select among
the options. Many IVR solutions also allow the caller to speak simple answers such as yes, no, or
numbers in response to the prompts. These systems have grown more sophisticated in the past
few years, and do a much better job of recognizing human voice than they did when they first
gained popularity.
Newer systems use natural language speech recognition to interpret the questions that the person
wants answered. There is also a growing trend called guided speech IVR that integrates live
human agents into the design and workflow of the application to help the speech recognition
with human context.
Other innovations include the ability for the system to “read out” complex and dynamic
information such as email messages, news reports, and weather information using Text-To-
Speech (TTS) conversion tools. TTS is computer-generated synthesized speech that has
advanced well beyond the robotic voice people may associate with computerized systems of the
past. Human voices are used to create the speech in very small fragments that are assembled to
create very real-sounding responses before being played to the caller.
IVR systems are used to create service solutions such as airline ticket booking and banking by
phone. Unlike voicemail systems, which are one-way communications, IVR systems provide
some level of two-way information exchange between the caller and the company systems. An
ACD group in a call center may be the customer’s first point of contact. IVRs are often used to
provide the primary front-end to a call center, automating the most common calls for account
balances or other easily delivered information.
IVR systems today are built with scripting languages such as VoiceXML or Speech Application
Language Tags (SALT), not unlike the way Web pages are constructed. In the case of an IVR,
the Web server also acts as the application server, allowing the developer to focus on call flows
rather than graphic presentation. Typically, Web developers understand and are familiar with
tools in this environment and often don’t require additional programming skills.

CTI
CTI provides for interaction between the telephone and computer systems. As technology has
matured, CTI has expanded to include the integration of all customer contact channels, including
voice, email, Web, and fax.

CTI History
CTI evolved from relatively simple screen population (or screen pop) technology. This technique
allows data collected from the telephone systems, most often via the touch pad, to be used as
input data to query databases with customer information. That data can then be populated
instantly to the customer service representative’s screen. When the agent already has the required
information on his or her terminal screen before speaking with the customer, overall transaction
time is reduced.

24
Chapter 2

This technology began in the closed, proprietary roots of every PBX/ACD vendor in the market
space. Most vendors eventually adopted the Computer Supported Telecommunications
Applications (CSTA) standard. CSTA originated in the ITU and was adopted as an OSI standard
in 2000. Some other accepted service and application integration standards in CTI include:
• Java Telephony API (JTAPI) promoted by Sun
• TSAPI and TAPI—TSAPI was originally promoted by AT&T (later Lucent, then Avaya)
and is by far the most widely adopted in large-scale contact centers; Microsoft pushed a
separate initiative, thus TAPI was born, with support predominantly from Windows-
based applications
Some commonly implemented functions in the CTI environment include:
• Information about the caller—This may include the caller’s number, the telephone
number called, and in many cases, which prompt in an IVR system the caller selected.
• Screen population (screen pop)—This provides the agent with information about the
caller before the agent picks up the call. One example is prompting a customer to enter an
account number on the telephone dialpad. This allows the CTI system to retrieve the
account and present that data to the agent, who can then be fully prepared to talk to the
customer immediately.
• Outbound call centers use on-screen dialing tools to increase productivity. Speed dial
options and predictive dialing features increase the rate of outbound call placement.
• A software screen control, such as a VoIP softphone, provides access to telephone
features through the computer. Functions such as answering a call, hanging up, placing a
caller on hold, and initiating a conference might be performed with the mouse or
keyboard shortcuts.
• Call transfer and transfer of the data screen provide the ability to transfer a call to a more
senior agent or supervisor when the agent involved needs assistance in resolving a
problem.
• Administrative functions and duties, such as logging in or out of an ACD group or
completing wrap-up work are common.

25
Chapter 2

CTI takes on two basic forms. The following excerpts from the Wikipedia online encyclopedia
article about CTI provide helpful descriptions (Source:
http://en.wikipedia.org/wiki/Computer_telephony_integration#Forms_of_CTI):

First-Party Call Control


First party call control operates as if there is a direct connection between the user's
computer and the phone set. An example of this would be a modem card in a desktop
computer, or a phone plugged directly into the computer. Typically, only the computer
associated with the phone can control it, by sending command directly to the phone. The
computer can control all the functions of the phone, normally at the computer user's
direction. First party call control is the easiest to implement but is not suited to large
scale applications such as call centers.

Third-Party Call Control


Third-party call control is more difficult to implement and often requires a dedicated
telephony server to interface between the telephone network and the computer network.
Third party call control works by sending commands from a user's computer to a
telephony server, which in turn controls the phone centrally. Specifically, the user's
computer has no direct connection to the phone set, which is actually controlled by an
external device. Information about a phone call can be displayed on the corresponding
computer workstation's screen while instructions to control the phone can be sent from
the computer to the telephone network. Any computer in the network has the potential to
control any phone in the telephone system. The phone does not need to be attached
directly to the user's computer, although it may physically be integrated into the
computer (such as a VoIP soft phone), requiring only a microphone and headset in the
circuit, without even a keypad, to connect to the telephone network.
Like many telecommunications technologies, CTI has made great advances over the past 10
years of evolution. The barrier to entry has lowered to a point that CTI is now available to small
and midsized businesses and is widely adopted as a competitive tool to improve business
processes.

26
Chapter 2

Application Integration with CRM and Enterprise Resource Planning Systems


As business advances, there has been enormous acceptance and adoption of an Enterprise
Resource Planning (ERP)-driven, Web-centered collaboration in business-to-business (B2B)
interaction. For many businesses, this is a competitive move. Companies around the world have
focused on improving business processes for many years, spurred early on by W. Edwards
Deming’s Out of the Crisis and his 14 Points for Management.
Those works, and others, drove business to improve processes, but with advances during the
same time period in IT, we’ve seen growth in the use of ERP systems across businesses of every
size and shape, creating more integrated enterprises. Today, efficiency in process drives many
organizations to focus on teamwork. ERP systems help dissolve the barriers that exist between
different business units or departments. They help break down the silo effect so common in the
past and change how people work together.
Web-centric business platforms and the success of e-commerce bring new levels of integration,
allowing easier support for inter-organization business process integration between trading
partners. ERP systems’ goal is to integrate all data and process within an organization into a
single unified system using computer hardware and software components. One key ingredient to
most ERP systems is the use of a single, unified, master database to store data for all the
disparate modules involved.
ERP originally implied a system for planning the use of resources across the enterprise, and
although it originated in the manufacturing sector, today it is much broader in scope. Today’s
ERP systems strive to encompass all the basic functions of an organization regardless of the
business. ERP systems have spread beyond manufacturing to business, non-profits,
governmental organizations, and other large enterprises.
The term ERP system generally refers to an application that replaces two or more independent
applications within an organization, eliminating the need for interfaces between systems. This
approach provides benefits ranging from standardization and lower maintenance (a single system
rather than multiple systems) to improved, dynamic reporting capabilities and enables managers
to better monitor the business.
Some examples of different modules that might be available in an ERP system include:
• Manufacturing
• Supply chain
• Financials
• CRM
• Human Resources
• Warehouse management

27
Chapter 2

Each of these discrete application modules support processes require both voice and data
communications. Integrating voice services into a CRM module provides a unified customer
database, easing customer contact to build relationships. Supply chain modules can build tighter
communications processes with vendor partners. Manufacturing and warehouse communications
integration can speed time to market. ERP represents the nervous system of an organization and
provides an integration point that can tightly couple business process with communications tools
to increase productivity and enhance efficiency.

Customer Relations and CRM


CRM encompasses a broad set of capabilities, methodologies, and technologies that support an
enterprise in managing customer relationships. The general purpose of CRM is to enable
organizations to better manage their customers through the introduction of reliable systems,
processes, and procedures.
CRM is a corporate-level strategy that focuses on creating and maintaining lasting relationships
with its customers. CRM tools look at the relationship between a business and its’ customers and
help manage relationship building and process improvement. Although there are several
commercial CRM software packages on the market that support CRM strategy, building value is
not about the technologies. CRM often represents a holistic change in an organizational culture
and philosophy, placing emphasis on the customer.
To be effective, the CRM process needs to be integrated end to end across marketing, sales, and
customer service. An effective CRM program needs to:
• Identify customer success factors
• Create a customer-based culture
• Adopt customer-based measures
• Develop an end-to-end process to serve customers
• Recommend what questions to ask to help a customer solve a problem
• Recommend what to tell a customer with a complaint about a purchase
• Track all aspects of selling to customers and prospects as well as customer support
As Figure 2.1 shows, CRM becomes central to both business process and IT systems. IT systems
become a valuable repository of information about the enterprise business, providing insights
into further improvements to increase productivity and profitability.

28
Chapter 2

Business Data
Activity IT Systems Data Mining
Warehouse

Knowledge Information
CRM Management

Informed Technology Intelligent Resource The Right


Choices Procurement/Allocation Service Bundles

= VALUE

Figure 2.1: Building value with CRM.

Customer relationships are managed by a variety of communications tools. There is an old adage
in sales that says “People buy from people.” CRM places the focus on the customer, without
whom no business can survive. Integration of CRM systems with VoIP technologies provides
tighter coupling of several aspects of relationship management.
ERP and CRM solution vendors recognize the criticality of voice communications and are
actively working on integration solutions. Because these systems are based on IP technologies,
there is a natural synergy with VoIP technologies, which have matured to the point that creative
convergence between the network infrastructure, services (voice and data communications), and
applications can now provide a seamless working environment.

29
Chapter 2

Delivering Call Quality with VoIP


Call quality is the single biggest factor leading to user acceptance of VoIP solutions. Later, this
guide will talk about managing call quality; this section will compare quality in the public
switched telephone network (PSTN) with VoIP, and see how voice conversations are handled.

Traditional Voice Characteristics—the PSTN


When you implement VoIP, you’re trying to use technologies that have evolved in new ways
over the past 30 years, specifically, IP. As you start adding voice traffic, a multimedia service, to
an IP-based network, you encounter a challenge to networking in general. In the past, networks
such as the PSTN were designed to perform specific functions and support a single type of
traffic. With VoIP, you’re using the Internet, or an IP network, to carry voice traffic.
The problem you face isn’t trivial. It actually has ramifications that ripple through every facet of
network engineering in both the Internet and in the PSTN. It’s a new problem that wasn’t
anticipated. In the past, when you had a task that required a form of networking, you designed a
new network to handle the task because different applications have completely different
requirements. IP changed all that because anything that can be digitized can be carried in the
payload of an IP packet.
Everything we understand about voice is based on how it’s been handled in the PSTN, so let’s
examine some basic facts about voice transmission:
• Voice calls require long duration for a conversation. A typical phone conversation lasts 3
to 4 minutes. Voice has historically been connection-oriented.
• Voice calls don’t tolerate network delay well. 50 to 100 milliseconds of delay is the
norm. Voice traffic is generally what is referred to as real-time traffic.
• Voice calls have traditionally been carried in a 4KHz voice channel, with the actual
sound frequency being carried ranging between 300 and 3300Hz.
• Converting analog voice traffic to a digital bit stream, you sample the voice at twice the
maximum frequency of the channel (8000 times per second).
• Each sample is coded into an 8-bit word.
• These 8000 samples multiplied by 8 bits give us 64Kbps of bandwidth, which matches
the telephone network architecture in the PSTN.
These basic facts bring us to the root of a number of the technical problems encountered today.
The public phone network has been consciously designed and optimized over a hundred years to
deliver voice traffic in the best way possible. This traffic engineering (in telco-speak) has been a
critical component of the management and growth of the PSTN. The telecommunications
industry has spent years analyzing the network, monitoring its growth, and designing into the
network countless optimizations to carry voice traffic effectively. It is a voice network that
guarantees delivery of the traffic within the users’ expectations.

30
Chapter 2

IP Traffic Characteristics
There is a packet switched public data network (PSPDN) that was specifically designed to carry
packet data. Private data networks have been built around the world to deliver enterprise data
applications. If you compare data networks, specifically IP networks like the Internet or the
PSPDN, you will find that they have very different characteristics than the PSTN:
• Long duration isn’t necessary because transactions are short in duration, or bursty in
nature. Packets are small in size and can be routed over different paths. IP packets carry
delivery information (addressing) in each individual packet. It is a connectionless
environment. At higher layers, TCP may be used to layer on connection capability, but
this requires the overhead of a three-way handshake and is commonly avoided unless
necessary.
• IP packets aren’t generally delay sensitive. Email messages sent over a network can be
delayed 30 seconds with no problems whatsoever. Web traffic delivered to a browser can
be delayed and loaded in the background. IP traffic is often non-real-time traffic. Delays
are expected in an IP network. IP itself doesn’t even guarantee delivery of the packet.
• IP uses the available bandwidth when it has data to deliver. There are 300bps modems
that send data and data transmitted over an Ethernet LAN at 100Mbps. IP doesn’t require
dedicated bandwidth. Again, IP comes with no guarantees of performance or delivery.
Table 2.1 makes the differences even clearer.
Voice Traffic on the PSTN IP Network Traffic
Connection-oriented—A dedicated path is Connectionless—Conversations are
established for each telephone call; calls packetized and transmitted over the
are long duration (4 minutes on average) best route based on routing protocols;
packets are small, so conversations are
cut up into many packets
Delivery is guaranteed once the call path Best efforts are made to deliver traffic
is established but there are no guarantees
Designed to use specific bandwidth; the Uses the bandwidth that is available
PSTN uses a 64Kbps voice channel
Real-time voice traffic is very sensitive to IP data traffic is delay insensitive
delay

Table 1: Voice and data traffic comparison.

31
Chapter 2

The chart is small, but it makes one very obvious point. IP was not designed to carry voice
traffic. IP was designed specifically to carry bursty data over diverse paths, and make a best
efforts attempt at delivery. Data can be freely discarded along the way if there are any problems.
This is nothing like the PSTN in the very design of the technology.
IP networks today are designed, redesigned, and modified to support carrying any kind of traffic.
Most often that is what is today called multimedia traffic—a combination of data, voice, and
video. Using digitization, any media can be carried inside an IP packet. The challenge in IPT and
video networks is delivering business-quality voice conversations with all the characteristics
users have come to expect, such as fidelity, clarity, and near-instantaneous delivery.
If users detect what seems to be unacceptable quality, time has shown they will most likely hang
up and try again. If this happens frequently, users become dissatisfied and quit using the service.
It’s clear that to effectively deliver voice services, tools and mechanisms are required for
measuring the quality of service (QoS) and guaranteeing that network resources can support the
traffic load without degradation of call quality.
One method to address call quality has been to overprovision the network. In many cases, this
simply means adding more bandwidth. Although this approach might work initially, in the long
run, it’s a road to ruin. First, it requires capital investment. Upgrading the capacity of
connections, switches, and routers can be a very expensive undertaking. This approach might
work for a time in a small local area network (LAN), but in a large enterprise network, it’s often
too expensive to be a practical solution. In addition, experience makes it clear that if the
bandwidth is available, users will fill it. Data applications can quickly consume the additional
bandwidth, leaving you with the same congestion and problems you started with in delivering
voice service.
To achieve acceptable voice quality over an IP network, factors such as noise, delay, echo, and
jitter must be managed to tolerable thresholds. Delay is present in every IP network due to the
statistical multiplexing used in routers and the generally bursty nature of packet data. There will
be delay. Jitter is nothing more than variation in delay. As different packets can easily take
different paths through a large network such as the Internet, variable delay is fairly common.
When transmitting voice, jitter can render the audio signal unintelligible to the human ear. Jitter
needs to be tightly controlled. Delay can also be caused by the time required to perform the
sampling in the codec used to digitize voice. Each router in the network can also induce
packetizing and routing delays as decisions are made about how to route each packet. This is also
referred to as nodal delay. When designing and managing an IPT service network, it’s important
to remember that delay is cumulative. Every place delay occurs, it adds to other delay in an end-
to-end service between two people.

32
Chapter 2

Design Considerations and Class of Service


Networks today are large and complex systems, with a variety of applications running
concurrently. They evolve and grow so quickly that they are almost organic in nature. There is a
concern for any enterprise about the spiraling complexity of network design. When you design a
specific set of criteria to support a given application, you create a class of service for that
application. The danger in doing so is that every new class of service you create increases
network complexity. An enterprise with numerous applications could quickly create an
unmanageable set of service classes. Many network designers lean toward simplicity and suggest
only a few critical service classes:
• Quick delivery can provide for real-time traffic such as voice and video—Real-time
traffic such as voice and video require quick delivery or the services are rendered
unusable. Video collaboration tools and VoIP telephone calls don’t work if the data
experiences needless delays. This class of service is used to tag those services requiring
quick delivery for usable service. It’s important to note that streaming video from a
server, such as watching a stored training video, is not real-time traffic. This traffic can
be buffered and delayed with lower QoS requirements. Real-time traffic is most often
person-to-person traffic between people. Person-to-system traffic generally doesn’t
possess the same requirements.
• Guaranteed delivery is suitable for mission-critical traffic—Guaranteed delivery is most
often used to support “mission critical” data. The CEO’s email isn’t what is meant by
mission critical. The data is what determines its criticality. Mainframe systems running
IBM’s SNA architecture require delivery but may be tolerant of delay. If packets are
delayed, within reason, SNA continues to function just fine. If they’re lost entirely,
session timers fail and applications no longer work properly.
• Best efforts delivery is what IP provides for everything else—This is the same delivery
mechanism used for all IP traffic today. Email, Web browsing, file transfers, and most
applications function perfectly well using best efforts delivery. In networks, most of the
network traffic falls into this class.
It’s easy to identify more classes of service in almost any network. For the sake of simplicity,
these three classes give you adequate prioritization capability in this guide and in many
networks. Although the debate continues over how many classes of service are really necessary,
the bottom line is that any implementation of a QoS mechanism that distinguishes service classes
is a vast improvement over the best efforts approach used by IP.

33
Chapter 2

Classes of service are only one aspect of QoS, but they aren’t the only factor involved. Let’s
think about how enterprise networks evolved over time:
• Businesses implemented standalone mainframes in the past, often to process accounting
information.
• Computing power grew, and user terminals were added to connect users to the
mainframe.
• With the advent of the PC in the 1980s, businesses started deploying PCs and
constructing small LANs. These LANs were disconnected. They were essentially islands
of information. Information was shared by carrying a disk from system to system (what
used to be referred to as Sneakernet).
• The Internet grew in popularity in the 1990s. Companies added routers to connect LANs
to each other and to the Internet.
• Many companies also developed an internal network, or intranet, to share company
resources and information.
• Public Web servers were added as the World Wide Web became a widespread resource.
E-commerce systems were implemented in many business sectors. Business automation
tools such as enterprise resource planning (ERP), customer relationship management
(CRM), and sales force automation were implemented.
• Today, real-time streaming voice and video traffic are being added to an already busy
network.
There was a long-held perception among vendors and service providers that customers’ networks
were poorly designed. In truth, they often weren’t designed at all. Like some unplanned cyber
“big bang,” they exploded and grew over time based on the needs of the business they supported.
It is important to understand that this happened for good reason. Companies in growth mode
were successful. This often led to mergers and acquisitions, producing exponential growth. The
economy boomed, and networks connected and merged quickly to meet the needs of rapidly
growing business. Sometimes networks were redesigned and new technologies were integrated.
In other cases, pressing business needs led to cobbling together the best that the beleaguered IT
staff could manage. Some of these networks were kept operating only through long hours of
frustration and ongoing reconfiguration.
Earlier, this guide noted that networks have typically been designed and optimized for specific
types of services. Voice traffic was the first networked application. The PSTN was tuned and
optimized to support narrowband voice channels with very low delay.

34
Chapter 2

Traffic engineering for voice is a highly developed field. Its foundation is built on traffic
patterns, the busy hour of the day, and statistical mathematics using something called the Law of
Large Numbers. This law basically says that large groups are easier to predict than small groups.
The larger the group, the more group members are likely to be near the average. For example, if
you measure 10 people’s height, you may find several people (a high percentage of the group) 5
feet tall. If you measure one thousand people’s heights, you’ll find that the statistical accuracy of
the sample is no longer skewed towards shorter people because the sample size provides for
greater statistical accuracy. Both telephone company central offices and enterprise PBX systems
are engineered for trunking requirements using a mathematical model known as the Erlang-B
distribution. Using Erlang-B, call loads are measured in centi-call seconds (CCS), which is a unit
equivalent to a 100-second call. 36CCS represents a telephony circuit at maximum occupancy
with zero idle time.

This guide won’t explore these formulas in any depth. The focus is the practical business of managing
IPT. For those interested in further study of traffic engineering, the basic formula is:
Calls / Hour × AvgMinutes / Call
OfferedLoa d = erlangs
60
The offered load in the formula isn’t a problem in a network such as the PSTN because the network
has been optimized to support the load. QoS is a manageable issue when the network is designed to
serve a single purpose. In today’s multimedia environment, there has been a phenomenon often
referred to as convergence grab center stage. Separate parallel networks don’t make either technical
or economic sense. As all of today’s data applications converge onto the IP infrastructure, the logical
next step in that integration is, for many, to integrate voice, video, and data on a single, multi-service
network.

As you look to IP networks as the next-generation architecture for delivery of multimedia


services, the fundamental design of IP leads to some complex issues for engineers. IP, and all
packet switched networks in use today, utilize time division multiplexing (TDM). This is a
statistical multiplexing technology used inside the switches and routers in IP networks. The IP
Suite protocols and hardware were developed to hand all traffic types, regardless of payload, in
the same manner. Routers and other network nodes have traditionally handled traffic not only on
a best efforts basis but often using a first-in first-out (FIFO) approach to processing the packet
flow.

35
Chapter 2

QoS
Classes of service describe what’s needed in the network. QoS characteristics look at the
technical aspects that make it possible to support those classes of service. There are several
characteristics within a network that may need to meet certain performance levels for any given
service or application:
• Availability—Availability is usually represented as the uptime percentage. It’s based on
the simple premise that the network is there and available for use whenever it’s needed.
In commercial telecommunications networks, the industry standard is referred to as 5
nines reliability or 99.999% uptime. This availability percentage equates to roughly 5
minutes of downtime per year. For many business networks, that is a challenging
availability threshold to meet.
• Reliability—Reliability is tied closely to availability, but it’s also a design factor.
Reliability includes network architecture features such as redundant paths and duplicated
or fault-tolerant equipment to ensure that in the event of a failure, the network remains
accessible in the event of an outage.

How necessary is reliability? A consideration is how much effort to put into building reliability into the
network. Does the network need redundant paths and fault-tolerant or high-availability equipment to
guarantee availability in the event of a failure? For businesses that only operate during the 9-to-5
workday, reliability may be an area of less importance.

• Throughput—Bandwidth is the most common measure used for throughput. This is


simply a measure of how much data inserted into the network at any given source is
successfully transmitted through to the destination on the far end over a specific period of
time.
• Error rate—Many applications are reasonably tolerant of data loss because the TCP/IP
suite relies on higher-layer protocols such as Transmission Control Protocol (TCP) to
request retransmission and ensure delivery in the event of errors. Not all applications are
delay tolerant.
• Delay—As described earlier, delay is simply a reality in IP networks because routers use
statistical multiplexing to process traffic and much of the transmitted data on any network
has a bursty characteristic. There will be some delay in any transmission network due to
the laws of physics. Naturally, a lightly loaded or over-engineered network will have
lower delay.

36
Chapter 2

• Jitter—Jitter is variation in the delay. Because packets can take different routes across the
network, delay variation in IP networks is common. Jitter in VoIP conversations results
in unintelligible conversations that sound “jerky.” In video transmission, the video stream
can break up and show visual signs of “jerkiness” as well. One common test technicians
have used for years is the jumping jack tests. The constant motion of a person doing a
few jumping jacks gives a good indication of the perceived quality of a video stream.
• Scalability—As companies grow and businesses change, the ability of the network to
grow with increased needs is an important consideration.
• Manageability—There have been many studies demonstrating that the most expensive
component of network services is the ongoing operational support of network
management and administration. Additions and changes to the existing network can be
very labor intensive, and many system designers now consider this factor very early in
the development process. A network with a 5-year life cycle will often cost far more to
manage over the course of its lifetime than the initial capital cost to build.

QoS Approaches in IP Networks


As noted several times, IP is a best efforts protocol that provides no guarantees. That is not to be
misconstrued to mean that the various parameters discussed so far cannot be obtained using IP.
However, you must remember that network design becomes a vital factor in any implementation
of converged services.
Although the techniques used vary widely, all approaches to QoS share a common characteristic.
Regardless of approach, with the exception of over-provisioning, QoS is simply a means to
implementing a prioritization scheme. In some cases, this prioritization might involve a routing
mechanism that aggregates similar traffic types, the routes each traffic type takes over a path
suited to the needs of the traffic. There are many approaches with intricate differences, but every
approach, save one, adds some form of overhead to the traffic to provide either a prioritization
function, or a “traffic cop” function to direct traffic appropriately. The one exception commonly
referred to as gigabandwidth has been seen by many to be the ultimate solution, but this fallacy
only defers the need for proper engineering to a future point in time.

37
Chapter 2

Ver IHL Type of Service Total Length

M D
F F
Identification Fragment Offset

Time to Live Protocol Header Checksum

Source Address

Destination Address

Options Padding

Data/User Payload

Figure 2.2: The TCP/IP packet structure.

Figure 2.2 serves as a reminder of the format of an IP packet, but it’s also shown to review one
field in particular. There is a Type of Service (TOS) field in the structure of the IP packet that
can and often is used as a QoS mechanism. This prioritization tool was provided in the original
specifications for IP.
The TOS field, expanded in Figure 2.3, is one octet, 8 bits in length. It consists of several
components:
• Precedence is used purely as a raw prioritization mechanism. The first three bits in binary
can represent a precedence value from 0 through 7. These bits provide eight possible
levels of prioritization for an IP packet. The higher the precedence value, the higher
priority assigned to the traffic.

IP still doesn’t guarantee a level of acceptability but merely a prioritization scheme.

• Delay (flagged as D in the figure) is a field to indicate whether the packet requires low
delay or can tolerate higher delay. A one would indicate low delay is required. A zero
would indicate more delay is tolerable.
• Throughput (flagged as T in the figure) is a relative indicator, with a one indicating the
need for higher throughput or more bandwidth.
• Reliability (flagged as R in the figure) is signified with a one to indicate that a more
reliable path is needed.
• Cost (flagged as C in the figure) remains generally undefined and misunderstood in use
or intent.
• Unused (flagged as U in the figure) is the last bit that remains available for future use.

38
Chapter 2

Precedence D T R C U

8 levels of precedence or
prioritization
-----
4 1-bit fields to further identify
packet requirements

Figure 2.3: The IP TOS field.

Although the designers of IP built this prioritization capability into the protocol, vendors rarely
implemented or took advantage of this feature in the past. For many years, router vendors created
OSs that didn’t even read the field when processing packets. This was due mainly to a complete
lack of standardization. The Internet Engineering Task Force (IETF), the IP standards body,
never assigned values to any of the sub-fields in the TOS field. Thus, a precedence of zero might
be the highest precedence to one vendor, and the lowest to another. Real-world implementations
vary greatly in the absence of a standardized approach, but the field has been used as part of a
prioritization scheme more in the past couple of years.
Once you accept that IP provides no guarantees of any kind for either prioritization or delivery,
the next step becomes clear. If a user needs to specify some particular network requirement, or
QoS, the user must add something to IP to provide it—add it in some other layer of the protocol
stack. The following methods are all in use today. They implement QoS in IP by layering some
other overhead into the data stream. There are other techniques used to provide QoS, but the
following sections explore three common approaches.

39
Chapter 2

QoS—the Signaled Approach


Integrated Services (IntServ) introduces a signaling protocol to IP. Using IntServ, the user’s
application sends some sort of call setup signal to the network. This signal is basically a request
for a set of service delivery parameters required to complete the call. Using the approach, the
network can check resource availability and deliver traffic accordingly. This approach is very
similar in process to the circuit setup and teardown associated with a telephone call on the PSTN.
Under IntServ operations, users, through automated software processes, request a particular type
of service of the network using Resource Reservation Protocol. RSVP has two key aspects—
policy control and admission control. Policy control is used to determine whether the user has
authorization to request the specified resources from the network. Assuming the user has the
necessary permission, admission control governs the process for allocating and reserving the
requested resources or setting up the path for the connection. RSVP passes the user requirements
from node to node through the network, requesting resources. If any node along the way is
unable to comply, the request is denied and no connection is established.
Real-Time Transport Protocol (RTP) is a widely used supporting protocol in IntServ
deployment. It’s used to assure the integrity of a session through timestamping and sequencing
of UDP segments. Real-time Transport Control Protocol (RTCP) is used in conjunction with
RTP and provides some level of monitoring capability.
Although IP theoretically has a rather comprehensive prioritization scheme available, IntServ
provides three levels of prioritization. They’re familiar because the classes of service identified
earlier have been derived from the development of IntServ:
• Best effort—This is exactly what IP provides in existing networks.
• Controlled load, which is described in the standard as providing “data flow with a quality
of service closely approximating the QoS that same flow would receive from an unloaded
network element.” This is equivalent to the guaranteed delivery class of service described
earlier.
• Guaranteed service, which provides a higher level of assurance, which might be required
for real-time traffic such as voice or video. This, although called guaranteed delivery in
IntServ, mirrors the quick delivery described earlier in this chapter.

40
Chapter 2

RSVP
RSVP provides a signaling capability that allows a user’s application to send a request to the
network reserving a particular set of requirements for a voice or video session. These
requirements are referred to as the template.

User applications are referred to rater than users because it’s important to understand that the person
using the computer will not be required to understand QoS or application requirements. The
applications will have requirements written into the code to automatically provide the service levels
needed.

RSVP allows user applications to identify three of the parameters reviewed earlier that might be
necessary for a given application to work properly:
• Throughput or bandwidth
• Delay
• Jitter
In Figure 2.4, the user application makes a request of the network. RSVP policy-control software
then evaluates the request against a permissions table and either confirms that the user has the
necessary permission to reserve the requested resources or denies access. Assuming the user has
permission, RSVP then engages admission control software to determine whether the network
has the necessary resources available to complete the request. The network passes the template
along the network from node to node in a PATH message until it reaches the receiver. Each node
must run an RSVP daemon to validate the request for QoS.
In simple form, this specification is a request passed from router to router identifying a need. If
an intermediate router can meet the need, the request gets passed to the next node. If not, the
request is denied. Intermediate nodes play a vital role in RSVP because every intermediate node
must be able to provide the necessary QoS PATH requested in the template. Lack of resources at
any point along the path can cause the call to be dropped.

41
Chapter 2

Sender Receiver
Router Router Router

Template
Template
PATH Template
Template
PATH
PATH
PATH

FlowSpec
FlowSpec
FlowSpec RESV
FlowSpec RESV
RESV
RESV

Figure 2.4: Using RSVP to reserve network resources.

Once a path has been identified, each node must pass a flow specification in a reservation
(RESV) message back along the network from node to node. As you’ve gone to the trouble of
reserving resources across the network, you now must ensure that you use the resource allocated.
This flowspec reservation is each router’s way of knowing where the next router in the reserved
path is to pass packets associated with the call or session.
There are clearly a several factors to consider when implementing RSVP for QoS. These factors
become crucial when using the Internet, as those routers are not all under the control of the
implementer.
• RSVP is overhead intensive. Call setup isn’t something IP was designed for. Adding a
signaling protocol in the mix increases the processing CPU cycles required at each router
in the path.
• Routing protocols such as RIP and OSPF support only a single routing metric; they don’t
understand the concept of reservations. RSVP does not provide any solution for this
issue. The implication is that to be successful, the network may have to be over-
engineered beyond expected carrying capacity.
• RSVP scales very well in the multicast network environment (one-to-many
transmissions) but does not scale well for unicast (one-to-one) traffic. As telephony is
primarily a one-to-one connection for an end-to-end service, scalability to support VoIP
and video conferencing services can quickly cause problems in a large enterprise
network. In a multi-provider network such as the Internet, where nobody “owns all the
routers, scalability may be unattainable.

42
Chapter 2

• As all the intermediate nodes must be able to comply with requests, every router in the
Internet would have to be either upgraded to support RSVP or replaced. This would
require universal support and acceptance that just doesn’t exist. Routers have to maintain
state tables containing information about every session. The processing load and memory
requirements can drive router cost up significantly.
• RSVP doesn’t provide any QoS whatsoever; it’s simply a mechanism for sending
requests to the network for some specific requirements. It’s a signaling protocol, and it
still requires help from other protocols or other methods to truly implement QoS.
Many network designers agree that IntServ and RSVP provide a potential solution but only in a
network that is completely under the implementer’s control. Each and every intermediate router
still presents a potential single point of failure. In the Internet, data crosses many provider
networks en route to its destination. IntServ can work well in a fully controlled environment, like
an enterprise-owned private network. It may also prove useful at the edges of the Internet, either
in customer networks or the local metropolitan portion of the Internet provider’s network.
Internet service providers have almost universally chosen not to adopt IntServ as a solution to the
QoS problem. The value provided just doesn’t offset the cost of conversion and implementation.
IntServ fails the ROI case for service providers. IntServ has found some application in private
networks and is still supported by all the major router vendors. Resource Reservation Protocol
for Traffic Engineering (RSVP-TE), a variation, has been deployed with success in conjunction
with Multiprotocol Label Switching (MPLS).

QoS—the Provisioned Approach


There is another approach that requires specific routes through the network that are predefined
and available for each type or class of traffic called Differentiated Services (DiffServ—also
called DiffServ Code Point or DSCP). These paths might be pre-existing as part of the network
design or they might be set up on demand in some manner. This second option might include
some type of signaling protocol such as RSVP to establish the paths. The DiffServ method is
often used as a traffic aggregation approach in order to direct similar traffic types on similar
network routes.
This guide won’t consider the evolution to IPv6 but will make an observation: In the
development of IPv6, the TOS field was the subject of direct discussion. In IPv6, the protocol
has been expanded to provide improved granularity in the delivery of QoS. In the IPv6 packet,
there is a field referred to as the DiffServ field as a replacement for the current TOS field. It uses
six bits of this field as the DSCP to identify how the nodes in the network should handle each
packet. Routers handle packets based on a set of forwarding treatments or Per Hop Behaviors
(PHB). These rules must be predefined in each network element or node. This is the provisioned
approach to QoS in the network.

43
Chapter 2

DiffServ is also a development of the IETF. The DiffServ working group charter and information is
available at http://www.ietf.org/html.charters/DiffServ-charter.html. The objective of this working group
is to employ “a small, well-defined set of building blocks from which a variety of aggregate behaviors
may be built.” This suite is also defined by a series of RFCs:
RFC 2386—Per Hop Behavior Identification Codes
RFC 2474—Definition of the Differentiated Services Field (DS Field) in the IPv4 and IPv6 Headers
RFC 2475—An Architecture for Differentiated Services
RFC 2597—Assured Forwarding PHB Group
RFC 2598—An Expedited Forwarding PHB
RFC 2983—Differentiated Services and Tunnels
RFC 3086—Definition of Differentiated Services Per Domain Behaviors and Rules for their
Specification
RFC 3140—Per Hop Behavior Identification Codes
RFC 3246—An Expedited Forwarding PHB
RFC 3247—Supplemental Information for the New Definition of the EF PHB
RFC 3248—A Delay Bound alternative revision of RFC2598
RFC 3260—New Terminology and Clarification for DiffServ

The DiffServ approach is to categorize traffic into classes of service. Similar services are
aggregated and treated the same way in the network. Thus, network paths have to be
preconfigured to support each class of service. Packets are “tagged” at the edge of the network
and the appropriate forwarding treatment for that class of service tag is applied. This approach
results in a much coarser granularity at each router and reduces the need for large state tables.
This helps control the need for CPU processing power.

DiffServ has two primary components described in RFC 2475. Packet marking redefines the TOS
field of the packet and uses six bits as a coding scheme to classify packets into a class of service.
The use of six bits provides for a prioritization scheme that can identify 64 different types of traffic or
aggregates.
PHBs govern how an individual class or aggregate is handled. This is defined via behavior
aggregates. In essence, the PHB describes the scheduling, queuing, and traffic shaping policies used
at a specific node for routing the traffic.

DiffServ can scale to very large enterprise and provider networks. It is widely supported by
manufacturers and has been deployed by several large service providers. There is far more to
DiffServ than can be addressed in this guide, but information and technical specifications are
easily located on the Web.

44
Chapter 2

QoS—the Bypass/Shim Approach


MPLS is a method of providing QoS that removes the usual hop-by-hop routing in IP from the
equation. MPLS adds a “tag” to every packet. This tag shortcuts the delivery by directing the
packet to the best available path for the type of traffic associated with the MPLS tag. MPLS is
compatible with frame relay and ATM networks. It has been widely adopted in enterprise and
service provider networks for adding QoS capabilities to support VoIP and video traffic. MPLS
has often been referred to as a bypass or “shim” approach to QoS. Because of the addition of a
tag into the data stream, it’s also often referred to as a Layer 2½ protocol.

Using MPLS for QoS


Implementing VoIP makes companies re-evaluate their service networks. Often, it brings some
fundamental changes to the way these IP networks operate. As has already been reviewed, IP is a
best efforts protocol with no assurances of delivery or quality. When organizations implement
VoIP, many businesses discover a compelling need to implement some QoS methodology in
order to provide acceptable voice call quality. MPLS is a method that allows packet-based
networks to essentially emulate some of the behavioral properties of a circuit-switched network,
such as the PSTN. The following section provides a look at MPLS as one method that’s
frequently used to deliver QoS assurance in a large-scale enterprise environment.
MPLS evolved from several different but similar techniques. One of the leading approaches was
developed by a group of engineers at Cisco Systems. Initially, this proprietary protocol was
called tag switching. As the technology evolved, industry leaders came together in a consolidated
effort under the auspices of an IETF working group pursuing an open standard that became
MPLS.

How MPLS Works


MPLS works by pre-pending packets with an MPLS shim header, or tag, to the beginning of a
packet. Although this guide is focused on IP networks, MPLS works equally well with frame-
relay frames and ATM cells in those networks. This shim header contains one or more labels and
is often called a label stack.

45
Chapter 2

Layer 3
IP Packet

4 byte generic “Shim Header” (1 to n)

MPLS Label Exp.(COS) S TTL


(20 bits) (3 bits) (1) (8 bits)

Layer2 VPI VCI DLCI “Shim Header”

ATM FR Ethernet PPP

Figure 2.5: The MPLS packet and MPLS encapsulation.

As Figure 2.5 shows, each label stack entry contains four fields:
• A 20-bit label value
• An experimental field, often used to denote class of service
• A 1-bit flag to signify whether this label is the last label in a stack
• An 8-bit time to live (TTL) field
These MPLS-labeled packets are switched by performing a label lookup/switch instead of a
lookup into an IP routing table. Because label switching can be performed within the switching
fabric of the hardware, it processes much faster than traditional router-based IP address lookups
process.
In an MPLS network, routers become label switching nodes that add and remove labels as
needed. You’ll hear the term label popping from time to time. Using this method, a large routed
network looks like one single routing hop at Layer 3 of the TCP/IP stack. Path selection is based,
not on a routing metric, but on the MPLS label. Routers may be thought of as Label Switching
Routers (LSR) through the core of the network, with Label Edge Routers (LER) at the border
points of the MPLS domain. Routers that provide ingress and egress to the MPLS are often
called Provider Edge (PE) routers.

46
Chapter 2

The Class of Service (COS) field is used to assign classes of service much as described earlier in
the chapter in a section that looked at three classes. MPLS often adds a fourth class for
management traffic. Typical MPLS classes of service are:
• Real-time traffic, such as voice and interactive video, is often given the highest priority to
ensure that adequate bandwidth can be provided. The class of service is also designed to
provide the delay, packet loss, and jitter requirements suitable for delivery of real-time
traffic such as VoIP and video.
• Mission-critical data traffic, such as that from mainframe computers, is often classified in
a guaranteed delivery class of service. Timing of delivery is often a more stringent
requirement than bandwidth for this type of data.
• Management traffic is commonly aggregated into a management class of its own.
Management traffic requires some assurances that it can still be passed regardless of
congestion in the network to ensure QoS is maintained.
• All remaining traffic is generally lumped into a best efforts class of service that mirrors
how IP networks deliver traffic normally.

The Experimental Bits and QoS


Using this approach, similar traffic types can be aggregated into the same class of service, not
unlike DiffServ. When addressing is applied to the packets, they might be labeled by an
application, a router, a switch, or some other mechanism at ingress to the network. This COS
header is used to aggregate packets into what is called a forwarding equivalency class (FEC) for
switching throughout the network. The undefined experimental bits are used by most MPLS
vendors to carry the latest 3 of 5 DSCP priority bits in the IPv4 header, as a simple emulation of
traffic prioritization.
When a labeled packet arrives at an MPLS router, the topmost label is “popped” and examined.
Based on the contents of that label, a swap, push, or pop operation is performed on the packet’s
label stack. Routers can have predefined lookup tables so that they can process the packet very
quickly. In a swap operation, the label is swapped with a new label, and the packet is forwarded
along whatever network path is associated with that new label.
In a push operation, a new label is pushed on top of the existing label, effectively encapsulating
the packet in another layer of MPLS. This allows the hierarchical routing of MPLS packets.
Notably, this is used by MPLS VPNs.
In a pop operation, the label is removed from the packet, which may reveal an inner label below.
If the popped label was the last on the label stack, the packet leaves the MPLS tunnel or domain.
This is usually done by the network egress router.

47
Chapter 2

LSR:
InPort: 3
InLabel: 34.22
Dest: 114.1 Egress LSR:
OutPort: 1 InPort: 3
Outlabel: 43.18 InLabel: 43.18
Dest: 114.1
REQ: OutPort: 1
Ingress LSR: 114.1
InPort: 3
Dest: 114.1 LSR IP Network
OutPort: 1 MAP: 114.1
43.18 LSR
1

Outlabel: 34.22
4.

2
.2
11

34
Q:
RE

:
AP
M

IP Network
114.3 LSR
LSR IP Network
LSR 114.2

TO: 114.1.3.2
LSR

Figure 2.6: Label switching with MPLS.

During these operations, the contents of the packet below the MPLS label stack, the IP packet
and payload, are not opened or processed. Intermediate transit routers only need to read the
topmost label on the stack. Packet forwarding decisions are based on the contents of the labels,
providing protocol independent packet forwarding. There is no need to look at a protocol-
dependent routing table. This eliminates the cumbersome IP longest prefix match routing
protocols require at each hop.
When the packet leaves the network at an edge router, and the last label has been popped, only
the payload remains. This can be an IP packet or any of a number of other kinds of payload
packet in other networks. The edge router must therefore have routing information for the
packet’s payload because the router must forward the packet using traditional routing methods.
An MPLS transit router has no such requirement.

For more information about MPLS, check out the following sites:
MPLS Resource Center at http://www.mplsrc.com/index.shtml
MPLS MFA Forum at http://www.mplsforum.org/
IETF MPLS Working Group at http://www.ietf.org/html.charters/mpls-charter.html
IETF RFC 3031 at http://www.ietf.org/rfc/rfc3031.txt

48
Chapter 2

Comparing MPLS to IP
MPLS can’t be compared directly with IP as a separate entity. They are complementary
protocols. MPLS works with IP and IP’s interior gateway (IGP) routing protocols. MPLS brings
a level of basic traffic engineering capability to IP networks. MPLS relies on traditional IGP
routing protocols to construct the label forwarding table, and the scope of any IGP is usually
restricted to a single service provider for stability and policy reasons. There isn’t a current
standard for interoperable carrier-to-carrier MPLS, so it’s not yet practical to span one MPLS
service across more than one carrier.

MPLS Deployment
MPLS is currently in use in large IP only networks and is standardized by IETF in RFC 3031. In
practice, MPLS is mainly used to forward IP packets and Ethernet traffic. Major applications of
MPLS are telecommunications traffic engineering and MPLS VPNs.

Traffic engineering considerations with MPLS include:


IGPs mainly use “shortest path” algorithms
Path “overlap” causes congestion in the network
Traffic can overwhelm a short path while others are underutilized
IGPs in large networks present challenges because equal-cost multi-paths (ECMPs) share loads
when they really should not, load sharing doesn’t happen on multiple paths with different costs, and
IGP metrics modified for traffic shifts tend to have side effects

Summary
This chapter began by considering high-level the business models and processes because they’re
most often the ones directly affected by VoIP. They’re also central focal areas as the concept of
unified communications brings network services, such as voice and video, and applications
together. The chapter then moved into factors crucial to providing call quality. The network can’t
effectively converge and thrive if suitable quality isn’t present. The next chapter will delve into
drivers for change in three areas: business drivers, const reduction drivers, and strategic drivers.

49
Chapter 3

Chapter 3: Business Drivers and Justification


In business, there are many factors to consider when addressing the shift in networking
paradigms to bring about what many today call Web 2.0 integration. Even the phrase “Web 2.0”
has spun off variations like Voice 2.0 and Office 2.0. The next generation of network-centric
solutions is a key business driver today.
In the earlier days of these technologies, convergence was viewed as a cost-reduction technique
for business. Although cost reduction is important and remains a driving factor, it has become a
minor factor in voice and data integration for many businesses. For large enterprises,
convergence brings about a unified, single-network bill from the carrier. Integrating voice and
data can lead to consolidation of staff as telephony and data services converge onto a single
infrastructure. This early driver has proven a factor only in the largest of enterprises.
Today, what integration brings about is a competitive edge. And sometimes it brings revenue to
the bottom line because it enables new revenue streams that couldn’t be fully captured in the
past. VoIP services coupled with Customer Relationship Management (CRM) tools in business
bring responsiveness, speed, and knowledge about customers that can provide a measurable
differentiator in customer service delivery. Years ago, Bill Gates articulated a strategy of
business knowledge being central to a company’s digital nervous system. Today, service
integration couples knowledge with responsiveness to deliver business solutions more quickly
than ever.
Convergence of voice and data tightly couples business intelligence with business relationships.
The real success of Web 2.0 lies not in technology alone but in integrated, comprehensive
business services. Chapter 2 looked at some of the technology drivers for convergence. Building
on that foundation, this chapter will explore business motivations for service integration and how
they may mirror the technological drivers.

Vertical Market Business Drivers for Change


In his book Telecosm, George Guilder coined the phrase disruptive technologies. In the
telecommunications industry as a whole, there have been many disruption points, including the
migration from circuit-switched technologies to packet-switching and advances in optical
networking in the core and broadband access technologies such as DSL, cable, and EVDO
wireless broadband services.
Let’s take a brief look at disruptive technologies and variations that aren’t directly related to
unified communications. Technology change enables business process change, but these changes
often occur at different rates. The adoption cycle for new technologies is fueled by success
stories of early adopters. In the Internet today, how you learn about new technologies, advances
in existing methods, and the success stories that drive business process has changed.

50
Chapter 3

A Side Note on Blogs and Wikis


A Weblog or blog is a Web site presented in something of a journal form. Entries are most frequently
displayed from newest to oldest (in a reverse chronological order). Blogs are often focused on a specific
subject or interest area, such as technology, a hobby, or regional events. Some blogs are simply online
diaries, but others represent citizen journalism or focus on tightly defined interests, such as VoIP. Many
blogs use a combination of text, photos, videos, and recorded podcasts as information-sharing methods.
Blogs have become a key component of the Web 2.0 evolution in unified communications. They are a
way to communicate. Blogs began as a hobbyist’s tool for sharing journal-like information. Today, many
companies and corporate executives use blogs as a tool for sharing information with readers and
customers.
A wiki is another kind of Web site that allows visitors to add, edit, and delete information. They can be
easy to use, simplifying interaction and delivering an effective collaboration tool for multiple contributors.
One of the most useful examples is the online encyclopedia, Wikepedia (http://www.wikipedia.org/).
Wikis are rapidly becoming another tool used in business to interact with customers and business
partners. They can provide a framework for collaboration among developers or early adopters as well as
a brainstorming mechanism when working with partners in developing new products and solutions.

In the leading edge of net-centric business, the buzz-phrase Web 2.0 captures the attention of the
most competitive business leaders. Blogs and wikis permeate the business environment with one
root motivation—the need to share information and support relationships. Many companies,
especially in the technology sector, use these tools to partner with customers through online
discussion, dialog, and collaboration. This revitalized interest in interactivity is recognized as a
highly competitive edge. Open dialog often brings a degree of personalization to an otherwise
faceless corporation. Done well, these new tools couple with more traditional business systems
to improve results.
These tools also enable vendors to share their new technologies. Blogs and syndication tools
such as Really Simple Syndication (RSS) let businesses follow trends and advances more closely
than ever. These Web 2.0 trends disrupt every type of business in some way. For some
companies, they’re a tool for sharing information; for others, they present a tool for learning
about how to be more nimble and use emerging technology trends to the best competitive
advantage.
Much of the business disruption in the market has been driven by record numbers of small
businesses adopting the Internet as an asset. This adoption allows small companies to level the
playing field with larger corporations. There is a famous cartoon from the Saturday Evening Post
that portrays a key truth that “on the Internet, nobody knows you’re a dog.” In more practical
terms, on the Internet, the fact that you’re operating a small business, or solo enterprise from a
home office, may well be completely invisible.
What was learned from the small business trend is that by 2003, roughly 70 percent of small
businesses were leveraging network technologies, according to a report from IDC. And
according to the “American Business Journal,” small businesses that use the Internet have grown
46 percent faster than those who don’t. Although these numbers don’t carry across to larger
businesses in linear fashion, Internet technologies have clearly become a vital business tool. In
this chapter about business drivers and justifications, it’s important to remember tools such as
blogs and wikis as business tools that can be used to integrate voice and data solutions more
tightly.

51
Chapter 3

Implementing new unified communications technology is similar to implementing anything new.


Success requires a methodical approach. Integration occurs not just in technology but in business
process as well. What these new tools enable is process change, but for many businesses, this
change ripples through the entire company, forcing many companies to reinvent themselves in
new ways. To succeed, you need to lay out a roadmap you can follow to guide the evolution
from the past to the future. You need an integrated roadmap to bring unified communications
technologies of today and the next generation into your business process. Figure 3.1 provides a
high-level view of one such integrated roadmap.

The Integrated Roadmap


Busin
Business Vision
ess Implementation
Current Business Strategic Tactical
Strategy Desig Deliverables Deliverables
n
I
Business Fram
Value Assess Bus Strategic Plan m
ewor Justification

Business Strategy
Model Proposition p
k Requirements (Refine Bus
Strategy) l
Target Choose e
Industry & Markets Business & Info Design m
Modify
Market Components Management,e
Assessment Capabilities & Business n

Roadmap
Services Choose Information, t
People & Management & OperationalP
Resources Products & Operational Resources & Solution l
Staff
Assessment Solution Components a
Components n
Technology Historical and
Scan Review
Forecasted
Performance

Funding Review Key Strategic Cost


Assessment & Timelines & Governance
Processes & Controls Controls
Bus Case Policies

IT Model
Assess IT & E-Business I
Value E-Bus E-Bus m

IT Strategy
Proposition CRM / BI Locations &
Requirements Data WarehouseDMZ pl
Knowledge e
IT & Determine IT m
Value E-Bus Vision Management
& E-Business Connectivity & e
Proposition (B2B, B2C, Topology
Solution nt
Intranet, Components Architectures Pl

Roadmap
Extranet, aModify
Data, S/W
Internet) Determine n
People & Applications Transition
Resources Management Plans
Targeted & Operational
Assessment Infrastructure
Capabilities Solution
& SLAs Components
Components Design
Technology including
Scan Key Projects Systems Operational
Historical and
Components
Forecasted
Performance
Review
Funding Resources & Cost
Review Key Staff Controls
Assessment & Processes &
Timelines & Governance
Bus Case Policies

E-Business,
Update Bus
Case

Business Intelligence Network


Model Value
Proposition
Assess
Network
Refine
Network
Locations Im
pl
e Network Strategy
Customer Relationship
Requirements Strategy m
Network en

Roadmap
Value Vision Review Technology Connectivity &
Solutions Topology t
Proposition Network
Pl
Targeted Strategy an

Management,
Capabilities,
Services Review Modify
People & Network Design Network
& SLAs Network
Resources Architecture Mgmt. & Oper.
Components
Assessment Components
Key Projects
Review

Knowledge Management
Resources & Performance
Technology Operational
Scan Components Staff

Key Processes
and Policies

Portals
Funding Update Cost Review
Assessment Controls
Timelines & Governance
Update
Bus Case

Master Project
Plan

Figure 3.1: Building an integrated roadmap.

Figure 3.1 shows a very high-level view of the cradle-to-grave process for implementing new
technologies. Although this sort of process isn’t the focal point of this guide, it provides a good
foundation for thinking about the upcoming discussion in this and later chapters. Let’s briefly
touch on the following four key components to provide some framework and food for thought in
developing strategies for implementing unified communications technologies in business efforts:
• Business strategy roadmap
• IT strategy roadmap
• Network strategy roadmap
• Master project plan
Each plays a key role in the success of leveraging unified communications for success.

52
Chapter 3

Figure 3.2 doesn’t present anything new but shows the many factors that come into play when
mapping out a business strategy. This is really just fundamental sound business planning, but it
also sets the stage for how a scan of technology—through watching the vendors, reading blogs,
and exploring new ideas—must complement the larger corporate vision. Technologies alone
don’t provide sustainable business processes. They complement your strategy, mission, and
vision as a corporation.

The Business Strategy Roadmap


Current Business Business Vision Business Strategic Tactical Implementation
Strategy Design Framework Deliverables Deliverables

Business Value Assess Bus Strategic Plan


Model Tactical Plan Implement
Proposition Requirements (Refine Bus Plan
Strategy)
Target Choose
Markets Business & Info
Industry & Components Justification Design
Market Capabilities & Management,
Assessment Services Choose Business
Management Information,
Products & Operational & Operational
People & Solution Solution Modify
Resources Components Resources &
Staff Components
Assessment
Historical and
Forecasted
Performance
Technology
Scan Review Key
Processes &
Policies

Funding Strategic Cost Review


Assessment & Controls Controls
Bus Case
Timelines & Governance

Figure 3.2: The business strategy roadmap.

The business strategy roadmap is the fundamental baseline. It ensures that you focus on the core
business. It’s crucial that you use the business strategy to drive your other strategies rather than
allowing the fanciful attraction of new technologies to drive your business direction. Unified
communications must enhance and support the business strategy, for companies both inside and
outside of the tech sector. The key for many companies is to remember that new technologies
may present more of a temptation than exist as a true business driver. Instead, they should be
viewed as tools to enhance what you do.
The IT strategy roadmap in Figure 3.3 is a natural progression from the business strategy. Again,
a scan of technologies is a key piece, but now you must really look at how these technologies
add value to your business strategy. This area is where you determine your approach to e-
business, CRM, Enterprise Resource Management (ERP), and other systems tools.

53
Chapter 3

The IT Strategy Roadmap


Current IT IT & E-Business IT & E-Bus Refine IT E-Business Implementation
Strategy Vision Design Framework Solutions Strategy Design

Assess IT & E-Business


IT Model Value E-Bus E-Bus Implement,
Proposition Business Locations & Deploy and
Requirements Intelligence DMZ Realize
IT & Portals Solutions
Determine IT CRM
E-Bus Vision & E-Business
Value (B2B, B2C, Knowledge
Solution Management Transition
Proposition Intranet, Components Plans
Extranet,
Architectures
Internet) Determine Data, S/W
Management Applications
People & Targeted & Operational Asset Mgmt,
Resources Capabilities Support Modify
Solution
Assessment & SLAs Components Services
Infrastructure
Key Projects Historical and Components
Forecasted including Design
Technology Systems Operational
Performance
Scan Services
Review Key Components
Processes &
Policies
Funding Resources & Cost Review
Assessment & Update Bus Staff Controls
Bus Case Case
Timelines & Governance

Figure 3.3: The IT strategy roadmap.

When delving into unified communications, as with several other areas, overlap beyond the IT
strategy into the network strategy (see Figure 3.4).

The Network Strategy Roadmap


Current Network Network Network Network Network Implementation
Strategy Vision Design Framework Solutions Design

Network Value Assess Refine Locations Implement


Model Proposition Network Network Plan
Requirements Strategy
Network
Vision Review
Value Network Technology Connectivity &
Proposition Targeted Strategy Solutions & Topology
Capabilities, Project
Services Review Initiatives
& SLAs Network
People & Components Design Network
Resources Visionary Network Modify
Architecture Mgmt. & Oper.
Assessment Initiatives Review Components
Operational
Components

Technology Resources & Performance


Key Processes Staff
Scan and Policies
Update

Update
Bus Case Cost
Funding Review
Assessment Controls

Timelines & Governance

Figure 3.4: The network strategy roadmap.

54
Chapter 3

The network strategy really drives how you will manage internal network services. In a small or
midsized business, the IT and network strategies may blend into one unified plan, but for a large
enterprise, the IT infrastructure of application services might best be separated from network
planning. The two complement each other but often have different characteristics. They are
separated here because in most organizations, the IT strategy has already been deployed to
support the business strategy. Unified communications, especially in a large enterprise, is most
often deployed to leverage existing IT resources. This roadmap discussion simply helps focus on
the context of why explore convergence and integration to support the entire chain of business
processes.
The last piece of the high-level view is the master project plan. The key is that the project plan
comes last in the strategic planning cycle. Whether the roadmap planning described here is
followed in excruciating detail, through formal or informal processes, is really a decision for
each organization to make based on existing management methods.

The Master Project Plan

Figure 3.5: The master project plan

The purpose of this review has really been to caution those who are enamored of unified
communications technologies because they appear, on the surface, attractive. In order to leverage
the convergence of voice, video, and data communications, it’s absolutely critical that your eye
always be focused on answering the question: What business are you in? You’re pursuing
integration to improve your processes, create new efficiencies, and better compete in the core
business.
Let’s move forward and take a look at some of the different business areas and why the
converged network really brings value. Later chapters will explore the operation and
management of the converged network; when exploring these areas, keep in mind that you
manage the network to support the needs of the business while avoiding the danger of letting
technology consume more mindshare in your attention than is necessary.

55
Chapter 3

Business Sales
Business sales have evolved into what is today B2B electronic commerce. Business partner
transactions often take place in much higher volumes than consumer transactions. Business
process automation has driven process change for increased efficiency across all sectors. In the
business world, everybody is a customer and everybody is a provider.
This chapter is about the changing landscape of voice and data integration, but those changes
have been explosive in nature since VoIP first came on the scene 10 years ago. The Web has
caused a “big bang” explosion in managerial and business process. The enterprise has exploded
from a compressed mass into many little bits. There are continuing advances in distributed
computing. Business units in enterprises are embracing peer-to-peer functionality rather than the
historical silos of total separation. And the top-down managerial and business hierarchies of the
past are vanishing, giving way to meshed, distributed departments and processes.
This e-business has become more standards-based, with UN/EDIFACT being one widely
adopted standard; other standards, such as E-Business XML (ebXML), continue to emerge. The
impact in the converged network is now lapping into many areas of business as the idea of a
service-oriented architecture (SOA) begins to couple with other business models. SOA really
focuses on loosely coupling software services, such as CRM and ERP, with one another,
allowing convergence between services on the network. Another view describes this approach as
Software as a Service (SaaS).
SaaS treats software applications of all types as network services. It’s more a delivery model
than anything else. One unifying trend in converged communications today is that voice services
are rapidly moving from the legacy technologies of the PBX to the IP network using VoIP. In the
converged network, voice is just another service, an application like ERP or CRM. Video is
quickly following voice. Treating communications technologies as a service on the network,
enables coupling them with other business applications.
What began as e-commerce has evolved into a broader e-business. Why? You can look to a
number of case studies for the details, but operating cost savings and net gains to the bottom line
have been huge drivers. Internet procurement of both goods and services continues to rise.
Eliminating the middleman in the process reduces overhead and bridges the gaps between
suppliers and customers.
Technology has also enabled the creation of virtual teams and workgroups on the fly and on
demand. In short, value is being redefined in the world of e-business. Communications
technologies can blur the lines between both vertical markets and lateral players within a market.
E-business drives new, flatter, dispersed business hierarchy. This drives efficiency, and
improved efficiency always helps the bottom line.
Integrating business sales systems with communications systems can bring a new competitive
edge by simplifying and tightening processes. For example, an ERP system integrated with a
VoIP system allows easy features such as click-to-call. There are many places within the entire
business vendor management chain of events that converged communications can streamline and
enhance.

56
Chapter 3

There are some obstacles to network convergence. They include:


• Regulatory and legislative issues
• Suitability of network infrastructure to support business needs
• Network reliability and security
Technology always advances faster than legislation can keep up, yet there have been regulatory
advances. Security will always be a concern, and will always require constant attention.

Later chapters will dig into security more deeply.

As customers, we want control. We want technologies that adapt, evolve, and offer convergence.
We want independence to use the technologies, both networks and devices, in ways that support
our core business. And with the evolution of the Web 2.0 mindset, we require the ability to create
global communities of interest on demand.
From a technology perspective we want powerful servers and workstations with unlimited
bandwidth and universal access to support survivable, secure networks that are easy to install and
maintain, and cheap to operate. When we step back to consider those demands, where we are
today in the convergence life cycle shows extraordinary growth from where our technology tools
were even 3 or 4 years ago.
Convergence technologies often focus on business sales—selling either products or services. The
enablers include:
• E-commerce—payment mechanisms
• Network architectures for Internet connectivity
• Voice and Video over IP
• Client/server architecture advances
• Security methods
• Business process modeling
These components have all undergone continued advances that are being leveraged to create a
converged environment of unifying communications tools with data systems and business
processes.
As Figure 3.6 shows, convergence brings you into a new life cycle that drives business activity
into your information systems, which help you better understand your customers. As your
technology converges with your CRM resources, the company converges to an ever-evolving
service or product for market, which, in turn, drives new business activities into the cycle.

57
Chapter 3

Business Activities Database Content Data Warehousing

Knowledge
Management Information Data Mining

Understanding of Technology
CRM
Customer Business Convergence

Services Convergence Company Convergence

Figure 3.6: The business-technology convergence life cycle.

Another way of looking at this is to look at the historical evolution of business. In the
industrial/manufacturing era, business was driven by supply and demand. What has changed?
Supply was a need for physical resources in the manufacturing process. Mass production could
only be fueled by quicker delivery of “widgets.” The industrial-age corporation survived by
striving for efficiency in manual processes. As information systems improved, automation led to
“just in time delivery” as an effort to cut costs and improve time to market. These improvements
brought the idea of virtual organizations, even virtual companies, into play—tightly coupling
companies with both supplier/vendors and customers.
Figure 3.7 shows the current step in the evolution of technology convergence: a thriving e-
business community that is global in nature, fueled by an abundance of knowledge. Today, we
know more about our customers, our partners, and our competitors than ever thanks to the
advances in computing and networking power. To win in the market, it is necessary to couple the
convergence power of integration voice and data to enhance the services delivered to customers.
The converged business becomes a leader in the global e-business community.

58
Chapter 3

E-Business Community

•Customer-Driven
•Service-Enhanced
•Customization

Creating
Value Virtual Corporation
•Extended
•Tightly Coupled

Industrial Age
Corporation
•Vertical
•Supplier-Driven
•Fully Integrated
•Mass production

Scarce Resources Abundance of


Physical Knowledge
Resources

Figure 3.7: The e-business evolution.

Web-Enabled Business
The Web enabled a new kind of business. Web services are often used to implement SOAs. They
enhance machine-to-machine integration, using tools such as SOAP-XML, and can couple
trusted business partner systems to speed business transactions, which speeds time to market or
delivery.
The Web has allowed outreach to a new customer base for many companies. Existing customers
may have stuck with existing tools, but the Web created many “little wins” for businesses that
didn’t wait for large breakthroughs. Recombining existing technologies with new technologies
can provide enhancements that increase the competitive edge. These incremental improvements
have assured that small successes offset small failures, so major fatal failure didn’t occur as
technologies advanced.
Many Web-enabled businesses also use some form of call center technology. Call centers
provide a tool for answering or originating high volumes of telephone calls. Earlier, this chapter
discussed how technology can’t be the driver for the business, but VoIP call centers give an
excellent example of why technology can and should drive change in the processes you use to
support the business. Call centers provide a great example because, in varying degrees, they
reach across every vertical segment of business.

59
Chapter 3

Why an Internet Call Center?


With the advent of near-ubiquitous broadband Internet, there has been a rise in the number of
teleworkers. Technology has allowed people to focus on family and friends in new ways and to change
how they participate in their work environment. Internet call centers provide an approach that offers a
work-at-home job but with far more connectivity that we’ve previously seen.
Call centers have historically moved jobs from metropolitan areas to other locations that offer a lower tax
base to businesses. Many US companies now use offshore call centers in other countries where the labor
rate is lower, but the distributed call center approach alters the cost structure and provides an effective
method to hire domestic staff around the country.
The leading approach for US-based home agents has historically been the Integrated Services Digital
Network (ISDN), which really doesn’t provide effective integration of services. ISDN costs have proven
too high for most companies to invest in this architecture. VoIP is another story. With VoIP, companies
leverage significantly reduced cost and far better integration of services than ISDN has ever offered.
Many companies now rely on VoIP coupled with consumer broadband to build an Internet call center.
The focus for any company doing business on the Internet has become customer service. There is
nothing customers appreciate more than talking to a live person in real time. The call center, in any form,
provides that seamless integration capability.
Customers might be anywhere in the world. They might be based in a large corporate campus or at
home. They may contact the provider via a Web site or simply by telephone. According to the Gartner
Group, more than 70 percent of transactions take place over the telephone. Web sites that offer live voice
as a support option for customers report as much as 50 percent increases in sales.
With VoIP softphone technology, the provider of products or services can receive a query for customer
support, and through distributed call center technology, seamlessly redirect the call and all data
information to a service representative working remotely, often from home. Today’s convergence of voice
and data allow not just handing off the telephone call. We can pass the customer account information,
populating the home-agent’s screen with all the necessary information to provide quick and accurate
customer support.
Staffing of call centers has always been a challenge. Hiring remote staff, perhaps even part-time remote
staff, lets the provider staff the distributed call center with qualified employees. Special services, such as
service representatives with special linguistic skills, become more obtainable. This solution also makes
available a pool of workforce candidates that may have been previously overlooked. Stay-at-home
mothers, retirees, and people without transportation now become potential job candidates, participating in
the workforce in new ways that were inaccessible previously.
Distributed call centers do not require VoIP to provide the convergence, but VoIP brings the tightest
coupling of services at the lowest cost. In the past, distributed call centers were implemented using PBX
solutions and off-premise stations or ISDN lines. Current converged technology solutions make the
distributed call center far more cost effective to implement today. IP technologies that integrate VoIP with
data systems simply use the network as a “PBX extender,” creating a virtual call center environment that
can physically be anywhere or everywhere. Most importantly, this distributed call center is transparent to
customers, who see a single, unified point of presence for the company.
Distributed call centers’ use of a technology for job performance requires that managers take a more
“hands off” approach to supervising workers than traditional workflow methods. The call center supervisor
does require a comprehensive set of tools to monitor both service delivery and employee productivity.
Supervisors can quickly become comfortable relying on integrated systems, both telephone and computer
networks, to measure and monitor productivity and worker activity. The idea of a worker being in the
corporate office where work can be directly observed is transformed into a measurement of productivity
and results rather than oversight of activity.
Although there has been a trend to move call centers offshore, today many companies have become very
security conscious and are more reluctant to engage in offshore arrangements. The Internet call center
based on VoIP technology can provide for substantial savings above traditional costs without sending
jobs outside the US. For those companies using offshore resources, IP telephony works as seamlessly
around the world as it does around a city.

60
Chapter 3

Product Sales
VoIP convergence brings creative new solutions to the retail sales market that are limited only by
imagination. Converged VoIP phones can easily interact with data network service, providing
SKU information about products either via a scanner or RFID tag reader. It’s also important to
note that VoIP phones today provide kiosk-like screen services and may not always provide
voice as a service. A VGA screen on a VoIP phone provides a powerful data sharing mechanism.
One retailer in Japan has leveraged this to maximize customer service. They’ve learned that if a
customer has to leave the changing room to find another size for an item, they are most likely to
abandon the sale and leave. This retailer equipped each changing room with a VoIP telephone
that can scan the SKU of a clothing item. If the customer finds the item doesn’t fit, they can
simply scan the tag. The system can then provide feedback to the customer on a touch-screen.
For example, “you have blue jeans in a size 12.” At that point, the customer can be presented
with a variety of menu options, allowing for another size, and even another color. Simple touch-
screen interaction allows them to make their selections.
A salesperson on the floor is alerted via a WiFi VoIP telephone that the customer in a particular
changing room requires black jeans in a size 10, for example. The salesperson can now take the
customer’s selection directly to the customer and hand it over the privacy screen. VoIP
technology provides information sharing that keeps the customer in the store and helps close the
sale.
Business selling products know well that making the act of buying as easy as possible for
customers is a huge success factor. VoIP alone is just a way of making voice calls. Convergence
of services and applications facilitates the delivery of completely new services and applications.

Service Sales
On the Web, customer service and support have always posed an interesting challenge.
Companies struggle with providing help and online support through screens that say “contact us”
and through frequently asked questions (FAQs). Some companies now provide interactive text
chat with service representatives. To date, most of what we’ve seen has been rudimentary, but
there is huge potential for innovation. The Web is powerful business tool that can indeed be
voice enabled to better serve the customer.
Figure 3.8 shows a Web page concept that brings voice directly into the Web page to allow
customer interaction directly with the support center service agent. Today, VoIP often requires a
softphone client to work—an application similar to the one show in Figure 3.8 requires that the
customer install some type of browser plug-in. But that is changing quickly. Development tools
such as AJAX, Ruby on Rails, and Flash are beginning to deliver new VoIP connections that
embed the VoIP softphone within a Web page or application.

61
Chapter 3

Enabling the Web with VoIP

Figure 3.8: The VoIP-enabled Web.

Convergence isn’t just about supporting the core services you provide; it’s about providing your
own business with services to support delivery of your customer-facing services. As Figure 3.9
shows, there is an inverse pyramid that makes up the services supply chain. At the top, is the
widest audience—the end clients. These may be retail consumers or business customers, but they
make up the largest piece of the service market.
Those clients interact with providers of services. But service doesn’t end there. Manufacturers
provide service to the providers and component suppliers provide services to them. And although
this looks like a supply chain management flow, what you have to remember is that in the
information age, although the product has changed, the process flows for business remain much
the same. Information is made up of component parts just as manufactured products. Component
suppliers in the converged network are data systems and those of business partners. The Web
page provides data input about customer trends that feed business intelligence. You need to
capture, process, and analyze information at every level to succeed.

62
Chapter 3

Clients

Services Providers

Manufacturers

Component
Suppliers

Figure 3.9: The services food chain.

In supporting the services food chain, you can see the potential for a supply chain of information.
As information passes through this supply chain stack, it evolves to become more valuable and
increase your competitive edge.
Raw data is the basic input—Ken, 360-555-1234, 6 feet 3 inches, Washington. As data is
processed through the stack, it becomes information—Ken is 6 feet 3 inches tall and lives in
Washington and he has a telephone. Further processing in the service stack matures information
into knowledge—It’s easier to reach Ken by phone than by email (we don’t know his email
address). At the highest end of services, we’ll know our customer well enough that we’ll turn
that system knowledge into human wisdom. Our systems gather and synthesize information so
that we can they make human judgments to guide our service interactions with clients.

Financial Services
Financial services present an opportunity to use every approach and technique described thus far,
with one simple variation. The financial services sector provides an incredibly high volume of
data inputs. Transaction data can present an enormous wealth of information for analysis. In the
converged network, there is a tendency to be very data-centric when working in financial
services, yet the call center provides a key human interaction with customers. Achieving balance
between the focus on the data and the focus on service remains a prime success factor.

63
Chapter 3

Health Care
In health care, patients become a focus of activity. Data systems provide information about them
in painstaking detail. This information ranges from address and Social Security number to
prescription drugs taken, medical history, and financial information. In the health care sector,
there is the added responsibility to protect this information and comply with the Health Insurance
Portability and Accountability Act (HIPAA), which addresses the security and privacy of health
data.
Health care offers many uncharted opportunities to use convergence technologies in new ways.
Patient record information can be retrieved automatically with the scan of a health plan
membership card. Hospitals rely more and more on wireless devices carried by staff, connected
to VoIP phone systems that are tightly coupled with data networking resources. The health care
professional needs a wealth of information readily accessible to ensure the best decisions
regarding the health and welfare of patients.
The converged network presents the best integration of information, technology, and efficiency
in delivering health care. This sector has proven to be far more than a user of converged
solutions. Health care providers lead the way in developing new solutions for integrating voice,
video, and data services, and delivering it to the health care worker immediately.

Manufacturing
The traditional supply chain flow of manufacturing is also impacted by the convergence of
services and applications. Much like the evolution to an e-business community, the linear supply
chain is slowly vanishing. It’s becoming a meshed supply chain model in which all players in the
supply chain interact as peers. This provides immediacy in interaction and redefines many
traditional enterprise functions, creating new opportunities for innovation.
Two models emerge from this new meshed supply chain model. One model focuses on high
volume and low price, creating economy of scale through volume sales. The other is a
differentiated model that uses the “higher-in-price-but-delivers-the-value” approach.
To maintain and keep customer relationship in the converged network supply chain, a
technology-based information and knowledge derivation infrastructure is critical. The supply
chain player has to gather and synthesize as much information as possible about both the
upstream inputs and the downstream outputs.
The supplier who has tightly coupled data with converged services will have ERP and CRM
systems integrated to provide customer service. An upstream supplier’s inability to deliver a
particular widget can effectively trigger alerts in the supplier’s CRM system to every key
customer who might be impacted downstream. Imagine the customer service impact of being
alerted that your deliverables will be a day late because of a component, and being able to notify
your key customers immediately.
Convergence gives supply chain players the tools to move quickly, be flexible, and adapt to
change. Being nimble and adapting is a competitive edge. It lets you use knowledge about your
environment as a differentiator. In short, the supply chain adds value, delivering not just widgets
or technologies, but solutions.

64
Chapter 3

Financial Cost Reduction Drivers for Change


The beginning of this chapter mentioned cost reduction as an early driver for convergence and
migration to VoIP technologies. Reducing costs in the corporate enterprise can come in several
different forms, and telecommunications costs have many different factors to consider.
Infrastructure for voice and data may be entirely separate networks that can now fold into one.
Although cost reduction is seldom the primary driver for convergence with VoIP, it’s a business
factor that cannot be overlooked. Cost reduction, coupled with delivering new converged
services, can drive revenue directly to the bottom line.

Local Telephone Expenses


One cost reduction area of focus in the past has been local telephone charges. For large
companies, using an architecture of PBXs distributed geographically across the enterprise, an
approach called least cost routing, has been widely deployed. In this architecture, local access
lines in remote offices, generally T1 circuits, presented a huge cost. Companies offset the
expense by routing calls across the internal network of T1 circuits interconnecting all the PBXs
to carry the calls internally, then drop them off on the Public Switched Telephone Network
(PSTN) as close to the destination as possible. Using this approach, companies circumvented
telephone company local access charges, effectively by using their own telephone network to
carry what might otherwise be “toll” traffic. Convergence of voice and data further reduces costs
by leveraging a unified IP infrastructure to carry all traffic, eliminating costly voice T1 circuits.

Long-Distance Expenses
Local and toll traffic expense wasn’t the only cost that could be reduced through local access
circuits. Long-distance calling could leverage the same method. Through creative deployment of
trunk circuits between PBXs in remote locations, companies achieved control over call routing.
This let many large enterprises control how calls were routed, using the cheapest rate available.
Many large enterprises built telephone networks rivaling the complexity of small telephone
companies in order to control costs.
Businesses meshed their PBX systems together using a combination of point-to-point circuits
and, often proprietary, PBX networking protocols. Automatic route selection intelligence was
then built on top of the connections, based on dialed digits. Although this approach could
effectively reduce telephony expense in a large enterprise, it created a very complex environment
requiring increase in staff to manage the voice network. In short, enterprises became their own
long-distance company.
This complexity was never a viable solution for many companies. The costs often failed to
sustain savings over time. Convergence again provides an architecture that puts all traffic on the
switched and routed IP network, eliminating the need for dedicated voice circuits and the
specialist telecommunications staff.

65
Chapter 3

Inside the Enterprise


Internal calls within the enterprise were no longer a problem given the approach to local and
long-distance costs. Companies used the same internal trunks between locations to carry inside
traffic. Typically, this was achieved via access codes as part of the dialing plan. To reach a user
in one city, you might dial a “5” followed by the four-digit extension, whereas to reach another
office in another city, you might dial a “6.” As technologies advanced, many of these interoffice
trunks also evolved. PBX manufacturers began shifting away from proprietary protocols, using
IP interface cards to move these trunk offices onto the IP network. In many cases, enterprises
simply used the bandwidth available in the data network to also carry voice traffic between PBX
systems.
This early step toward convergence often became a competitive issue between vendors who
touted “free VoIP” on the data network. This hyped approach caught some customer attention as
companies derived cost savings from using a single carrier to provide both data and voice over a
unified circuit infrastructure. This beginning of convergence is really only infrastructure
convergence——a small piece in the total unification of applications and services.

Outside the Enterprise


The telephone is the lifeline for most businesses. With recognition of the importance of the data
network and the value brought by tools such as email, the telephone network has a history of
reliability and availability that the data network, in many organizations, still doesn’t quite
provide. Telephone service is mission-critical.
When companies evaluate the costs of handling inbound and outbound calling, prudent fiscal
management requires business managers to dig deep into the pool of resources to develop new
cost-saving approaches. VoIP, tightly coupled with the IP network, offers a compelling cost-
savings potential that integrates with other advances in applications used today.

Support Costs
Looking at the techniques described earlier, one fact is obvious: enterprise business accepted the
cost of adding a telephone support staff to support creative methods to control costs. Although
the savings were often justified, maintaining a full data network staff along with a telephone
network staff duplicates many costs. It’s inefficient.
Many companies saw convergence as an opportunity to reduce technical staff by eliminating one
of the support groups, but reality dictates otherwise. VoIP allows convergence of the
architecture. It can allow for blending of staff, but don’t assume the VoIP service can be
designed and implemented solely by the IP network group in a large organization—or that the
need for telephony expertise will completely vanish.

66
Chapter 3

It’s deceptively easy to overlook the complexity of telephony engineering. The enterprise voice
support staff understands traffic engineering for the voice service in ways the data networking
team doesn’t. Peak telephone calling periods, the busy hour of the day, or the busy day of the
week, are an example of an understanding the data networking group won’t have. Beyond that,
interoffice connections of site-to-site networking for local calling, long distance, and interoffice
communications may be very different in voice traffic patterns than those supporting data flows.
There aren’t many data networking engineers who fully understand the inner workings of voice
traffic in an enterprise call center. Although VoIP technology lets you move the traffic over the
IP network, the underlying service is still voice, and it remains a mission-critical service.
Rather than eliminate a workgroup entirely, converge staff as you converge services and
applications. There may indeed be reduction, but to cut staff prematurely can lead to disastrous
consequences. Don’t be lured into cutting vital institutional knowledge about your services,
network, and business needs in a misguided effort to eliminate cost too aggressively. Cost
savings come over time. They don’t come immediately, and they aren’t necessarily a large factor
in convergence for every enterprise.

Help Desk Support


Help desk support in the telephone network seems simple. Most people don’t need training on
how to use a telephone. When businesses deployed a new PBX, training was often conducted the
day before implementation and consisted mostly of dialing shortcuts, how the new voicemail
system worked, and an introduction to pretty basic features.
When integrating VoIP into the existing IP data network, features can work in completely
different ways. VoIP phones often don’t have the same buttons as traditional telephones beyond
the basic dialpad. Most VoIP solutions introduce some data workstation, or computer, interaction
with the system. Features such as speed dial lists, call pickup groups, and voicemail controls are
often reached through a browser interface, coupling the PC with the telephone in a new way that
people aren’t familiar with.
As services converge onto this new unified communications infrastructure, the company Help
desk will get calls about new features, functions, and capabilities that they aren’t familiar with.
When planning a VoIP deployment, it’s important that both users and the entire support staff are
involved throughout the project. It’s the only way they’ll be able to fully support your employees
to minimize disruption to normal business operations.

67
Chapter 3

Adds, Moves, and Changes


In the traditional world of telephony, the most expensive part of ongoing operations has always
been adds, moves and changes to the environment. To move a computer, in many companies,
simply unplugging it in one cubicle and plugging it in to another worked. Many companies use
Dynamic Host Configuration Protocol (DHCP) to configure the IP network connectivity.
Telephone moves have been more problematic. Because a telephone number has historically
been assigned to the physical jack the phone plugged into, everything required reconfiguration
when an employee moved to a new office. Changing work groups might mean complete
reconfiguration of telephony services like hunt groups, pickup groups and call coverage.
VoIP convergence may offer tremendous savings here as the IP network is now linked to
telephony services. The telephone number and feature might now easily follow the workstation.
Moving an employee to another department might be as simple as drag and drop of the
employees name in a GUI interfaces to update telephone requirements.
Convergence can mean that when a new employee joins the company, the human resources
system provides all the input for the unified communications network so that telephone service is
tied directly to network login. When fully implemented, a new employee might be assigned to a
work cubicle and log in on their first morning to find all the voice and data services associated
with their position are preconfigured and available for use.

Remote and Mobile Workers


The advances in broadband access to the Internet have fueled an exponential rise in remote and
mobile workers over the past 10 years. VoIP convergence adds another dimension to the remote
workforce. As described earlier, call centers can be staffed by people who are anywhere in the
world. Previously inaccessible staffing resources from stay-at-home mothers to contracted
multilingual translators to those without transportation have become viable staff resources.
Enhancing the widespread availability of broadband Internet with the proliferation of Virtual
Private Network (VPN) technologies and VoIP services, employees and contractors can now be
anywhere in the world and appear to customers and business partners as if they’re at the
corporate office. Location has been removed as a barrier to getting work done in the information
age.
Teleworkers aren’t the only employees who benefit. Road warriors, such as sales people, on-site
support engineers, and traveling executives, also benefit from convergence. The company
business extension can easily be delivered anywhere with VoIP, meaning these people might
work transparently from a hotel room, a client site, or any remote location. All that is required is
Internet connectivity.
When a customer calls their sales representative, they don’t need to know that he or she is
halfway around the world attending a conference in a hotel. The office telephone number
becomes as transportable as an IP address, working anywhere. New advances in unified
communications, particularly in fixed mobile convergence, will bring this transparency even
further as the multi-mode handset becomes a device that can connect transparently anywhere
using ubiquitous wireless services.

68
Chapter 3

Strategic Drivers for Change


The strategic reasons for business change are many and varied. Some enterprises have unique
applications that apply only to a particular vertical market segment. Inventory and supply chain
management tools vary from market to market in manufacturing, but may not play at all in a
financial services or health care environment.
Shifts in the market drive changes in business strategies. Earlier, this chapter delved into
strategic roadmaps for business, IT, and network strategies. Continual process re-engineering
goes on in all businesses. New strategic applications appear. CRM systems evolve and grow to
offer new functions and features. Many companies are exploring how to adopt CRM and ERP
systems in a blended fashion, while smaller companies may look to application service providers
to integrate new strategic applications on a pay-as-you-go basis.
Whether you’re deploying a major SAP platform, developing an in-house Web application, or
integrating something like salesforce.com into business flows, any strategic change in data
applications will drive other impacts into your voice services. For many companies, the act of
integrating voice and data to a converged service network will be a key enabler for business
process change.

New Applications
New applications emerge daily, and many present new features previously unavailable. With
programming tools today—such as the .NET framework, Java, AJAX, and Ruby on Rails—the
speed to market for new applications is increasing. Things we couldn’t dream of doing in
applications last year are the realities of this year.
Application developers in today’s convergence market are often seduced by the lure of Instant
Messaging (IM), syndication (RSS), the idea of online presence (an indicator of a user’s online
status and availability), and how these all complement VoIP services. Certainly there is a danger
that this Web 2.0 development mentality of slapping together components will create problems.
As developers embrace these technologies, they quickly migrate to the more mature view of
SaaS. The SaaS model isn’t new, but speed with which converged applications move through the
adoption cycle from early adopters to broad consumer acceptance is increasing. We’re seeing
shorter time to market and shorter time to adoption in many unified communications
technologies. We can expect to see this trend continue, and for many businesses, these new
applications and service combinations will provide a strategic impetus for change.

69
Chapter 3

Obsolescence of Legacy Systems


Older telephone systems were based on Time Division Multiplexing (TDM), just like the public
telephone network. Many of these systems had a planned lifetime of about 10 years and were
amortized as such within the corporate financial systems. Although advances in technology have
left many of these systems appearing obsolete, they were still within a planned life cycle.
Data networks have been managed differently. Typically, the network elements for business
were smaller than a PBX. Data technologies in switching and routing have advanced rapidly.
Data networks are often redesigned every 3 to 5 years, and data networking equipment is
commonly amortized over 3 years in corporate financial systems.
What this means is that many traditional telephone systems are now reaching the end of their
viable life at a point in time when convergence and the benefits of moving to VoIP makes good
sense from both a technical and financial viewpoint.

Stop Investing in Legacy Technologies


Although obsolescence of a system or the fact that it’s amortized off the books in terms of
financial value may drive some companies, others, especially leaders within their market, simply
tend to look to the future. For these companies, the need to migrate to newer technologies may
be driven by other factors. Many companies stop investing in older technologies prior to the end
of life just to ensure they’re future-proofing the technological assets of the enterprise. Seeing the
end of life for a technology on the horizon may be enough incentive to pursue a change for many
organizations.

Manufacturer Discontinuation of Product and/or Support


Equipment vendors and manufacturers announce they are discontinuing support for old products
in the voice and data marketplace daily. Often, products have become costly to manufacture,
difficult to support, or parts have simply become scarce, complicating hardware repair.
In many cases, a vendor will make a conscious decision to declare end of life for a particular
product in order to motivate customers to purchase a new product the vendor is promoting. In the
traditional telephony PBX environment, this resulted in what was often referred to as a “forklift
upgrade.”
There is a danger for vendors to consider when taking this approach. Motivating a customer to
replace a system with something new might also motivate them to look elsewhere. Many an
incumbent vendor has found themselves losing a customer using this approach. Whether
telephony customers upgrade with their incumbent voice vendor or migrate to a new vendor,
some migration to VoIP will be driven by manufacturers who are discontinuing existing
products.

70
Chapter 3

Rising Support and Maintenance Costs


Aging systems cost more to support and maintain than new ones. Availability of parts may drive
what were once reasonable prices upward. The technical skills needed to support older systems
become less available as fewer and fewer of these systems are in operation. Sticking with a
technology, like a phone system, that is 10 years old, when the market has shifted to a newer
technology can become an expensive proposition. Several studies have shown that telephony
systems reaching 10 years of age may cost as much as 30 percent more to support and maintain
today that they did when they were new.
New systems, based on new component technologies, bring another operational support factor
into play. Every new technology also needs to be updated through its life cycle. When you
deploy a voice or data networking solution, a series of patches and upgrades follows for the life
of the equipment. New components allow for more efficient patching and upgrades, driving
support costs down on new technologies, while costs continue to rise on older systems.

Inability of Legacy Solutions to Support Business Needs


Another strategic driver for change is simply the practical limitation of hardware and software.
The hardware providing voice services today may not be able to support new software because
of limitations in CPU processing power. Port capacity may be constrained to an internal
architecture that can’t be expanded further. In short, the system may do what it was designed for
several years ago perfectly well. When telephony systems hit the market, they do so with a
combination of features, functions, and price point to serve a defined set of needs. For businesses
that evolve quickly, needs often outrun the ability of a system to evolve. Older systems may
simply be unable to support features and functions your business requires.
That brings up an important point about system evaluation overall. When evaluating a new
solution, remember that you’re making a strategic investment to support the corporate roadmap.
Supporting the operational needs of today only serves the present. For some companies, system
extensibility may be a key factor over 3 or 4 years, while others may strategically plan further
ahead. Look at the enterprise tactical plans for today and strategic plans for tomorrow and
evaluate solutions based on a full set of requirements for the present and the future.

Summary
Network convergence, unified communications, integrated services—whatever you call the
impact of VoIP, video, mobility, and the combination of voice and data communications, it’s
clear we’re faced with changes in how we manage the environment. Information flow is both
vertical and horizontal between systems. And it flows not just inside the enterprise but outside to
customers and strategic partners.
Linear thinking and development taught us to build and sell, but today, you need to leverage
information so that you can probe the market and respond. Knowledge, or intellectual capital, is
for many, the largest asset; thus, we build databases and pull information from them. However,
collaborative technologies that integrate voice and data aid in tying information assets to people.

71
Chapter 3

Convergence will make the network indistinguishable from the enterprise. It will create, over
time, a virtual community of the enterprise. The technology, converged and blended, becomes
completely transparent, yet absolutely necessary.
Convergence isn’t simply about voice and data blending into a single network architecture. It’s
the full integration of voice, video, and data into the core of the network, the access options,
myriad network-connected devices, and the applications running over everything. It’s a four-
layered convergence of the network, services, applications, and end points. The next chapter
explores how to use this combination of sustaining, disruptive, and emerging technologies to
improve productivity and differentiate your business in delivering products and services to
customers.

72
Chapter 4

Chapter 4: Productivity Advantages of Unified


Communications
In years past, cost reduction was a primary driver for integrating services. Convergence was seen
as a cost reduction enabler first and foremost. As companies worked to consolidate to a single
cable plant, voice and data converged onto one Cat5 or Cat6 wiring infrastructure. Further
consolidation to a single wide area network (WAN) circuit infrastructure based on IP has slowly
followed. In practice, cost reduction has proven to be a secondary benefit. The real benefit to
converged communications is in productivity gains and as an enabler of new business operations.
Although process re-engineering itself can be a major effort, convergence further strengthens the
competitive edge of increased efficiency and productivity.
This chapter will extend the reach of unified communications beyond cost into specific business
areas and interests. It will provide real world examples of how business operations and industry
segments can realize tangible productivity gains, and will not only touch on the current
convergence of data, voice, and video but also see how they set the stage for other advances
coming to the converged network.
Worker productivity will rise and fall with the integration of data, voice, and video
communications. This natural ebb and flow is representative of gradual process changes coupled
with workers overcoming the learning curve as they adapt work habits to best utilize new tools
and resources. It’s important to remember that basic telephone usage is something very natural to
working adults. It’s something learned at a fairly young age. In the workplace, changing how
people work—how they use the telephone—will have unexpected impacts. For example, a sales
team that has been working from individual PC-based contact calling programs such as Act! or
Goldmine will encounter a learning and adaptation curve with the enterprise shifts to a company-
wide Customer Relationship Management (CRM) system with Sales Force Automation (SFA)
features. The work paradigm changes dramatically. This paradigm shift provides the catalyst for
a change in corporate culture within the organization.

73
Chapter 4

Convergence is a buzzword that has garnered a lot of play in the past 8 years or so, but
convergence means many different things, all of which apply to the enterprise:
• Network convergence is really the first phase of a long evolutionary process. Converging
voice and data onto a single infrastructure provides opportunities to reduce operating
expenses (OPEX). It reduces billing complexity with service providers. It provides for
early workforce consolidation. It sets the stage.
• Service and application convergence is the hot topic of the market today. The idea of a
Service Oriented Architecture (SOA), or deploying Software as a Service (SaaS) on the
network, sets the stage for radical change in how work gets done. This convergence of
data, voice, and video as services, coupled with the convergence of applications as
services on the enterprise network, completely changes the basic steps and procedures of
performing even some of the most basic work tasks throughout the day.
• Many organizations employ convergence between the telephone set and the desktop PC.
Desktop real estate is at a premium for workers in the information economy, and the
ability to use a single device on the desktop for all communications activity provides new
integration enabling new workflow efficiencies.
• Further evolution in fixed mobile convergence will more tightly couple the telephone, the
mobile phone or PDA, and the desktop workstation, providing a convergence that offers
device independence with the freedom of mobility and choice of the best available device
for communications at a given point in time.

How Service Convergence Drives Productivity and Enables New


Business Operational Processes
Service convergence as a productivity enabler has become a root motivator for many companies
pursuing unified networks. It’s important that organizations not pursue new technologies in
unified communications solely for the sake of their novelty. The key for any enterprise is how
the convergence of the network supports the established business strategies.
As unified communications technologies develop, companies around the world are discovering
innovative business benefits provided by unifying data, voice, and video onto a single service
infrastructure. High-level business processes can be heavily impacted by convergence,
underscoring the reality that convergence is far more than cost consolidation of multiple,
separate networks.

74
Chapter 4

For many enterprises, the question of one network versus several networks remains unaddressed
today. Voice and data networks were developed in isolation, each being tuned for the
performance requirements of the specific traffic type supported. As separate services, integration
between the two wasn’t anticipated or planned. VoIP introduced the idea of change, but the
complexity of integration and rapidly evolving network capabilities made convergence a costly
proposition for many companies. During the earlier years of convergence, organizations focused
on the cost without fully understanding the value of business process efficiency and change that
could evolve from integrated services.
Today, it’s both feasible and practical to integrate business service solutions onto a single
network infrastructure. Convergence also enables implementation of business solutions that were
just too costly to deploy on non-converged, dedicated voice and data networks.
One example of a service solution that is far more costly to implement in separate dedicated
networks is Computer Telephony Integration (CTI) as described in Chapter 2. Call centers were
once large, densely populated work centers deploying expensive integration solutions. It was
difficult to justify the extreme costs in smaller business operations. Shaving a few seconds off a
telephone call provided justification in the high-end call center processing thousand of calls per
day, but that integration hasn’t always transferred down to smaller business processes. Why not?
Workers in today’s information economy agree that there are productivity gains that can be
achieved by integrating the telephone and PC workstation. The main hurdle in deploying call
center technologies more broadly across the enterprise has been the negative cost ratio. The cost
associated with separate networks has, for a number of years, outweighed the business benefits
of implementing call center applications. In the fully converged network, many of the
complexities of integration are eliminated. This reduces the implementation cost of call center
solutions and creates a positive cost ratio. Figure 4.1 illustrates the importance of evaluating the
business benefits over the cost ratio.

Business Benefit

Data Network + Voice Network + Video Network

Lowering the cost of the network will improve the business case and
will show a higher return on investment for the same business benefit

>1 <1

When the converged When the sum of separate


network cost is less that networks is more than the
the business benefit business benefit

Figure 4.1: Business benefits and cost ratio evaluation.

75
Chapter 4

What we find are some vital areas of business benefit. Convergence can deliver both financial
and operational benefits derived from the integration of applications and services, while enabling
improved productivity. For many organizations, these benefits appear in business facets that
weren’t fully considered in the past. Although this chapter is primarily focused on productivity
gains, let’s briefly touch on these business facet benefits as a reminder that convergence delivers
a broad range of gains to the enterprise. They all need to be considered in the deployment and
management of the converged service network.

Management and Utilization of Property


Service network convergence can enable property modifications that provide better working
facilities for employees while reducing the enterprise needs for real estate ownership. Given
distribution of work within an enterprise, the converged network can provide collaboration tools
and a work environment that eliminates the requirement for all employees to work in a large,
centrally located facility. Different work functions can be distributed to locations, taking
advantage of tax benefits and other financial drivers. For many companies, placing the
employees closer to the customer geographically provides an added value.

Cost Reduction and Management


Network consolidation leads to more efficient use of technical staff, leading to further cost
reduction. Housing of technology resources, such as network server farms, and reducing the
number of vendors or leasing requirements bring further reduction in costs. Additionally, a
single, consolidated network provides greater clarity and visibility of future costs as the
enterprise plans technology enhancements to support a unified business strategy.

Enterprise Agility
Enterprise agility is difficult to measure and quantify. One competitive factor that has been
widely recognized by many companies is that Internet technologies have allowed small,
entrepreneurial, nimble companies to compete against large enterprises effectively. Making the
large enterprise a nimble, agile business through convergence enables quicker response to new
opportunities in the market. Leveraging the property and asset management values, large
enterprises can quickly move resources where they are needed or where favorable tax and labor
rates prevail. Adaptation to change is simpler and less costly with a fully integrated network.

76
Chapter 4

Staff Productivity
The converged network, integrating data, voice, and video services, allows employees a wider
range of information media and network accessibility options. Tools such as unified messaging,
voice, and video conferencing, and Web-based productivity tools create an environment for
employees to more effectively accomplish business goals.
New applications can be deployed more quickly in a unified network. Interoperability concerns
are reduced or eliminated. Data, voice, and video applications no longer have to be developed
and tested on separate networks, then cobbled together to create new service applications.
To enhance productivity, the enterprise needs to improve employee flexibility and mobility while
ensuring secure access to corporate information resources. Some examples include:
• Campus roaming
• Itinerant workers who routinely work out of different corporate offices
• Mobile telephone users—Cellular users today and VoIP users through fixed mobile
convergence tomorrow
• Relocations and moves of departments and within departments
The convergence of data and voice allows employees to be as fully productive outside the office
as they are when they’re in the office at their desk. VoIP, unified messaging solutions, and call
center technologies can all be used today to support the increasingly mobile enterprise
workforce. Teleworkers, sales people, and consultants who spend much of their productive work
time away from the office still require access to corporate services and information resources.
One added value with VoIP services in the converged network is the accessibility for workers in
transitory locations such as hotels and airports.
As we look at the productivity advantages of unified communications, there are a number of
other transitional events that help the enterprise determine whether to adopt convergence as part
of the overall business strategy. For many organizations, these events act not as catalysts on their
own but as accelerators to convergence adoption. In many cases, they can bring quicker benefits,
both financial and via productivity, to the enterprise.

The Non-PC Workstation Environment


For many enterprises, particularly those whose core business is not based solely on information
technology, there is a need to provide real-time data in a variety of different environments.
Manufacturing environments may be too dirty to deploy PCs. Refineries, chemical facilities, and
the shipping industry often present hazardous environments. Remote locations can present
special security requirements. In addition, PCs may just be too expensive. In many cases, there is
a need to provide a cheaper end user device. A converged network may be able to utilize a single
handset device to provide both voice and data services in a cost-effective manner, reducing the
equipment exposed to risk.

77
Chapter 4

Evaluating the Call Center Strategy


In organizations with existing call centers, agent turnover is a universal concern. Alternate
facilities might be considered for reduction in operating expenses (OPEX). Optimizing
technologies onto a single platform can offer both cost savings and productivity increases. The
converged call center can bring crisp and dynamic management to call flow balance between
multiple call centers to ensure peak agent utilization.
Implementing efficient call transfer technologies in the unified service network facilitates
movement of both voice and CRM data across any customer “touch point” within the enterprise.
This multi-channel call center approach lets the enterprise humanize business relationships with
customers, eases access to critical information resources, and makes agents more productive by
reducing handling times. These factors all lead to increased customer satisfaction. For sales
channels, increased customer satisfaction can lead to cross-sell and/or up-sell opportunities,
increasing the bottom line revenue potential.

Business Process Optimization


Enterprises constantly tune and optimize business processes for increased productivity. Remote
collaboration tools can reduce time lost to travel. Video conferencing can eliminate travel time
for internal company meetings. With converged services, giving every employee universal
access to necessary resources, the enterprise can deliver a consistent user experience.

Productivity in Resource Management


The overhead associated with adds, moves, and changes is a significant burden. It costs time and
money, and for many organizations, this ongoing administration OPEX is far more costly than
the initial capital expenditure (CAPEX). Convergence reduces OPEX by creating efficiency in
the administration process.
Facilities managers who oversee a combination of office space and services such as voice and
data in large enterprises often use what is referred to as “swing space.” Swing space is extra
capacity to allow for the ongoing movement of individuals or groups of employees through
constant organizational changes. It’s essentially a buffer. For many large enterprises, swing
space may comprise as much as 5 to 10 percent of the building.
For many companies, the need for swing space was driven by the inability for real-time adds,
moves, and changes in the traditional business PBX platform. Moving a telephone user in the
legacy PBX world required time and effort to reconfigure wiring and implement programmatic
changes.
In the converged network, administrative intervention can be eliminated, enabling real-time
reconfiguration of data, voice, and video services. Network cabling requirements have been cut
in half, using the unified IP cabling plant.
WiFi technologies coupled with VoIP softphones may completely eliminate the need for cabling
to the individual desk. Other fixed space assets, such as conference rooms, can become viable
temporary work spaces during reorganization or to support special projects with significantly
reduced administrative overhead to support adds, moves, and changes.

78
Chapter 4

Focusing on Employee Productivity


Email, voicemail, fax, and other messaging tools have provided increased ability to communicate
both within the enterprise and with business partners and customers. These tools have also
created inefficiencies. Some studies indicate that employees spend an average of 2½ hours each
day either reading and responding to email or listening to voicemail messages. To increase
productivity, communications need to be managed more efficiently.
The converged network provides a number of benefits to the organization. Perhaps the most
significant is the ability to quickly integrate and deploy a wide range of services and
applications. These service and applications can help streamline administrative tasks, letting
employees focus on business goals, customer care, and revenue generation. The following three
sections offer examples of application services that increase productivity in the converged data,
voice, and video network.

Unified Messaging
Unified messaging platforms can give users immediate, integrated access to voice, email, and fax
messages from any workstation in the enterprise, whether it’s a VoIP phone or a PC. The time
reduction in using a phone for voicemail versus a PC for email can be significant. Unified
messaging is a natural lead-in to device convergence, enabling any device on the corporate
network to act as the workstation of choice at any point in time. Traveling employees can access
all messages from a single device, speeding response times.

Personal Communications Assistants


Enterprise employees have many different contact points:
• Desk phones—In some cases, in multiple offices
• Mobile phones—In many cases, multiple mobile phones
• Home office phone
• Email—To PC workstations and to handheld devices or mobile phones
• Video collaboration tools
• Instant messaging tools—Both inside the enterprise and with outside contacts
It’s becoming increasingly difficult to know which number to call or how to contact any one
individual at any point in the work day. Multiple phone numbers lead to counterproductive
“phone tag,” or worse, “voicemail tag.” Missed calls and multiple voicemail messages present a
resource drain on productivity. Communications assistant tools in a converged network can
provide information about presence and availability to help workers prioritize who can contact
them and via what communications devices. Critical calls, or contacts, can be automatically
routed to multiple devices to ensure efficient communications. Another advantage of these
personal communications assistants is the ability to set up conference calls on demand,
increasing collaboration efficiency both inside and outside the enterprise.

79
Chapter 4

IP Video Solutions
The fully integrated network doesn’t just converge data and voice; it brings videoconferencing
power in a cost-effective, ubiquitous way. Many organizations see the power of
videoconferencing as a means to reduce travel expenses. Beyond the travel cost, video can save
time and provide a rich user experience in communications.
In a converged network, the enterprise can provide both video on demand and videoconferencing
capability to every desktop. Many organizations today are very laptop- or notebook-computer–
oriented. Today’s systems often come equipped with a video camera (or Webcam) built in. For
companies needing to invest in Webcam technologies, high-quality cameras have become low-
cost commodities and can easily be bought in volume for less than $50 apiece.
Video provides an array of business communications tools that are quickly becoming a normal
part of the day for many organizations. Distance learning via video allows employees around the
world to continue learning without the headache of travel to a centralized training facility or
classroom. Crucial business information, board meeting updates, product announcements, even
staff meetings, can all be viewed in either broadcast or interactive modes depending on the tools
used for video. For teleworkers and remote staff, IP video provides face-to-face communications
in real time, maintaining strong working relationships with coworkers in other remote locations
and in corporate offices.

Web-Centric Businesses
For Web-centric businesses, or those tightly coupled to information technologies, the evolution
of voice to next-generation voice and video services holds great potential. Although some of the
content in this section may not apply directly to the enterprise implementation of unified
communications today, it certainly paints a clear picture of where integrated voice and video
technologies are quickly headed.
Presence and availability information are crucial, but they’re also not well understood by the
enterprise business world today. That is changing as society evolves. The established leadership
in enterprise business today is comprised of people from the Baby Boomer Generation, but in the
high-tech market, GenXers have stepped into many leadership roles. The next-generation
workforce is coming from what is now being called the Millennial Generation. For many users,
particularly younger users who have known Internet technologies their entire lives, tools such as
instant messaging are vital communications tools. Adults often learn and adapt to new
technologies from their children.

For an interesting perspective on the generation of children that has grown up with a computer mouse
in hand, check out Homo Zappiens - Growing Up in a Digital Age by Wim Veen and Ben Vrakking.

80
Chapter 4

In the world of instant messaging (IM), presence and availability provide key pieces of
information for the enterprise. These concepts are a follow-on to the simple IM buddy list.
Although presence and availability are the key buzzwords used in the unified communications
space today, relevance and context are basic concepts behind them both. The buddy list provides
a window into online contacts’ availability, simply showing that they are online, available, busy,
at lunch, on the phone, or some other simplistic status indicator.
The real unified communications vision is far, far broader than the concept of relevance. Today,
communications assistants (that is, software tools) offer the ability to give users control over who
contacts them—when, where, and how. In short, users can define the context they are working in
and control how they are contacted. This idea is a piece of the future, but only one piece.
Service integration is crucial—voice, data, and video converged on a single set of tools.
Collaboration tools today are many and varied, but most currently lack extensive video services
coupled with widespread application sharing. That’s a collaboration component developers are
just beginning to fully understand.
Fixed mobile convergence (FMC) presents another piece of the puzzle, and it’s sorely lacking
today. In its earliest stages, FMC is to many the ability to pass a phone call from the mobile
network to a locally managed WiFi. That’s really nothing more than arbitraging the cost of
mobile minute airtime. It’s an achievable technology that simply needs the right protocols to
make it work. If the market demand existed today, this view of FMC would be fairly easily
accomplished. Every technology piece needed exists in some form today.
As unified communications evolves, FMC will enable the initiation of a call from a PC in a home
office in the morning. The converged network can make this a multimedia call with video and
application sharing. When the time comes to leave the home office and drive in to the corporate
facilities, the voice stream can be handed off to a mobile handset on a home WiFi network. As
you get in the car and drive away, the call will seamlessly hand off to the mobile carrier. Upon
arrival at the office, it will be handed over to the corporate WiFi network. And when you arrive
in your office cubicle, you’ll be able to hand the call back off to the phone on your desk, a
softphone on your PC, or even to re-engage a full video collaboration client on the desktop to
rejoin the collaboration call. That’s the unified communications path for FMC. It’s more than
just the network. The network, or networks, is still nothing more than a transport mechanism.
Presence, availability, and relevance technologies today are barely the tip of the iceberg. Barely a
toddler in terms of what the mature model will become. Today, you can share presence with your
buddy list with your contacts but that doesn’t begin to describe the enterprise value of
convergence.

Integrating Voice with Sales, Service, and Support


Combine data, voice, and video with relevance, collaboration and FMC advances and think about
how these services bundle together. Imagine an enterprise customer service team that is beyond
relevant. A customer can call their sales rep, but the relevant enterprise will have every
employee’s presence and availability information. Integrate the CRM system to enable advanced
customer choice. The customer doesn’t have to simply leave voicemail because they can’t reach
their designated account rep. Why not let the customer choose whether to leave voicemail or ring
through the next relevant member of an account team who is available automatically? With
converged systems, you not only pass the call but can easily send all the pertinent customer
information as well.

81
Chapter 4

Call Centers—Localized, Distributed, Offshore


Now consider the ramifications of all this enterprise relevance, presence, and availability
capability tied into the call center philosophy. Why not make the entire enterprise a business
relevant call center. You can know where every employee is, and their availability. You also
know what tools they have available to communicate in the moment. “Voicemail jail” disappears
from the landscape. Nobody ever needs to leave a voicemail message for their account rep except
by choice.
This evolution of convergence will redefine the “easy-to-do-business-with” enterprise! The
relevant enterprise. But at the core, it’s not relevance. It’s not presence. It’s not availability. It’s
responsiveness. It enables nimble adaptation to the tactical needs of day-to-day business—
business communications at the speed of thought.
When you think of the global telephone network, one of the features that has made it so valuable
for worldwide business is its ubiquitous presence. The telephone is everywhere. Why not
leverage convergence and unified communications technologies for the ubiquitous enterprise.
The enterprise that is always on, always accessible, always responsive. That is where the next
generation of communications convergence is headed. Integrating data, voice, and video today is
the enabling foundation.

IVR Systems
As Chapter 2 noted, interactive voice response (IVR) systems are computerized systems that let
callers choose options from a voice menu. With advances in voice recognition technologies,
you’ll not only see new IVR applications, you will begin to see costs drop significantly. Simple
IVR solutions let the caller speak simple answers such as “yes,” “no,” or numbers in response to
the prompts, but they continue to grow in sophistication.
IVR systems today can also “read out” complex and dynamic information such as email
messages, news reports, weather information, and faxes using complex Text-To-Speech (TTS)
conversion tools. These TTS systems use human voices creating speech in very small fragments
that are assembled to create very lifelike voice.
Although IVR systems have been used to create service solutions such as airline ticket booking,
banking by phone, balance inquiry, and so forth, in the converged network, they present a new
set of services for internal use. They allow employees out of the office to call in and have email
messages or fax messages read back as part of a unified solution. IVR technologies enhance the
ability of any employee from any location to receive and respond to important business calls and
email messages.

82
Chapter 4

Computer Telephony Integration


Chapter 2 also reviewed how Computer Telephony Integration (CTI) can facilitate interaction
between the telephone system and enterprise computer systems. The converged network
integrates data and voice onto a single IP infrastructure, drastically reducing the cost and
complexity of CTI.
In the converged service network, caller information, screen pops, call control tools, and
outbound calling features are integrated into the VoIP system more tightly than was possible
when integrating legacy voice and data network services. Call transfer, call hold, and conference
calling features often become one-click operations.
In Internet services today, the idea of click-to-call gets a lot of attention. On the Internet, it’s a
nice idea that’s forming. In the enterprise converged network with CTI, it’s a reality today. The
converged network of unified communications tools makes CTI features readily accessible by
business operations groups that previously couldn’t implement these efficiencies.

Application Integration with CRM and ERP Systems


Enterprise Resource Planning (ERP) driven, Web-centered collaboration in business-to-business
(B2B) interaction has grown significantly in recent years as business partners found ways to
leverage Web services in Internet technologies. ERP systems are often implemented during
process re-engineering within enterprise businesses to help break down the legacy “silo”
mentality that compartmentalized large companies into smaller fiefdoms, often struggling
internally within the organization. ERP systems dissolve many barriers by unifying all data
resources and business processes under a single umbrella solution. This unified approach
facilitates, and even encourages, collaboration between different business work groups.
Although ERP systems frequently began as supply chain monitoring tools in the manufacturing
sector, today they’re widespread across every business environment. Today’s ERP systems
support manufacturing, supply chain management, customer relationship management (CRM),
sales force automation, human resources, and more. All of these components of ERP require
unfettered access to voice services and data resources. As with CTI, the convergence to a single
network infrastructure for data, voice, and video reduces the cost and complexity for
implementing ERP systems.
For many organizations, the ERP system represents what Bill Gates has referred to as the Digital
Nervous System of the enterprise. ERP systems can be very tightly coupled as a service on the
converged enterprise network. Communications tools, integrated with process management and
monitoring, provide a level of integration that speeds countless business operations, reducing
costs and increasing efficiency within the company.

83
Chapter 4

Customer Relations and CRM


CRM is used to describe a wide array of business capabilities, methodologies, and technologies
that support how an enterprise manages day-to-day relationships with customers. CRM systems
bring reliability and consistency to customer interactions, enriching the customer experience and
increasing customer satisfaction overall.
As a corporate strategy, CRM is often implemented to create and maintain lasting customer
relationships. As Chapter 2 noted, CRM is often a cultural shift for organizations, moving to a
holistic view of managing the entire lifetime of the customer relationship.
CRM systems are typically implemented within the marketing, sales, and customer service
groups who have the most frequent and direct customer contact. One success key to CRM in the
converged network is simply the ability to capture every single touch point and every customer
interaction that occurs, regardless of what group within the enterprise is involved.
CRM’s focal points are to create a customer-based culture of end-to-end service. Convergence
integrates data and voice services to effectively capture this information for continuous analysis
and improvement. Figure 4.2 revisits a process flow introduced earlier in this guide. CRM may
be viewed as analogous to ERP for many enterprises. What the CRM system enables is capturing
every piece of information the organization has about every interaction with every customer,
throughout the life of the customer relationship. This information repository becomes enterprise
metadata; that is data about your enterprise business that can be analyzed and leveraged to
further streamline and improve customer service processes.

Business Data
Activity IT Systems Data Mining
Warehouse

Knowledge Information
CRM Management

Informed Technology Intelligent Resource The Right


Choices Procurement/Allocation Service Bundles

= VALUE

Figure 4.2: Building value with CRM.

84
Chapter 4

To fully use the strength of the converged network, you need to capture every business activity
into an IT system—every one. This data warehouse of information becomes the repository of
what happened at every step of the way in the business flow. This information can be mined and
processed against customer interaction data from the CRM system. This entire process of
knowledge management gives business leaders fully developed information about the enterprise
so that they can make informed technology choices for the future, intelligently procure and
allocate resources within the enterprise, and develop the right new product and service offerings
to increase revenue by filling customer needs, both present and future. That’s the value of CRM.
Where the rubber meets the road, at customer interaction, you use a variety of communications
tools—data, voice, and increasingly video. The converged network simplifies the ability to let
employees focus on customers, while leveraging the strength of the technologies to automate
capturing the business intelligence information that will help the enterprise evolve.
CRM systems today provide comprehensive order history tracking, click-to-email, click-to-call,
and analysis reporting tools that arm employees with comprehensive CRM tools to manage
larger sets of customers more efficiently than could ever be achieved in the legacy voice and data
networks.

Vertical Market Business Drivers for Change


Although there are numerous small drivers within each vertical segment of the market, one key
driver across all sectors is the hyper-connected nature of workers. Workers today are information
professionals regardless of the business sector they work in. Everyone in the work force is an
information professional. Everyone is connected. Many workers are hyper-connected.
Workers use email, office phones, mobile phones, pagers, Blackberries, smart phones, home
computers, and laptops—in short, they are almost always online. In this hyper-connected state,
workers have also achieved a virtual Masters degree in multitasking.
Using today’s communications tools, workers manage more projects, cultivate more customers,
and complete more work than ever before. This hyper-connected, all-inclusive toolset also means
that work and personal life blur. Customers today often involve personal relationships as the
work day becomes an artificial constraint for business in many areas. People work from home or
from wherever they are at the point in time they’re needed. Convergence technologies and the
unified communications evolution not only leverage this ability but also bring business tools that
help employees ensure that business and personal lives don’t blur entirely.
Convergence isn’t just VoIP. Convergence of data, voice, and video today are the foundational
elements for unifying all communications technologies. The power of convergence doesn’t lie in
VoIP. The power of convergence lies in integrating voice and data services of all kinds with
business applications in the enterprise, tightly coupling services and applications to support
business strategies.

85
Chapter 4

An Example of Convergence Beyond VoIP


Voice isn't always VoIP. The real value of convergence is the broader unified communications that
embraces all data voice and video technologies. There is a new startup company today that can provide a
single “virtual” number that works on a mobile user’s existing mobile phone. Users can both dial out from
and receive calls on their virtual number, so they can easily separate their business calls from their
personal calls, for example.
Not only can a mobile user have two mobile numbers on a single phone, but the virtual number can be
selected from almost anywhere that the user desires. In this way, professionals who conduct business in
multiple regions can use their virtual number to give the impression of a local presence because callers
can see that they are reaching or being reached by a local number. This virtual number is another
telephone number. It can be assigned to your cell phone. The mobile phone can support two numbers,
with CallerID working on both. The virtual number can be anywhere. Consider foreign students attending
college in the U.S. Recent reports say there are 674,000 students with active VISAs at present. They
could have a phone number on their cell phone in England, Japan, India, and Australia. From home—
their real home where family live. Two huge benefits brought about by this development are the ability to
leverage international long distance arbitrage to obtain the cheapest per minute cost to call home. But this
also means that mom can call back and reach her student with a local phone call.
Foreign students are just an example. How many workers travel to other parts of the world from their
home to work? Some parts of the U.S. are heavily populated with migrant workers in agriculture. Many
professionals in the medical field come to the U.S. from the Philippines. How many people have corporate
offices in their home country and another part of the world? What about a consultant in Idaho who wants
a Washington DC number to work on contracts with the federal government. Local presence, via a local
phone number, is easily accomplished. And Caller ID follows that number. Place a call from the UK
number, and that is what the person receiving the call will see on their display.
Beyond that there is another aspect of aliasing existing telephone numbers. Let me give a personal
example for aliasing. I have phone numbers for home, home office, office, personal cell phone (Treo),
business cell phone (Blackberry), and a couple of others that I use daily. With telephone aliasing
technology, I can make them all appear on my cell phone, just like the virtual number described a
moment ago. I can make and receive calls with the full presence of my telephone number. If I call on work
business, I can place the call from my business number and that is what the called party will see on their
display. In short, the identity of my telephone number is extended to my mobile phone, complete with
Caller ID information. I know I keep mentioning Caller ID, but it's important.
Let's look at a broader vision. Consider professional services workers—consultants, doctors, and
lawyers—as examples. Consultants and lawyers make their living on billable hours. There is a law firm in
Silicon Valley that has privately estimated that they lose $1.5 million dollars a year in billable time from
attorneys talking on their cell phones while they are driving. If those attorneys had virtual numbers with an
account code, that time could be tracked, and billed, driving billable time back into corporate revenue
stream. Consultants may provide a block of hours or billable service to special clients. Many provide a
dedicated phone number for the client to reach them any time. What a perfect fit! A virtual telephone
number, dedicated to a client with account code tracking for billing built right in.
Doctors have different constraints. When a doctor calls a patient with test results, the Health Insurance
Portability and Accountability Act (HIPAA) regulations forbid patient test results being left in a voicemail
message. The medical professional has to speak with the patient. Doctors don't call patients from their
cell phones because they don't want patients calling them back there. If you get a call from your doctor's
cell phone, you're likely to let it go to voicemail as an unfamiliar number anyway. But if you see the
medical center number on Caller ID, you know they're calling with test results. Everyone is assured the
privacy and confidentiality of patient information, but through Caller ID, you leverage the telephone
network for better efficiency.
Beyond North America, these technologies mean that global boundaries don't matter. If I do business in
six countries, I expect I'll be able to get virtual telephone numbers in every country at some point. All on
my mobile phone. All at once. Whether the underlying network is VoIP, PSTN, or cellular is irrelevant to
the user. Convergence eliminates the technology and simply makes it a voice service.

86
Chapter 4

Business Sales and the Web, or Net-Enabled Business


Earlier, it was pointed out that the real challenge of convergence is in supporting the enterprise
business strategies. The fundamental goal isn’t to implement convergence. It isn’t to shift to
VoIP. The real objective is solving business problems. In a recent article entitled “VoIP Outside
of the Box—A New Way of Thinking,” Telephony World editor Don Panek had this to say:
Business owners and executives today are not sitting around board room tables
discussing their existing phone systems, services, or expenses. They’re not thinking about
VoIP or Telephony and what it can do for their bottom lines. They’re discussing real
business problems and looking for real solutions to those problems.
His closing drove the point home yet again:
I think the time has come to stop talking about VoIP and Telephony and start talking
about applications and solutions. Solving problems. And at the end of the day after the
order has been signed, we can then mention that all the phone calls will be free or
virtually free! And how is that done? Well the solution has VoIP built in!
Communications technologies manufacturers, vendors, and developers often get sidetracked into
the idea of how their products can interoperate…how they can build a platform. Customers don’t
care about a platform, and don’t care what vendors are doing beneath the hood. Customers care
about their business problems and solutions to those business problems.
For business sales, Web-enabled businesses and the hyper-connected workers of today, access to
corporate business resources and instant, easy communications are vital to success. Convergence
of network voice and data services provides a first step. Integration with ERP and CRM systems
takes convergence further along the road to productivity enhancement and business success.

Product and Services Sales


Convergence to the fully integrated network provides access across the enterprise to a full set of
business tools. Business sales, whether products or services, all have a time interval—a sales
cycle. Outside this sales cycle, business information travels at the speed of light, or electrons on
the wire. Converged services in the hyper-connected, always-on business provide differentiators
that didn’t previously exist, in part because the converged business is a nimble business able to
respond quickly to market demands. Some examples of how this always-on enterprise can
leverage the converged data, voice, and video services of the future include:
• All markets are up for grabs—The marketing paradigm changes dramatically. The
always-on enterprise leverages communications tools to reframe the discussion and win
customers while the less-empowered competitor is left behind.
• Difference not differentiation—The converged enterprise minimizes the behavioral
changes in customers by embracing change within the enterprise. This can give your
customers a tangible set of reasons to love your products and services.
• Don’t disappoint—By leveraging reliable new technologies, you ensure that everything
works and the organization can react instantly to a situation.

87
Chapter 4

• Make your marketing sociable—The enterprise can’t control the customer conversation,
but can leverage best-in-breed convergence technologies and refined business processes
to build genuine relationships with established and potential customers rather than white-
noise relationships. CRM becomes not a buzzword, not a catch phrase, but a corporate
culture of nurturing the business.
• Interaction requires iteration—It’s not enough to listen and respond to customers.
Business success requires a long-term, sustainable dialogue that convergence
technologies support. Meaningful long-term connections with customers come from
community, co-operation, and co-creation—all collaborative efforts that the integration
of services and applications enhance.
• Don’t forget to sell—Engagement is great but it doesn’t pay the bills. Remember that
every touch point throughout the customer relationship cycle is an opportunity to up-sell,
cross-sell, or lose business. Selling is responding to the customers. It’s about making it
easy to do business with you. Convergence technologies integrate all your corporate
intellectual capital into a suite of services and applications that make you easy to do
business with.

Financial Services
Beyond the basics of efficiency and productivity, there is another set of drivers in the financial
services sector that are tied to regulatory compliance. Whether it’s the Sarbanes-Oxley Act
(SOX), the Gramm-Leach-Bliley Act (GLBA), ISO-17799, or IT Infrastructure Library (ITIL),
there is a set of practices that are ever widening across business sectors. The financial services
sector has always been tied to close auditing and scrutiny.
Beyond audit or compliance considerations, the financial services sector is perhaps the closest
sector to a pure information economy. In financial services, the ledgers and paper trails of old
have become a stream of information on the network. The paper is gone and finance is all about
moving information quickly, accurately, and securely.
Convergence of data, voice, and video onto a single infrastructure will, for many financial
institutions, provide a consolidated approach to monitoring and management, simplifying the
entire data capture, warehousing, and analysis process for regulatory compliance. Figure 4.2
showed the importance of CRM in the business flow, but that same chart provides an overview
of touch points in the flow that eases compliance reporting and documentation for companies
involved in financial services.
For financial services businesses, the reduction in cost of call center technologies may play a key
role in redefining the financial services business. Deploying call center methodologies in smaller
volumes now affordable with convergence can drive measurable gains in productivity down into
smaller workgroups.

88
Chapter 4

Health Care
Like financial services, the health care environment brings a unique set of compliance
requirements related to HIPPA. This legislation established standards for transactions in the
healthcare sector, but also established requirements for the security and privacy of patient health
data.
Converged services again provide a single infrastructure to secure and bring into compliance.
The complexity of HIPAA has spawned numerous supporting, consulting service markets.
Network consolidation can allow health care providers to take a single, unified approach to
HIPAA compliance, thereby focusing on their core business—health care.
Beyond compliance, the converged network brings the model of CRM to patient care in the
health care environment. This integration of voice and data services means that doctors,
physician’s assistants, nurses, and other health care professionals can leverage advances such as
screen pop CTI and click-to-call technologies to better serve patient needs.

Manufacturing
For manufacturing, the advantages of convergence may seem hidden. The manufacturing sector
doesn’t appear at first to be tied to network services the way purer information-driven companies
might be. But in manufacturing, there are several key areas where convergence brings real value
to the integration of data, voice, and video.
Perhaps the most tangible value in convergence lies in the broad spectrum of supply chain
management and the vendor/supplier relationships that make up an integral part of the
manufacturing process. Inventory control data systems can easily prompt via an enterprise ERP
system to contact a supplier and ensure inventory restocking is timely. The integration of data
and voice in the converged network has for many manufacturing businesses proven the key to
success in following the precepts of “just in time” component delivery.

Summary
In a recent paper entitled “Voice and Video over IP: Leveraging Network Convergence for
Collaboration,” Melanie Turek, Senior Vice President & Founding Partner at Nemertes
Research, had this to say:
For several years, the big question with Voice over IP (VoIP) was whether it actually
worked, and if so, whether it worked well enough for corporate ears. Well, the answer is
in: Yes! As long as the network is architected properly, VoIP is definitely ready for
enterprise use, and convergence projects are running strong in the vast majority of
organizations.
Better still, while voice typically is the first application implemented on a converged IP
backbone, Nemertes is starting to see IT executives explore new applications—such as
video, unified communications, and other collaborative tools—that can also leverage the
IP network. The benefits can be great, including cost savings and increased productivity
in the virtual workplace.

89
Chapter 4

As they deploy these and other technologies, companies are starting to recognize the need
for network optimization, enhanced management tools and tight security. And although
most IT executives don’t spend a lot of time worrying about specific standards, they like
what standards get them—easy integration and interoperability among vendors and
networks, both of which are important when it comes to communications technologies.
As SIP grows more robust and more common, companies will have more vendor options
open to them—and they’ll start to take advantage of the benefits convergence brings:
Integration, interoperability, and the ability to stay agile in an increasingly global world.
Most companies that have already deployed VoIP, when queried, identify plans to extend
converged voice service to teleworkers. Although today the focus is predominantly voice service,
other collaborative applications, such as voice conference services, desktop videoconferencing,
and Web conferencing/collaboration are expected to follow quickly.
Many organizations are already running some kind of video over IP. Business managers show
increasing interest in leveraging the technology for more than just voice communications. Many
enterprises are running trials and developing applications using video to address business needs
in their emerging converged network environment.
General industry predictions are for significant growth in both desktop and room-to-room video
deployments in 2007 and 2008. IP networking and advances in broadband technologies have
made video an affordable and practical business tool. Although desktop videoconferencing is
viewed as a relatively new technology for almost everyone, the barrier to entry is very low.
Business managers and strategists see video as one tool in a larger suite of collaborative
applications, and show their interest in using it where no video exists today.
According to a report by Nemertes, video is one of the leading drivers for converged networks
(28 percent of participants in Nemertes’ latest benchmark name video as a key driver). That’s
very important because of the large number of enterprise business employees who either work
remotely from their direct managers or are geographically dispersed in remote offices. Some
broad industry projections show that teleworkers have increased as much as 800 percent in
number over the past 5 years. As companies become more global in nature, and more widely
dispersed, the value of real-time communications over converged networks becomes apparent.
Citing Nemertes again, in a Convergence benchmarking study, participants were asked to rate on
a 1-to-5 scale (where 1=unimportant; 2=somewhat important; 3=important; 4=very important;
5=vital) how important the following drivers were in approaching convergence:
• Growing revenue
• Boosting employee productivity
• Gaining competitive advantage
• Reducing costs
• Meet regulatory/legal requirements

90
Chapter 4

Results showed, not unexpectedly, that growing revenue is the foremost business driver across
all surveyed organizations with a mean score of 4.38 out of 5. Second in importance was meeting
regulatory and legal requirements, with a 4.07. Cost reduction placed third in the survey with a
4.02 mean score. Gaining competitive advantage scored 3.84. Of these five categories, boosting
employee productivity came in with the lowest rating at 3.8. When you compare these strategic
business drivers with IT services and network convergence projects, it becomes apparent that
demonstrating benefits at the top line (revenue) increase the adoption rate of convergence. Cost
reduction is also important, but gains in productivity are still seen as a soft cost and are very
difficult to quantify. Given the difficulty in proving these benefits, it’s not surprising that
improving employee productivity is the lowest importance of the business drivers in Nemertes’
survey on convergence.
Organizations that have gone through the evolution of adopting convergence technologies have
identified several lessons learned:
• When methodically implemented, the converged service network can lower OPEX and
increase employee productivity.
• The converged service network typically creates a scalable infrastructure capable of
supporting new business applications in a dynamic environment.
• To achieve cost savings and productivity benefits, a holistic view of business services and
applications is required. It’s critical to look at the full picture and not just focus on either
cost savings or employee productivity.
• The full converged network can provide greater visibility into granular cost controls.
• Industry studies have shown that 80 percent to 85 percent of the enterprises that have
already implemented a converged network determine that the quality, resiliency, and
scalability that these technologies provide either meet or exceed their expectations.
Convergence of data, voice, and video is a viable technology. It’s available in the market and can
be implemented today. Through integration of unified data, voice, and video onto a single IP-
based infrastructure, organizations can lower their total cost of ownership (TCO). They can
lower expenses for equipment and maintenance, reduce administrative costs, and lower carrier
charges. The converged services network can also increase productivity and enterprise
communications capabilities by facilitating employee mobility and providing a solid foundation
for the deployment of advanced, feature-rich services and solutions.

91
Chapter 5

Chapter 5: Key Steps in VoIP Deployment and Management


for the Enterprise
Different applications have different requirements. The introduction of voice and streaming
video onto the existing IP network presents a completely new set of requirements to the
operational performance envelope of the network. This chapter will examine the importance of
assessing the readiness of the network and fully evaluating design considerations related to
delivering integrated service in the enterprise. To address new service integration, the chapter
will explore a methodology, called the “performance envelope,” for mapping the characteristics
of the corporate network.

Network Readiness Assessment


As businesses delve into VoIP and video solutions in the enterprise, their focus is on business
drivers. For many companies, cost reduction is a key business driver for convergence. Cost
cannot be the only driver, and for practical purposes, shouldn’t be the primary driver. It’s
important to set reasonable expectations and to fully understand all the business drivers behind
service integration. A new application service or an integrated VoIP and CRM solution will
require consideration of very different factors than cost reduction alone. It’s vital to fully
understand the business motivation for converging video, voice, and data so that the project
planning and implementation teams can address the true success factors when mapping out
project milestones. Communicating the objectives clearly with everyone involved helps maintain
a clear view of the expected results.

Ensuring Network Readiness for Converged Services


Ensuring network readiness is a significant task and not to be taken lightly. Over the past 4 or 5
years, both VoIP vendors and systems integrators have learned the hard lesson of failure because
of inadequate preparation and planning. The existing data network has been optimized over time
to support the existing business requirements. As new applications and services are added,
network tuning takes place and a variety of parameters are often tweaked. Historically, these
optimization efforts have been driven by packet data applications such as CRM, ERP, and Web
services solutions—“normal” IP applications. Voice and video present new challenges as they
add streaming, real-time traffic to an existing service network. VoIP call quality requires low
latency in the network. Jitter needs to be low, and to avoid complicated jitter buffering
requirements, it needs to be consistent. Bandwidth and packet loss need to support the new,
integrated services.

92
Chapter 5

Testing and documenting the parameters and characteristics of the existing networks before
extensive planning will help ensure the network is capable of supporting VoIP, video, or both.
It’s important to understand aspects of the traffic needed to support all the different service types
that will coexist on the new network. Traffic types, frame size, prioritization schemes for quality
of service (QoS), jitter, latency, and packet loss are all crucial factors. Because consistent
network performance is so important, it’s also prudent to evaluate utilization in the network both
at peak and normal times.
If the network will include PSTN gateways for global connectivity, there may be requirements
specific to the gateway demarcation between networks. Both the IP network and PSTN need to
be considered. The PSTN side of the gateway will have specific trunking requirements. These
have traditionally been calculated using Erlang-B formulas. The IP network must support the
bandwidth and other network characteristics to ensure a clean handoff of calls from one network
to the other.
Erlang-B Calculation
Telephony carrier central offices and enterprise PBX systems are engineered for trunking requirements
using the Erlang-B traffic tables and hundred calling seconds per hour, or CCS. 36 CCS would represent
a circuit at maximum occupancy with zero idle time. I won’t explore the formulas in any depth as we
investigate managing the converged services network, but it’s worthwhile to understand the complexity
and sophistication of traffic engineering in the traditional telephony environment. The formula for
determining voice traffic load that’s most widely used is:
OfferedLoad = Calls/Hour × AvgMinutes/Call × erlangs
60
Erlangs are really a simple concept identifying the number of calling minutes per hour a system can
handle. The key differentiatior is between the traditional circuit world, where minutes equate to busy
circuits, against the VoIP world, where minutes of use may not be applicable. In VoIP services, it’s really
the carrying capacity of the network in packets that’s the key indicator.

The enterprise IP network must complement the trunking requirements when connecting to the
PSTN. Overlooking details like this may result in performance bottlenecks that impact call
quality.

Plan to Succeed—Don’t Fail to Plan


Whenever new technologies are implemented in the corporate network, there are some basic
planning steps that should not be overlooked. These basics will improve the likelihood of success
in the deployment of converged voice and video services.
Although this brief section seems to be focused on the planning and implementation of integrated
services, that is the crucial starting point of holistic management of the total integrated services
network. When an organization decides to implement VoIP, that decision commonly leads to
acceptance of a new, holistic approach to managing the entire service network.

93
Chapter 5

Analysis
Before undertaking a convergence project, fully identify the business drivers for change. It’s
important to identify the business objectives and set everyone’s expectations accordingly. List
the benefits you expect to derive from integrating services so that each benefit can be addressed
individually. Broad, sweeping objectives such as “save money” aren’t measurable. The success
of VoIP implementations really depend on advance data gathering and analysis.
The more you can document prior to any implementation work, the greater your likelihood of
success. It’s important to balance objectives and expectations with analysis of your network’s
capabilities. The determination that your network can support the new, converged services, or
learning that it can’t and will require upgrades in routing and switching, will influence both the
budget and project timelines.

Planning
There’s an old adage that says “people don’t plan to fail, they fail to plan.” It’s vital that your
project team be methodical and document every facet of the entire planning process. One
widespread complaint heard from both systems integrators and VoIP vendors has been that the
assessment and planning has been the weak link in a high percentage of failed projects.
Resources invested in planning, whether spent on a project manager or the hours spent by a
technical team evaluating requirements, will bring benefits at cutover time. There is no substitute
for preparation and planning. For an enterprise mission-critical service such as voice, failure can
be especially painful, both in stress and in business impact. The more meticulous the planning
process, the higher the chance of success.

Testing
Once the network requirements have been documented and are fully understood, test to make
sure the existing network can support your VoIP requirements. Comprehensive analysis up front
helps ensure that you know what to test and what parameters or characteristics your network
requires. It’s important to approach testing with the specific aim of supporting the fully
converged network. Testing VoIP services alone is not enough. Existing applications should be
tested while simulating VoIP. There is a hidden danger of integrating VoIP services and doing all
the right things to support call quality, only to starve network resources for other mission-critical
applications. It is absolutely vital to take a holistic view of all network services and applications
when testing.

Acquire the Right Resources


The “right” resources will vary from organization to organization. For some, acquiring resources
might mean dedicating a project manager. Most organizations will need some training to bring
existing staff up to speed on VoIP resources. For many companies, this will entail bringing in a
VAR partner or systems integrator to bolster the expertise available. There is no shame in
admitting you need expertise from outside and working with a consulting partner. Leverage
expertise everywhere you can.

94
Chapter 5

Look to the Future—Think Long Term


For several years, VoIP has been viewed somewhat as the final objective. It’s important now to
recognize that this isn’t the case. VoIP is a foundation or building block technology. It’s
important to look at scalability in terms of users and raw numbers, but that’s only scaling
vertically. An earlier chapter looked briefly at the Software Oriented Architecture (SOA) and
Software as a Service (SaaS) concepts. This guide has also talked briefly about emerging
characteristics of unified communications such as presence and availability. These are horizontal
scalability factors.
To future-proof your network and ensure it can continue to grow to support the enterprise
business, a holistic approach that scales both vertically and horizontally will help enable future
growth at an acceptable pace. It will enhance readiness at anticipating that next new application
requirement that lies just beyond the horizon now. The one you don’t yet know about. The VoIP
integration project presents an opportunity to review the overall future capacity of your network.

Understanding Existing Voice Needs


Once you’ve developed all your business needs and determined that VoIP service integration is
the right business move, you need to gather data to fully understand voice services requirements.
In telecommunications, there are patterns of calling that can be identified. Call usage reports
from either the existing PBX or the incumbent telephony services provider offer a wealth of
information. Some variation of these reports should be readily available from the current
telephony technology, whether it’s carrier-based Centrex-like service or a locally administered
PBX.
There are several key elements of voice service that need to be well understood. Business
telephone systems often focus on the “busy hour,” or that hour of the day the most telephone
traffic occurs. This is a starting point, but there are many nuances to consider. The busy hour for
inbound calls may be different than the busy hour for outbound calls. The aggregated
inbound/outbound busy hour may be different than either taken individually. These busy hours
may vary on different days of the week.
Most companies also have an identifiable busy day in the average week. And business managers
understand the cyclical nature of their respective businesses. The busy month of the year is
generally well known in any business sector.
For large enterprises with multiple divisions, or business units, calling patterns may vary and
some may have different requirements and calling patters than other divisions of the corporate
headquarters. Flexibility is key to understanding the needs of all business units.
To identify specific business needs, you must understand the calling flows and volumes. Some
enterprise calling patterns are very inbound in nature. This might be easily identified by a
company using a call center, but other companies establish large outbound call centers as well.
The patterns for inbound and outbound calls are often very different. It’s vital to know what is
required to support your particular business. It’s also important to identify upper call volumes, or
absolute peak traffic.

95
Chapter 5

Most enterprises are somehow cyclical in nature. When analyzing calling patterns, it’s
worthwhile to account for several months or even for the past year. Peak periods vary widely by
industry. For example, a retail services company may experience much higher call volumes
during the Christmas season than any other time of year, whereas a financial services company
might expect the highest call volumes at tax preparation time approaches. It’s important to take
these peak activity periods into account and not simply design the VoIP services to support day-
to-day business needs.
As you develop a strategy for integrating VoIP into the network, it’s also a good time to
recognize that the PSTN has become a mission-critical component of business operations. The
PSTN network has been engineered to deliver 99.999 percent uptime. That is what the traditional
carriers tout as their benchmark. This service level equates to roughly 5 minutes of downtime per
year. Corporate data networks rarely provide this level of reliability. For many organizations,
increasing network uptime will provide incentive for network enhancement and redesign
requirements during preparation and implementation of a converged services network.

One Voice Factor—Compression Methods


One factor to consider in VoIP deployment is the encoding method used to convert analog
human speech into a digital signal for packetized transmission. There are several encoding
techniques used in VoIP solutions. Each uses different compression levels. These compression
levels affect the bandwidth needed to transmit the voice packets, the quality of the voice, and the
processing time. Pulse Code Modulation (PCM), or G.711 encoding, is widely used. It’s the
standard encoding used in the PSTN, and is generally assumed to provide the best voice quality,
although new wideband codecs are now starting to surpass PCM in some implementations. PCM
generates a 64Kbps voice stream. For some companies, bandwidth requirements, weighed
against quality considerations, might make a different encoding scheme more appropriate.
Increasing the compression of the voice signal reduces the overall bandwidth requirement, but
voice quality may suffer as a result.
Even though the PSTN is more than 100 years old, voice quality is still often measured by
having a group of people listen to sounds in headphones and subjectively rate the quality. These
evaluators rate the sound quality on a scale of 1 to 5. This rating, from 1 to 5, is referred to as the
mean opinion score (MOS). A rating of 1 roughly equates to the scratchy sound quality of the
intercom speaker in a fast food drive through or a warehouse environment. A 5 is the highest
rating and is considered “perfect,” or the theoretically highest-grade voice quality achievable. It
has always been the target for conducting corporate business calls, but as you’ll see in the chart,
not actually achieved. A score of 4.4, garnered by the G.711 codec used in the public telephone
network, was described in the past as “toll quality” voice. This has been the traditional carriers’
target for business-class voice services.
Although the statistical validity of this human sampling may be questionable to some, this
approach has been used for many years and is generally accepted globally. Today, MOS is often
determined using ITU Recommendation P.800, which details methods for subjective
determination of transmission quality. In the real world, the human ear can clearly distinguish
between a 4 MOS and a 4.5 MOS.

96
Chapter 5

Table 5.1 provides a comparison of several widely adopted encoding schemes. This table
describes the codec types and algorithms used, the sample size, bit rate, and encoding delay for
the algorithm used. The last column identifies the MOS for these codecs. When selecting a codec
for VoIP implementations, it’s important to remember that delay is cumulative from end to end.
Encoding delays can become real call quality concerns in a large enterprise network. If the total
delay exceeds 250 milliseconds (ms), call quality will suffer. The ITU-T standards
(http://www.itu.int/) provide extensive details about the operation of these codecs. Vendors and
consulting partners will also be a valuable resource in identifying the best approach for each
specific implementation.
ITU Codec Sample Encoding
Encoding Schemes Descriptor Size Bit Rate Delay MOS
PCM G.711 8 bits 64kbps <1 ms 4.4
Adaptive Differential Pulse G.726 4 bits 32kbps 1ms 4.2
Code Modulation (ADPCM)
Sub-Band Adaptive G.722 8 bits 64kbps 4ms N/A
Differential Pulse Mode
Modulation (SB-ADPCM)
Low-Delay Code Excited G.728 40 bits 16kbps 2ms 4.2
Linear Predictive (LC-CELP)
Conjugated Structure G.729/G.729a 80 bits 6kbps/8/kbps 15ms/10ms 4.2
Algebraic Code-Excited
Linear Predictive (CS-
ACELP)
Algebraic Code-Excited G.723.1 160 bits 5.3kbps 37.5ms 3.5
Linear Predictive (ACELP)

Table 5.1: Encoding schemes.

PCM, or G.711, is the method that has been used in the PSTN. It’s perhaps the most widely used
codec in VoIP systems because it provides a well-known call quality rating. Every VoIP
equipment manufacturer supports G.711. G.726, known as Adaptive Differential Pulse Code
Modulation (ADPCM), may reduce the bandwidth requirements by 50 percent while only posing
minimal degradation (.2 of a point) on the MOS rating.
G.722 is mostly used in FM radio, but is included here because FM radio quality is a good audio
fidelity comparison. It’s rarely used in VoIP solutions. G.722.1 offers lower bit-rate
compressions. A more recent variant, G.722.2, also known as AMR-WB (Adaptive Multirate
Wideband), offers even greater compressions and the ability to adapt quickly to varying
compressions if needed due to network changes. Bandwidth is conserved during periods of high
network congestion, then returned to a normal level when congestion is alleviated. Because
G.722 and its variants sample audio data at a rate of 16kHz, rather than the PSTN PCM method
at 8KHz, these codecs result in far better sound quality, clarity, and fidelity.
G.728, or Low-Delay Code Excited Linear Predictive (LC-CELP), coding is widely used in
voicemail systems for digitizing stored voice messages. G.729 can deliver an 8-kilobit sample
with less than 16ms of processing time. This G.729 codec standard is often used in Voice over
Frame Relay (VoFR) and is supported by many frame relay equipment vendors.

97
Chapter 5

The G.729a codec is a variation on CS-ACELP that is rapidly growing in popularity. It


compresses the voice stream into 10ms frames. It’s less CPU intensive that many of the others,
making it a good choice for bandwidth-constrained implementations. It offers a blend of
bandwidth and delay performance characteristics that are making it quite popular for VoIP
services. It has the added advantage of interoperating with G.729 codecs.
G.723.1, or Algebraic Code-Excited Linear Predictive (ACELP), codecs work differently. They
create models of the human voice and “predict” what the next sound will be. These codecs
encode the differences between the predicted sound and the actual sound. Only this difference is
transmitted to the receiving end, reducing bandwidth requirements. Although bandwidth needs
may be reduced, this codec has often resulted in complaints that women and children’s voices
were not represented accurately.
The codecs described earlier are narrowband codecs, initially developed for supporting narrow
voice channels. The PSTN voice channel is a 64kbps circuit carrying audio signals in the range
from 300Hz to 3000Hz. Some pure IP voice solutions are now beginning to adopt broadband
codecs, mostly proprietary, for VoIP call processing. It’s expected that the use of these, and
other, new broadband codecs will increase as more and more voice calls are carried over IP in
the future.

Voice Call Quality—The PSTN vs. VoIP


As mentioned earlier, the PSTN has been in operation for well over 100 years. During that
lifetime, it has been tuned and optimized to deliver one type of traffic, voice, as efficiently as
possible with acceptable call quality. There are quality assurances in the traditional PSTN that
are sometimes inferred, based on past use, and for enterprise business, sometimes contractually
guaranteed. In short, voice service on the PSTN is a known quantity. You know exactly what
you expect for call quality when you pick up the telephone.
The PSTN is a “connection oriented” network. When you make a telephone call, the network
establishes a voice path, or circuit, across the network that is dedicated to that specific telephone
call, and meets all the basic quality requirements you’ve come to expect. Over the years, the
telecommunications carriers have optimized the PSTN to support the delay, jitter, and loss
characteristics that guarantee the expected voice call quality. In this connection-oriented, or
circuit-switched, network, each circuit is dedicated to a single conversation, so security
requirements are easily met. The circuit resources aren’t shared as they are in an IP network.
Once the call is completed, the circuit is disconnected, or “torn down.”
IP networks such as the Internet share resources all the time. They were designed to carry
multiple traffic types. That’s the fundamental shift being embraced with convergence. In the
legacy past, we designed and optimized a specific network for each type of traffic. A large
enterprise might well have a voice network, a data network, and a video network, each with
different performance characteristics. Convergence is about leveraging all these technologies—
data, voice, and video—onto one single network infrastructure. To succeed, it’s important to
understand how these different networks behave.

98
Chapter 5

IP is commonly referred to as a “best efforts” protocol. The protocol itself provides routing
delivery, but if delivery attempts fail, for any reason, IP will discard the packets. IP was designed
to use available network resources in real-time, to deliver data, but without any assurances of
delivery and with no quality guarantees. IP was initially designed to support data traffic that is
inconsistent and unpredictable in volume, or “bursty.” IP data is described as bursty in nature
because different data types have different characteristics. The data flow for email messages or
Web browsing is quite different than the data flow for streaming media such as voice or video.
IP is a “connectionless,” or packet-switched, network. Dedicated circuits aren’t established.
Messages are divided up and placed into packets. Each packet is somewhat like an envelope in
the postal system. It has header fields that contain the source address and a destination address.
IP uses routing protocols to identify paths through the network. These packets take the best path
available at the time, based on whatever metric a particular routing protocol uses to determine
the best path. Thus, as congestion and other factors impact the performance of the network, IP
packets can easily take different paths across the network, arriving at different times. A large
message is often made up of many IP packets, so they need to be buffered and reassembled at the
receiving end.
QoS includes characteristics such as delay, jitter, and packet loss. IP doesn’t provide any
assurances for quality or delivery, so performance requirements can only be achieved by adding
some other feature, or protocol, to the network. In practical terms, we increase bandwidth,
optimize switching and routing paths, and incorporate other, higher-layer protocols to meet VoIP
call quality requirements. We also add a layer of security by incorporating industry best practices
in network design, incorporating firewalls and user access controls, and including monitoring
and response mechanisms throughout the network. Table 5.2 highlights important differences in
characteristics of the PSTN when compared with typical IP networks.
Characteristics of the PSTN Characteristics of IP Networks
Designed and optimized to support voice as a Designed to carry any type of traffic
single traffic type
Uses dedicated circuit for guaranteed delivery No delivery guarantees using packet-based
routing
Connection-oriented—A dedicated circuit path, Connectionless in nature—Conversations
or connection, is created for each voice are digitized and inserted in many packets
conversation
Phone calls are typically long duration (4 IP packets are limited to 65,536 bits;
minutes on average) packets are small and only “live” on the
network for a very short time; a
conversation is made up of many packets
PSTN circuit is dedicated to one 64Kbps voice The IP network uses all the available
channel resources of the network
Delay, jitter, and loss have been designed to Delay, jitter, and loss must be overcome by
meet specific levels for voice calls adding resources to the network
Dedicated circuit provides generally acceptable Security requires special resources in the
security for voice calls network

Table5.2: Performance characteristics in the PSTN vs. IP networks.

99
Chapter 5

Because IP packets are small in size and data traffic is bursty in nature, IP networks use routing
protocols to pick the best path for transmission. In a typical IP network, this bursty nature of data
means that the traffic load on the network is constantly changing throughout the day. This
constant state of flux means that the best path identified by routing protocols may change often.
Routing protocols are dynamic and constantly update to identify the best path through the
network based on the current traffic load. The best path one minute may not be available the
next. For VoIP calls, this means that the packets comprising a conversation can potentially take
many different paths through the network.
Traditional data use of IP is often not particularly sensitive to network delay. If an email message
is broken into many packets that can’t be delivered immediately, there is likely no problem. The
message can be placed in many different packets, then routed across any number of paths
available in the network. The entire message can be reassembled and delivered to the recipient,
with no repercussions. Unlike email, voice calls are real-time interactions between people. Even
a half-second of network delay can be so detrimental to voice quality that VoIP service is
rendered unusable.
The PSTN dedicates bandwidth, a 64Kbps voice channel, to each telephone call. It dedicates this
bandwidth for the duration of the call. It does this consistently and predictably for every phone
call, every time. In an IP network, all the resources of the network are available to every user or
node on the network. This is necessary for routing protocols to work and so that IP can make
“best efforts” at delivery. Thus, every voice call can be impacted by changing traffic load and
network conditions.
Over time, we keep adding new traffic types to IP networks. Today, these networks support
email, file transfer, Web surfing, and data sharing. As we add new traffic types to the network,
we must redesign and modify the networks to support not just the new traffic type but all traffic
types in the aggregate as well. The challenge with implementing and managing a converged
VoIP service network is meeting user expectations such as call clarity and fidelity without delay
or jitter impairments. Voice conversations must be clear and intelligible or uses won’t trust the
integrity of the network for voice calls. Beyond the quality of the voice itself, the basics of
delivering dial tone, successfully completing calls, and always being available are key quality
concerns.
Managing call quality perceptions is a bit like walking the razor’s edge. Because voice service
and telephone calling are so widely used and accepted, user expectations are high, based on their
own experiences. When using a new VoIP service, if users feel that call quality is unacceptable,
they will simply hang up and use alternative methods. In consumer VoIP solutions, this
perception problem is a constant battle. In the enterprise environment, you must take proactive
steps during the implementation process to level-set user expectations with what the integrated
service can functionally support. If users develop mistrust of the network, even based on
misperceptions of capabilities, it is very difficult to regain their confidence. You need to be able
to measure, guarantee, and deliver suitable call quality. You also need to ensure that the network
can continue to support the pre-existing data traffic without service degradation to existing
services.

100
Chapter 5

Network Design Considerations


An often-stated (by consultants, systems integrators, and equipment vendors) misperception is
that many corporate networks were poorly designed. It’s simply not true. It’s important to
recognize the truth so that we aren’t doomed to repeat the mistakes of the past.
Existing networks in business were not poorly designed, but in many cases, they weren’t
designed at all. For many organizations, PCs were implemented way back in the past. Windows
95 brought in the concept of the workgroup, and local area networks (LANs) emerged, as small,
isolated pockets of information, typically for a workgroup. With routing and internetworking,
organizations began interconnecting LANs, and the explosive growth of the Internet was just one
more connection.
Today, we have business partner connections, VPN links, Web services, and more. We have the
Internet, the intranet, and extranets. We have Web servers, offering Web services, supporting e-
business. And in many cases, each new feature or function was brought about by tweaking the
service network with some incremental upgrade. We didn’t design a large number of our
networks, they just happened.
Today’s networks have evolved and grown over time. Many have been carefully and
methodically redesigned every few years. Like the PSTN has been tuned and optimized for voice
traffic, there are IP networks that have been designed and tuned to support known and
understood data applications that exist within the enterprise. The enterprise applications in use
today may have very different network performance requirements than converged video and/or
VoIP services.
VoIP projects begin with an implied assumption that some network upgrades may be necessary.
Equipment in the network, such as routers, may already be running at high CPU utilization. In
many cases, the router OSs may not be able to support VoIP services. Existing wide area
network (WAN) links may already be taxed to their limits supporting existing data applications.
In many cases, these links may be using frame relay connections that might not be voice-
friendly. Frame relay is very much a data services. The overall capacity in terms of bandwidth
and processing power are vital factors to assess before implementing any new network service.
There is a pitfall to watch out for when implementing VoIP services. It’s important to avoid the
added expense associated with multiple upgrades. Given that for most organizations, network
upgrades represent capital expenditures (CAPEX), no technical manager should be put in the
awkward position of justifying multiple upgrades. A smart approach is to assess and understand
the voice calling and quality requirements in conjunction with all existing data services. In short,
invest the effort to do a complete, holistic assessment of the entire network services suite across
the enterprise. This is also a good time to consider any planned new data services as well as the
VoIP requirements. Rather than upgrading prematurely, focus early in the process on data
gathering and information analysis. Taking a methodical approach will not only ensure
successful pilot testing, it will help yield a more future-proof design with the ability to grow
smoothly and support future, anticipated business needs.

101
Chapter 5

Trunking capacity is always a consideration in voice networks. It’s tied directly to carrying
capacity. In the VoIP world, how and where gateways are connected to the outside voice
network, or PSTN, are important performance and management factors. T1 circuits, whether
primary rate (PRI) or standard channelized connections, can provide a standardized approach
that traditional telecom providers support. This approach is simply building T1 trunks in the
same way you connect an enterprise PBX to the PSTN today. In the traditional T1 environment,
a T1 circuit can carry 24 voice calls.
IP trunking presents a different approach, using the full bandwidth of a network connection to
the carrier. SIP trunking over IP is becoming very popular and quite common. SIP trunking is
easy to implement in most current vendor offerings, and provides an easily quantifiable return on
investment. In short, it’s proving to be quick and easy to implement, and a powerful cost-savings
approach. Given the codecs mentioned earlier, a 100Mb SIP trunk to the carrier, using a codec
such as G.729a, can carry a high volume of voice traffic easily.

Chapter 7 will explore trunking and other services, such as E-911, in more detail in Chapter 7.

The Performance Envelope


There’s an old networking paradigm that has been used by many IT managers. When network
performance seems sluggish, add bandwidth. This approach works in many cases, but it’s usually
a temporary fix. It’s also an approach that, if followed over time, guarantees that network circuit
costs will spiral ever upward. Bandwidth is an important characteristic in network performance,
but it’s not the single most important factor. Networks have a wide-ranging set of performance
characteristics that determine how smoothly they operate.
There’s a principle in network management, often related to security, which identifies three
important network requirements—confidentiality, integrity, and availability (often called CIA).
There are many different components and characteristics that play into achieving complete
confidentiality, integrity, and availability in the enterprise network. These characteristics make
up what I’ll call the network performance envelope. The shape of this envelope gives a view into
the “personality” of the network.
We’ve touched on the importance of knowledge about your network operating environment up to
this point. This section will put together the pieces so that you can utilize a knowledge-based
approach to ensuring success in deploying integrated VoIP services. The more you know about
the network, the more accurate your planning will be, increasing your ability to meet call quality
expectations for end users. Better data improves your ability to provide better total network
performance.
The basic concept of confidentiality, integrity and availability doesn’t offer enough granularity to
fully assess the requirements for successfully delivering a converged VoIP service. To
accomplish that, you’ll need to look at a broader set of data inputs. You’ll need specific,
measurable data that you can assess to gain a complete understanding of both the existing
network and the new requirements.

102
Chapter 5

There are a nearly unlimited number of different data elements an organization might use or
consider when defining the network performance envelope. This guide will use the factors shown
in Figure 5.1 as an example and will explore some characteristics in more detail than others.
These chosen factors provide a good representative sample of the types of characteristics
enterprises might consider when working through this process.

Availability
Throughput
Manageability
Scalability

Integrity Cost

Response
Time CPU Utilization

Reliability Security
Network Segment
Utilization

Figure 5.1: Graphing performance characteristics.

This figure shows the basics of CIA in a granular form and brings some other network
performance characteristics into the mix to ensure you can fully support converged VoIP
services.
Reliability overlaps both the availability and the integrity of the network. Because it’s so
important, we’ll treat reliability as a discrete evaluation factor here. Throughput also plays a dual
role, providing both availability and integrity. We’ll include three facets of throughput in the
example:
• Because bandwidth is a major contributor to throughput, we’ll measure it as a standalone
factor.
• Response time provides a measurement of delay or latency in the network. As VoIP is an
end-to-end service and delay is cumulative, we’ll look at response time as a broad
indicator.
• CPU utilization may be a performance factor across every element of the network.

103
Chapter 5

We’ve added cost, which is unrelated to CIA in every way, but a business consideration that
must be taken into account with every network change. Some enterprises might substitute a more
comprehensive ROI model as part of this assessment. No matter how simple or complex your
assessment mechanism, managers will have to answer for profitability and cost recovery of
changes to network services.
You must asses the manageability of the network. Deploying new services that cannot be
effectively managed is courting disaster and ensures failure. For many service delivery
organizations, this correlates into a service level agreement (SLA), or contractual commitment.
Scalability of the network has been included here as a broad measure for future-proofing your
design. Every business enterprise plans for business growth. Business growth inherently drives
network growth as almost every business sector becomes more and more dependent on data
networking resources. You’re striving to future-proof the network for some period of time,
typically 3 to 5 years. No network designer or engineer wants to implement a major new service
such as VoIP on a network, only to discover a year later that the network was under-powered
during a major project and now requires a complete redesign.
Integrity encompasses network reliability requirements. Security is included here as an aspect of
integrity, encompassing three facets for consideration:
• Packet loss, if minimal, might not adversely impact VoIP services; however, it’s a
performance component that should be measured and included in your analysis.
• Jitter, or variability in delay, degrades the quality of voice services if not controlled.
• Security can be broken down into many subcomponents. These might include firewalls,
intrusion detection solutions, antivirus software, and other tools. For the purposes of this
model, security will be treated as a single component.
Availability may have many different meanings to many different people. To some enterprises,
and most of the telecommunications carriers, it generally indicates a guaranteed uptime at the
five nines, or 99.999 percent level. Rather than deeply explore network design issues such as
high availability, redundancy, and business continuity, availability will be used here as another
data point on a graph for analyzing the network performance envelope.
Figure 5.2 gives you the next step in the process. It presents the framework you can use for
mapping each operational characteristic of the larger network performance envelope.
Network services and applications all have unique and different requirements. Email can be
delayed several minutes without any measurable negative impact. Delays in Web browsing
might be barely noticeable as the browser screen gradually loads information. Its use of
Transmission Control Protocol (TCP) at a higher layer guarantees information delivery, although
the overhead might add delay. Web traffic is non real-time traffic, so between a person and a
system, performance impact may be negligible. Integrating VoIP into the often already overtaxed
data network introduces a whole new set of requirements into the network performance
envelope. As these new characteristics needed for new network services are identified, they can
be mapped onto the graph. As you do so, notice that the personality of your network, the
performance envelope, begins to take shape.

104
Chapter 5

Availability
Throughput
High
Manageability High

Easy
High
Scalability

Integrity (Errors, <.1


%
Moderate
Cost
Collisions)
<100ms
75%
Response
Time 70% CPU Utilization
High
High
Reliability Security
Network Segment Utilization

Figure 5.2: Identifying performance characteristic requirements.

Every network service and application in operation needs to be evaluated and factored into the
performance envelope data set you’re evaluating. Remember, you’re identifying the
requirements for successful service delivery—your success factors. You are planning to succeed.
For existing services such as email and Web traffic, you might make some simple assumptions
about acceptable performance. For mainframe applications, or those that were custom-developed
in-house, it might be prudent to overlay the specific requirements to support each. The key is that
each requirement be identified using some measurable value. The data points or elements used
by any organization can be established as appropriate.

This model is demonstrated in a very simple format. Each network has different and unique
requirements depending on the services provided and applications in use. If your network employs
QoS mechanisms for different traffic types, don’t forget to account for each of them when graphing
out your network requirements. The more granular the data used, the more accurate the assessment
of performance requirements becomes. As always, the more you know about the network, the better
your decision will be.

Every performance characteristic you asses becomes a data point on the graph lines. In the next
step, which Figure 5.3 shows, “connect the dots” to gain visual representation of the shape, or
personality, of your network. You know what the network needs to look like to successfully
deliver the existing service and applications in addition to VoIP and other new services needed.

105
Chapter 5

Availability
High
Throughput
High
Manageability
High Easy Scalability

Integrity (Errors, <.1% Moderate


Cost
Collisions)
<100ms
75%
Response Time
CPU Utilization
70% High
Legend High
Envelope
Limits
Reliability Security
Network Segment Utilization

Figure 5.3: Giving shape to the network performance envelope requirements.

For many organizations, now the easy part is completed. The next step in the process is to
physically measure each data point on the graph. Some elements, such as cost and security, may
be relative assessments rather than technical measurements. The key to success is to be
methodical and thorough. .This step will quantify network performance capabilities today. As
Figure 5.4 shows, when you map the real-world measurements from your network with the
performance envelope you’ve established as a requirement, you have a gap analysis mechanism.
This gap analysis can now help you focus directly on those areas of network performance that
fail to meet your established requirements. This helps ensure a methodical approach to upgrading
the network.
As part of the overall operational requirements of the new, converged network, security and
performance management often mesh to become one overarching facet of design consideration.
Achieving the delicate balance between service delivery and security requirements means that
compromises and tradeoffs will be necessary. The right set of management tools will allow you
to continually monitor network performance and security, assess risks, and measure performance
around the clock.

106
Chapter 5

Neither performance nor security can be effectively monitored as a single element within the
network. The network is a large, almost organic environment. A systematic and holistic view of
the health and welfare of the entire system requires vigilance and constant review. Beyond this
methodical approach to assessing the network performance, requirements implementing
repeatable, sustainable processes will help ensure consistent network performance that delivers
both the quality and security necessary for widespread enterprise success. There are a variety of
Network Management Systems (NMSs) in use in business networks today. These range from
expensive commercial products to freeware and open source tools. Beyond network and security
monitoring tools, there are a number of VoIP-specific solutions to aid in constant monitoring of
the VoIP service. The best resource for identifying the tools that will work in your environment
is generally the VoIP solution provider.

Availability
High
Throughput
High
Manageability
High Easy
Scalability

Integrity (Errors, <.1% Moderate


Cost
Collisions)
<100ms
75%
Response Time
CPU Utilization
High
70%
High
Reliability Security
Network Segment Utilization
Legend
Ideal
Actual

Figure 5.4: Gap analysis—Mapping existing performance against the requirements.

You can use this performance envelope graph to overlay the performance characteristics the
network provides today with the requirements you’ve already documented. You know your
requirements for successful service delivery and you know your network capabilities. Figure 5.4
demonstrates a common occurrence in many networks. The reality of network measurements and
the requirements don’t align in all areas. In some characteristics, the network provides better
service than needed. This may mean that you’re paying a premium price to a carrier
unnecessarily. In other areas, there are gaps to address in order to meet service delivery needs.
Implementing a methodical approach such as the performance envelope when analyzing service
requirements also helps in defining VOIP service delivery expectations. Clear expectations are a
key factor to a successful implementation. This approach ensures that you know what is required.

107
Chapter 5

Completing a thorough assessment of the existing network has two benefits. First, it provides an
accurate and viable gap analysis to use in preparing the network for converged services. In
today’s frenzied network operations environment, it’s sometimes too easy to assume that adding
a switch here, a link there, and increasing the bandwidth somewhere else will meet service
requirements. It’s important to take the time to complete a thorough assessment and provide
comprehensive gap analysis information to achieve success. As a side benefit, especially in large,
distributed networks, some enterprises will identify areas in which networks have been over-
engineered in the past. This can lead to redesign and cost savings.
This performance envelope readiness assessment is an opportunity to re-evaluate the existing
network and determine whether current needs are being met. Many corporate enterprise networks
were not designed to be what they have become today. Networks often began as small islands of
information—isolated workgroups or departmental LANs. As the network gradually grew,
connections to other groups and organizations were added. In most organizations, new business
applications were also added over time. Corporate networks have grown from a simple beginning
into a complex and sophisticated mesh, weaving corporate operations together. Often they’ve
done so without being revisited from a holistic, service delivery perspective. This performance
envelope approach to network assessment can help lead to a network that delivers better
performance at a lower cost.

Choosing Performance Envelope Characteristics to Measure


Data, voice, and video services each place different demands on the network. They’re different
types of service. Each one is designed to provide a specific service. To support different types of
service in the network, you need to be able to offer some consistent and predictable QoS that you
can manage. This is often accomplished by creating a specific class of service for each type of
traffic supported in the network. For some network designers, determining the number of service
classes required can be a challenge. Although implementing QoS mechanisms isn’t a
convergence step that’s required, many organizations find that VoIP drives the need in order to
deliver the required performance characteristics.
QoS can become very complex. Implementing an overly complicated QoS scheme can lead to a
network that can’t be readily supported. Rather than add undue complexity, let’s look at just
some of the performance characteristics included in the performance envelope example.
As network convergence and emerging technologies gain momentum, be mindful of the danger
that each new application brings. It’s important to look toward the future and not create a new
class of service for every new application, or network complexity can spin out of control. For
manageability, most network engineers favor using only a few critical service classes.
VoIP and video collaboration represent real-time traffic. This is most often communications
information flowing from person to person, rather than interaction with a server or system. These
real-time services require quick delivery with quality assurances for delay, jitter, and loss.
IP networks use a best efforts approach to deliver all general traffic. This class of service is quite
suitable for email, Web browsing, and most other normal network traffic. Best effort simply uses
whatever network resources are available.

108
Chapter 5

In some networks, management traffic may warrant a dedicated class of service all by itself. This
approach is quite common in service provider networks. It provides a mechanism to ensure the
network can always be managed, regardless of congestion problems that might arise.
All QoS mechanisms provide some form of traffic prioritization scheme. In the converged
network, there are many different traffic types. Each may have different requirements and
different prioritization needs. Similar traffic types need to be identified so that they can be
handled the same way within the network. Most organizations choose to aggregate similar traffic
types. This approach allows a company to take advantage of network routes that are optimized to
provide the appropriate class of service.
In designing a VoIP service network, the focus is typically on call quality for the voice user, but
it’s important not to degrade pre-existing services when implementing VoIP. You need to
recognize all the different traffic types in use. If mission-critical data applications aren’t given
the necessary resources through QoS prioritization, the applications might “starve” for lack of
resources. Email and Web traffic will still need to be delivered, even if it’s a lower priority. It’s
vital to maintain balance across all traffic types when managing an integrated data, voice, and
video network.

Throughput
Throughput is frequently measured in terms of bandwidth. When evaluating throughput
requirements, due consideration must be given to traffic aggregation points. Don’t overlook the
congestion issues that can develop as a result of combining 10Megabit, 100Megabit, and Gigabit
Ethernet connections on the network. This traffic, all flowing to a centralized aggregation point,
may overwhelm a lower throughput link. This can aggravate network congestion and introduce
service delivery problems.

Bandwidth
Over-engineering or over-provisioning the IP network has been a common approach for many
network engineers. Increasing bandwidth, by ordering higher capacity links, has been the most
common technique. Adding bandwidth may alleviate short-term problems, but it’s important to
remember that IP uses all the available network resources. Bandwidth is one resource. Over-
engineering frequently proves to be a delaying tactic that simply stalls necessary redesign work.
This approach can be more costly in the long run. Adding bandwidth still requires investment—
upgrading equipment and increasing bandwidth of circuits. These can become very expensive
approaches, and they don’t solve the problem of design. IP data applications can quickly
consume all available bandwidth, leaving the same congestion problem to be addressed.

Response Time
Response time is one measure of network performance, most often measured using ping as a test
tool. One important nuance in the converged voice and data network is the fact that voice is an
end-to-end service between people. Ping can be an effective diagnostic tool, but the end-to-end
nature of voice service, coupled with the fact that delay is cumulative, necessitate comprehensive
management and testing to ensure service levels and call quality are maintained.

109
Chapter 5

CPU Utilization
CPU utilization in network nodes offers a good indicator of the overall health of the network. In
planning VoIP services, utilization might provide an indicator that network elements are
overtaxed and can’t effectively support VoIP. After VoIP services are live, ongoing monitoring
of CPU utilization can provide a benchmark and timeline trend analysis to monitor the health of
the network over the complete life cycle.

Network Segment Utilization


Ethernet is, at its roots, a shared media technology. LAN switching provided network technicians
a means to segment traffic into smaller broadcast domains. This increased granularity is now
often enhanced through the use of virtual LANs (VLANs). Like bandwidth, CPU utilization, and
other factors, network segment utilization can be monitored as part of the day-to-day
management operations to ensure adequate network performance to support the required
services.

Integrity/Reliability
The integrity and reliability of the network encompasses a number of different technical facets.
Each can be monitored, measured, and managed as a part of network operations. In the legacy IP
network, these may have provided acceptable service and been left unchecked. In the converged
service network, they warrant ongoing monitoring and proactive management

Packet Loss
A common measure is error rate and data loss. When IP networks are used to transmit normal
data—email, file transfer, Web browsing, and so on—some data loss is acceptable. The higher
layer protocols, such as TCP provide a measure of quality assurance and request retransmission
when needed. Other data types, mainframe data, as noted, may be very intolerant of packet loss.

Jitter
Jitter describes the variations in delay. As IP networks route traffic over the network using the
best path identified by routing protocols, it’s possible that every packet in a stream might take a
different route. A VoIP call could potentially traverse many network paths. Each route through
the network may have different delay characteristics. Jitter typically isn’t a concern for the
normal IP data traffic c. VoIP is far more sensitive to jitter than email or Web browsing because
it is a real-time service between people. High jitter can result in unintelligible conversations that
sound “jerky.” Users won’t trust or use VoIP services if the call quality is unacceptable.

Delay/Latency
Delay exists in all IP networks. It exists in several forms. Routers use statistical multiplexing
algorithms to process traffic. Assembling data into packets takes time. Checking the routing
protocol to identify the best route through the network takes time. These miniscule delays add up
and all impact the total end-to-end delay. Delay is cumulative.

110
Chapter 5

Cost
No business can operate without maintaining vigilance in controlling cost or expense. In
networking, you face not just CAPEX in hardware investment but also the operating expense
(OPEX) of keeping the service up and running on a daily basis. Some businesses, notably service
providers, focus on profitability and use ROI as a performance envelope measurement.
Enterprise businesses may treat their IT and network operations internally as either a profit
center or a cost center. For many, no profit is expected, but the costs associated with
implementing and managing the network still must be recovered. For most organizations, over a
3-to-5 year life cycle, network OPEX tends to be much greater than the initial CAPEX
investment. It’s important to always factor the appropriate cost analysis in both VoIP pre-
planning and the ongoing service management.

Availability
Availability, for many, is defined as reliability. It means that the network is fault tolerant and
services don’t degrade when problems do occur. To provide reliability in the PSTN, there are
millions of circuit paths available through hundreds of central offices. In an enterprise IP
network, redundant paths, alternative routes, load-balanced connections, and high-availability
equipment may all need to be incorporated into the design to ensure resilience in the network.

Uptime—99.999%
Availability in traditional telephony has often been measured as the uptime percentage. In
commercial networks, you’ve heard the term “five nines reliability” (or 99.999 percent uptime)
used as the target availability measure. For enterprise operations, it’s crucial to recognize that
this number equates to roughly 5 minutes of downtime per year. Although the commercial
telephone providers have widely met the five nines measure, there aren’t many corporate data
networks that can claim less than 5 minutes downtime in the past year. Introducing VoIP and
other real-times services into the network raises the bar for IP network availability and drives
designers to invest in robust, fault-tolerant design solutions.

Security
When implementing VoIP solutions, network security is as great a concern as reliability and call
quality. Corporate networks may include firewalls and multiple connection points. These
security devices can add nodal processing delay that may impact VoIP services. The more
complex a rule set in a firewall, the more latency it induces to the data flow. Remember that
delay is cumulative and counts toward the 250ms maximum tolerable end-to-end delay.
VPN services are deployed in two typical fashions. Point-to-point VPN solutions may be used to
connect remote offices over the Internet. Many companies also use VPN services allowing
employees to connect to network resources while telecommuting or away from the office.
Encryption algorithms consume processor power. A VPN device running DES or Triple DES
encryption as a VPN endpoint will add further latency as packets are encrypted and decrypted. If
the VPN endpoint is a firewall, this CPU load problem may be further compounded.
Security concerns reinforce the importance of assembling a comprehensive technical team for
network assessment, readiness testing, and managing the operational environment. It’s crucial
that the telecommunications, IT, and network security teams collaborate to be successful.

111
Chapter 5

Manageability
The human resources aspect of supporting network services may be the mostly costly
component. This is the biggest OPEX cost component. The more complex and difficult the
network is to manage, the more resources required. As network designers and service providers,
we must consider the ease of management we’re incorporating into VoIP network design.

Scalability
Scalability provides the best measure for future-proofing the network. Corporate networks
change, grow, and evolve continually. They become very organic in nature as business needs
change. The evolutions of the SOA and SaaS on the network are beginning to accelerate. The
demands placed on the network will grow as applications, services, and networks become more
tightly coupled with business processes. You must always consider the ease with which the
network can scale to support new business applications and services.

Summary
The key to successfully managing the converged network begins during the planning process
long before implementation. Conducting a needs analysis will help define the business needs in
measurable terms. Completing a methodical network readiness assessment will further develop a
holistic vision of what the network must support in order to successfully deliver integrated
services. Don’t sidestep these important steps early in the VoIP deployment process. They will
not only ensure that the converged network fulfills your business needs but also help formulate
the day-to-day management methods required to support the enterprise integrated services.
Perhaps a key differentiator in using this performance envelope approach to network
performance analysis is that it offers a tool to support all services in the network. VoIP is
important, but unified communications is only effective when it supports the larger unified
application environment that supports the complete business of the enterprise. It offers a holistic
view of not just data voice and video but also enterprise-specific applications and services. Sales
force automation, inventory management, customer relationship management, and Web-based e-
commerce are examples of vital application services that the enterprise needs in order to thrive.
Although the software-oriented architecture mindset will more tightly couple these applications
with services such as voice and data, using the performance envelope as a holistic approach is an
integral part of managing a converged network, encompassing the full spectrum of data, voice,
video, applications,

112
Chapter 5

This chapter only briefly touched on the idea of using outside resources. It’s worthwhile, in
summary, to review how outside integrators and consultants can work for you. They can help
look out for your best interests. There are many VoIP consulting firms that specialize in
converged technologies. Their reputation is built on how well they take care of client needs.
They build business through word of mouth referral and will be your advocate for success. They
can ensure that you are being well served by your vendors.
Your equipment vendors and service providers will gladly offer consulting services, and their
sales engineers will graciously come in and help you design a solution. It may be pertinent to
recognize that vendor staff may be specialists in their company-specific products. They may do a
very good job based on their own experience, but they might not have comprehensive training on
the workings of competitive solutions. Use your trusted and incumbent vendor partners, but
don’t give them free rein to design your network. You need to remain in control of the project
overall, because you’ll have to manage it in day-to-day operations.
The next chapter will delve into the impacts of events in the network, root cause analysis, and
how event correlation comes into play as a critical network management tool.

113
Chapter 6

Chapter 6: Impact Analysis, Root Cause, and Event


Correlation
To conduct full impact and root cause analysis, event correlation engines are often used to
provide data about what happened. This chapter begins by building an understanding of
protocols involved, examining their strengths and weaknesses. Key protocols include Simple
Network Management Protocol (SNMP), the Internet Control Message Protocol (ICMP) tools,
and even Network Time Protocol (NTP) for effective event correlation.
Beyond protocols, the chapter will explore syslog. Syslog servers provide a part of the picture,
but they really provide data collection mechanisms, not analysis engines. Correlating events
across an enterprise network of disparate systems presents a difficult challenge.
In Business @ the Speed of Thought (ISBN: 0446525685), Bill Gates described what he called
the “digital nervous system.” He said, “The most meaningful way to differentiate your company
from your competition...is to do an outstanding job with information. How you gather, manage,
and use information will determine whether you win or lose.” When deploying converged
networks over IP, you’re integrating voice technology with the critical data infrastructure.
Building monitoring and management processes into daily network operations provides the
information, or knowledge base, about the corporate nervous system that lets you manage a
complex, almost organic, business operating environment. Your management and monitoring
tools become a key part of your enterprise business intelligence.

SNMP
SNMP is a widely used protocol for monitoring the health and well-being of a network. It’s a
simple, text-based protocol that uses a database called a management information base (MIB) to
describe network device management data. Almost all network elements are SMNP-enabled.
Most equipment comes from the manufacturer with the community strings of public and private
enabled by default. Typically the public string provides read-only access. The private community
string often provides write access also, and is often used for managing devices remotely and
“pushing” updated configurations to routers and switches across the network.
SNMP was designed to ease monitoring and remote management of network elements. These
include servers, routers, switches, and even workstations. It can provide monitoring for
performance, utilization, and state information about the device. SNMP uses what are called
“traps” to capture this information, which is then often passed on to a centralized management
station in a network control center. These stations typically provide network maps, with icons
representing each node being monitored. In many systems, a simple green-yellow-red icon
allows easy monitoring of network element status from healthy and operational (green) to
potential problems (yellow) to out of service (red) conditions.

114
Chapter 6

What Is a MIB?
The MIB is a type of database, comprising a set of objects used to manage individual network
elements. MIBs are structured based on the OSI/ISO network management model. In the public
switched telephone network (PSTN), Abstract Notation One (ASN.1) has been used for years as
a mechanism for describing the object data structure of that network’s elements. The PSTN
elements include things like Class-5 central office switches, carrier trunking technologies, and
the SS7 signaling network elements. ASN.1 was jointly developed by the ISO and the ITU-T in
1984. Today’s network MIBs are developed as a subset of this larger standard. This subset is
defined in IETF RFC 2578.

IETF RFCs for MIBS


RFC 1156 - Management Information Base Network
RFC 1157 - A Simple Network Management Protocol
RFC 1441 - Introduction to SNMP v2
RFC 2579 - Textual Conventions for SNMP v2
RFC 2580 - Conformance Statements for SNMP v2
RFC 2578 - Structure of Management Information for SNMP v2
RFC 3416 - Protocol Operations for SNMP v2
RFC 3417 - Transport Mappings for SNMP v2
RFC 3418 - Management Information Base for SNMP v2
RFC 3410 - Introduction and Applicability Statements for Internet Standard Management Framework
RFC 3411 - Architecture for Describing SNMP Frameworks
RFC 3412 - Message Processing and Dispatching for the SNMP
RFC 3413 - SNMP Applications
RFC 3414 - User-based Security Model (USM) for SNMP v3
RFC 3415 - View-based Access Control Model for the SNMP
RFC 3584 - Coexistence between SNMP v1, v2 and v3

A MIB Object is one of any number of specific characteristics of a managed device. Examples of
MIB objects include:
• Output queue length, which has the name ifOutQLen
• Address translation table (like ARP tables) called atTable

115
Chapter 6

The Architecture of SNMP


There are three components needed for managing a network with SNMP:
• Network management systems—The network management system (NMS) runs the
applications that control and monitor the network devices. The NMS is an active system
that sends and receives SNOM queries. It accomplishes this by “setting” SNMP traps,
then “getting” the results.
• Agents—An agent is simply the SNMP component of the software running on the device
being monitored. This might be an integral part of the OS software or it might be another
process or daemon that is executed when the device boots. The agent has information
about the local operating characteristics of the network device. This information makes
up the MIB for that specific device. The agent translates that information and provides
the communication with the NMS.
• Managed devices—These are the network elements that are monitored by the NMS.
These devices are typically routers, switches, servers, printers, and other service delivery
elements of the network. In a unified communications design, these also often include the
gateways, session border controllers, signaling and voice servers, and voicemail systems.
Although workstation OSs—such as Windows—include SNMP capability, they are
generally not monitored except when required for some very defined requirement. These
devices often collect and store some form of management information locally in either
event logs or syslog files.
SNMP provides a standards-based protocol and mechanism for remote monitoring and
management of the unified communications network on a large scale. SNMP currently exists in
versions 1, 2, and 3 in the real world. SNMPv2 was not widely adopted due to disagreements
over the security framework, but many networks are evolving to use SNMPv3.
Version 3 includes some important new features. The most notable is encryption of the data in
transit. Earlier versions send data in plaintext, which is easily read, making SNMP a prime tool
for a malicious intruder to learn about the network. Encryption ensures that only the NMS and
authorized personnel can read and evaluate this information. Different SNMP versions can
interoperate to a limited degree. Interoperability between versions is explained in IETF RFC
3584.

116
Chapter 6

Remote Monitoring MIBs


Remote Monitoring (RMON) is another technical specification that provides for a different variety of
network monitors and console systems. RMON is designed to support network probes and monitors
(often called sniffers). It allows the integration of diagnostic tools from multiple vendors, which may be
used for very specific diagnostics or analysis.
RMON was initially developed when LAN switching became popular. It allows for managing switched LAN
segments from a central monitoring facility or Network Operations Center (NOC). RMON is simply
another extension of the standards already described as part of the SNMP MIB.
Unlike SNMP, RMON uses only two components. The probe contains the agent and is inserted into the
network. One example would be a sniffer inserted into a specific network segment or VLAN for
troubleshooting purposes. The other component is a management station. This workstation is frequently a
network engineer’s workstation, used interactively in troubleshooting and problem diagnosis. Like SNMP,
RMON information uses the MIB found locally on the device, but the RMON agent is most commonly
embedded in the OS. RMON agents don’t monitor the entire system; only the traffic flowing through the
RMON device. An RMON sniffer placed in listening mode on a LAN segment can only report on traffic on
that LAN segment.
There are several variations in RMON MIBs. The Token Ring RMON MIB, for example, provides specific
objects for managing a Token Ring network. The SMON MIB extends RMON and provides support for
RMON analysis of a switched network.

SNMP is a very simple application protocol. Because it doesn’t require a full three-way
handshake or guaranteed communications, it’s encapsulated in User Datagram Protocol (UDP).
All three versions of SNMP contain the same message components:
• Version—The SNMP version number. It’s key that the agent software running on the
network device and the NMS use the same version of SNMP. Messages that arrive tagged
as being a different version are typically discarded by the NMS.
• Community—The community name, or string, is used to authenticate the management
system and grant either read or write access to the agent. The most common default
strings, described earlier, are public and private. Many vendors’ products come with
other SNMP community strings enabled by default.
• PDU (Protocol Data Unit)—The PDU types and formats are different for SNMPv1, v2,
and v3. A PDU is a descriptor of how information is packaged for a given technology.
For example, the PDU for LAN technologies such as Ethernet and Token Ring is called a
“frame.” Ethernet and Token Ring frames differ in format, but the PDU they transport is
a frame. IP packets are the PDU that IP transmits and differ slightly in format from other
packets.

117
Chapter 6

Using SNMP
SNMP uses IPv4 but also supports IPv6 for the future. The following list highlights the
capabilities provided by SNMP:
• Data gathering—Collect data from a device that is SNMP capable. Single requests can be
submitted using the snmpget and snmpgetnext requests. Multiple requests can be stacked
using the snmpwalk, snmptable, or snmpdelta commands.
• Configuration modification—The snmpset command provides for altering the
configuration information.
• Status checking—Commands such as snmpdf, snmpnetstat, and snmpstatus allow for
retrieval of status and other information.
• Translation—Converting MIB information, content, and structure from text and numeric
forms to other formats for use in a wide variety of analysis systems is accomplished using
snmptranslate.
Many current SNMP tools provide a graphical MIB browser. Most organizations use graphical
tools that provide some underlying, automated mechanism of implementing the snmptrapd
command to automate receiving of SNMP notifications. A GUI can provide a human-friendly
view that makes changes in the environment quickly observable. These notifications can also be
logged to a syslog server or an event log or exported to a plain-text file. They can also easily be
forwarded to other SNMP management systems and passed to external applications for event
correlation and further analysis.

snmpwalk
In UNIX systems, snmpwalk is a widely available application. An administrator can run a very
simple snmpwalk command
snmpwalk -c [good community string] [target host]
and learn a great deal of information about a device. Windows users can download a variety of
SNMP exploration tools from the Internet. These tools generally eliminate arcane command-line
interfaces, making basic exploration of networks and devices a simple point-and-click operation.
Figure 6.1 shows a simple snmpwalk of a print server on the author’s network using a GUI tool
from Solar Winds. In an enterprise network, routers, switches, servers, and VoIP service delivery
systems can yield routing information, user account information, performance information, and
details about TCP and UDP services running from this output information. When enabled,
SNMP can provide an administrator with extensive information about an enterprise network very
quickly.

SNMP is a reconnaissance tool. If SNMP must be enabled, it is absolutely critical that default
community strings be replaced. Just as a network administrator can use SNMP to perform quick
network reconnaissance and learn information about the network that must be kept private, so too
can an attacker. As a tool, SNMP is a double-edged sword, providing value while potentially exposing
vital information.

118
Chapter 6

Figure 6.1: An example of SNMP information.

Factors to Consider with SNMP


SNMP versions 1 and 2 do not encrypt the transmitted data. This means that management
information is passed in the clear and is quite readable by humans. There’s a security risk in
allowing critical management information to pass in the clear, even inside the enterprise
network.
Because the different versions of SNMP are not compatible, use of SNMP for network
management is often relegated to the lowest version supported in the network. For most
enterprises today, that is SNMPv2, which does not support encrypted messages. Upgrading an
enterprise network to SNMPv3 has often proven impractical. Existing routers and other network
elements often cannot support the newest version. Although upgrading the OS might seem like a
simple solution, often hardware replacement is the only viable means to upgrade to SNMPv3.
The benefits of the latest protocol standard may not be a powerful enough business driver to
warrant necessary hardware upgrades.

119
Chapter 6

Autodiscovery
SNMP tools are widely used by malicious intruders for reconnaissance purposes. Many SNMP
tools allow the simple use of subnet masking to run a scan across not just an individual network
element but also a subnet or full network to discover what devices are listening for SNMP
commands.
SNMP is a very simple network discovery tool. One of the features of SNMP tools is an
automatic discovery feature, through which new devices discovered in the network are polled
automatically. Most implementations will allow for a quick scan that yields tremendous
information. Even if the public and private community strings have been set to a secure string
that is not the default, the simple act of allowing SNMP enables discovery that quickly identifies
working IP addresses on the network and the domain or network names associated with each.

Negative Implications
SNMP may be the intruder’s easiest and most friendly tool. Software utilities are abundant for
free downloading. Many are point-and-click operations, requiring no technical skill. It’s quite
common within an enterprise for employees to be curious and use these simple tools for network
reconnaissance and exploration. Employees are typically within a trusted environment, so it’s
natural that they may have access to view a great deal of information. There is a danger of
network topography being mapped from within because of this implied trust relationship.
Vendor’s approaches to SNMP implementation vary widely. For some vendors, it isn’t an
element of the core product design but a feature that has been added or incorporated later in the
product development cycle. Since the tree structure and data indexing techniques may vary, the
internal data structures any particular vendor has implemented may vary. As a result, querying
the network equipment with SNMP can produce in unwanted problems, like increased CPU
utilization. Large routing tables, like those often found in Border Gateway Protocol (BGP) or
Interior Gateway Protocol (IGP), are one example of a situation where this problem is likely to
occur.
The lack of encryption capability in versions 1 and 2 introduce the threat of simple packet
sniffing/capture, easily revealing the plain-text SNMP community string. No versions of SNMP
use a challenge/response approach to authentication. That leaves all versions vulnerable to both
brute-force and dictionary attacks. An assortment of both free and commercial software tools to
instigate these attacks are readily available.
Because SNMP is UDP-based, it’s connectionless in nature. This leaves SNMP vulnerable to IP
spoofing attacks. Effectively restricting access to SNMP requires extensive access control list
implementation across multiple network elements in many corporate networks.
It’s noteworthy that SNMP has frequently surfaced in the SANS Institute’s Top 10 Most Critical
Security Threats as a result of the default community strings being set to public and private.

SNMP has frequently surfaced in the SANS Institute’s Top 10 Most Critical Security Threats as a
result of the default community strings being set to public and private.

For more information about SNMP security implications, the US-CERT maintains an excellent SNMP
Vulnerabilities FAQ at http://www.cert.org/tech_tips/snmp_faq.html.

120
Chapter 6

ICMP
ICMP is a foundation of the TCP/IP suite. It is mainly used by networked computers’ OSs to
send error messages—indicating, for instance, that a requested service is not available or that a
host or router could not be reached. In the connectionless, packet environment of IP, each host
and router acts autonomously. Packet delivery is on a best-effort basis. Everything functions just
fine as long as the network is working correctly, but what happens when something goes wrong
within the subnet? As a connectionless service, IP has no direct mechanism to tell higher-layer
protocols that something has gone awry. Furthermore, IP does not even have a method for peer
IP entities to exchange information; if an IP host receives a packet, it attempts to hand it off to a
higher-layer protocol. ICMP has been defined for exactly this purpose—IP-to-IP communication,
usually about some abnormal event within the network. ICMP messages are carried in IPv4
packets with a protocol value of 1.
ICMP is defined in RFC 792 and is part of STD 5, which defines IP; this strongly suggests that
ICMP is an integral part of IP. There are several types of ICMP messages. The following list
highlights the most commonly used ICMP messages:
• Destination unreachable—Indicates that packets cannot be delivered because the
destination cannot be reached. The reason is also provided. Examples include:
• Host or network unreachable or unknown
• Protocol or port is unknown or unusable
• Fragmentation is required but not allowed (DF-flag is set)
• Network or host is unreachable for this type of service
• Time exceeded—The packet has been discarded because the Time to Live (TTL) field
decremented to 0 or because all fragments of a packet were not received before the
fragmentation timer expired.
• Parameter problem—There was a problem with something in the packer header
preventing a router or host from processing the packet.
• Source quench—Indicates that a router along the path is experiencing congestion and is
discarding packets. This is usually caused by limitations in buffer space.
• Redirect—If a router receives a packet that should have been sent to another router, the
router will forward the packet appropriately and let the sending host know the address of
the appropriate router for the next packet.

121
Chapter 6

The remaining ICMP messages are used to query the network for information:
• Echo and Echo Reply—Is used to confirm whether systems are active. One host sends an
Echo message to the other. The destination system must respond with an Echo Reply
with the same data that it received. These messages are the basis for the TCP/IP ping
command.
• Timestamp and Timestamp Reply—These messages provide more information than the
simple Echo messages. A timestamp, with granularity to the millisecond, is inserted in
the messages. This provides a mechanism for measuring how long remote systems spend
buffering and processing packets. It can also be used as a clock synchronization tool
between hosts.
• Address Mask Request and Address Mask Reply—Can be used by network nodes to
determine their address mask when assigned an IP address.
• Information Request and Information Reply—This field is now obsolete.

ICMP
IP Packet IP Packet
Used to report general information
• Destination unreachable
• Time Exceeded
• Parameter Problem
• Source Quench
• Redirect
• Echo
• Timestamp
• Address Mask

IP Header ICMP Payload


ICMP Messages are carried directly within IP itself

Figure 6.2: ICMP.

122
Chapter 6

ICMP Message Format


Figure 6.3 shows the general format of an ICMP message. The following list highlights the first
four bytes of all ICMP (“error” and “query”) messages:
• Type—Indicates the type of ICMP message, including Echo Reply (0), Destination
Unreachable (3), Source Quench (4), Redirect (5), Echo (8), Time Exceeded (11),
Parameter Problem (12), Timestamp (13), Timestamp Reply (14), Address Mask Request
(18), and Address Mask Reply (19).
• Code—Additional information specific to the message type. In the Time Exceeded
message, for example, the Code field indicates whether the TTL counter was exceeded
(0) or if the fragment reassembly timer expired (1).
• Checksum—16-bit checksum similar to that used in IP.
The next four bytes are labeled miscellaneous. They’re used differently by different messages. In
most ICMP “error” messages (for example, Destination Unreachable, Source Quench, Redirect,
Time Exceeded, and Parameter Problem), these 32 bits are unused and set to 0. In the Parameter
Problem message, however, the first byte is used as a pointer to the byte where the parameter
problem was detected; in the Redirect message, these four bytes contain the address of the router
to which future traffic should be directed.
The final field shown in the diagram contains the IP packet header plus the first 64 bits of the
packet’s Data field (or payload) in the offending packet. The receiving host uses this information
to match the message to the appropriate CPU process. The 64 bits of user data are returned so
that at least part of the header of any upper-layer protocol, including any port numbers, gets back
to the original sender.

12 byte ICMP message format

Type Code Checksum

(Miscellaneous)

Internet Header + 64 bits of


Original Packet’s Data

General format for ICMP “error” messages

Figure 6.3: The ICMP message format.

123
Chapter 6

ICMP differs in purpose from TCP and UDP in that it is usually not used directly by user
network applications. One exception is the ping tool, which sends ICMP Echo Request messages
(and receives Echo Response messages) to determine whether a host is reachable and how long
packets take to get to and from that host.
ICMP Technical Details
ICMP is part of the TCP/IP suite as defined in RFC 792. ICMP messages are normally generated in
response to errors in IP packets, per RFC 1122 specifications, or for diagnostic or routing purposes. The
version of ICMP for IP version 4 is also known as ICMPv4, as it is part of IPv4. IPv6 has an equivalent
protocol, ICMPv6.
ICMP messages are constructed at the IP, or network, layer. They are usually built from a normal IP
packet that has generated some type of ICMP response. The appropriate ICMP message is encapsulated
in IP with an IP header in order to return the ICMP message to the originating host.
For example, every router in the network that forwards an IP packet must decrement the TTL field of the
packet header by 1. If the TTL reaches 0, an ICMP TTL Exceeded message will be sent to the source
from that router.
Every ICMP message is directly encapsulated in a single IP packet. Like UDP, ICMP does not provide
any delivery guarantees.
Although ICMP messages are contained inside standard IP packets, ICMP messages are usually
processed as a special case. They’re not normally treated as an IP sub-process because it’s often
necessary to inspect the contents of the ICMP message, then deliver the appropriate error message to
the originating host and application.

Reachability Testing
One of the most crucial tests for network monitoring is the simple determination as to whether a
system or network element is reachable via the network. The two most common tools for
determining reachability are ping and traceroute.

ping
Ping is perhaps the most widely utilized tool on all TCP/IP systems. It allows users to determine
the status of other systems. It also provides a tool for measuring the expected round-trip delay
between the local system and a remote network element. Ping is useful for many reasons. Prior to
attempting to establish a TCP virtual circuit, a local system might first ping the intended
destination to verify that it is up and reachable.
Ping uses ICMP Echo and Echo Reply messages. It has the following general format (where
items in square brackets [] are optional):
Usage: ping [-t] [-a] [-n count] [-l size] [-f] [-i TTL] [-v TOS]
[-r count] [-s count] [[-j host-list] | [-k host-list]]
[-w timeout] target_name

124
Chapter 6

In the first test that Figure 6.4 shows, the test pings the host www.yahoo.com to determine
whether it is up and running. This demonstrates the simplest use of the ping command and uses
none of the optional parameters.
The second test in the figure uses the optional -t parameter to tell the workstation to send an
ICMP Echo message continuously. The optional size and quantity parameters are not specified,
so ping uses the default values—64-byte messages. These are sent continuously until the
program is interrupted using Control-C to break the continuous cycle. The second test results in a
list of the round-trip delays experiences by each Echo message sent.

Figure 6.4: Ping.

Traceroute
Traceroute is another common TCP/IP tool that lets users learn about round-trip delays and the
network routing between systems. Traceroute works by sending a sequence of UDP packets with
an invalid port identifier to the destination system. The first three packets have the TTL field
value set to 1; this causes the first router in the path to send back an ICMP message reporting
that the TTL has expired. Then three more UDP messages are sent, each with the TTL value set
to 2, which causes the second router to send ICMP replies. This process repeats, incrementing
the number of router hops until the message actually reaches the destination. Traceroute
identifies a completed cycle when it detects an “invalid port” error reply.

125
Chapter 6

Figure 6.5 shows the route from a workstation on the author’s network to www.yahoo.com. The
route that is displayed tells the following:
• The first hop is through a system called GE-1-1-ur01.olympia.wa.seattle.comcast.net.
This is the first point on the path in the Comcast provider network. The second hop is
further along the way on the Comcast network.
• Hop 3 shows a hop across the AT&T provided backbone network in Seattle.
• Hops 7 – 10 traverse the Level3 backbone network from Seattle to San Jose.
• Hop 11 hits a network node that only provides an IP address, 4.71.112.14.
• Finally, on the Yahoo network, we hop across to the final destination at 209.131.36.158,
which has the DNS name f1.www.vip.sp1.yahoo.com. This is the system responding to
the request for www.yahoo.com.

Figure 6.5: Traceroute.

For more information about traceroute, see RFC 1393.

126
Chapter 6

Syslog, Data Logging, and the Console


The term syslog is used to describe both the application that sends syslog messages and the
syslog protocol itself. The syslog protocol, defined in RFC 3164, is very simple. A syslog sender
transmits a small text message to the syslog receiver or server.
Syslog is available almost universally to aid in systems management and security auditing.
Although syslog has several shortfalls, it is widely supported by almost every element of the
network. Because it’s nearly ubiquitous, syslog can be used to integrate log data from many
different types of systems into a central data store for event correlation and analysis.
Syslog is used for network management and security auditing. Syslog itself is quite simple. It
may seem simple for auditing use, but its broad availability is a great advantage. It allows a
centralized, corporate syslog server to become the central data repository for audit and event
correlation information. Syslog data is in plain-text format, so it’s easy to manipulate with
standard simple tools and scripts. Most organizations start with scripts and spreadsheets for
analyzing syslog data. Larger organizations, monitoring many devices, may find that this
approach is too labor intensive to be effective. Large log files and large numbers of log files may
require adopting scalable commercial tools and developing automated processes to ease the work
involved.
Technology alone can’t solve the anomaly detection problem. A great deal of syslog and event
monitoring is tied not just to performance but also to network security. What gets monitored,
how log data is used, and how the organization responds to events at the time of detection are all
a critical part of the cycle of network management, monitoring, and defense. Administrators
employ detection mechanisms because they offer notification as quickly as possible when a
network anomaly, intrusion, or other malicious event occurs. Network threats mutate quickly.
Worms spread almost instantaneously. The threat of zero-day attacks will not allow for weak
incident management prevention and detection processes. Effective incident management tools
and processes ensure quick reaction and recovery when an event does occur.
The syslog protocol provides a transport mechanism that allows devices to send event
notification messages over the network to syslog servers. These servers are often simply message
collectors that don’t return any acknowledgement.
The syslog protocol is very simple. The sender transmits a text message that is less than 1024
bytes. The syslog server (often referred to as syslogd or the syslog daemon) appends the message
to the file. These messages can be transmitted using either UDP or TCP. Normally, syslog data is
transmitted as plaintext, but there are tools that use an SSL wrapper to add encryption for
increased security.
Although TCP can be used, syslog doesn’t require a three-way handshake. Given the small size
of the messages, UDP port 514 is the most commonly used communication. As UDP is
connectionless, no acknowledgments are provided. At the application layer, syslog servers
normally don’t send any acknowledgments back to the sender either. Thus, devices transmitting
syslog messages never know whether the syslog server has received the messages. Most sending
devices will send syslog messages even if there is no syslog server in place.

127
Chapter 6

Syslog packets are limited to 1024 bytes and carry the following information:
• Facility
• Severity
• Hostname
• Timestamp
• Message
Syslog messages are categorized based on the generating source. These sources can be the OS
running a device, a syslog process (or service), or an application. To learn more about syslog, see
the IETF document at http://www.ietf.org/internet-drafts/draft-ietf-syslog-protocol-19.txt for
further technical details.

Integrating Tools for Event Correlation


An NMS is a combination of hardware and software used to monitor and administer the
addressable and manageable elements of the network. In converged service networks, VoIP and
video services introduce a new set of manageable network elements that perform
telecommunications service functions. These elements typically include gateways, call
management servers, emergency responders, voicemail servers, media gateways or servers, and
so on.
General network management involves functions such as network planning, traffic routing, user
authorization, configuration management, fault management, security management, performance
management, and accounting management. Many protocols exist to support network and
network device monitoring and management. As we discussed, SNMP is a common network
protocol, but others that may come into play include Common Management Information
Protocol (CMIP), Web-Based Enterprise Management (WBEM), Common Information Model
(CIM), Transaction Language 1 (TL1), and Java Management Extensions (JMX). We won’t
probe these protocols in depth here.
When implementing the converged network, NMSs take on a new, crucial role in enterprise
service delivery. Enterprises need to bolster their management capabilities to test and manage
QoS, performance, and availability in performance metrics, especially with VOIP services. To
get started, companies should analyze their business requirements and determine key
performance and QoS metrics.
A comprehensive, enterprise-wide data collection mechanism is required to provide effective
service assurances. Collecting as much data about the network as possible will aid in the ability
to ensure call quality and consistency of service.

128
Chapter 6

Network Monitoring
An NMS constantly monitors and notifies the network administrator via email, pager, or other
alarms in the event of outages or anomalies that exceed defined thresholds. Monitoring is vital to
service assurance and VoIP management. An NMS continually monitors the network for
problems that result from overloaded and/or crashed servers, network connections, or other
devices. For example, to determine the status of a Web server, monitoring software may simply
ping the server periodically to check for a response. A more comprehensive NMS technique is to
send an HTTP request to fetch a specific Web page; testing email servers might involve sending
a periodic test messages to ensure the email services (SMTP, POP3, and IMAP, for example) are
up and running properly.
Status request failures—like those found when a ping fails, the Web page can’t be retrieved, or
another unexpected condition is encountered—can be configured with most NMS platforms to
activate some predefined response. These responses can vary from event to event. In some cases,
an alarm might be sent to the systems administrator’s email, pager, or mobile phone so that
human intervention can follow. Highly evolved systems might trigger some automatic failover
system mechanism for continuity of operations. Or a non-critical server experiencing problems
might simply be removed from service until a suitable time is available for repair.
Some of the most important characteristics of network elements monitored in the IP network
include CPU utilization, physical memory, disk space usage, virtual memory, and fans and
power supplies. Many systems monitor temperature to ensure a proper operating environment is
maintained. Monitoring of system backups is incorporated to ensure positive confirmation that
backup jobs run as scheduled. Many organizations monitor Web server software (typically
Apache or Internet Information Services—IIS), directory services systems, and Domain Name
Service (DNS) servers. Security monitoring is often incorporated into the monitoring performed
in this network operations center environment.
Managing VoIP services raises the need to monitor both the VoIP service elements and QoS
facets of network performance to ensure acceptable call quality. In traditional IP networks, the
infrastructure elements are monitored. VoIP introduces new infrastructure elements including
voice processing systems, signaling servers, gateways to other networks, border controllers
supporting SIP trunking, and voicemail systems.

VoIP Service Elements to Monitor


Voice traffic carries a set of performance expectations users have come to expect through years
of telephone use. VoIP services introduce a new range of network elements to monitor.
Whenever a device (for example, phone, gateway, and gatekeeper) registers with the network,
there will be an auditing entry to review. Problems with device registration, for any reason, can
impact service availability. You’ll want to be alerted when the number of registration attempts or
failures exceeds predefined thresholds. If the number of registered telephones changes
dramatically, it could be a signal that there is a problem with the VoIP network. Gateway
registration monitoring will help identify new or missing gateway servers.

129
Chapter 6

Call monitoring isn’t eavesdropping on individual calls. It’s really call-traffic monitoring. It
involves monitoring incoming and outgoing call volumes to identify failures. If your VoIP
system supports fax calling, attempted fax calls also need to be monitored. Call monitoring
typically focuses on four specific areas:
• Calls in progress—When a VoIP phone goes “off hook,” a call is deemed in progress
until it goes back “on hook.” If every call in progress connects successfully, the number
of calls in progress will equal the number of active calls. When designing the VoIP
network, you’ll need to establish an upper-limit threshold for the number of calls that can
be in progress at any given time.
• Active calls—Active calls have successfully connected a voice path. Again, when
designing the VoIP network, you’ll need to establish an upper-limit threshold for the
number of active calls that can be handled at any point in time.
• Attempted calls—Designers strive to ensure that all calls attempted will be completed
successfully, but such isn’t the case in the real world. Monitoring calls attempted over
time yields data that aids in identifying peak periods and the busy hour call attempt
(BHCA) value.
• Completed calls—A completed call is any successful active call that completes without
an abnormal termination code. Monitoring completed calls over time is also useful in
identifying peaks periods and the BHCA value.
VoIP services need to interconnect to the PSTN through gateways. In addition to gateway
monitoring, it is vital to monitor the PSTN side of the VoIP service network. PSTN connections
are frequently established using ISDN Primary Rate Interface (PRI) channels over T-1 circuits.
Monitoring active PRI channels, especially over time, can help identify call patterns and busy
hour peak call volumes. Baseline data can also be used to identify underutilization of circuits.
Data trending helps in capacity planning and the growth and maturation of the VoIP service.
One benefit in deploying VoIP services is the conference bridging capabilities. If your
deployment supports conferencing, you must configure the maximum number of audio streams
that will be supported. Monitoring will ensure that the number of available audio streams meets
acceptable service levels for your organization.
IP phone functionality requires continual monitoring for service assurance. You should monitor
IP phones for their registration status, the validity of their dial tones, jitter, latency, and lost
packet count. These QoS parameters directly affect service delivery.

130
Chapter 6

Monitoring Bandwidth and QoS


Voice traffic requires specific bandwidth based on the codec used in the VoIP design. G.711
requires about 64Kbps for each direction of a bidirectional call. G.723 and G.729 require
significantly less bandwidth due to compression, but congestion can severely impact call quality.
When you add applications to your network, there is always a risk of oversubscribing links.
Oversubscription leads to congestion, and congestion may introduce a negative impact on call
quality. Packet loss and increased latency are common side effects of congestion and can, when
left unchecked, render VoIP services unusable.
For VoIP users to receive an acceptable level of voice quality, VoIP traffic may need to be given
some kind of prioritization over other kinds of network traffic, such as data. The main objective
of QoS mechanisms is to ensure that each type of traffic—data, voice, and video—receives the
preferential treatment it deserves, thereby reducing or eliminating the delay of real-time
streaming voice or video packets crossing the network.
The following list highlights examples of metrics that are frequently monitored because of their
effect on VoIP call quality:
• Delay or latency is an estimate of the network delivery time expressed in milliseconds.
It’s measured as an average value of the difference between the timestamps noted by the
senders and the receivers of messages. It is measured when the messages are received.
The end-to-end delay, or latency, measured between endpoints is a key factor in
determining VoIP call quality.
• Jitter is also called delay variation. It indicates the variance of the arrival rate of packets.
Jitter points directly to the consistency or predictability of the network. It is a call quality
factor known to adversely affect call quality. Networks can compensate for jitter by
implementing jitter buffers to normalize the timing of the traffic flow. Jitter buffer loss
occurs when jitter exceeds that which the jitter buffer can hold. Jitter and jitter buffer loss
affect call clarity, which affects the overall call quality.
• Packet loss indicates a packet lost during transmission. In VoIP, this could mean the loss
of an entire syllable or word during the course of a conversation. Obviously, data loss can
severely impair call quality. Monitoring systems measure the number of packets that
were expected against the number actually received.
• Mean Opinion Score (MOS) is a subjective measure used in voice telephony, especially
when codecs are used to compress the bandwidth requirement of a digitized voice
connection from the standard 64Kbps PCM modulation. MOS is generated by averaging
the results of a set of standard, subjective tests. In the past, a number of listeners rate the
heard audio quality of test sentences read aloud by both male and female speakers then
rate each as follows: 1-bad, 2-poor, 3-fair, 4-good, 5-excellent. The MOS is the
arithmetic mean of all the individual scores. In current systems, MOS is often determined
through software algorithms.

131
Chapter 6

Measurements and Metrics for Voice Quality


Voice quality measurement as part of operational monitoring can be either non-intrusive or
intrusive. Non-intrusive tests are typically based on actual voice conversations taking place
during daily operations, whereas intrusive testing requires placing test calls across the network.
One approach to evaluating call and voice quality is to assemble a group of participants who will
act as judges. A common technique is to have them listen to test calls, and assign scores from 1
to 5, much like the MOS evaluation testing. There are a number of algorithms and methods that
might be used, including MOS, Perceptual Analysis Measurement System (PAMS), Perceptual
Speech Quality Measurement (PSQM/PSQM+), and Perceptual Evaluation of Speech Quality
(PESQ/PESQ-LQ):
• MOS has been adopted from the PSTN and traditional telephony. This historical measure
of voice call quality was judged in the US by eight men and eight women who rated
voice quality on a scale from 1 at the low end to 5 at the high end. Although this human
evaluation of MOS was useful for identifying the quality of experience with calls in the
PSTN, it hasn’t proven useful for network monitoring or SLA compliance measurement.
• PAMS was developed by British Telecom. It doesn’t replace MOS but offers a
reasonable first attempt at automating the MOS scoring. PAMS compares the original
analog voice wave with reproduced speech using a complex weighting method that
purports to take into account specific characteristics that are important to the human ear.
The PAMS scale is from 0, which represents a perfect match between samples, to 6.5.
PAMS values estimate MOS scores with a ±10 to 20% level of accuracy that is
inadequate for use in VoIP networks. PAMS values are excellent for benchmarking
purposes.
• PSQM/PSQM+ is defined in ITU-T Recommendation P.861. PSQM estimates MOS with
a greater level of accuracy than PAMS at ±10%. PSQM and PSQM+ are also good for
benchmarking and comparisons. They represent improvements on the earlier PAMS
algorithm. PSQM and PSQM+ use the same 0 to 6.5 scoring that PAMS uses.
• PESQ/PESQ-LQ is defined in ITU-T Recommendation P.862. Like other techniques
described here, PESQ doesn’t replace MOS. It’s proven excellent for benchmarking and
comparisons and is an enhancement and improvement to the PSQM algorithm. It also
uses the same 0 to 6.5 scoring.

132
Chapter 6

The Do-It-Yourself Approach


Many enterprises create their own management platform suite over time. For many small to mid-
sized businesses, a managed service has always been the primary option. These organizations are
often resources-constrained and simply don’t have staffing to do everything themselves. Larger
enterprises often take exactly the opposite approach. These organizations have provided their
own telephony services for many years and seem inclined to continue this approach in the
emerging multimedia converged networks of today. Although many enterprises view VoIP and
video as new service applications on the IP network, the trend seems to be a continued do-it-
yourself approach to data, VoIP, and video. Although this is common, it isn’t necessary. For
many large enterprises, migration to a converged service network may present a perfect
opportunity to rethink service delivery and develop partner relationships with service providers
for both delivery and management of these services.
The enterprise network is evolving and becoming somewhat organic in nature. This converged
network provides a shifting set of real-time, near real-time and non real-time voice, data, and
video services. Even an internal SLA for workgroups inside an enterprise needs to be developed
to support metrics that are relevant to each of the services provided.

The SLA
An SLA is essentially a contract mechanism that documents the level of service that a customer
should expect to receive. An SLA that has been thoroughly thought out will also describe actions
that will be taken by the service delivery organization. This is where different service classes will
be defined and delivery characteristics identified for each.
Data, voice, and video services will be differentiated and require different traffic/service flows.
Thus, the network provides differing Classes of Service (CoS). To assure service delivery,
management, monitoring, and analysis tools must be capable of monitoring the service
parameters for each service call. The alternative is to implement discrete tools for each service
time. This option is often unwieldy and expensive.

SLA Metrics to Consider


The SLA process leads to an informed user within the enterprise. Users will begin to understand
multimedia services much the way an automobile purchaser understands performance
characteristics of different cars. The MOS, PQSM, R-Value, and other performance metrics
described earlier might become as familiar to the integrated services users as MPG, horsepower,
and such are to motor vehicle owners. Embedding the right tools within the service delivery
organization that support comprehensive monitoring, measurement, and analysis of multimedia
traffic compared with established baselines are crucial service delivery support tools.

133
Chapter 6

It’s important to look beyond the metrics that might be specific to VoIP and video. You must
also monitor more traditional aspects of the network including loss, delay, jitter, and general
availability. These measurements give a common basis of comparison across data, voice, and
video services.
In a collaborative network environment between end user workgroups and the service delivery
group, the SLA can be used as a tool for continuous improvement as metrics evolve over time to
reflect actual network requirements and performance capabilities. The right tools for monitoring
and managing your evolving corporate multimedia networks move far beyond the mindset of the
Internet as “plumbing” or a simple common architecture for delivering any traffic type.
Corporate management must understand, and appropriate necessary funding to support
specialized tools to maintain the health of the converged network. Without proper funding of
personnel and systems, the IP network may fail with dire consequences on day-to-day
operations.

Beyond Implementation—Operational Support


One of the implementation dangers in integrating network services is the inclination to reduce
staff without necessary skills transfer. Organizations that are overly focused on cost reduction
run the risk of embracing staff cuts followed by the creation of services that remaining staff can’t
effectively support. Many organizations have moved forward on the voice integration path with
the IP network staff leading the initiative. This can lead to gaps in vision of telephony services
and elimination of the traditional telecom support team. This is a risky proposition and the
business impacts should be considered at every turn.
It’s surprisingly easy to overlook specialized skills that enterprise telephony engineers bring to
the table. IT network engineers often don’t fully appreciate the workings of Erlang-B
calculations and the importance of traffic engineering. These technical resources are also the
ones that design interoffice trunking facilities, understand call center requirements, and build the
organization’s automatic call distribution groups, hunt groups, and call pickup groups. It’s
important to remember that although providing these services over IP may be a new approach,
these are still core voice services necessary for daily business.
It’s important to partner voice and data specialists in support of new unified communications
service. This approach will let you make the best use of every technical resource within the
organization. Don’t make the mistake of allowing key institutional knowledge about your
services, your network, and your requirements escape through oversight while focusing on cost
reduction. In the unified communications network, reductions in cost are won over time. They
may not come immediately.

134
Chapter 6

Even Help desk support for telephones is generally viewed as something simple. We assume that
everybody knows how to use a phone. In the integrated data, voice, and video network, simple
telephony features may well be delivered like they were in the past. It’s more likely that the
integrated system will bring new facets of data workstation feature management into play. For
example, one common feature in VoIP systems is to provide a way for users to manage their own
telephone features. Button configuration and speed dial lists may now be managed via a Web
interface from the desktop. Some method of oversight, whether through management and
monitoring or through specialized vendor-provided tools, will help simplify management of even
the simplest changes.
Remember that services converging onto a single infrastructure also mean the corporate Help
desk may become the primary support point for all new services. They may begin receiving
questions they’ve never dealt with previously. They too may be unfamiliar with the new
integrated services. It’s crucial that all support staff, including the Help desk, get the training and
management and monitoring tools they need to support daily business operations effectively.

Pros and Cons of Rolling Out Your Own Management Platform


There are distinct advantages in custom-building an enterprise-specific management platform.
When an enterprise opts to utilize a suite of managed services, the service provider has some
commitment to meet established expectations. In the do-it-yourself approach, you become your
own service provider. You might be an internal service provider delivering services to internal
departments, divisions, regions, and employees, but end user expectations remain the same. An
organization embracing the do-it-yourself approach must implement not just the services but also
the tools for measurement, monitoring, and analysis that will ensure service availability.
Vendors and manufacturers often provide the most accurate, most granular, and what might
appear to be the most desirable systems embedded within their solutions. These integrated
service hardware and software solutions commonly provide a combination of syslog and SNMP-
like monitoring. These tools can provide both real-time monitoring and event correlation
capabilities.
Many vendors also provide product-specific, proprietary tools to aid in management and
monitoring. It’s important to identify these tools early in the process. Some may be provided
freely as part of the product suite. Some vendors will offer yet another management suite to ease
the process. Again, it’s important to recognize the costs of adding more management tools to a
service that seems to be growing in complexity daily.
Third-party tools for modeling, monitoring, measuring, and managing are widely available and
can provide incredible value to service managers. Some provide broad visibility across a variety
of platforms, bringing insight into the network performance aspects that far exceed human
capacity to observe or analyze. Some of these third-party tools are commercial products. Some
may come from open source libraries.
On the plus side, there is a wealth of available resources. On the negative side, the onus is now
kept within the enterprise to select the right tools to manage, monitor, and analyze every facet of
the services provided to ensure availability and sustainable service quality.

135
Chapter 6

Freeware and Open Source vs. Commercial Products


When taking the do-it-yourself approach, don’t overlook the sophisticated tools already
embedded in the routers, switches, gateways, end systems, and other components of the
integrated network. There is a feature-rich toolset built-in to the products your organization will
use.
Although manufactures may provide the most accurate tools embedded within their hardware,
they may raise a different area of concern. The tightly coupled vendor solutions provide great
management granularity and detailed analytical capacity. However, unless the entire network and
all services are provided by a single manufacturer, this granular approach may be unable to
provide a holistic view of broad services. Manufacture-provided, proprietary tools certainly have
their place. They often prove invaluable when optimizing a network or when troubleshooting
specific system problems. But they may lack the broader, more standardized and less platform-
specific capabilities of systems provided by third parties.
Open source and freeware solutions present a unique set of challenges. Although the price may
be right (they’re often free), there may be no support available. This can lead to enterprise
engineers providing their own support, developing their own patches, and creating their own
performance tweaks. In time, open source software and freeware can morph into a variation on
in-house software development.
Another consideration may be the security of these solutions. There are two considerations for
open source solutions. Although an enterprise might build a level of trust for a vendor providing
commercial products, freeware, and open source tools essentially come from often unknown
sources. In many cases, this code has been vetted by hundreds or thousands of well-qualified
independent developers. The open availability of the source code means that many people can
contribute, each incrementally improving and crafting the code. However, whether appropriate
secure coding practices were used in their development may remain completely unknown. And
while the danger of these tools being purposefully developed with malicious code elements
nested inside has not proven common, these are management platform tools that support the
daily enterprise operations. Can you trust the “crown jewels” of the enterprise service network to
unsupported software written by unknown third parties? It’s important to weigh the value against
any and all potential risks.

Integrated Commercial Solutions


An organization that selects a managed service approach may be able to ignore many of the
technical details of the inner workings of service delivery systems. These enterprises may simply
need to “check the pulse” and “take the temperature” of the integrated services network. Often
service providers will either provide access to these tools or the tools themselves as part of the
managed service offering. These tools may simply provide a snapshot of the network status at a
given point in time. Some tools may provide a more detailed, near real-time view of the health of
the overall service. It’s important to review SLAs, as these tools may only provide a customer
view that might not correlate directly with service delivery requirements identified in the SLA.

136
Chapter 6

Multimedia networks supporting voice applications are complex systems that require
sophisticated monitoring intelligence. Although individual manufacturer’s monitoring tools offer
a partial solution, they cannot account for the granular nuance of monitoring and managing a
complex, multi-vendor service environment. Generic monitoring and management tools are
available, but generally insufficient. To truly be effective, third-party tools have to be assembled
with some knowledge of each of the individual manufacturers’ systems. This may be the only
way to obtain a comprehensive, system-wide, end-to-end view of the converged services
network. The following list highlights key areas for monitoring and management.
• Pre-deployment simulation
• Manufacturer-specific monitoring and management
• Real-time business views
• Call detail records
• Calls in progress
• Delay-to-dial-tone rates
• BHCAs
• Busy Hour Call Connects (BHCCs)
• Gateway channel utilization and loading
• Real-time call monitoring
• Phone and multimedia device availability and monitoring
• Poorly performing components
• Service level breaches and SLA compliance
• Real-time interface to manager of managers
• Summary and exception reporting
• Utilization trends over time
• Managed devices by company, department, and location
• Asset tracking
• Capacity planning
• Incoming and outgoing calls
• Loading by route pattern, route group, route list, and gateway
• Bandwidth utilization
• Delay and delay variation
• Packet loss
• Route patterns, utilization, and availability
Using the right mix of sophisticated third-party tools will provide a toolset for documenting
results consistently while eliminating any vendor bias.

137
Chapter 6

Pros and Cons of a Packaged Management Platform


As earlier noted, when choosing the do-it-yourself approach, organizations will still have access
to the sophisticated tools embedded in network equipment. There are still other tools that can
provide insight to the big picture of service management and aid in troubleshooting granular
problem areas.
Managed service providers will generally allow end-user organizations, customers, to view a thin
slice of their network for performance monitoring. These views pertain directly to a specific
customer organization. Service providers often call this service feature Customer Network
Management (CNM), which allows a customer to see their own part of the larger, shared virtual
network. This approach has been in wide use for at least two decades. Early CNM systems were
periodic summary snapshots showing sets of statistics that bore little resemblance to the real
world or the SLAs. Today, providers offer services that are much closer to real-time views, with
granularity down to the individual connection. In some cases, they allow customers to adjust
certain characteristics, such as CoS or available bandwidth, in real time.

Summary
Without an NMS focused on the integrated data, VoIP and video services, companies will find
themselves in the dark with a lack of information to support necessary service assurance. VoIP
services that lack management will be prone to service delivery and quality problems that cannot
be tracked to any specific network elements or service delivery metrics. The appropriate QoS
and network performance metrics require constant management, monitoring, and analysis to
ensure acceptable service delivery.
Integrated services introduce new complexities and opportunities to simplify the process of
moves, adds, and changes to the network. In the past, employees would often move their
workstation easily, using Dynamic Host Configuration Protocol (DHCP) to retrieve data network
setup parameters. Telephones have always represented a more complex move process. Requiring
PBX reprogramming and intercession by a telephone services administrator; convergence to an
integrated IP service network may simplify this process. VoIP solutions may also provide new
productivity tools to remote workers, but again, management, monitoring, and analysis of
network services is crucial to the continuity of daily business operations.
Strategic new business applications appear on the horizon every day. We see network service
like voice and video beginning to couple tightly with enterprise resource planning and customer
relationship management systems. With a better understanding of how management and
monitoring play into the life cycle of integrated services, we have the foundation to move
forward to the next chapter, in which we’ll explore managing service availability and capacity
planning for converged services.

138
Chapter 7

Chapter 7: Effective Service Availability Management and


Capacity Planning
Life cycle management is crucial to the sustainability of any network. Deploying data voice and
video services is all well and good, but long-term success requires a repeatable methodology for
consistent measurement and planning. This chapter takes a look at an approach that has been
used for many years in the legacy telecommunications industry called FCAPS. The name is an
acronym for fault, configuration, accounting, performance, and security management. We look at
FCAPS because it has been tightly coupled with managing large voice and data networks for
many years. Later, the chapter will touch briefly on the IT Information Library and the ITIL
framework. These two well-know methodologies dovetail nicely as a foundation for managing
the life cycle of the network.
This chapter will delve into availability management, network optimization, and capacity
planning issues as the primary focal point. We’ll explore these areas using the FCAPS model as
a base for methodical techniques in delivering integrated services.
Network optimization is a vital part of life cycle management. Optimization supports both
availability management and capacity planning. It’s an ongoing process and aids in keeping
network costs in check. Network optimization can also help in managing the demands of
business users, who might be rapidly stretching the capabilities or capacity of the network as
they deploy new applications. Optimization includes trending, capacity planning, and ongoing
minimization of infrastructure costs. By decommissioning legacy services as they are no longer
needed and continually monitoring and maximizing the success rate of data voice and video
sessions, we are, in a sense, future proofing the network. This holistic approach to network
management also provides investment protection by continually evaluating network performance
and service delivery throughout the entire life of the network, maximizing the usable life of
equipment and leveraging technologies for full value.
Managing the life cycle includes continuous evaluation of both the return on investment (ROI)
and the return on effort (ROE). An important part of this process is being able to perform all the
following:
• Maximize business potential through performance and availability management
• Learn to decipher results from performance and availability testing
• Use lessons learned from testing to better utilize system resources; this includes assessing
capacity and application upgrade needs
• Recognize the risk factors of not testing network applications and capacity, including
slowdowns in network response time for your customers or users
• Learn how to monitor, analyze, and prioritize your business and network management
needs to ensure a responsive suite of converged data, voice, and video services
• Prepare for the convergence of performance and availability management

139
Chapter 7

These skills often aren’t put to use as part of the general IT staff’s daily routine. These tasks may
actually dilute the focus of the IT group from the daily core business and operational
requirements. Balance is crucial so that no single aspect of service delivery, management, or
assessment consumes excessive resources. These resources are often viewed as network
resources, but the human brainpower and time resources must also be taken into consideration.
Comprehensive network management requires the effective use of people and resources,
business processes, technology tools, and products, vendors, and service providers. It’s a never-
ending effort to provide reliable and consistent service levels for an array of converged services
to meet end user and customer expectations.

Introducing FCAPS: A Sustainable Model for Balanced Network


Management
As data, voice, and video services converge on the IP network, it’s prudent to examine and
incorporate the lessons learned by the legacy telecommunications carriers. In many cases, these
techniques have also been embraced by the major Internet service providers (ISPs). The telco
carriers have been successfully delivering voice services in the Public Switched Telephone
Network (PSTN) for more than a hundred years. Although new unified communications
technologies such as VoIP, video, and mobile voice are converging to shake up the landscape of
telephony, many long-standing service delivery practices are easily adopted to fit today’s
integrated networks.
The telecommunications industry adopted and followed the Telecommunications Management
Network (TMN) framework. This framework was originally defined by the International
Telecommunications Union (ITU at http://www.itu.org). Within this large management
framework, the general management functionality offered by telecommunications systems is
split into five key areas referred to as FCAPS. FCAPS is the International Organization for
Standards (ISO) model for network management. It provides a standards-based foundation for
network management.

140
Chapter 7

Service Layer

End-to-End Layer

Domain
Domain
Network Layer
Domain
Domain

Ethernet Layer

Figure 7.1: The TMN framework.

Digging inside the TMN, the ISO has provided FCAPS as a framework for the task of network
management. All forms of management impose elements of monitoring and control. The
management of service networks is no exception. For each aspect of network management, there
is a life cycle of some sort, corresponding to analysis, design, construction, monitoring, and back
to analysis again—endlessly repeating.
There are several areas of mission-critical business management that FCAPS overlooks. The
FCAPS model is centered on the technical management of the service network. Daily operational
management will include people and staffing issues, finance and purchase issues, process
documentation, and a number of other supporting roles. The network manager’s day is filled with
a variety of tasks and workflows. FCAPS simply provides one network management approach
for that aspect of the larger role of a network manager.

141
Chapter 7

Fault Management
Configuration/Change Management
Accounting/Asset Management
Performance Management
Security Management

By isolating the management challenge into


distinct areas, the FCAPS model allows us to
conceptualize solutions that make the most
sense for the challenges unique to each
functional area.

Figure 7.2: The OSI FCAPS model.

Fault Management
Every network will encounter faults originating in equipment, cabling, human error, or software
during operation. The concept behind fault management is that of formalizing a repeatable
process to identify these continually occurring faults. Faults may be recurring events that can be
easily recognized with repetition, or they can be discrete events that occur rarely. Smooth
network operations require timely isolation and resolution of faults as they occur. A Network
Management System, usually deployed inside the Network Operations Center (NOC), provides
both trapping and logging of network faults.

Chapter 6 reviewed some of these techniques—using SNMP for trapping and logging system events.

Fault management is not about resolution or restoration of service. Fault management is more a
process for intelligence gathering related to network performance. Faults impact performance,
and a high fault rate results in impaired or degraded network service delivery. Although
mitigations of problems and service restoration must occur in a timely fashion, the fault
management flow is oriented to identification and isolation first. Fault resolution and service
restore might employ workaround procedures where the root cause of a particular fault may
preclude comprehensive correction in a quick and timely manner.

Chapter 8 will explore fault management in greater depth.

142
Chapter 7

Configuration Management
Configuration management is a workflow centered on data gathering and storage. Configuration
files and data are collected from network elements. These network elements in the traditional IP
network include routers, switches, servers, and so on. In the unified communication network,
configuration management is concerned with configurations of gateways to other networks,
gatekeepers that authenticate and grant registration access to users, SIP signaling servers, media
servers, and even telephone sets.
Knowing where things are deployed in the network and how they are configured is central to
managing the total enterprise network. Many day-to-day operational problems are related to
improper configurations of one sort or another. Mistyped IP addresses or network masks are a
common problem. Patches or updates to OSs, antivirus programs, and business software create
gaps in the integrity of the network. You need to monitor and manage all aspects of
configuration, including:
• IP addresses
• Subnet masks
• DNS settings
• Frame size
• Directories
• Disk drives
• Drivers for network cards
• Video cards
• User groups and identities
• Domain structures
• Applications

The list goes on and on. Chapter 8 will explore configuration management in more detail.

143
Chapter 7

Accounting/Administration/Asset Management
Comprehensive accounting and effective control can save money and improve the allocation of
scarce network resources. How do you know whether an upgrade is needed without measuring
usage? Factors that might be measured include printer pages, use of disk space, processor time,
network bandwidth, use of resources (for example, the Internet), and so on. Choosing an
appropriate metric is often tricky. For instance, do you measure the number of pages or the
amount of ink used for printer accounting? The latter is where the cost is.
Accounting costs money and should be used only where appropriate. Costs can include computer
resources (such as disk space, processor time, network bandwidth) as well as the time and effort
spent enforcing restrictions and collecting funds. A large portion of the cost in your legacy phone
bill is taken up by the cost of collecting it.
The FCAPS model evolved from the legacy commercial telecommunications industry. Customer
billing for minutes of use is a core component of that industry segment. In the FCAPS model,
early documentation focused on these traditional customer billing, or call accounting, systems.
Efforts to mirror this accounting mindset in data networks have been implemented through tools
that collect usage statistics that may be tied to disk space, CPU utilization, minutes of use,
bandwidth, or some other metric. Bandwidth has become a simple metric that’s widely used.
Tools implementing protocols such as RADIUS and TACACS are commonly deployed in
enterprise networks to monitor usage. For most enterprises, auditing, rather than accounting and
billing, is the primary usage for these monitoring systems.
In enterprise networks whose core business is something other than telecommunications or voice
service delivery, administration, asset management, or auditing may provide a more accurate
frame of reference for this piece of the puzzle. Administration objectives include managing a set
of authorized users by establishing user IDs, passwords, and permissions. Administration can
also include support of systems and equipment such as performing software backups and
synchronizations between systems or network elements.
Asset management might simply encompass the financial aspects of capital investment and
system replacement as asset costs are amortized off the corporate books. Beyond physical
hardware assets, many enterprises adopt their asset management efforts to include software
licenses and oversight of version control.
Auditing, in the enterprise world, can range from simple usage auditing to enterprise policy
management. With the widespread focus on regulatory and compliance requirements—such as
the Sarbanes-Oxley Act (SOX), the Graham-Leach-Bliley Act (GLBA), and the Health
Insurance Portability and Accountability Act (HIPAA)—audit processes often expand. Business
sectors may have other auditable compliance requirements based on business sector.

144
Chapter 7

Performance Management
Performance management is another aspect of the network that looks solely at performance data.
Continual monitoring of network health and performance requires collection of performance
data. Although SNMP might be used for monitoring faults and errors as part of fault
management, performance management monitors interfaces on network elements, throughput,
responsiveness, and the holistic performance health. In a converged network that delivers unified
communications services, performance management drives the monitoring of trends in network
performance and utilization. These trends can raise the visibility of network capacity and
reliability issues before they impact delivery of the suite of unified communications services.
Performance thresholds can be set to trigger an alarm, which might then be handled through the
fault management process.
To monitor and improve the service, deploy hardware and software effectively, and to spot
trends and changes, monitor and control performance. Closely allied to fault management is the
task of measuring everything from the number of bits sent per second to the speed of transactions
to the refresh rate of your monitor. Understanding the things that can affect performance is a
technical issue, but knowing what to monitor and why is a management function. You need to
establish baselines and identify bottlenecks in the system. Bottlenecks are the slowest point in a
system. A lot of money can be wasted attempting to improve a system if you do not know the
part that is making it slow.

See Chapter 5 for more detail about performance management.

Security Management
The main objective of security management is to identify and mitigate risks. In the service
network, security helps:
• Maintain legitimate use of corporate resources
• Maintain confidentiality of enterprise intellectual capital
• Ensure integrity of the data integrity
• Provide consistent auditability
Security is a big issue and always comes with attached costs. These costs can involve extra
processing (for example, for encryption, authentication), extra administration (for example,
setting up user IDs, monitoring for security breaches and so on), extra hardware (for example,
firewalls and fiber optic cable instead of easily tapped wires), and a host of other less tangible
elements, such as convenience or lack thereof, that might impact the ability of users to work
effectively.

Chapter 9 is devoted to network security risk and incident management.

145
Chapter 7

FCAPS Simplified
Today in business network management, this FCAPS model exists in many forms and flavors.
As defined by the ITU, it is more complex than necessary for many organizations. It’s important
to remember that FCAPS was designed to support the telco carrier networks. It may be the best
choice for a VoIP service provider but not for a health care or financial institution. Rather than
undertake the complexity of a full-blown FCAPS methodology, many organizations implement a
four-layer approach for converged network service assurance focusing on:
• Performance management/availability management—Guarantee availability and
performance of the network infrastructure, VoIP, and other application or business
services
• Security management—Ensure protection from risks and response to security-related
incidents
• Configuration and vulnerability management—Manage network configurations to ensure
compliance with corporate guidelines; this could be an enforcement point to ensure
regulatory compliance where applicable
• Change control management—Ensure that only authorized and appropriate changes are
made to the network; audit controls provide knowledge-based tracking of every change
You’ve often heard the acronym TCO. Throughout, this guide touches on four facets of holistic
management for the converged services network. These are:
• Operational integrity
• Service management
• Policy compliance
• Risk management

146
Chapter 7

Performance
Configuration
Management— Security Change Control
and Vulnerability
Availability Management Management
Management
Management
Operational Monitor system Continually Identify vulnerable Ensure no
Integrity and network monitor threat or exploited unauthorized
performance environment systems changes are made
to operating
environment
Service Map performance Event correlation Benchmark and Delegate admin
Management indicators to of security baseline system change permission
business issues incidents with configurations only to approved
performance staff
issues
Policy Measure SLAs Protect audit trails, Ensure baselines Audit logs for all
Compliance and compliance monitor and report follow corporate changes to
security violations standards and production
policies environment
Risk Forecast and Identify, mitigate, Identify, assess, Test changes
Management avoid service and respond to document, and before
interruptions security incidents remediate or implementation to
accept risks minimize risk

Table 7.1: A four-layer variation on FCAPS.

The various capabilities identified in Table 7.1 can all be implemented at once. Many companies
today follow the ITIL framework as a root model. The ITIL framework is a series of industry
best practices. It’s a set of documents that focus on the core business areas of IT service
management (ITSM). ITIL is referred to as a library because it has been published in a book
series. ITIL and IT Infrastructure Library are registered trademarks of the UK Office of
Government Commerce (OGC), where this concept took root. The library has evolved to
encompass a wide set of best practices.
When you think about adopting best practices, there is one simple view that makes their benefit
clear. Best practices provide a practical substitute for conducting comprehensive risk assessment
and analysis. In most cases, the global business universe can leverage the shared knowledge
gained through others and lower the cost of progress by avoiding the high cost and effort
associated with comprehensive risk analysis best practices.
The ITIL framework supports the wide field of IT infrastructure, development, and infrastructure
for service delivery. One facet of ITIL is referred to as the maturity model. When organizations
embrace ITIL, they typically conduct a series of assessments of their own process maturity. Most
companies choose to work methodically toward a mature model that supports all the enterprise
business needs. They adapt and refine along an evolutionary path and adopt new methods and
techniques along the way.

147
Chapter 7

Optimization for Service Availability and Capacity


Approaching at service management, availability, and capacity planning from the traditional
carrier telco provider’s perspective wasn’t necessarily easy to adopt early on in the evolution
towards the converged networks of today. For most businesses that were early adopters of new
technologies such as VoIP, it seemed much simpler to add to the existing network and grow
incrementally. This controlled growth can be a good thing from an investment and impact
perspective. The danger is in cobbling together disparate pieces and continuing to add and grow
without stepping back to take a larger view of what is happening in the network. The
performance envelope discussion in Chapter 5 provides one technique for protecting against
growth pains through methodical evaluation.
The larger the enterprise organization, the closer the proximity to the carrier’s mindset. Many
large enterprises, especially in the Fortune 1000, find that over time they become their own
telephone company. They have become their own data network provider. These organizations
provide internal voice and data services to more employees than many small telcos and ISPs do.
And they add the complexity of business processes to support their core business, which might
be financial, health care, transportation, or some other facet of business.

Network
Management ALARM Tracking
System Database Workflow
Customer
Trouble
Tickets

Operation Service
Customer Inventory Billing
Support Records Database
Level
Agreement
Systems

Systems

Figure 7.3: The typical tools.

148
Chapter 7

Figure 7.3 shows the typical tools of the voice and data network service delivery business.
Network management systems monitor the service network, producing logs and alarms that feed
into databases. Some form of trouble resolution mechanism or Help desk tracking is needed.
Operational Support Systems (OSSs) provide repositories for other information—customer
records, inventory, contracts or SLAs, and billing. The cloud in the visual of workflow represents
all the rest of the core business and includes the processes and procedures around service
delivery in the corporate environment.
The key is in scaling the methods described here to the size and scope of your business.
Customer records alone can vary from a major Customer Relationship Management (CRM)
system, running on widely distributed servers, to a central contact management system in a mid-
sized organization to an individual salesperson’s customer database running on their desktop. For
small businesses, the contact list in the salesperson’s cell phone might be the customer database.
At some level, every component the carriers deploy to manage a global telecommunications
network correlate to similar functionality in every enterprise organization.

Trending and Capacity Planning


This guide is really aimed at service delivery organizations, whether they are a VoIP service
provider, the IT group in a small to mid-sized business, or the Chief Information Officer (CIO) in
a Fortune-100 corporation. Every business depends on business intelligence for survival and
success. A technology organization inside a large enterprise that is in the business of delivering
services to that enterprise has just as great a need for business intelligence as a provider of
services in the open market.
Monitoring trends and usage provide the supporting data for ongoing network expansion and
upgrade. Although a NOC might focus on monitoring for availability, uptime and performance,
the capacity planning staff needs comparable data. It’s not enough to manage utilization of
trunks and network connections. Capacity planners need to know what features are being used
too. Feature sets such as call hunt groups, call pickup groups, and automatic call distribution
groups may be tied to levels of licensing. Given that many VoIP systems are now servers
running on general-purpose hardware platforms, software licensing and upgrade has to be
considered in the capacity planning process.

Bandwidth
Bandwidth has become, for many, the holy grail of networks. The consumer end of the spectrum
has evolved from very slow dial-up connections to broadband in the home. The enterprise
network has gone from fractional T-1 circuits over frame relay to 100Mbps switched Ethernet to
the desktop and Gigabit Ethernet in the core. Large enterprises may even use SONET
technologies to provide even greater bandwidth across a wide area network (WAN). Dark fiber
deployment during construction and around the country continues to increase. Wireless
technologies are spreading throughout municipalities and are quickly becoming prevalent in
campus, warehouse, and office deployments.

149
Chapter 7

In the PSTN and Internet, high-bandwidth circuits have carried the backbone traffic for years.
Unified communications technologies have, over time, pulled the two clouds closer and closer to
the point that today lines of distinction between the two blur. This is a challenge for the
telecommunications and Internet providers, as they’ve often used bandwidth and transport
technology as a differentiator. Customers and end users care less about technology and more
about solutions today. As described earlier, what matters is that sales force automation or CRM
systems, and others, simply integrate and work.
Costs of fiber, higher-speed Ethernet switches and routers and advances in optical technology
have driven the cost of higher-capacity connections down to some of the lowest dollars-per-bit
prices ever imagined. In terms of VoIP deployments in particular, this low cost can induce
motivation for change and is something that needs to be evaluated, designed for, and monitored
for the life cycle of the network service. In the past, enterprises deployed Private Branch
eXchange (PBX) systems for their internal corporate voice services. These systems were
typically connected to the PSTN via T-1 circuits carrying 24 voice channels. An ISDN Primary
Rate T-1 was the most common PBX connection.

PSTN
PSTN Separate Networks Internet
Internet
No Problem

Touching Networks
PSTN
PSTN Internet
Internet
Early Dialup Access

Converging
Networks
PSTN
PSTN Internet
Internet
Integrated
Communications

The Fully Converged Complete


PSTN
PSTN Internet
Internet Service
Future Network
Transparency

Figure 7.4: Network convergence evolution.

This is a good point to revisit the network convergence evolution from Chapter 1. Over time,
completely separate networks, the PSTN and the Internet, have evolved to become tightly
coupled. Today, they complement one another to deliver an array of data, voice, and video
services.

150
Chapter 7

This tight coupling of the PSTN and Internet has led to choices in design for service delivery
today. When providing voice services, one decision will be selecting the appropriate trunking
technology between networks. For enterprises that continue to maintain a traditional PBX, T-1
trunks to the PSTN might represent the best solution. Some organizations will implement
gateways that convert from the IP network in the enterprise to the PSTN network via T-1
circuits.
As enterprise networks move forward to integrate communications services, SIP is quickly
gaining momentum as a peering and trunking technology between voice service networks. SIP
trunking may introduce some new costs, such as Session Border Controllers (SBCs). These
provide a gateway and firewall services between service networks. SIP trunking is rapidly
proving to be very cost effective.
Bandwidth comes into play in SIP trunking because they’re generally IP network connections. If
the cost of a 10Mbps or even a 100Mbps IP link using Ethernet is significantly lower than a T1,
many more voice calls can be carried at a fraction of the cost between connected networks. A
legacy voice T-1 circuit carried 24 phone calls. IP links can carry many, many more.

Softswitch
Trunking Gateway
PSTN T-1
Trunks
SIP
Signaling Gateway Trunks

Integrated
Access Device
Managed
Broadband Access Link
Network
IP/ATM

VoIP Server

Internet

Figure 7.5: The variety of convergence options.

A glance at Figure 7.5 reveals the variety of connection times coming into play as technologies
converge while they’re evolving. In some cases, separate signaling and trunking gateways are
apparent. In other cases, they’ll be integrated into a single softswitch.

151
Chapter 7

SIP trunking may be used between what was thought of in the data environment as extranet
business partners. For example, a health care provider might implement SIP trunks to connect o
insurance carriers and pharmaceutical providers as integral partners. SIP trunking is also on the
rise as a connection to both traditional telecommunications carriers and other voice and video
service providers. Like every other key component of the converged service network, bandwidth
must be monitored and managed to ensure service levels can be maintained.

Ports/Lines
Ports and lines are a primary consideration during the initial design phase of any VoIP network.
The number of trunk ports and telephone sets a system can support is rarely overlooked.
Capacity planning needs might seem simple, but can be deceptive.
The service delivery team must account for business plans as well as technical plans. Mergers
and acquisitions can radically alter the number of trunk ports or telephone sets a company needs.
It’s vital that the technology service delivery organization participate in business planning, or at
least have clear vision of where the enterprise is moving in order to support the wide range of
business needs.

Codec Planning
Chapter 5 explored codecs, primarily for their role in network performance. Codecs directly
impact the fidelity of the voice that the end user hears. Codecs have other impacts as well.
Remember that you hear an analog voice sound. The process of compressing, digitizing, and
packetizing voice requires hardware. The choice of codec can drive CPU utilization up or down.
Codec selection also impacts the bandwidth requirements inside the service network. Pulse Code
Modulation (PCM) was designed for the 64Kbps voice channel of the T-1 architecture of the
PSTN. However, the G.729a (CS-ACELP) codec that has been a rising star in VoIP solutions
recently can be optimized to produce a 6 to 8Kpbs bit rate. Although the real world doesn’t
generally produce a true 8-to-1 reduction in bandwidth required, this codec can dramatically
impact bandwidth requirements.
It’s a good idea to regularly re-evaluate codecs in use. One practice that is becoming more
widespread is an annual codec review with an eye to quality of service for end users. Although
converting to a different codec seems labor intensive, it may be one approach to improving
quality if the network is becoming saturated.

Optimizing the Infrastructure Optimization


Chapter 5 looked in depth at the performance envelope approach in the context of network
readiness assessment. Just as the life cycle is continuous, so is the evaluation or assessment
process.

152
Chapter 7

Optimizing Hardware
In legacy PBX technology, the life cycle of an investment was commonly projected to be 10
years. In IP networking, the refresh cycle has become much shorter.
Repair is most commonly viewed as trouble resolution. For PBX technology, it often means
swapping out parts. The same might remain true for a VoIP call processing system today,
especially for mechanical component failure. Disk drives and fans will fail. Power supplies will
burn out. Hardware will require repair.
Another perspective worth considering is that the integration of unified communications services
moves off traditional, dedicated, and often proprietary hardware to what might be viewed as
general-purpose systems. Most enterprises standardize on a server hardware platform that often
supports VoIP services as readily as Web services.
You need to not only consider the legacy telephone system at the time you initially deploy VoIP
services. Because voice and data are becoming integrated, the data network life cycle also comes
into play. For some organizations, obsolescence is only one driver. There may also be
compelling business reasons to migrate with newer technology. As discussed earlier in this
guide, the ongoing integration of software applications and network service can make it prudent
to move to new technology—not when the older systems are at the end of life but when the end
is visible on the horizon.
Another important consideration is that manufacturers discontinue support for older products
every day. Some do so because parts become scarce, making it more difficult to repair systems.
In many cases, vendors declare “end of life” of a product hoping that customers will migrate to
newer systems.
As systems age, maintenance and support tend to become more expensive. Newer hardware
brings new efficiencies in patching and upgrading, which may actually drive support costs down.
And older systems may just reach full capacity with no expansion capability to meet current and
future business needs. When evaluating the refresh cycle for systems approaching obsolescence,
it’s wise to remember that technology won’t do you any good if it can no longer support your
business needs.

Optimizing Software
The ongoing management process must also encompass software. In the old world of PBXs,
software was often proprietary. An upgrade or patch was tested and businesses deployed it into
the production phone system at night or over a weekend when the phones weren’t highly used.
VoIP service elements are often general-purpose hardware; maybe even the same platform as the
company’s data servers. Software optimization may include modifying configurations, adjusting
frame or packet size, or even some esoteric nuance like altering the TCP window size in the
registry setting

153
Chapter 7

Maintenance, Enhancements, Upgrades, and Patches


Many companies understand the need to keep the service processing software up to date but
overlook underlying operating systems (OS). Whether a voice service element is running an OS
from the UNIX/Linux family, the Windows family, or a vendor’s custom internal OS, it’s
important to incorporate a patch management process that ensures timely updates of both the OS
and the service software. For a large service network, this will often necessitate a lab
environment for testing all upgrades before they are deployed in the production network.

SLA Optimization
Companies that contract with external providers watch SLAs closely to ensure contractual
obligations are being met. In the enterprise network, SLAs might only be described in service
offerings and not actual contracts to business units. Internal SLAs may be quite informal.
In the enterprise world, VoIP service is often designed to meet specific criteria for successful
internal delivery. Characteristics for the quality of service needs—such as delay, jitter, loss, and
other QoS provisions—need to be handled just like contractual SLA commitments. As
conditions change, the network is regularly tuned to meet these SLA requirements, whether
they’re contractual or self-imposed to assure the service quality.
It’s also good to remember that the SLA approach isn’t just documentation of the minimum
service levels being provided. It’s also used to describe remedial actions to be taken. For internal
service delivery, the SLA can be a process document with specific details for life cycle
management. It may also be worthwhile to document what service classes, or bandwidth and
QoS commitments, are being delivered to end users, then monitor for over-usage. In the
enterprise network, it’s common for internal customers (in business units or divisions, for
example) to evolve and suddenly consume far more network resources than originally planned.
Although enterprises outside the service delivery sector may have no need for the SLA-driven
model, there may be other motivators for the larger enterprise. When SLA-based network
monitoring and management systems are fully integrated and automated, they can provide both
long- and short-term visibility to overall and service-specific operations. Adopting this “service
provider mentality” turns the SLA into a key tool for successful operation. The cycle of
continuous improvement driven by managing the SLA metrics can, and should, be modified over
time as network objectives and realities change and evolve.

Delay
Delay is simply a reality of IP networking. Delay is cumulative and all sorts of factors can
impact the time it takes for an IP packet to traverse the network. In addition to delay due to
transmission distance, remember that there are other processing delays. In most enterprise
networks, delay is usually a constant factor for any given communication. Network traffic in
most enterprises tends to follow the same path for everything of that media stream type.

154
Chapter 7

Call Setup
Delay in VoIP networks can lead to call setup failure. For most implementations, this is a factor
that doesn’t have an adverse impact. Rather, this is a consideration for ongoing management and
monitoring of the VoIP service network.

During Call
Regular monitoring of delay in the network and knowing the operating environment may lead to
quick trouble resolution later. If you recognize that data traffic is bursty by nature and that traffic
such as email and Web browsing are not real-time traffic, it’s easy to see that they might not
provide good early indicator of rising delay during day-to-day operations.
VoIP users may provide the “canary in the cage” sort of early warning system that helps in
managing the service as a whole. A change in delay patterns in the network, coupled with routine
monitoring and a few calls from users identifying call problems or call setup issues, can quickly
help identify problems and lead to mitigation.

Delay Variation
Delay variation is called jitter. Because packets can take different routes across the network,
delay variation in large IP networks is common. Jitter is measured in milliseconds. For most
enterprise networks, no extra engineering is needed to maintain a 1 to 2 millisecond level of
jitter. Large enterprise and service provider networks, spanning across a large geographic area,
are more likely to have diverse routes of varying distance that can increase the tendency for
jitter.
Jitter is generally not a problem during call setup. It’s more likely to arise during the media flow
associated with a conversation. In other words, jitter is more likely to impact the voice than it is
the signaling. Jitter in voice conversations results in unintelligible conversations that sound
“jerky.” Jitter buffers can be deployed to reduce this impact, but buffering can only do so much
to overcome impairment.
Jitter is managed in a number of ways. Many vendors’ equipment provides buffer management
algorithms for setting the size of ingress and egress buffers. Beyond buffering at the node level,
some vendors include jitter buffers. These are very small, often only 3-bits in size. They can be
used to compensate for the timing variations. These jitter buffers may be static, dynamic, or
adaptive. Static buffers are common in older routers, particularly in small branch office or home
office devices. Dynamic buffers calculate an optimum buffer size based on the first series of
packets received. The most advanced jitter buffers adapt to changing network conditions. There
are unified communications monitoring tools in both commercial and open source variations that
enterprises can use in the NOC to include jitter monitoring as part of the overall network health
watch.

155
Chapter 7

Packet Loss
Remember that IP is a best-efforts protocol. It relies on higher-layer protocols, like Transmission
Control Protocol (TCP), for delivery guarantees. Packet loss will occur. It’s also referred to as
error rate.

Blocking/Non-Blocking Access
In the traditional telecommunications history, engineers design central office and PBX systems
to be either blocking or non-blocking. A non-blocking system assumes that every end user will
be able to go off-hook at the telephone set and place a call at the same time. In the traditional
environment, non-blocking systems were very rare. Building a non-blocking system meant
investing in hardware and circuit connections that statistically would sit idle most of the time.
With the evolution to VoIP, and the shrinking cost of networking, some organizations take a
different view today. Non-blocking systems may be achieved at a lower cost, and may be the
appropriate solution for some businesses. If a non-blocking system is put in place, the monitoring
systems may need to evolve to track items such as successful versus failed calls and abandoned
calls to assure that a non-blocking service delivery network is being maintained.

Success Rates of Call Setup


During network readiness assessments, engineers evaluate current requirements of the data
network and compile existing voice services requirements. One aspect of ongoing oversight in
the voice services network is close monitoring of the call completion rate. Calls may fail to
complete for a number of reasons, but a low or declining call completion rate might be an
indicator of signaling problems, bandwidth consumption (or a network saturation problem), or a
number of other issues.
Another aspect of call completion rate to monitor is the number of abandoned calls. In a call
center environment, abandoned calls are generally tied to wait time for customer service agents.
A message telling a caller that the wait is 10 minutes may induce callers to hang up and retry
later. Interactive Voice Response (IVR) systems or dial-prompt systems with deeply nested
menus may drive callers to hang up in frustration. Call completion rate trends can provide
information about specific VoIP services that make capacity planning and overall service more
effectively support business processes.

Gateway Issues
Because you’re looking at a converged network of voice and data, you have to recognize that not
all organizations are going to move directly to new VoIP systems. Convergence is a drive to
integrate existing services that work together in new ways, then managing the new network
environment of voice and data.
Figure 7.6 shows a company with old and new. The location on the left has a PBX connected to
the VoIP gateway. In this case, the PBX programming determines which calls are directed to the
PSTN and which are directed to the VoIP gateway. This might be done via a special access code,
or it might be programmed into the dialing patterns of the PBX. For example, any time a 10-digit
number with area code is dialed, the call might be directed to the IP network to minimize long-
distance charges on the PSTN. The company might route only local and 800 number calls
directly to the PSTN.

156
Chapter 7

The gateway processes and packetizes the phone call, then passes it on to the router, which will
transfer it along to the corporate IP network. At the other end of the network, on the right side of
the figure, notice there is no PBX. This company has chosen to eliminate the cost associated with
buying, managing, and maintaining a PBX at the location on the right. There is a mix of VoIP
phones and softphones at this office.
The SBC on the right is using SIP trunking via the IP network. VoIP calls come into the SBC,
and are then routed through the network to the appropriate endpoint. When everything is
configured properly, every voice endpoint can talk with every other endpoint. Traditional voice
and VoIP have converged.

Router Router

IP Network

SIP Trunk

VoIP Session Border


Gateway Controller

PSTN
T-1 Trunks

Legacy
PBX VoIP
Server

Workstation Workstation

Figure 7.6: A converged company example setup.

This highly simplified view of the network might make convergence sound like a trivial matter.
It is not. This company is managing a complex multi-faceted network, and performance
monitoring will have to be undertaken diligently.
The gateway must have a database of telephone numbers associated with users, yet you must not
overlook that many LANs use Dynamic Host Configuration Protocol (DHCP) for address
assignment at startup. Thus, the gateway must maintain a dynamic database that allows a static
telephone number to correlate to a dynamic IP address every time the user connects. The SBC
simply represents another type of gateway where SIP trunking information and access
permission controls reside.

157
Chapter 7

Monitoring systems need to provide alarms for the NOC staff to alert when a number of new
situations arise.
• DHCP addressing is straightforward, but how do you monitor and alarm error conditions
when an unknown user tries to register with the VoIP system?
• Trunk utilization between traditional and VoIP systems, whether done in house or with a
telco provider, must be monitored closely. Trunking or media gateways often prove to be
the choke point in voice call services.
• SBCs require the same sort of monitoring as a major enterprise router or firewall. They
represent a new perimeter point in the network.
Although convergence presents many wonderful opportunities to bring services together onto a
single infrastructure, it also brings several added degrees of complexity to the network. The
existing IT staff may not have a deep enough understanding of telephony requirements. The
telecom staff may not be familiar with data networking technologies. The skill set required to
design and manage this network is a blended skill set that has not been necessary in the past
when the two networks remained distinctly separate. As networks and services converge, the
pool of talent with the right skills to design, manage, and maintain those networks shrinks. It’s
important to ensure the support staff receives the proper tools and training to keep these mission-
critical services operating at acceptable performance levels.
Network management is one of the critical and most challenging aspects of service delivery. In
the past, many network designers followed the simple rule of “better safe than sorry.” This often
led to over-engineered and over-priced networks.

s
ines
s r
Bu able
- n
r v i ce d E
Se ente
i
ive Or
ct
r oa
eP
c tiv
a
Re
y
a rch
An

Figure 7.7: The network service maturity path.

158
Chapter 7

Figure 7.7 shows a network services evolutionary path that is actually quite common in business
organizations. The underlying goal for focusing on managing the integrated data, voice, and
video services is to turn them into business enablers for the enterprise. This evolution of maturity
follows a parallel track with the ITIL maturity mentioned earlier.
If there is a complete lack of business process, anarchy and chaos reign. No business manager
plans this, but sometimes explosive growth, especially in a highly successful start-up company,
can cause total chaos. Early business process development is often reactive in nature. As
business processes mature, you find they move into a more proactive, forward-looking model.
For the most successful information-oriented enterprise, network management becomes a part of
the service. It becomes a differentiator. And in the most advanced cases, network management
information is such powerful business intelligence that it is leveraged to enable completely new
lines of business.

Summary
Knowledge about your network is a crucial element to effective service delivery. Service
delivery managers need know and understand every aspect of network service delivery. Without
a knowledge-based framework, service assurance efforts are hit or miss at best. Managers
overseeing services networks need to know as much as possible about performance trends and
service requirements to guarantee availability of mission-critical services in the enterprise. As
you integrate voice, through VoIP, with the IP data network, you migrate what is for many
organizations the single most mission-critical service. Capacity planners also need utilization
information to provide consistent, systematic growth across the suite of unified communications
services as business requirements evolve. Everyone involved in service delivery needs
information about the user experience to identify bottlenecks in service availability and delivery.
Correlating information about failure points is vital to effective troubleshooting in daily
operations.
Industry best practice today, whether via FCAPS, ITIL, or some other model, all encourage
taking a holistic approach to network life cycle management. That holistic view is widely
supported by a wide-ranging data gathering and information analysis strategy. To ensure service
delivery, you must align service delivery efforts with the needs of customers, internal or external,
while finding a cost-effective approach to ensuring capacity, availability, and security of the
data, VoIP, and video services.
Two of the most treasured assets any information-based company has are its people and its
intellectual capital. Data, information, and knowledge are frequently far more precious than
inventory and cash reserves. The latter are generally easier to replace or regenerate. Raw data is
easily collected. In years past, log data often either purged from systems after some time period
or went into archival storage. For many organizations, historical information that you now
perceive as key business intelligence was stored on tape or microfiche and buried in a data tomb.
In more recent years, attentions have focused on data warehousing concepts. This quickly
spawned another industry subset promoting knowledge management solutions. Today, many of
these knowledge management theories and techniques are used to arm managers with
information needed to ensure service levels meet user needs.

159
Chapter 7

The enterprise service delivery workforce requires appropriate data gathering and analysis tools
to facilitate making informed decisions about network services in daily operations. The more
accessible and readily available these tools are, the more efficient and productive the
organization can be. One way to bring the tools closer to hand is to create that converged
environment where access to any piece of information and any communication need is available
instantly. This knowledge-based approach ensures the highest level of confidence in all network
services.
A well-managed converged network, tightly integrating data, VoIP, and video services may soon
prove to be a huge differentiator between competitors in business. The next chapter will dig into
configuration, fault, and performance management.

160
Chapter 8

Chapter 8: Effective Network Configuration, Network Fault,


and Network Performance Management
Chapter 7 introduced the FCAPS model and examined service availability and capacity planning
management. This chapter will continue that theme of using a methodology for consistent
management of network faults or problems, configuration of network devices, and performance.
Network management means different things to different people. For some organizations, it
simply means a network consultant is monitoring network activity with some tool. In larger
enterprises, network management involves continuous polling of network devices to monitor
status, distributed databases containing logs and error reports, and graphical representations of
the network topology to present a high-level view of the overall health condition of the network.
All network management can be viewed as a service that uses tools, devices, and software
applications to assist network managers in monitoring and maintaining the quality of service
(QoS) being provided.

Fault Management
As a part of holistic network management, fault management is the term used to describe the set
of tools and functions used to detect and isolate and then remediate problems in the service
network. These malfunctions may be technical, such as equipment failures, or caused by human
error. The central theme is that something failed in the network. Fault management sometimes
includes environmental control systems or monitoring.
Faults are detected through monitoring system events. In many organizations, event monitoring
occurs in three ways:
• The Network Management System (NMS) in an enterprise command center monitors the
status and health of network elements. Often an icon representing an element in the
network will simply turn from green to red, indicating a problem. The NMS typically also
functions as a fault management system.
• Event correlation and analysis systems are designed to process syslog and event log files
from a number of systems. Many of these systems include an engine for detecting
anomalies as part of event correlation.
• Human analysis, while often effective, is also inefficient. The sheer volume of log data
generated in large networks makes human analysis impractical for many purposes. Yet
when network performance issues arise, most groups have people begin to review logs.
Humans apply different logic than NMS and event correlation engines apply, and may of
those systems spot trends, patterns, or unique items that fall below automated thresholds.

161
Chapter 8

When a failure or fault occurs, elements of the network send information about the problem to
the NMS. SNMP is widely used for this purpose. Elements transmit alarm information or
indicators. These alarm indicators remain in alert state until the error condition or problem is
fixed.
Many fault management systems are configurable to allow prioritization of network faults into
different levels of severity. Some alarms may simply represent debugging information, while
others might indicate emergency problems. The following list highlights the most common
severity levels used in fault monitoring and management systems today
• Warning
• Minor
• Major
• Critical
• Indeterminate
• Cleared

Note the “cleared” severity level. It is considered good practice to have network elements send
positive alerting information when an error has been cleared to ensure all monitoring systems note
problem resolution.

The fault management console, often referred to as the “dashboard,” allows the operations center
staff to monitor events across the network from a large number of systems. A large enterprise
network approach might be to overlay the network map on top of a geographic map to give a
sense of where in the network events are occurring. A fully robust fault management system may
be automated to the point that corrective action is handled automatically. Although many
network elements include protocols for automatic failover and resilience, other elements may
require creative approaches. For example, detection of a failed server at the NMS might trigger
running a script that moves active services to a backup system. Scripting reconfiguration of this
type of response helps ensure quick activation and eliminates the potential for keyboarding input
errors.

162
Chapter 8

In many cases, the automated system will simply be used to alert network support staff. Today,
email messages, pager alerts, and SMS messages to mobile phones are widely supported
approaches used in the NMS. It may also be prudent to implement practices that include
escalation mechanisms. Notifying a single person is probably inadequate for a very large
enterprise network. Sometimes a group of supporting personnel is notified as part of the incident
response team.
At a high level, fault management can be performed in active mode or passive mode. SNMP
monitoring and data gathering are a passive approach. Data is gathered as part of the normal flow
of traffic. If the network element recognizes some error condition exceeding thresholds, an error
message is sent to the NMS. This also presumes the network node is able to send the appropriate
alarm. If, for example, a power supply were to fail, often no alarm can be sent. The element
simply powers down.
Active fault management engages other tools to monitor the health of systems. These tools
include PING, Web page retrieval, testing of email responses, port scans, and a number of other
methods. If a Web server fails to serve up a Web page, an active monitoring system can send an
alert based on that failure. The server hardware might well be fully operational with no
detectable malfunction. In this case, perhaps Web services failed, but the server is still operating.
No hardware components have failed. Even the OS and Web application may be running
properly. Active monitoring can help catch faults that would otherwise pass undetected.
One key to successful fault management is to deploy tools that support processes rather than the
reverse. Established processes shouldn’t be bent to fit new tools without careful consideration.
On the flip side, tools may lead to the development of new network management processes
where none previously existed. When deploying network management tools, it’s helpful to focus
on the fault tolerance and resilience of the service delivery network, with an eye on both
redundancy and security.

163
Chapter 8

Network and Fault Management—An Integration Strategy


There are some basic rules of thumb to follow when integrating the NMSs into the converged
services network:
• Identify the best sources for data. Not every network element warrants monitoring, but
it’s vital to monitor key service delivery elements.
• Use the best tools available and the best features of each. A mediocre tool might seem
cost effective initially, but if it doesn’t provide the necessary capabilities, it won’t help
maintain an optimal network for delivery of integrated data, voice, and video.
• Wherever possible, use the dashboard approach to deploy a “Manager of Managers.” This
centralized dashboard will provide a complete holistic view for at-a-glance snapshots of
the overall network status and health.
• Follow the tried and true KISS (Keep It Simple Stupid) approach wherever possible.
Don’t introduce undue complexity.
• Use modular management components that integrate well together. Stay focused on out-
of-the box functionality. An NMS that requires customer script development and
programming to implement is likely to prove a problem to sustain. The GUI approach is
easier to learn and adopt than a system using command lines. Where practical, specific
workgroup GUIs can provide individual teams with the tools they need from a single
source (the dashboard or “Manager of Managers”). The network support group, Help
desk, voice services staff, and security administrators may all need slightly different GUI
views of the same underlying network infrastructure to operate effectively.

Caveats of Implementation Strategies for NMSs


When deploying the NMS, there are some basic caveats to keep in mind. A complex and
convoluted system can be less effective and more labor intensive than a manual system of work
procedures. Again, keep it simple.
Avoid over-automating. System automation is a wonderful boon to productivity, but the delivery
of integrated data, voice, and video services requires human oversight. Automating in the wrong
place can lead to a false-positive system event that triggers automated network reconfiguration or
failover to backup systems. Look carefully at these mechanisms to ensure that qualified support
staff is in control. Remember these are tools for staff, not replacements.

164
Chapter 8

When evaluating NMS solutions, here are some basics to consider:


• Does it have a simple interface? It should be easy to access everything you need. Users
shouldn’t have to toggle back and forth between screens to perform basic tasks. A Web
browser based interface may allow for easy customization for different workgroup views.
• Does it provide the ability to set a baseline? Simply put, the value in an NMS is in the
ease of setting baseline thresholds at normal network performance and operation levels.
How easy is it to establish the baseline and set thresholds than trigger notifications?
• Does the reporting capability meet your needs? If the NMS can report an event, whether
it’s a failed element or a spike in traffic, does it also provide enough historical tracking
and analysis capability to provide management reports that will be useful to the service
delivery manager? An NMS tool that requires knowledge of another programming or
report writing language will make it more difficult to extract useful information in a
format that helps make informed business decisions.

Later, this chapter will talk about configuration management, which is tightly coupled with change
control processes. Throughout this chapter, the goal is to maintain a broad holistic view, keep
perspective, and link all the various components of network management together.

Recognize, Isolate, and Detect Faults


Past studies have given way to anecdotal evidence in many areas of networking. One example is
commonly referred to as the “80-20 Rule.” Sometimes it seems that there is an 80-20 Rule to fit
every situation. Fault management is no different. The commonly held belief, supported by past
studies, is that when faults do occur, 80% of the time is spent looking for the problem, while
only 20% of the time is spent fixing the problem. Recognizing that faults exist, detecting them,
and isolating them is also referred to as anomaly detection and correction. What we’re striving
for is near real-time event correlation information to gather a timely root cause analysis of the
fault for timely remediation of the problem.
NMS vendors provide this root cause analysis functionality in a number of ways. Some use
mathematical/statistical modeling techniques. Others base solutions on complex academic
systems models. Many take a much simpler approach using rudimentary event filtering schemes.
This last category is often the approach implemented by fault monitoring systems that
organizations develop in-house.

165
Chapter 8

The Role of the NMS


Regardless of the methodology used internally in the NMS, event filters sort through the flood of
SNMP traps that are constantly being processed. In its most basic form, which Figure 8.1 shows,
the NMS provides two key roles—correlation and an event watcher. Most organizations
configure the NMS to ignore trivial events and closely monitor areas of concern. As events are
processed, they’re typically handled in one of the following ways:
• Traps reporting many failures are analyzed, leading to a verification, event correlation,
and a notification process. These are typically warnings and minor alarms.
• Traps reported “clear” or restored may also trigger notifications to cancel other work
tasks.
• Some traps, especially those deemed as major or critical, may trigger immediate
notifications.
• Information and debugging traps might simply be logged for future reference

Server Server Server

Switch Router
X
Router Switch

Printer

Server

Verification Paging
of
Failure Correlation
Email
Router Event Watcher Notification

Popups
Trouble
On Screen
Tickets
Log Files
NMS

Figure 8.1: A basic NMS.

166
Chapter 8

SNMP and Fault Management


Chapter 6 reviewed SNMP, which is a crucial protocol in holistic fault management and network
health monitoring. SNMP is the most widely used protocol for managing devices in the network.
It’s popular because it’s flexible and easy for vendors to implement. SNMP management
includes three components: managed devices, agents, and the NMS.
A managed device is any hardware element in the network that implements SNMP and is
capable of reporting information. This includes routers, switches, servers, workstations, printers,
and other devices. In the converged services network, the media and signaling gateways,
voicemail systems, and other call processing elements will all probably provide SNMP
monitoring and management capabilities. They may also include vendor proprietary
mechanisms.
An SNMP agent is simply the software that provides the SNMP information. The agent might be
a daemon process running within the OS kernel or some additional software that is installed at
the time the network device is set up. The agent software collects information and passes it to the
NMS.
The NMS is the centralized monitoring overseer. The NMS sends requests to monitored devices
in the network, and the agent software running on the device sends a reply. The NMS sends five
types of messages:
• GetRequest is used to retrieve a specific value from a network device.
• SetRequest is used to set a defined value in a network device.
• GetNextRequest is used by the NMS when building a table of responses. It’s used to
collect multiple inputs.
• GetResponse is used to return error codes and other responses to requests from the NMS.
• Trap is an unsolicited message. It’s sent from the agent to the manager. This is the error
reporting mechanism that is used to provide immediate information to the NMS about the
status of a network device.

167
Chapter 8

For technical details on SNMP see the following resources.


The SNMP Version 1 RFCs are:
* RFC 1155. Structure and Identification of Management Information for TCP/IP-based internets
* RFC 1157. Simple Network Management Protocol
* RFC 1212. Concise MIB Definitions
* RFC 1213. Management Information Base for Network Management of TCP/IP-based internets
The SNMP Version 2 RFCs are:
* RFC 1901. Introduction to Community-based SNMPv2
* RFC 2578. Structure of Management Information Version 2 (SMIv2)
* RFC 2579. Textual Conventions for SMIv2
* RFC 2580. Conformance Statements for SMIv2
The SNMP Version 3 RFCs are:
* RFC 2576 (PROPOSED STANDARD). Coexistence between SNMP Version 1, Version 2, and
Version 3 (March 2000)
* RFC 3410 (Informational). Introduction and Applicability Statements for Internet Standard
Management Framework (December 2002)
* RFC 3411. An Architecture for Describing SNMP Management Frameworks (December 2002)
* RFC 3412. Message Processing and Dispatching (December 2002)
* RFC 3413. SNMP Applications (December 2002)
* RFC 3414. User-based Security Model (December 2002)
* RFC 3415. View-based Access Control Model (December 2002)
* RFC 3416. Version 2 of SNMP Protocol Operations (December 2002)
* RFC 3417. Transport Mappings (December 2002)
* RFC 3418. Management Information Base (MIB) for the Simple Network Management Protocol
(SNMP) (December 2002)
* RFC 3584. Coexistence between Version 1, Version 2, and Version 3 of the Internet-standard
Network Management Framework
* RFC 3826. The Advanced Encryption Standard (AES) Cipher Algorithm in the SNMP User-based
Security Model

168
Chapter 8

Syslog and Fault Management


As described in Chapter 6, syslog is another important tool in fault management. SNMP is
widely used for status and health monitoring. It provides real-time information about elements in
the network.
Syslog provides more comprehensive log data, including error messages with extensive data
about what is occurring internally within the network device. As syslog data comes from an array
of devices throughout the network, it’s used in event correlation to understand the impact a
failure in one part of the network may have in other areas.
Because syslog data is in plaintext, it’s easily readable, even when not easily understood. This
makes it easy to manipulate for analysis. Syslog data is crucial for verification and validation of
details associated with many network faults.

Chapter 9 will explore the value of syslog information in assessing security issues.

Fault Management and ROI


There are tangible and measurable return on the investment (ROI) values in fault management
methods. The level of service delivered to users is higher when fault management and
monitoring are applied. NMS tools provide immediate impact assessment capabilities for
response to faults and problems arising in daily operations. This leads to better decision making,
both in troubleshooting and network planning. This improved operational stability will result in
fewer outages and errors and improve the QoS delivery overall. Consistency in service delivery
engenders user confidence in the integrated services. Satisfied users will leverage the converged
services into other enterprise business process flows. Knowing and understanding the details of
problems in the service network provides not just historical information but also crucial business
intelligence that aid in future capacity planning and new services for consideration in the future.

169
Chapter 8

Configuration Management
The primary objective of configuration management is to monitor network and system
configuration information to provide an audit trail, or tracking mechanism, of changes made to
the network. Every device in the network has some associated configuration information. Given
the variety of network elements in the converged services network, it’s critical that configuration
changes be coupled tightly with some formalized change control process.
As with every other aspect of managing the FCAPS model, the NMS is the brains of the overall
process. NMS solutions might be single-vendor, single-system solutions or they might be
distributed systems that work together and roll information to the previously mentioned
“Manager of Managers.” NMS is a system that may have several components.
Some key features to consider when investigating configuration management include:
• Auto-discovery—Is the system capable of automatically probing your network and
learning what the components of the infrastructure? Can it determine topology and layout
of the existing network, and detect new elements as they’re added? Can it be configured
to automatically retrieve configuration information and store it in the configuration
management database?
• Ability to import configurations—Will the system support all the various configurations
of different vendors’ solutions into a single database schema? Will different
configuration databases be required for servers and workstations than for switches and
routers? Are all the elements of the converged services network supported (call
managers, media/trunking gateways, signaling gateways, voicemail systems, and
traditional telephony elements such as PBXs)?
• Configuration analysis—This is a new development in NMS capabilities that has arisen
in the past few years. Is the NMS able to compare existing configurations with other
similar devices in the network? Does it include a standardized comparison of known best-
practice configuration issues? Routers may have common configuration errors that don’t
impact prior operations but might impact converged services. Servers and workstations
may have unnecessary services running that should be disabled. OS versions and patch
levels need to be monitored, and in most networks, maintained with consistency across
similar network devices. Will the NMS provide this comparison capability?
• Policy-based configuration—Does the NMS support a single policy-based solution and
provide for easy deployment across all devices in the network? Policy-based
configuration management allows precious staff time and resources to be used more
effectively. The ability to write a policy template that can then be universally “pushed”
ensures consistency of operations and unifies compliance with enterprise standards.

170
Chapter 8

Collecting and Storing Configuration Data


In unified communications, network configuration management provides control of changes in
the network. The most common changes noted are to hardware, firmware, and software, but
documentation changes also apply. And these changes continue throughout the lifespan of the
network. Change is constant, driven by upgrades, patches, and equipment replacement. A
comprehensive configuration management approach will collect all the changes made to a
network element throughout its entire life cycle.
Once a device is deployed in the network, change control processes guide the evaluation and
approval of implementing configuration changes. This function is vital for two reasons:
• The interrelationships between network elements become more complex as the network
incorporates new services. In the past, a simple change to a router might have only
impacted traditional data flows. In the converged network, a simple router change can
easily alter the QoS for a traffic flow. This simple change might disrupt or degrade voice
services unexpectedly.
• Fallback to previous configurations may be required quickly. One of the great benefits of
comprehensive configuration management is the ability to undo changes and revert to a
prior, known operational state.

Configuration and Change Management


Configuration and change management thought processes in the industry today are loosely based
on guidelines for software configuration management principles described by Roger Pressman in
his book Software Engineering: A Practitioner’s Approach (McGraw Hill, 2004, ISBN:
007301933X). SCM is a widely accepted methodology for controlling and managing change in
the software development environment, but the core principles apply equally to networking. The
central focus is to identify what changed, who changed it, and how the change can either be
reproduced or undone.

Configuration Management Resources


The Institute of Configuration Management at http://www.icmhq.com/index.html
Configuration Management Training Foundation at http://www.cmtf.com/
Configuration Management Body of Knowledge at http://www.cmbok.com/
The Configuration Management Wiki-Web at http://www.cmcrossroads.com/cgi-bin/cmwiki/view/CM/
Configuration Management Principles and Practice by Anne Mette Jonassen Hass (ISBN: 0-321-
11766-2)

171
Chapter 8

Configuration management provides for the identification of changes, a controlled storage


environment for archival data, change control, and status reporting of every change activity
throughout the life cycle. Figure 8.2 shows a simplified view of configuration management
activity areas and offers a peek into process flow. The configuration data itself is shown as
metadata (data about the configuration information). In configuration management, the metadata
for a configuration item may include:
• The name of the change
• The name of the person initiating the change
• The name of the approver of the change
• Text description of the change
• Date change was placed into production
• References to other configuration items
In most NMS solutions, the metadata is often stored on the same system as the configurations
themselves.

Production
Configuration Audit
Network
Item

Configuration
Item

Configuration Controlled
Item Item Storage
Approved

Metadata Change Control

Metadatabase

Metadata

Change
Data

Status Reporting

Reports

Figure 8.2: Configuration Management Activities.

172
Chapter 8

Configuration items that are different versions of the same original item are obviously strongly
related, but each one is an individual item, which will be identified and may be extracted and
used independently. This is one of the main points of configuration management: to be able to
revert to an earlier version of an item class.
At every step of the FCAPS methodology there is a documented, methodical process being
reinforced as a commonly accepted best practice. In Chapter 7, Figure 7.7 shows an evolutionary
path of an enterprise network moving from anarchy through reactive and proactive modes as part
of the maturation to a service-oriented business enabler. The maturation cycle of a network often
mirrors the evolution of the business itself. Configuration management is a vital part of
collecting and managing business intelligence about the services supporting the core business.

Performance Management
One key to the emerging integrated services network is the constant appearance of new
technologies. Today, 10 years into deployment, VoIP isn’t really new or emerging. Within the
broad telecommunications sector, many service providers are now stepping back and thinking
more purely in terms of voice services, with VoIP being one voice delivery mechanism.
For enterprise networks, the impetus is much greater to integrate technologies with business
processes. You’ll recall that Chapter 1 looks at convergence from several views:
• Infrastructure convergence of wiring and circuits
• Service convergence using IP as the delivery mechanism
• Device convergence of the physical tools used in everyday business
• Application convergence of enterprise business applications such as sales force
automation (SFA), supply chain management, and customer relationship management
(CRM)
Although commercial providers of voice services might be starting to take a narrower view of
voice as simply a service, that is a luxury the enterprise business cannot afford. The value in
convergence is found in the total integration. Complete integration can reduce cost and increase
productivity. Once integration has become inculcated into the corporate culture, it can lead to
creative new solutions to old business problems and even spawn ideas for new business tactics.
Chapter 5 looked at the Network Performance Envelope methodology for assessing
requirements. One tangible benefit to this model is that it supports many phases of the life cycle
of the enterprise converged services network. Figure 8.3 touches on some of the most basic
principles of this model.

173
Chapter 8

Availability
Throughput
Manageability
Scalability

Integrity Cost

Response
Time CPU Utilization

Reliability Security
Network Segment
Utilization

Figure 8.3: Revisiting the Network Performance Envelope.

There are many factors that impact performance of the network. Prior to implementation, you
assess network readiness—you determine whether the network can support converged data,
voice, and video. As you implement those services, you reassess to ensure that your assumptions
have all been proven correct, that any requisite upgrade activity was successful, and that the
network does indeed deliver the integrated services as planned.
In the operational phase, you continually monitor performance to ensure those standards are
being upheld. Operational success is driven by information. The more you know about the
network, the better your service delivery consistency. Although network managers all strive to
deliver good service, sometimes consistency of service gets overlooked. Vigilant performance
monitoring helps ensure a consistent, reliable network.
Performance management is a series of checks and balances. We often visualize a counter-
balance scale in weighing performance against other factors. That simplistic view doesn’t serve
the needs of managing a converged service network. Some managers view this multi-service
network in terms of the trade-off triangle (see Figure 8.4).

174
Chapter 8

Data

Trade-off
Triangle

Voice/VoIP Video

Figure 8.4: The misleading trade-off triangle.

The danger in taking this view of balancing performance and service is that it pits network
services against one another for resources. This approach may be simple, but the danger of
giving preference to one of the new emerging services, VoIP or video, is that existing data
services—Web services, financial applications, customer service tools—may suffer from an
exhaustion of resources. Business-critical applications may starve for resources while the new
technologies take center stage. The holistic truth is more akin to balancing a high-performance
automobile tire.

175
Chapter 8

A binary decision of either this or that


will lead decisions about holistic network
management to ineffective conclusions

Throughput
Availability
Manageability
Scalability

Integrity Cost

Response
Time CPU Utilization

Reliability Security
Network Segment
Utilization

Figure 8.5: Managing holistic balance.

The illustration in Figure 8.5 has simply taken the network performance parameters used in the
performance envelope methodology and shown them as balance points around a tire.
Performance monitoring and management is not a binary decision in any way. Trying to balance
security against manageability, error rate against response time, or throughput against reliability
are just a few examples of how binary decisions fail.
Because much of the focus at this stage of convergence is on VoIP services, look first at the
VoIP service elements that must be monitored. Remember that years of telephone usage have
socialized us to expect a base performance and quality profile akin to the traditional telephone
network.
When VoIP devices register with the network, whether they are phones, media/trunking
gateways, signaling servers, or other devices, registration problems can adversely impact the
delivery of service. If a VoIP phone on a desktop can’t pull down a profile and phone number for
any reason, problems follow. You’ll want to establish a monitoring and alerting threshold in the
event that something as simple as registration attempts or failures exceeds predefined levels. If
there is growth in the network, or the number of registered telephones changes dramatically, it
could be a signal that there is a problem with the VoIP services themselves. Gateway registration
monitoring will help identify new or missing trunk capacity to other networks.

176
Chapter 8

Call monitoring is often an inflammatory topic because of privacy concerns. At this level of
providing service in the network, call monitoring doesn’t mean eavesdropping on individual
calls. It’s truly call traffic monitoring. This function involves monitoring incoming and outgoing
call volumes to identify failures. If the VoIP system supports fax calling, attempted fax calls
should be monitored as well. In call monitoring, look at four distinct areas:
• Calls in progress—The instant a VoIP phone or softphone goes “off hook,” the call is
deemed in progress until the caller or called party hangs up. Calls in progress include
those in the process of dialing, off-hook getting busy signal, and so on. If every call
attempted completes successfully, the number of calls in progress will be the same as the
next time, active calls. As part of capacity planning, decisions have to be made at
implementation time as to the number of active calls the system can support. Performance
monitoring assists with capacity planning. If the percentage of available capacity runs
normally at 70% utilization, then rises to 95%, performance monitors can alert capacity
planners to a change in the network; a change that must be addressed before calls are
blocked due to demand exceeding the capacity.
• Active calls—Once a call has successfully connected, it’s deemed an active call. It has
completed the setup signaling and now has a voice media path connected through the
network. The same capacity considerations apply as for calls in progress.
• Attempted calls—In general, designers try, within available resources, to provide non-
blocking systems. That is, they try to guarantee that all calls attempted will be completed
successfully, but such isn’t the case in the real world. Calls encounter busy conditions
and go unanswered or are abandoned. Monitoring calls attempted over time yields data
used to identify peak calling traffic periods and what is referred to as the busy hour call
attempt (BHCA) value.
• Completed calls—Any phone call that completes and terminates without an abnormal
termination code is counted as a completed call. Monitoring completed calls over time is
also helpful in determining peak traffic periods and the BHCA value.
For most businesses, using VoIP strictly as an internal calling system isn’t practical. Corporate
VoIP services have to interconnect with the Public Switched Telephone Network (PSTN)
through either SIP trunks or standard voice trunks. This is done through some form of gateway.
A gateway connecting to the PSTN via traditional T1 trunks needs to be monitored on the PSTN
side of the VoIP service network not just on the IP network. Monitoring active channels on these
trunks over time can help identify calling patterns and busy hour peak call volumes. Baseline
data can also be used to identify underutilization of circuits, leading to downsized capacity and
reduced operating costs. Data trending helps in capacity planning and the growth and maturation
of the converged services.

177
Chapter 8

Another benefit to implementing VoIP services is the conference-bridging capabilities.


Conferencing adds yet another monitoring aspect. If your VoIP implementation supports
conferencing, the maximum number of audio streams that can be supported for conferencing has
to be configured. Monitoring will help ensure that the number of available audio streams meets
the service levels needed for day-to-day business requirements.
IP phones, whether they’re physical hardware phones or softphones running on workstations,
require continual monitoring for service assurance. IP phones should be monitored for
registration status, dial tone validity, jitter, latency, and packet loss. These key QoS metrics
directly impact both call quality and the user experience.

QoS and Bandwidth Monitoring


Earlier chapters talked about codecs and QoS. Bandwidth requirements for VoIP traffic are
driven in large part by codec selection. The PCM codec (G.711) requires about 64Kbps to
support a bi-directional phone call. The G.723 and G.729 require significantly less bandwidth
due to compression, but network congestion may have a greater impact on call quality.
Whenever new applications are introduced into the business network, there is a risk of
oversubscribing individual links. Oversubscription leads to congestion that can, in turn, degrade
call quality. Packet loss and increased latency are common side effects of congestion. In the
worst case, they can disrupt call setup and voice media transmission to the point that VoIP
services become unusable.
To guarantee that VoIP users receive an acceptable level of voice call quality, VoIP traffic
generally needs to be given priority over other types of traffic on the network, although video
traffic may warrant even higher priority that voice. Data traffic is bursty in nature and often does
not involve real-time communications between people. The primary objective of QoS techniques
it to provide a prioritization mechanism that can ensure that every type of packet on the network
is handled appropriately for the content inside. Because of its real-time nature, VoIP traffic
typically receives the preferential treatment to reduce or eliminate delay. Commonly monitored
voice performance metrics include:
• Delay—Latency or delay is an estimate of the packet delivery time across the network.
It’s expressed in milliseconds and measured as the average value of the difference
between a sender’s and receiver’s timestamps on messages. It’s measured when the
messages are received. Remember that delay is cumulative. End-to-end delay is a key
factor in determining the overall VoIP call quality.
• Jitter—Variation in delay is called jitter. We’ve mentioned consistency in performance,
and jitter is one area that is crucial in delivery of real-time services such as voice and
video. It indicates a variance in the arrival rate of packets at the destination. Jitter is a
predictability factor that is often used in discussion of the overall reliability of a service
network. Jitter problems are well known to adversely impact call quality. Networks can
compensate for jitter by implementing jitter buffers to normalize the timing of the traffic
flow. Jitter buffer loss occurs when jitter exceeds that which the jitter buffer can hold.
Jitter and jitter buffer loss affect call clarity, which affects the overall call quality and
user experience.

178
Chapter 8

• Packet loss—Loss simply indicates packets lost or discarded in transmission. In VoIP


services, this could mean the loss of an entire syllable or word during the course of a
conversation; more importantly, it might mean the loss of a dialed digit, preventing a call
from ever completing successful. Obviously, data loss can severely impair call quality.
Monitoring systems measure the number of packets that were expected against the
number actually received.
• Mean Opinion Score (MOS)—MOS is a subjective quality measure that was discussed in
earlier chapters. MOS testing is non-intrusive and provides a way to monitor and measure
call quality in ongoing network operations. Historically, MOS was derived from panels of
judges listening and scoring call quality. Although it’s useful for human users to interpret
quality, it’s a useless metric in delivering quality assurances under any form of Service
Level Agreement (SLA). It doesn’t fit well in operational monitoring systems; but there
are alternatives that are gaining popularity, especially with large enterprises and service
providers.
The E-Model is tool that can predict the average voice quality of a call using a mathematical
model. E-Model accounts for the estimated impact of delay, jitter, loss, and codec performance.
The output result of an E-Model calculation is called the R Factor or R Value. These values
estimate voice quality on a scale from 0 (the lowest quality) to 100 (the highest quality). Like an
MOS score of 5, an R Value of 100 is, in theory, unattainable, but it’s the standard goal network
engineers shoot for when designing voice service networks.
Because E-Model scores are based on measurable parameters, monitoring tools are becoming
more common that can be incorporated into the enterprise NMS strategy for performance
monitoring. E-Model monitoring tools evaluate the Real-Time Protocol (RTP) streams based on
information found in the traffic flows (source address, destination address, TCP.UDP port
numbers, and packet sequence numbers) to create what is called a “jitter profile.” E-Model then
creates a score that, in testing, correlates to traditional MOS with 80 to 90 percent accuracy.
There is another non-intrusive performance measurement technique that is emerging called the
ITU-T Calculated Planning Impairment Factor (ICPIF) score. It’s based on the ITU-T G.113
standard. Driven by increasing sophistication required to accurately assess voice quality, some
vendors are beginning to adopt this technique. With increased attention to VoIP quality in
standards bodies such as the IEEE and ITU-T, this method, which has been around for nearly 10
years, is gaining traction across the industry.

179
Chapter 8

ICPIF takes some factors into account that the E-Model does not:
• Signal attenuation distortion
• Circuit nose
• Codec encoding distortion
• Delay distortion
• One-way transmission time
• Echo
Although ICPIF may be slow moving into enterprise business network monitoring, it will
certainly play a role among service providers in delivering commercial voice services using
VoIP.

Collecting and Analyzing the Data


Throughout, this guide has repeated a common theme: Information gathering is absolutely
necessary to successfully manage a converged service network. The more extensive and
comprehensive the data collection and analysis tools, the better armed network managers are to
make decisions about present daily operations and future capacity and growth plans.

Monitoring the Health of the Network


Ongoing health monitoring provides the service delivery organization information with which to
make informed decisions. Perhaps more importantly, it provides a reportable mechanism to
demonstrate service delivery assurances to end users and customers.

Performance and Utilization Trends


Close monitoring of network services brings the tools to hand to proactively measure
performance and utilization for multiple reasons. The enterprise network—particularly the
integrated services network providing delivery of a mix of data, voice, and video services—is a
finely tuned machine. Just as a high-performance vehicle will get better mileage, achieve faster
acceleration, and handle better on the race course, a finely tuned network will handle peaks in
traffic and unplanned events more smoothly than a network just left to run.
The dashboard of the NMS provides a window into the performance levels and utilization of the
network at any point in time. Changes in operating conditions can be detected quickly and
proactively, ensuring that business-critical network services are always available.

180
Chapter 8

Administration Management for Performance and Planning


In addition to daily operations, monitoring all these factors provides trending information on
utilization, traffic patterns, and capacity. This enables network planners to accurately anticipate
future network needs. This added knowledge factors into business decision in other ways.
Knowing that capacity planning needs are being closely managed helps ensure that a company
doesn’t overspend in future capacity that isn’t required to support the business. Information acts
as a quality assurance mechanism in future network planning.

Gathering Usage Statistics


Beyond the call quality and user experience measures and metrics, the infrastructure itself needs
to be monitored closely to ensure the critical service delivery elements are operating at
appropriate levels and can handle the load of traffic being passed. Minutes of use in the
traditional telephony environment were used as an indicator that enough T1 trunks and circuits
were available. In the converged network, minutes of use correlates to bandwidth and network
resource consumption. A trend indicating growing minutes of usage means more network
resources are being used. It can provide an early indicator that SIP or TDM gateways, signaling
servers, voicemail systems, and other elements of VoIP service delivery need to be evaluated to
ensure they aren’t being stretched beyond tolerable capacity levels.
Disk space and CPU utilization are very simple indicators to watch and good barometers of
changes in the environment. If any vital network element spikes in CPU utilization, or suddenly
runs out of disk space, monitoring helps ensure an early warning system can engage immediate
response before service delivery is impacted.

Managing Backups and Synchronization for Performance


In enterprise business, backup systems are widely deployed to ensure all mission-critical data is
protected. When delivering converged services, some new considerations may come into play.
It’s imperative that you remember to back up not just the data systems but also the management
systems. The NMS in an integrated network is a mission-critical piece of service delivery inside
the enterprise. Backup and restoration plans need to be developed and rigidly followed. After
investing all the time and effort required to baseline the network, collect configuration
information, and build a historical library of past performance metrics and trends, it would be
disastrous to lose the information simply because the NMS wasn’t part of the information backup
strategy.
For large enterprises and geographically dispersed organizations, continuity of operations may
involve a different approach. Many companies have a business continuity or disaster recovery
center located at a remote location. As part of the NMS strategy, it’s wise to consider a mirrored
NMS at the remote facility to ensure uninterrupted service in the event of a disaster.

181
Chapter 8

Summary
Networks are increasing in both breadth and depth. The increase in breadth relates to an increase
in size. More users connect to enterprise networks. The number of connected devices is also
growing. There is a parallel increase in depth driven by the set of services delivered to any
endpoint on the network. In earlier networks, the service delivery footprint was small, relegated
to FTP, email, and Web browsing. Converged networks are rapidly changing that. The integrated
network is now bringing data, voice, and video together. From an infrastructure perspective, that
drives the need for more vigilant monitoring of faults, capacity, assets, performance, and
security.

Chapter 9 will address security in the converged services network in more detail.

The converged network is also rapidly growing in other ways. Voice and video are viewed as
network services. Application services are also quickly integrating, both into the network and
with voice and video. There are multiple concurrent convergences underway, making it vital that
you monitor closely to protect your business services. Providing users with a consistent QoS and
reliable network experience in increasingly complex networks requires the use of increasingly
sophisticated tools.
Legacy NMSs don’t do a good job of managing emerging, converged networks. The old tools
focused on a GUI interface with a network status indicator. In today’s multi-service network, that
may not be enough. New tools are based on algorithms for network discovery, real-time
monitoring, visualization tools for modeling, simulation, anomaly detection, and event
correlation.
This guide has touched on several views of life cycle management. Figure 8.6 presents another—
a simple network performance management life cycle.

Topology
Discovery

Configure New
Systems

Collect
Statistics
(Monitoring)

Remediate
Problems

Identify
Problems

Figure 8.6: Network performance management life cycle.

182
Chapter 8

Many legacy NMS approaches only deal with discovering network topology and collecting
statistics. When planning for the ongoing life cycle management of an integrated data, voice, and
video network, it’s important to look at NMSs that can incorporate problem identification and
remediation along with the configuration of new systems.

As a note on topology, remember that with current networking equipment, topology is more than just
ports and cables. The logical topology of VLAN configuration, MPLS domains, and IP subnetting is an
important facet of topology discovery in depth.

As networks evolve, QoS guarantees for bandwidth and delay characteristics for audio are a
starting point, but they aren’t enough. The technologies are advancing rapidly and this is a good
time to consider broader performance attributes such as variations in connectivity, robustness of
redundancy and failover, and the overall performance of the traditional best-efforts approach
used by IP. For future networks, these tools will need to provide for:
• Performance monitoring and measurement
• Forensic analysis (beyond packet sniffing)
• Capacity planning (to project impacts of new services)
• Load generation (for scenario testing)
The only way to can achieve an ideal NMS is to proactively incorporate components of
performance monitoring into the system. Metrics for availability and responsiveness need to be
included in any notification system. Baselining support is intrinsic to success, and routinely
scheduled baselines should be run across the corporate network.
Without an NMS that is oriented to the complete array of services provided, including data,
voice, and video, companies will fly blind as to service assurance. VoIP services that lack
comprehensive management and monitoring procedures are likely to deliver poor quality
services. Continual monitoring of documented service-level metrics (appropriate QoS metrics) is
the only way to truly ensure and demonstrate enterprise-class service delivery. These metrics
should include, at minimum, some combination of jitter, latency, call completion and quality,
and voice quality (MOS, and so on).
Chapter 9 will wrap up the dive into the FCAPS model and will highlight security management
issues facing enterprise deployments today. In addition, the chapter will identify common
industry best practices for creating an effective VoIP security plan that balances securing the
network against the VoIP requirements for availability, reliability, and performance.

183
Chapter 9

Chapter 9: Effective Security Management


The convergence of voice, video, and data networks has been evolving and gaining momentum
for the past several years. Many organizations have undertaken VoIP implementation to
converge networks for cost reduction. Others work to achieve the competitive advantage of
integrated services. Whatever the reason for network service integration, you cannot overlook the
security risks that arise as technologies converge. VoIP implementers often focus on issues of
voice quality, and interoperability. These are truly important factors in the delivery of voice
services. In many ways inside the converged service network, voice security needs to be treated
as data security. And data security needs to be treated as voice security. Both technologies bring
issues and management techniques that benefit the other. This chapter will highlight security
management issues facing enterprise deployments today and identify common industry best
practices for creating an effective and comprehensive security plan that balances securing the
network against the VoIP requirements for availability, reliability, and performance.
Security methods can adversely impact network performance. Firewalls induce delay by
inspecting each packet in the data stream. This will add delay to packet delivery. Congestion at
the firewall can lead to variable processing time within the firewall. This will increase the
problem of jitter. A systematic and holistic approach to managing integrated network
performance and security includes working with vendors, services providers, and trusted
business partners to ensure a comprehensive approach to security is followed.
As previous chapters illustrate, successful operations are driven by knowledge and information.
The more you know about the network, the better you’re able to analyze problems. Solid
knowledge and understanding of the network leads to an approach that balances all aspects of
network management. Building this base knowledge helps you effectively manage the entire life
cycle of network services and applications to ensure you’re delivering the services needed today
and able to meet the needs of tomorrow.

184
Chapter 9

FCAPS and Security Management


Recent reading online brought my attention to an article from Professor Eugene Spafford on the
Center for Education and Research in Information Assurance and Security (CERIAS) at Purdue
University. CERIAS is viewed as one of the world’s leading centers for research and education
in areas of information security that are crucial to the protection of critical computing and
communication infrastructure. In regard to best practices, Professor Spafford said:
In the practice of security we have accumulated a number of “rules of thumb” that many
people accept without careful consideration. Some of these get included in policies, and
thus may get propagated to environments they were not meant to address. It is also the
case that as technology changes, the underlying (and unstated) assumptions underlying
these bits of conventional wisdom also change. The result is a stale policy that may no
longer be effective…or possibly even dangerous.
Policies requiring regular password changes (e.g., monthly) are an example of exactly
this form of infosec folk wisdom.
From a high-level perspective, let me observe that one problem with any widespread
change policy is that it fails to take into account the various threats and other defenses
that may be in place. Policies should always be based on a sound understanding of risks,
vulnerabilities, and defenses. “Best practice” is intended as a default policy for those
who don’t have the necessary data or training to do a reasonable risk assessment.
In small and midsized businesses, and even large enterprises, often the resources aren’t available
to perform comprehensive or even reasonable risk assessment. Perhaps more importantly, we
create deliberate change control processes that don’t provide the luxury of time to reassess risk
each time we make a change in the dynamic networked environment we support. To perform a
reasonable risk assessment each time your environment changes would stall all forward progress
in technological growth.
Industry best practices are a tool employed because they afford us the collective institutional
knowledge of the entire technology sector in adopting processes that have been proven to work.
Best practices leverage, to your benefit, the common body of knowledge that IT managers share
based on experience.
The key to the FCAPS model is in establishing processes for managing each of the five service
areas—fault management, configuration management, accounting management, performance
management, and security management. This chapter will delve into security management
processes to protect and sustain the converged video, voice, and data service network.

185
Chapter 9

Identifying Risks
Best practices provide a tool to support operations, yet some basic risk assessment is always
prudent. Risk is simply something that is outside normal operations. When voice, video, and data
networks converge, every risk associated with each service is incorporated into the newly
converged network. This additive risk compounds the necessity for building a layered defense to
protect not just the network infrastructure but also each of the services running on the network.
Figure 9.1 shows a simple risk model called Jacobson’s Window that is often used in considering
the range of risks to which networks are exposed.
Jacobson’s Window present a simple, two-factor risk assessment model, distinguished by the
frequency or probability of an event occurring coupled with the impact or consequences of the
event. This two-factor matrix simplifies all risk into one of four categories.

Consequences
LOW HIGH

L
O DISREGARD
W
Occurrence
Rate of

H
I
G DISREGARD
H

Figure 9.1: The Jacobson’s Window risk model.

The four risk classes identified using this approach are low-low, high-low, low-high, and high-
high. This approach is further simplified by breaking risks into two broad classes:
inconsequential or significant.

186
Chapter 9

Inconsequential Risk Classes


One risk that is considered inconsequential is the low-low class. It can generally be ignored
because it doesn’t matter statistically. The likelihood of occurrence is low and the consequences
of these risks are deemed low, so they represent minimal impact on the organization. A risk that
occurs once a year and has the impact of costing $1 is just too trivial to worry about.
The opposite end of the spectrum, high-high risks, can also be ignored for the most part. In his
writing, Jacobson suggests these risks just don’t ever occur in the real world. He uses the
example of a 50-ton meteorite crashing into your computer room on a daily basis. Although this
example is extreme, it demonstrates why the high-high risk class is inconsequential. If this event
occurred, you wouldn’t build your computer rooms in the first place. In the real world, the high-
probability and high-loss risks are generally immediately mitigated or we simply couldn’t
conduct business. These risks demonstrate lessons that we have already learned and for which
we’ve developed remediation strategies. A new high-high risk is typically spawned out of some
disruptive, radical shift either in technology or business practice. They’re typically addressed as
they’re created.

Significant Risk Classes


If you’re looking for a simplified approach to risk management, you’ve just eliminated half the
classes in the entire risk spectrum. That only leaves two categories or classes of risk you really
need to be concerned with: high-low and low-high. Spam email represents a perfect example of a
high-low risk. There is a very high likelihood of occurrence. Current studies indicate that more
than 50% of the email received is spam. In most cases, the resulting loss is a loss of productivity
as people delete the spam messages.
Jacobson uses a fire destroying a telephone company central office as an example of a low-high
risk. It demonstrates an event that has low probability of occurrence but a high consequential
loss.

On May 9, 1988, an electrical fault started a fire in the Hinsdale, IL central office operated by Illinois
Bell. Early during the fire, telephone services failed. The fire department didn’t arrive on site for 45
minutes. Because of the dense, black smoke, firefighters had difficulty entering the building and
locating the source of the fire. Emergency power was automatically provided by generators and
batteries and could not be shut off easily. Neither standard dry chemical nor halon extinguishers were
effective against the fire. Water had to be used, which exposed the firefighters to electrical shock
danger. It took firefighters more than 2 hours to shut down power, enabling them to control and
extinguish the fire. It was more than 6 hours after the first fire crews had arrived on the scene that the
fire was declared under control.
This fire was confined to an area roughly 30 feet by 40 feet on the ground floor. Cables were burned
to various degrees and smoke residue covered most of the ground and parts of the first floor. The
most severe damage away from the fire was caused, not by flames, but by corrosive gases in the
smoke. These corrosives damaged the equipment that survived the fire. Although the existing
equipment was cleaned and used to provide interim service, it was deemed unreliable. This
equipment all had to be replaced over time after the fire.

187
Chapter 9

There is a broad spectrum of real-world risk that ranges from low-high to high-low. Human
nature drives another facet of low-occurrence risks. Jacobson postulates that “People tend to
dismiss risks that they have not experienced themselves within the last 30 years.” Although 30
years may seem arbitrary, psychologists believe it’s accurate and related to the range of human
life experience. We generally remember events that happened during our lifetime. There are
plenty of examples of Jacobson’s “30 Year Law” at work, but they’re easily identified. How
often have you heard senior managers or public officials say, “How could we expect anyone to
anticipate this event? We’ve now taken appropriate measures to ensure that this will never
happen again.” Human nature shows us that some lessons may only be learned through direct
experience.
When providing security for network services, managers need to visualize every possible risk
that might occur in the enterprise. You’re often called on to consider events you haven’t
personally experienced. It’s often useful to pull together incident response teams and think well
outside the box of routine business. Natural disasters, cyber security incidents, and unrelated
networking problems, such as a cut fiber, all impact service delivery. Sometimes a good planning
session involves discussion of “all hazards” risk. This approach, undertaken with a broader set of
staff members and managers, can help bring other physical and misconfiguration incidents into
clearer view as potential problems. The next step is to identify optimal mitigation strategies for
each risk.
Some organizations lean in favor of ROI-based risk management. This approach is often met
with limited success. The ROI-based approach can work well with events that are high-
likelihood, low-consequence because managers believe the risk exists, and therefore, it should be
addressed. Low-probability, high-consequence risks present a different problem. It’s difficult to
quantify the likelihood of occurrence, making it difficult to achieve management concern that the
risk is tangible.
Every approach to risk assessment brings benefits and each has drawbacks. ROI-based decisions
aren’t appropriate for every situation. When implementing security measures, it’s often wise to
take a step back and evaluate why the security measure is necessary in the first place. Generally,
security measures are implemented due to one of four reasons:
• The value of a security measure is significant but the cost is trivial. For example, the lock
on the front door of an office costs little yet provides the very first measure of security. If
the door is left unlocked, the consequences may be quite high but the cost of
implementing a policy to lock the door is quite small.
• In many cases, the potential reduction in future losses will more than offset the cost of
security. In this case, the security measure has a demonstrable positive ROI. This is the
justification widely used for many security measures related to high-low risks in
business. Card key access controls to buildings, password change policies, and basic
security training for employees are good examples.

188
Chapter 9

• For many enterprises, legislation or regulatory requirements necessitate specific security


measures. The Health Insurance Portability and Accountability Act (HIPAA), Graham-
Leach-Bliley Act (GLBA), Sarbanes-Oxley Act (SOX), and compliance with ISO17799
standards are good examples of this justification for security measures.
• In many cases, security measures are implemented to address a low-high risk that has an
unacceptable consequence for any one single occurrence. This single occurrence loss
(SOL) might be something that exceeds the owner equity or net worth of the company.
These measures represent protection against risks that bring devastating consequences.
Low-high risks often require senior management involvement in decision making. ROI analysis
may not justify implementing protective measures. Sometimes good judgment is required to
decide what’s acceptable to feel “safe enough” for the needs of the business.

SOLs and Reducing the Risk


There are two tactics to reduce overall risk using this two-factor approach to risk assessment.
Because they offer tangible risks with a heightened sense of urgency, we have a natural tendency
to focus on risks that generate a large SOL. You can typically mitigate these risks by:
• Reducing your vulnerability—Enterprises have often thought in terms of disaster
recovery—how you might restore operations after a major event. There has been a shift
in recent years to think more in terms of business continuity, or continuity of operations.
Robust enterprise business continuity and resumption plans reduce any adverse
consequences by shifting some aspect of the consequences to a resilient business
continuity environment.
• Spread the risk across the enterprise—Distributing work across multiple geographic
locations or business units can help spread the impact of any event across Help desk,
training, and other organizations. The risk to each business area might be smaller and
more easily addressed than a large risk to the entire enterprise.
• Transfer the risk—One solution may be obtaining insurance. The insurance premiums
will factor in the exposure against a deductible amount, so the impact might be lowered
to a tolerable level. A $50 million dollar risk reduced to a $5 million dollar potential
single loss through insurance may provide an acceptable alternative.
After the potential risks have been identified, each enterprise will need to make unique decisions
relative to their own business needs. High-low risks may be analyzed using an ROI-based
calculation. Low-high risks often need to be reviewed in a management strategy session. It’s
wise to use the team approach to estimate the SOL potential and gain a collective sense of the
likelihood of occurrence. Management will make a decision and draw a line somewhere along
the spectrum. Usually this line is drawn based on the largest acceptable SOL.
Risks in the enterprise tend to spread out in a graph like the one shown in Figure 9.2. Viewed in
this context, a conscious decision can be made about each risk and each type of risk. Low-low
risks might be deemed inconsequential and addressed in another venue. As the risk increases,
some risks may be transferred. Higher risks that occur infrequently might be areas for insurance
considerations. Toward the middle of the low-to-high consequence range may be risks that can
be addressed through policy and training efforts.

189
Chapter 9

LOW – Frequency of Occurrence - HIGH Happens too


rarely to
Accept address
Risks
Max Tolerable
Single
Occurrence
Loss

Mitigate
Risks

Don’t Transfer
Care Risks

LOW - Consequences of Occurrence - HIGH

Figure 9.2: Modeling risks.

Although this four-quadrant view of risk provides a simple tool, for most businesses, the risk
mitigation focus falls toward the middle area of the graph. The lower left quadrant represents
things that happen so rarely and cost so little you might not need to factor them into daily
operations. The upper left quadrant represents things that occur more frequently but cost little.
These are often viewed as acceptable risks. They’re typically treated as a factor of the cost of
doing business. High-occurrence risks that rise in consequence show up in the lower right. These
may be transferred to other parts of the enterprise, insurance carriers, and when appropriate,
outsourced to a business partner. The upper right portion represents that area of high-high risk
and consequence events that, at least in theory, have been addressed before you started doing
business.
Jacobson’s Window isn’t always the best approach to risk assessment for an enterprise. It’s a
simple tool for quantifying and analyzing risk to help make sound business decisions that is easy
to use and provides some method for basic risk assessment.

190
Chapter 9

Risk Management Life Cycle


Like every process touched on throughout this guide, risk management is a never-ending
process. Figure 9.3 shows the ongoing cycle of identifying, analyzing and assessing, mitigating,
or transferring risks. Throughout the risk management life cycle, there are four key questions to
consider:
• What could happen?
• If it does happen, what is the impact?
• How often could it happen?
• How accurate are the answers to the first three questions (measure of uncertainty)?
Risk management includes risk assessment and mitigation as two elements of a much broader set
of activities. Other elements include establishing focal point for management, implementing
controls and policies, promoting awareness, and continuously monitoring effectiveness.

Assess Risk
And
Determine Needs

Implement Monitor and


Controls and Central
Focal Point Evaluate
Policies

Build
Awareness

Figure 9.3: The risk management life cycle.

191
Chapter 9

Risk assessment provides the foundation to all the other components of the risk management life
cycle. It provides a basis for enterprise policies and helps maintain a focus on cost-effective
approaches to implementing these policies. Risk assessment is an ongoing process that never
ends. New services, video, voice, and data are constantly being incorporated into the service
network. New vulnerabilities and exploits in operating systems (OSs) and applications are
discovered every day. Reconnaissance attacks through firewall probing, port scanning, worms,
and viruses are an ever-increasing nuisance. Threats and risks evolve over time and require
vigilant monitoring.
Enterprise information security needs a management focus. This central focus point comes about
through policy recommendations, business controls, awareness, and other areas that shape the
overall corporate culture of stewardship toward business intelligence. Security awareness
supports both daily operations and the risk management life cycle. Educating users on acceptable
use and security risks is a key factor in mitigating security problems.

Introducing Incident Response Planning: CSIRT Basics


One vital component of how any enterprise responds to a security incident is the maturity and
structure of the incident response team. In network security, there are three key management
factors: detection, prevention, and response. Although detection and prevention are the ideals,
you cannot detect and prevent every incident. Problems will arise. Security incidents will occur.
Detection is crucial; it provides the situational awareness as to what is happening in the service
network. Your reaction determines how you mitigate problems and restore business to normal
operations.
Most large enterprises have some form of Cyber Incident Response Team (CIRT) or Computer
Security Incident Response Team (CSIRT). For many companies, these working groups emerge
from a de facto team that engages when an incident occurs. Organizations that don’t think ahead
and fail to build a team find themselves caught in the throes of ad hoc response. This causes
delays in mitigation, often leaving critical services impaired for an extended period of time. A
well-organized and trained incident response team can provide a range of services that help
maintain service delivery capabilities. Typical roles include:
• Notification of the organization’s management when an incident occurs
• Activation of a documented response plan to resolve the problem
• Record keeping of the steps taken to reach resolution
• Problem containment
• Evidence collection and gathering, including making backups
• Problem mitigation and service restoration to maintain continuity of operations
• Coordination of incident post-mortem reviews
These steps collectively gather lessons learned from every incident that occurs in the enterprise.
These lessons learned feed process improvement plans, which leading to a continually rising
ability to provide services with a minimum of disruption.

192
Chapter 9

Many enterprises treat incident response as a value-added service within the organization,
coupling prevention, detention, and response services to internal measure or service level
agreements (SLAs). As Figure 9.4 shows, the CSIRT team provides incident response services.

Some organizations are now shifting to a model of building a cross-divisional response teams that
can provide a wider variety of services to multiple customer organizations inside the enterprise. Later
this chapter will address some of these services in more depth.

Figure 9.4: Reactive, proactive, and quality management services in CSIRT.

Building Processes for Defense in Depth


A primary driver for a defense-in-depth, or layered, defense strategy is to create a holistic
framework for end-to-end security across the enterprise. This comprehensive approach
encompasses layers of security and multiple points of enforcement throughout the architecture. If
the worst happens and there is a serious security breach, the organization must detect, contain,
and correct the problem. Defense in depth not only reduces the likelihood of a successful attack;
it significantly increases the chances of problem detection.
Confidentiality, integrity, and availability (CIA) of corporate data and network resources are
paramount for any organization. This CIA acronym has become a watchword for the security
industry. One goal of a layered defense is to build a service network infrastructure that can
withstand attacks or disasters and protect the enterprise from unacceptable losses.

193
Chapter 9

One approach that is often successful is to implement a series of routine technologies and
procedures than embrace a complex and elaborate scheme that might contain single failure
points. The simpler the infrastructure, the easier it is to protect. It’s also easier to maintain.
Service networks are large and complex enough without making them more so. And although
simplicity isn’t always a realistic goal in a large enterprise, it’s important to build a service
network infrastructure that is easily understood and maintained.
Any single control point can too easily become a single point of failure. Every layer of a strategy
for defense in depth supports the other layers. When breaches occur at the perimeter, this layered
approach limits the organization’s exposure. Security through layers is often compared to an
onion, as Figure 9.5 illustrates. There is no single layer that can be breached to get to the heart of
the onion. Several layers must be either penetrated or stripped away. In network security, each
layer provides another barrier, protecting the valuable business resources.

Policy, Procedures, and Awareness Physical Security

Perimeter Security

Network Security

System Security

Data Security Application Security

Figure 9.5: Layered security; a defense-in-depth strategy.

The best strategies provide effective protection of critical assets while allowing business to
flourish. The layers or concentric rings of a layered defense allow for controlled and monitored
access at each layer.

194
Chapter 9

User Authentication and the Gold Standard


The primary focus of enterprise security efforts is to protect not only the services network
infrastructure but also corporate data. Corporate data is a form of business intelligence, and as
such, it’s a vital asset to any business. Au is the symbol representing gold in the periodic table of
elements. This reference to gold provides an easy mnemonic tool to help remember three of the
most critical elements of network and data access security. Figure 9.6 shows these three “golden
rules” in data access protection.

• Authentication –
Verify that users
are who they say
they are
• Authorization –
Ensure users only
access things they
have permission to
access
• Auditing – Know
who did what, and
when they did it

Figure 9.6: The golden rules of data access protection.

Authentication
Authentication is simply the process of users proving that they are who they say they are. It’s
important to authenticate all users connecting to the network through some form of credential.
User authentication may occur many times throughout the work day. The most common form of
user authentication is simply an ID and password, but much stronger mechanisms exist and are in
widespread use. There are some considerations to be mindful of with regard to the authentication
process itself and how you deploy the technologies used:
• Do you employ a single authentication data store or are multiple resources involved?
• Are IDs and passwords encrypted during transmission or sent in the clear?
• Is simple user ID and password authentication sufficient or should you use some form of
strong or two-factor authentication?
• How secure is the authentication and user validation process?

195
Chapter 9

Authentication is simply the process of verification that users are who they claim to be. It’s
important that you authenticate users when they connect to the network and when they access
business information that is restricted in any way. Many networks use simple password
authentication technologies based on either Password Authentication Protocol (PAP) or
Challenge Handshake Authentication Protocol (CHAP). These methods require only that the user
know a user name and password. This is single-factor authentication.
Two-factor, or strong, authentication is typically based on a combination of “something you
have” and “something you know.” One widely recognized example of this is the key fob
technology that uses a time-based token to generate a number (something you have) coupled
with a PIN (something you know).
Biometrics may be used to introduce a third authentication factor: “something you are.” This
addition of a fingerprint/palm scan, facial recognition, voice recognition, or retinal scan adds
another layer of depth to the authentication process. Three-factor authentication is common
today online in the highest security environments and many research facilities. Although this
technology works, the cost is generally still prohibitive to deploy in the typical business network.
Two-Factor Authentication Risk Analysis
Strong, or two-factor, authentication cannot be considered an inconsequential risk under Jacobson’s
Window analysis. Access control cannot be treated as a low-low inconsequential risk due to its high
likelihood of occurrence. In a recent US-CERT Analysis Report, unauthorized access attempts rate in the
highest category of network probes. This category encompasses 85.6% of attacks on networks in the
latest quarterly evaluation.
Access control cannot be treated as a high-high inconsequential risk either. These exposures occur daily,
constantly in the stream of questionable activity inbound to state networks. Unauthorized access attempts
inbound to state government resources number in the millions each month. The rate of occurrence is
high, but the success rate is very low. The initial consequence of breach to a user account may well be
considered low.
Two-factor authentication concerns clearly fall into the category of significant risk. Significant risks are
those that have either a high rate of occurrence or present a high consequence risk. Given millions of
attempts per month against state government network resources, the rate of occurrence for login
attempts through password dictionary attacks and other methods is clearly a high occurrence risk.
Whether the consequences of a breach are high is dependent on which system(s) are breached. The
danger is that one breach easily leads to another, allowing deeper penetration into the network. These
cascading attacks are among the most dangerous, as the attacker gains increasing ability to masquerade
as an authorized user with each layer of penetration.
Single user passwords have long been proven as the weakest form of security in widespread use. In a
recent survey, 70% of users will trade their password for a chocolate bar (see
http://www.securityfocus.com/columnists/245 and http://www.securityfocus.com/columnists/245).

There is research underway on many fronts to shift the technology from two-factor and three-
factor authentication to a newer mindset of continuous authentication. This might be
accomplished through a variety of heuristic information, such as keyboard typing and other
behavioral patterns. As networking technologies have improved, so have security options, and
many networks implement a variety of these technologies.

196
Chapter 9

Authorization
After a user provides his or her credentials and has authenticated, the next step is authorization.
This is simply a matter of matching the authenticated user against a stored profile and applying
the appropriate access permissions for that individual. For some users, network access may be
the only action permitted, perhaps to the entire network, but in some cases only granting access
to a small subset. Different employees will typically have permission to work with different sets
of data. This is commonly referred to as role-based authorization. (for example, Human
Resources, sales, and engineering staff all need access to different resources but not to each
another’s resources).
Other authorization examples in a unified communications environment include:
• The ability to establish a remote VPN tunnel. This is generally predefined by user group
and establishes another set of user controls based on rules in a VPN concentrator or
firewall.
• The ability to program speed dial lists in the VoIP system may be restricted to only users
with specific privileges.
• Conference call reservation and management is often a restricted service.
• Access to log and reporting servers is often employed to ensure that appropriate
management or support staff are the only employees with access privileges.

Auditing
Auditing has become an important factor for enterprise business. In the traditional telephony
provider, this focus was on accounting to support customer billing. As enterprise services grow
and mature, auditing becomes a more necessary tool.
Auditing is often used in testing security configurations. It’s also important for troubleshooting.
Audit records will let you see whether the network permits multiple, simultaneous logins from
the same user. Some organizations permit this by design, while others deny this capability.
Multiple logins from different locations might provide an indicator that an employee’s account
has been breached or their password has been stolen.
A comprehensive audit trail provides a record of what every user did, and when. It can provide
tremendous value in event correlation analysis when some incident occurs. Knowing when a
VoIP call was placed, when a router configuration was changed, or when a server patch was
applied can all provide crucial business intelligence about the service network.

197
Chapter 9

Prepare to Respond
You and your network are a target, whether you think you are or not. Security incidents will
occur regardless of whether you plan for them. The odds are high that when an incident does
occur, it will be broader in scope and more deleterious to your network services that you’ve
planned for. The time to prepare is now; and tomorrow; and the next day. When your network is
under attack and services are failing, you will not have the luxury of thinking about a response
plan. It’s wise to take preparatory steps in anticipation of the inevitable.

Treating the network like a doctor treats a patient, managing overall health care, provides a robust
and sustainable network service delivery environment. Gathering documentation and consolidating
knowledge about the unified communications service delivery environment is vital to understanding
the healthy profile of the network. Strong, industry-adopted security practices, like good managed
health care, require monitoring the health of the network on a continuous basis. Quick response to
abnormal conditions will ensure early diagnosis and treatment of problems. Prevention, detection,
and timely response are cornerstones to maintaining the good operating health of the service delivery
network.

Policies, Procedures, and Awareness


The most secure system in the world is only as good as the people administering it and the users.
People are the foundation for any good security model, and historically they are the weakest link.
Policies and procedures provide the foundation for any layered defense strategy. People form the
first layer of defense. A corporate culture of stewardship for protecting the company’s
proprietary information is fundamental to good network security.
It’s not enough to put security policies in place. Employees need to understand them. Security
policies should be reviewed with all employees annually. Every employee should be trained and
understand their individual role in protecting network services and corporate information.
These processes and procedures are more important than security technologies. It’s absolutely
vital that you account for the human factor. People are the weakest link, and it’s widely accepted
across the information security community that the greatest threat to security lies within. People
can all too easily make mistakes or become careless. A disgruntled, unhappy, or careless
employee can be a threat to the security of the entire enterprise. Employee awareness and
training are essential to building a sense of ownership into the corporate culture.

198
Chapter 9

Physical Access and Environment Controls


The physical layer security of a services network seems easy to protect. Grant appropriate
physical access only to systems providing services. Often, physical access grants administrative
access via console ports and the like. Environmental controls, such as power and HVAC, should
also implement controlled access mechanisms. Drastic changes in the operating environment,
like an overheated server farm room, can result in damage and unexpected system behavior.
Physical access controls also mean controlling access to server farms, wiring closets, and other
services equipment locations. You build layers of defense in the network. It’s important to build
layers of defense in your physical security as well. You use monitoring systems to monitor
network performance and security. For many businesses, physical access monitoring may also be
appropriate.

There is a simple security review you can easily conduct that often yields some surprising results.
Simply walk around your enterprise imagining that you’re a burglar “casing the joint.” Pay close
attention to the server farm rooms and data center areas. Make sure you visit every room containing
critical systems—mainframes, servers, voicemail systems, telephony systems, routers, and switches.
Check out the door hinges. Are they on the inside or the outside? Often doors are built to swing
outward for specific fire or building code logistics. But a door that swings outward often leaves the
hinges exposed on the outside. How easy would it be for an intruder to gain entry by removing the
pins from the hinges on a locked door? Wiring closets housing equipment are often built with doors
that swing outward because they’re built in shallow recesses. They’re particularly vulnerable to this
problem. It’s worth the time and effort to conduct a complete physical inventory and identify wiring
closets that should have reinforced hinges and locks installed.

Perimeter: First Line of Defense Between the Internet and Internal Networks
The network perimeter used to be easy to identify. It was a single point of network ingress and
egress, typically protected by a firewall. It’s certainly true that the enterprise connection to the
Internet is one perimeter point. But every extranet connection to a business partner is a perimeter
point. Every VPN tunnel linking to another network is another perimeter point. De-
perimeterization of the network has become a large issue for security managers to address.
Not only do you have external perimeter points but there are numerous internal perimeters within
any large organization. Sales and marketing need to be cordoned off from research and
development groups. Human Resources operations require special access controls to protect
confidential personnel records. The network management resources need to be isolated from
everyone except authorized users.
It’s important to recognize that the goal of security perimeters isn’t to keep everyone out. If that
was the goal, you would simply disconnect the network and isolate it, eliminating all danger.
Security mechanisms are implemented to support business need and allow appropriate access
permissions to users who are trusted to do certain things in the network.
Enterprise networks today provide a layered defense system. Firewalls provide a hard outer shell,
but without access permissions to demilitarized zones (DMZs) in the network and other
resources as required, business would grind to a halt. The network services infrastructure and
network operations center represent vital organs that sustain the health and continuous operation
of the corporate organism.

199
Chapter 9

The external hard perimeter is most often a firewall or series of firewalls. You implement
firewalls to protect corporate assets from the malicious traffic on the Internet. They are vital to
network safety but they also must allow business to take place. Traffic must get in and out of the
network while you reduce the threats to an acceptable level. Some tactics used include:
• Using internal and external firewalls in combination
• Logging, analyzing, and reporting all access to the network
• Eliminating use of plaintext or unencrypted data that might expose internal systems to
external threats—FTP and Telnet are cleartext protocols that can expose vital business
information to anyone who might be capturing traffic; SIP is a widespread standard
protocol in VoIP services today that is also a cleartext protocol that could potentially
expose information about the enterprise network
• Encrypting incoming network traffic that is directed toward secure internal systems with
VPN technologies, Secure Sockets Layer (SSL for Web services), and Secure Shell
(SSH) instead of Telnet can provide an added measure of protection
• Proxy servers can provide traffic aggregation and funnel services through a single point
to limit exposure; reverse proxy technology can hide all the inner workings of the
corporate network
• Checking the identity of everyone accessing the network with strong user authentication
as described earlier will help ensure that only known users gain access to corporate
resources
• DMZs and filtered networks can be used to constrain external traffic to specified areas or
segments of the network
• Routine penetration testing and vulnerability assessments can help identify weaknesses in
the enterprise security posture

Network: Internal Network Layer


Layered defense can map conveniently to the OSI Reference Model in many ways. At the
network layer are tools you can bring to bear to deliver security. You can use network
monitoring tools as well as segmentation in the network design to help provide another layer of
security.
Intrusion detection systems (IDSs) can provide alerts and warnings to assist the security team in
quickly responding to incidents as they’re detected. The faster a problem or security incident can
be identified, the more quickly it can be mitigated. Quick problem resolution helps keep security
incidents from cascading into a larger chain of events. In short, early detection improves reaction
time.

200
Chapter 9

Intrusion prevention systems (IPSs) go a step further. An IPS may be configured so that it can
reconfigure other network elements in an automated response to some trigger event. In early
2004, the Bagle and Netsky worms spread widely and caused major network problems for many
companies. Today, years later, both Bagle and Netsky variants are still being distributed daily.
For most companies, some antivirus engine, either running on the mail gateway server or
individual workstations frequently mitigates this problem. But Bagle and Netsky are both easily
recognized by their digital signature. Intrusion detection devices routinely identify this malicious
traffic. These worms have been around for so long that they are easily recognized. An IPS might
not simply discard the packets at network ingress. The IPS might reconfigure border routers to
shun all traffic from the sender. This might be a rule that is implemented automatically with a
pre-set period of time. It might also be configured to make a permanent change to reduce risk
and eliminate the CPU processing and bandwidth consumption that follows from allowing these
malicious packets onto the enterprise network.
Traffic-shaping tools are used in large networks to provide trending information about network
traffic. For example, assume that your services network traffic is normally made up of 30% Web
traffic, 40% VoIP/video traffic, 20% email traffic, and 10% other mixed traffic. Traffic-shaping
tools can provide indicators of pattern changes in the traffic. Although patterns evolve and
change over time, making this useful planning information, a sudden or unexpected change in
traffic patterns is often indicative of a network security problem. A new worm propagating via
the corporate email system might cause email traffic to spike to 40% of all network traffic or
more. Monitoring traffic patterns can provide an early indicator of a problem of this sort.
Access control lists (ACLs) can help block traffic and allow only specific IP addresses to
communicate across network segments and subnets. Although they are basic and simple, ACLs
can help protect critical VoIP systems by only allowing access to authorized system
administrators from predefined network management workstations.
Internet Engineering Task Force (IETF) RFC1918 (http://rfc.net/rfc1918.html) defines non-
routable addresses. A well-developed network strategy for deploying RFC1918 addresses can
increase the services network security and further segregate traffic. In the VoIP environment, a
design consideration might include using only RFC1918 addresses for VoIP phones and
constraining voice service traffic to VLANs and MPLS domains dedicated to voice. Simply not
allowing voice traffic on data VLANs and subnets provides a degree of granularity that offers
some level of service protection by preventing intermingling of voice and data traffic.
Implementing a network-based approach to antivirus technologies may help eradicate viruses
and worms before they spread to affect other layers. One layered approach to antivirus solutions
might be to deploy one vendor’s solution in the network, at the email gateway, and another
vendor’s solution on the desktop workstations. This two-fold approach might also circumvent
any single vendor being slower than another to release updated virus signature files.
Conducting routine vulnerability assessment and testing of network services helps provide
insight from the attackers’ point of view. During normal operations, network personnel always
strive to follow best practices but the demands of business and heavy workloads can lead to
potential errors. In some cases, a solution is implemented quickly just to “get the job done.”
Regularly scheduled assessments can help ensure that any suboptimal techniques used to meet
business needs today don’t become the services network security vulnerabilities of tomorrow.

201
Chapter 9

Systems: Server and Client OS Hardening Practices


Chapter 8 looked at configuration management, but server and client OS hardening goes beyond
the network routers and switches to include all systems. You need to build processes for
standardized installation and configuration of OSs. Services that aren’t specifically required
should be disabled on servers and workstations. Programs that aren’t needed should be removed.
All configuration changes to enterprise systems should be traced through audit logs. Change
management procedures are necessary to ensure that only authorized changes are made to
production systems.
Establishing a standardized set of configurations, across all platforms, is a widely accepted
practice. Software programs that have been approved and accepted for use within the enterprise
should be installed in standardized configurations. Some divisions or workgroups within any
company may have specific software requirements, but only tested, approved, and authorized
software should be installed. In many organizations, users are not allowed administrative rights
to their local desktop machine. This approach can prevent the installation of unknown software
and help enforce the use of standard enterprise configurations.
Strong user authentication and password technologies need to be implemented. When a system is
prepared to move into network services production, special precautions should be taken to
eliminate all guest accounts and vendor default accounts and passwords. Keep in mind that
default password information is easily obtained over the Internet directly from vendor
documentation.

When setting up new systems in the pre-production environment, it is vital that all OS patches be
applied and the system scanned prior to connecting it to the production network. Systems built from a
6-month-old CD will be built from an image that isn’t at current OS patch levels. It’s a sure bet that
these older patch versions have vulnerabilities that are known. It’s highly likely that there are exploits
“in the wild” to take advantage of those weaknesses. These systems, once connected to the network,
can be infected in minutes; well before OS updates from the vendor can be applied. Several industry
studies have shown that an unpatched machine is typically compromised within 4 to 6 minutes of
being connected to the Internet. A rigorous process of building, patching, scanning, and then
revalidating patch levels will help ensure that new systems are not infected during the process of
being brought online in production.

202
Chapter 9

Application: Application Hardening Practices


The applications running on hosts often have control of how corporate data is handled. It’s
important to see that corporate applications don’t introduce security vulnerabilities. A poorly
written or malfunctioning application might allow for unauthorized access or undesirable
manipulation of data. Secure coding techniques and stringent quality control processes are vital
for any in-house application development.
You implement change control mechanisms in the IP network. You must take the same steps
with your business applications. If a vendor partner releases a software patch for a VoIP gateway
in the services network, diligent testing and documentation should be completed as the patch is
applied in a methodical fashion.
Just as network hardware has a life cycle, so do business applications. Part of the application life
cycle should include code auditing and peer review of the programming. Doing so will help
ensure the integrity of enterprise applications. Authentication, authorization, and audit
capabilities at the application level should be a consideration in all enterprise software, whether
purchased commercially or developed in-house.

Data: Protection of Customer and Private Information


The single most valuable asset within the services network is the data. This data provides
business intelligence information about the enterprise. It provides transaction records and history
with customers. It provided details about products, services, employees, and finances. Every
layer of a strong defense-in-depth strategy is implemented with one primary goal—to protect the
data.
One common approach in businesses is to follow traditional military strategies and classify data
in terms of the intended audience for distribution. When customer data is involved, privacy laws
and regulations to protect personally identifying information will often dictate the protective
measures your company must take.

Chapter 10 will delve more deeply into regulatory compliance issues.

Summarizing Defense in Depth


The basics of a defense-in-depth strategy needn’t be difficult to remember. Policies, procedures,
and awareness provide the foundation of a secure network. Physical security ensures that only
authorized personnel can touch systems. Perimeter security segments traffic to protect the
boundaries between untrusted, semi-trusted, and trusted areas of the network. Network security
adds a layer of intrusion detection and prevention solutions and provides traffic trend analysis to
identify changes in patterns. Systems security ensures standardized configurations that adhere to
company policies. Application security ensures best practices in coding and configuration of
applications are followed. Data protection safeguards vital business intelligence information.

203
Chapter 9

Prevention
Networks are vulnerable to a number of known problems. Perhaps the most feared, classic
problem is a denial of service (DoS) attack. When under a DoS, the network is reduced to a state
in which it can no longer carry legitimate users’ traffic. Attacking the routers and flooding the
network with extraneous traffic have been the historical approach malicious attackers have used
to accomplish this. As malware has increased, today a worm attempting to replicate across the
corporate network may present the greatest DoS problem as it consumes all available network
bandwidth. Another classic problem is IP address spoofing. To protect the network against
increasingly sophisticated attacks, you need to build stronger layers of defense. Just as layers of
clothing are the best protection from the cold, layers of defense eliminate single points of failure
and provide broad network security.

Digital Common Sense


In an article published in Computer World in 2002, Thornton May said “There can be little argument
that the digital world would be much improved if all senior executives were required to enroll in some
kind of information protection program. And we must hold each employee responsible for protecting
intellectual property. But this will be difficult because many executives lack a digital common sense.”
Although business executives understand that protecting intellectual property is a key protective
measure, perhaps their greatest responsibility is to lead by example and establish the climate of the
corporate culture within any company. What you take as common sense in the real world often
translates poorly to the digital networked world. It’s vital that the enterprise leadership team foster a
culture of awareness, stewardship for corporate information, and heightened digital common sense.
You must develop and hone your digital common sense so that you can implement smart, cost-
effective approaches that protect the safety of your corporate assets in the networked digital world.

In the integrated services network, each service you provide may have its own security
requirements. These requirements often vary based on the intended use of the service. Services
that are only intended for use inside the corporate network are likely to require very different
protection methods than services provided for external use. For example, if VoIP is used only for
internal calls, completely blocking all external access to the server may be adequate.
Servers that provide internal service to the organization shouldn’t normally be co-located on the
same server hardware that is used to provide external services. One common best practice is to
isolate external traffic to a set of outside-facing DMZ segments while using another inside-facing
set of segments for access from inside. These segments are often protected by implementing
firewalls to allow only authorized communications between the inside and outside.
Internal communications services still require design review considerations:
• Will every employee use VoIP services?
• Do you allow vendors, contractors, or visitors to connect to the internal trusted network?
• Will these vendors, contractors, or visitors ever be authorized VoIP users?
If not, special care must be taken to ensure that only authorized VoIP users can gain access to
servers, gateways, and other VoIP service delivery elements.

204
Chapter 9

Protect the Protection


Perhaps the most critical component in establishing a network security posture is protection of
the network services security and management platforms themselves. It’s a fairly common
network design practice to implement a dedicated management segment within the network.
Because this management segment has oversight of the whole network, systems on this segment
may require unfettered access to the entire network. The reverse is not true. This segment
represents in some ways, the “keys to the kingdom” of the enterprise services network.
Management servers should be accessible only to trusted employees in authorized work groups.
Only authorized staff should be permitted to view log data for analysis. Intrusion detection and
prevention systems shouldn’t be seen or touched by anyone outside the network security staff.
It is vital that network management and security protection services themselves be protected.
Firewall rules and ACLs can define authorized network locations by IP addresses that are
granted access. Strong user authentication techniques like those described earlier can ensure that
only authorized users have permission. Discovery protocols such as Simple Network
Management Protocol (SMNP) and Internet Control Message Protocol (ICMP) should be
disabled if they aren’t necessary.

Detection
Although prevention tools are valuable aids to securing the integrated services network, you
can’t prevent everything bad that might occur. Tools help with prevention, but ultimately your
survival depends on detection. When prevention fails, and it will fail, you must have strong
detection and notification mechanisms. You need to know as much as possible about every event
in the network. Comprehensive detection tools can shorten response time and lead to quicker
remediation. Reliable incident management techniques help shift the security posture from a
reactive mode to a proactive one. These measures make pre-emptive security changes part of the
corporate culture. Like every process, good incident management involves continuous review
and assessment.
It’s important to take measures that will allow detection of not only situations in which
information has been damaged, altered, or stolen, but also about how it was damaged, altered, or
stolen, and who has caused the damage. This can be accomplished through procedures and
implementing technologies discussed earlier that can detect intrusions, damage or alterations,
and viruses.

Reaction
Just as every facet of network service delivery has an accompanying life cycle, everything you
do in network security relies on layers of protection. This chapter has looked at some of the steps
you can take to prevent security breaches, but you know breaches will happen. You establish
procedures and implement tools to detect security breaches as they occur. You do so to enable a
quick reaction. By building comprehensive processes into your day-to-day operations, you can
be primarily either reactive or proactive. How an organization responds to an incident is driven
by how well prepared everyone is.

205
Chapter 9

Proactive and Reactive Strategies


A comprehensive security response plan should include both proactive and reactive strategies.
The proactive strategy is really a pre-attack strategy. It documents a series of steps taken to
reduce any existing security policy vulnerabilities. It also involves developing contingency plans.
Determining the damage that an attack can cause, along with the vulnerabilities exploited during
an attack, will aid in building a complete proactive strategy.
The reactive strategy is a post-attack strategy. It provides tools for assessing any damage caused
by the attack and identifies steps to repair the damage. A reactive strategy also engages any
contingency plans developed as part of the proactive strategy. During reaction, you restore
normal business operations, then document and learn from the experience.
Incident management is really a matter of addressing a few simple questions:
• What happened?
• Where did it originate?
• Who did the incident impact?
• What steps are needed to mitigate the problem?
• How can it be prevented in the future?
Reacting to a Security Incident
We spend a lot of time trying to prevent security incidents from occurring. What sometimes gets lost in all
of this preparation is plans for dealing with an incident should it actually occur. This is a very brief
synopsis of several incident-handling guides that provides a high-level framework for dealing with either
the realization that a system has been compromised or the recognition that a system is under active
attack. The SANS Reading Room has a large set of papers about incident response at
http://www.sans.org/reading_room/.
Remain Calm
To successfully handle any perceived emergency situation, you must remain calm so that you can assess
what is going on around you and react in a methodical manner. A compromised system/network or an
attacker on the loose demands well-thought out action; and frankly, the bad guys have probably been in
your computer for days or maybe even longer, and another few minutes won’t make much difference.
You’re probably going to have to rebuild your compromised servers anyway.
Notify Your Organization’s Management and Activate Your Response Plan to Get Help
Your security policies should identify the pecking order of who gets called when if there’s a security event.
Individuals with particular responsibility for the affected server(s) and/or network(s) should be notified as
well as any information security personnel. The severity of the incident (and your own policies) will dictate
who else is brought in—your Internet Service Provider (ISP), department head, corporate officers, the
press, law enforcement, consultants, response centers, and so on. Notify whoever is necessary to assess
the situation and get it under control, but it is generally best to maintain a "need to know" stance and
communicate, at least initially, with only the necessary parties.
Whenever possible, use telephones and faxes during a computer security incident. If the attackers have
full access to your computer, they can possibly (probably?) read your mail. If you use your computer, this
allows them to know when you report the incident and what response you got. There is a real possibility
that other systems at your site have also been compromised and one or more packet sniffers are running
on your network. Thus, if you absolutely must use a computer to communicate and you are fairly certain it
can’t be intercepted, use a different system and/or dial-up ISP access if possible.

206
Chapter 9

Take Good Notes


This cannot be stressed enough—document, document, document!! Maintain a log of everything you see
and do, everyone you speak with, and the team working with you. This will not only help you in criminal
cases (and in remembering the events at a trial that might take place a year or two down the road) but
also in the investigation/forensics process, post-event analysis, and as an educational/intelligence
gathering vehicle for others in the InfoSec community. Notes should be detailed, organized, and
complete, and should reflect the basic who, what, where, when, and how ("why" might be left for later on).
Keep copies of any altered files before restoring your system.
Contain the Problem
Take any necessary steps to keep the problem contained and prevent it from spreading to other systems
and/or networks. This may well involve disconnecting the compromised system from your network and/or
disconnecting your network from the Internet. Containment may require a physical disconnect or might be
accomplished while you clean up and recover; circumstances, including whether you are dealing with an
active attack or the aftermath of an attack, will dictate what is a prudent action. Note that the latter
approach (containing the problem while still online) might well leave you vulnerable to additional attacks.
Gather Evidence and Make Backups
For purposes of learning what happened and to have evidence for future analysis (and possible
prosecution), make backups of OS and file system information as well as any state and network
information (for example, output from netstat or route). Keep a detailed history of this activity if you have
even the least suspicion that this information will be used in a criminal or civil trial; digital signatures and
file timestamps are part of the procedures you should follow to maintain the custody chain. If possible,
coordinate your evidence gathering with that of a second source, such as an ISP or another network (if
you detect another network that is involved in this incident). Finally, as you make the backups, consider
where they are going and who will be using them; if possible, make multiple copies and secure one for
historical purposes while analyzing/sharing the other(s).
Get Rid of the Problem and Maintain your Business
This step might be easier said than done. If your server has been compromised, you should totally rebuild
it from scratch unless you are 100% certain what the entire problem is. Before you can eliminate the
problem, you need to be sure that you understand the cause of the incident. What vulnerability did the
intruder use to gain access and what have you done to prevent another attack? You should rebuild the
server and applications from original media.
The next issue, of course, is that of re-installing content. Note that some files that were exploited might
have been on your system for some time already and, therefore, might have been backed up as part of
your regular operations. So, although you’ve rebuilt the OS and applications software, you might very well
reinstall files that can be exploited over and over. Again, it is imperative that you understand how an
incident happened to avoid this eventuality.
Finally, business continuity is a major issue. Get rid of the problem and get your server back online as
soon as possible.
Perform a Post Mortem
Once the situation is resolved and you’re back in operation, get all relevant parties together to review the
incident and the response. Review your security policies and operational procedures to see what
changes, if any, are required. To the extent possible, contact appropriate incident response agencies,
such as US-CERT, and share your knowledge with them.
Hold this meeting a day or two after the incident is deemed "over," when everyone is rested and has had
time to reflect on what happened and why, and what went well and what didn’t go so well. Don’t do it
immediately while people are still tired and don’t wait weeks when people will forget and will have moved
onto other things.

207
Chapter 9

Reactive Strategies
A reactive strategy is implemented when the proactive strategy for the attack has failed. The
reactive strategy defines the steps that must be taken after or during an attack. It helps to identify
the damage that was caused and the vulnerabilities that were exploited in the attack, determine
why it took place, repair the damage that was caused by it, and implement a contingency plan if
one exists. Both the reactive and proactive strategies work together to develop security policies
and controls to minimize attacks and the damage caused during them.
The enterprise incident response team should be included in the steps taken before, during, and
after any attack to help assess the incident, document what happened, and to learn from the
event.
Reactive strategies include responding to calls for help and reporting of incidents, threats, or
attacks. These steps might be manually driven processes or triggered through technology in IDS
and firewall monitoring systems.
From an operational perspective, incident handling involves taking reports, dealing with
information requests, assisting with triage, and analyzing incidents. The incident response team
can coordinate triage efforts between different divisions or work groups within the company and
help share appropriate mitigation strategies. This team can help monitor for malicious activity in
other parts of the network. They may be directly involved in rebuilding systems, applying
patches, and developing any workarounds. The incident response team provides communications
support, notifying work groups across the enterprise about vulnerabilities and sharing
information on remediation tactics.
Vulnerability response might also involve researching to find the necessary patches, hotfixes, or
workarounds. Response also involves notifying other workgroups of the mitigation strategy
through advisories or alerts. In some enterprises, this service will include patch management
procedures to install appropriate patches, fixes, or workarounds.

Proactive Strategies
Proactive strategies are designed to avoid incidents in the first place. They also help control the
scope of impact when incidents inevitably do occur. The technologies used by the incident
response team to handle incidents are often used to watch new technical developments, intruder
activities, and trends that might indicate future threats. Members of the incident response team
need to be afforded the time to read security mailing lists, security Web sites, and current news
articles to mine relevant information from the broader security community. This intelligence-
gathering function can be combined with past lessons learned to provide a powerful defensive
resource to the enterprise.

208
Chapter 9

Ongoing, regularly scheduled security assessments can provide detailed review and analysis of
an organization’s security infrastructure. These assessments might be comprehensive system
audits or simply desktop reviews of security practices. There are several types that can be
performed:
• Reviews of hardware and software configurations, routers, firewalls, switches, servers,
and workstations. These reviews can confirm that systems all meet enterprise standards
or follow industry best practices.
• Interviews with employees can provide insight into how well actual security practices
match the enterprise security policy.
• Vulnerability or virus scans can be performed against network segments to identify
vulnerable systems and networks.
• Penetration tests can be conducted in a controlled, methodical environment, including
social engineering, physical, and network attacks. This information can be leveraged to
provide a stronger security posture before a malicious outsider takes advantage of any
weaknesses that are discovered.
Security tools vary in scope and functionality. Because of their focused experience and
understanding of security issues, the incident response team may provide valuable guidance on
how to harden systems and develop a set of approved security tools.
Incident Management In a Nutshell
Assess the Damage
Determine the damage that was caused during the attack. This should be done as swiftly as possible so
that restore operations can begin. If it is not possible to assess the damage in a timely manner, a
contingency plan should be implemented so that normal business operations and productivity can
continue.
Determine the Cause of the Damage
To determine the cause of the damage, it is necessary to understand what resources the attack was
aimed at and what vulnerabilities were exploited to gain access or disrupt services. Review system logs,
audit logs, and audit trails. These reviews often help in discovering where the attack originated in the
system and what other resources were affected.
Repair the Damage
It is very important that the damage be repaired as quickly as possible in order to restore normal business
operations and any data lost during the attack. The organization’s disaster recovery plans and procedures
should cover the restore strategy. The incident response team should also be available to handle the
restore and recovery process and to provide guidance on the recovery process.
Document and Learn
It is important that once the attack has taken place, it is documented. Documentation should cover all
aspects of the attack that are known, including the damage that is caused (hardware, software, data loss,
loss in productivity), the vulnerabilities and weaknesses that were exploited during the attack, the amount
of production time lost, and the procedures taken to repair the damage. Documentation will help to modify
proactive strategies for preventing future attacks or minimizing damages.

209
Chapter 9

Testing the Strategy


The last element of a security strategy, testing and reviewing the test outcomes, is carried out
after the reactive and proactive strategies have been put into place. Performing simulation attacks
on a test or lab system makes it possible to assess where the various vulnerabilities exist and
adjust security policies and controls accordingly.
These tests should not be performed on a live production system because the outcome could be
disastrous. Yet the absence of labs and test computers due to budget restrictions might preclude
simulating attacks. To secure the necessary funds for testing, it is important to make management
aware of the risks and consequences of an attack as well as the security measures that can be
taken to protect the system, including testing procedures. If possible, all attack scenarios should
be physically tested and documented to determine the best possible security policies and controls
to be implemented.
To be effective, both proactive and reactive strategies need to be tested or exercised. Regular
simulation exercise can help maintain a vigilant security posture.
It’s altogether too easy to work through an incident that only has minor repercussions, and gloss
over the lessons learned. Remember that lessons learned are only learned when they feed into
some process improvement loop and a change is actually put into practice. Don’t let your
organization get trapped into collecting a set of lessons observed but never acted on. Learn the
lessons well and incorporate them into the every day work flow to maintain a vigilant security
posture.

Summary
In the converged services network, security goals will be largely determined by the following
key tradeoffs:
• Services Offered vs. Security Provided—Each service offered to users carries its own
security risks. For some services, the risk outweighs the benefit of the service and the
administrator may choose to eliminate the service rather than try to secure it.
• Ease of Use vs. Security—The easiest system to use would allow access to any user and
require no passwords; that is, there would be no security. Requiring passwords makes the
system a little less convenient, but more secure. Requiring a device-generated one-time
password makes the system even more difficult to use, but much more secure.
• Cost of Security vs. Risk of Loss—There are many costs to security: monetary (the cost
of purchasing security hardware and software such as firewalls and one-time password
generators), performance (encryption and decryption take time), and ease of use. There
are also many levels of risk: loss of privacy (the reading of information by unauthorized
individuals), loss of data (the corruption or erasure of information), and the loss of
service (the filling of data storage space, usage of computational resources, and denial of
network access). Each type of cost must be weighed against each type of loss.
The threats to the security of your service network environment have reached staggering
proportions from both within and without. This chapter has identified a number of widely
accepted industry best practices to help secure and closely monitor your environment. The next
and final chapter will delve further into compliance, asset management and reporting, and
documenting the whole environment.

210
Chapter 10

Chapter 10: Asset Reporting, Audit Compliance, and IT


Documentation
This final chapter will explore asset and audit compliance methods and procedures. They’re the
last component of the FCAPS model. As this guide comes to a close, it will examine why
documenting IT process is critical to business operations. Earlier chapters have made brief
mention of approaches such as ITIL and ISO 17799 as best practice methodologies. This chapter
will probe into those a bit deeper.
For many enterprises, regulatory requirements to comply with SOX, GLBA, and HIPAA raise
concerns about the impact on managing an integrated service network with VoIP. This chapter
will close with a review of managing the network life cycle with an eye toward holistic risk
management.

FCAPS and Asset/Administration/Accounting Management


Over time, and depending on what business the enterprise is engaged in, the “A” in FCAPS has
stood for accounting, administration, auditing, and asset management. That variation is perhaps a
good reinforcement that the FCAPS approach is a model, not a process. FCAPS is simply a
holistic model that suggests you look at multiple domains within the service network—fault,
configuration, accounting, performance, and security—to provide comprehensive service
management.

Accounting/Administration/Asset Management
In the commercial services market, whether delivering traditional telephony services or enhanced
VoIP services, accounting becomes vital to successful service delivery. A Call Detail Record
(CDR) in telecommunications contains information about system usage. This includes the
identities of call originators, the identities of call recipients, call duration, any billing information
about the call, information about time used during the billing period, and other usage-related
information.

211
Chapter 10

Managing Billing
CDRs are created constantly throughout the business day, with every telephone call. The
telephone calling patterns within an enterprise can provide a great deal of information about
customers, business partners, and internal relationships. In many VoIP systems, CDRs are stored
in databases on the VoIP servers. A SQL server in the VoIP segment may be just as vulnerable
as one sitting elsewhere in the network.

Internal Billing
For many organizations, IT and telecommunications services represent a business within the
business. For enterprises that treat these services as a cost center, billing within the organization
may drive either real dollar transfers or cost center cross charges for tracking the cost of doing
business. CDRs and call accounting systems provide the means to generate and control this
recordkeeping function within the service delivery network.

Billing as a Service Provider


Advances in unified communications technologies have dramatically lowered the barrier to entry
to the VoIP services business. It’s become relatively easy and cost effective to become a voice
service provider in the converged network environment. As a result, hundreds of service
providers offer a wide range of integrated services, many focused in niche and vertical markets.
For many global enterprises, the deployment of self-managed infrastructure makes sense. In
these enterprises, affiliates and business units are often paying customers of the services delivery
organization.
Whether CDRs are used to bill paying customers or to track system utilization doesn’t matter.
The data collection and analysis of usage statistics provides the same baseline knowledge about
what’s going on in the network. In traditional telephony, carriers billed for minutes of usage. In
the integrated services environment, there are several new factors that need to be accounted for.
Bandwidth utilization is the primary accounting factor for many enterprises. Although it might
not provide the granularity of detail that some facets of the network offer, bandwidth can be used
to account for WAN links, based on the speed of the link. In large enterprise networks, it helps
delineate the cost differences in delivering 10Mbps, 100Mpbs, or Gigabit Ethernet connections
within the LAN. These port speeds become more critical factors as the enterprise evolves from a
legacy LAN environment into an integrated voice, data, and video service network.
Disk space and CPU utilization become increasingly important in the converged service
environment. Adding a voice service introduces codecs, signaling, and control processing that
consumes resources, not just in the network, but sometimes in nodes across the network.
Minutes of use may seem like an obsolete billing mechanism in the new integrated services
network, but that isn’t always the case. Many enterprises rely on telecommunications services to
derive information about peak business hours during the day or week. For businesses that are
very voice-centric in nature, minutes of use, and when they are used, continue to provide key
business intelligence information about the ebb and flow of work throughout the business day.
Collected over time, this information is used by dynamic organizations to assist in managing
staffing requirements.

212
Chapter 10

User Accounting
As we touch on a variety of regulatory and compliance issues, user accounting becomes a vital
recordkeeping and auditing component of the service network. With heightened requirements for
audit trails and accountability, the integration of VoIP and video into the enterprise operational
network raises the need for effective user accounting mechanisms.
User administration tools have widely permeated the IT and telecom services environment.
They’re common enough that today, reference to an “AAA server” or service may not raise any
questions, but for many organizations, the adoption of new services for authentication,
authorization, and accounting/auditing may be new.

Chapter 9 discusses the importance of authentication, authorization, and accounting/auditing.

Tracking UserIDs, passwords, and user permissions becomes more important as service
networks mature. As a more diverse suite of services is available, it’s important to track which
users have permission to use which services. From a regulatory and compliance viewpoint, you
might focus more on who the services were used by and what activity was performed.
When crafting the enterprise converged services network, one factor to consider is a backup
methodology. Recovery of auditing or accounting information impacts compliance for many
organizations.

The Asset Management View


Another facet of the FCAPS approach is the challenge of asset management. Current unified
communications technologies introduce a plethora of new devices and equipment into the
network. Chapter 8 discussed configuration management. In the integrated services network,
management and monitoring tools are vital service delivery support mechanisms.
As part of configuration management, the NMS commonly incorporates an asset management
module. These tools in the NMS help in managing the IT asset life cycle. As this guide has noted
throughout, every facet of network and services management has its own life cycle component.
The integrated services network is a series of life cycles within life cycles. In this area, we focus
on:
• Asset evaluation and selection
• Asset procurement
• Asset maintenance and support
• Asset disposal at end of life
As you adopt a comprehensive FCAPS model, many organizations are inclined to take a singular
view on asset management, auditing, or accounting for billing purposes. It’s wise to consider a
more holistic view and incorporate all three into the enterprise management model.

213
Chapter 10

Why Documentation? Why Process?


Throughout this guide, there has been a recurring thread reinforcing the need for process and
documentation. As we wrap up in this closing chapter, let’s touch on why these are so important
in the integrated service network of voice, data, and video.

Creating Repeatable Processes


When we deliver services of any kind, we strive for a predictable and consistent service quality;
VoIP and video services in the converged network present some technical challenges. Traditional
networks dealt with bursty non-real-time data. Voice and video services place new demands for
real-time services on the network. Real-time services can be designed to overcome a number of
network impediments, but doing so requires predictability.
We put great effort into repeatable processes in every facet of the service delivery network
because repetition brings consistency. With consistency, we can achieve the quality we commit
to, whether those commitments are via service level agreements (SLAs) or through tacit user
expectations in the enterprise.
Additionally, repeatable processes can free time and resources in many organizations; those
resources can be better spent on more creative work that requires focused attention. Not every
project or task in the enterprise requires repeatable processes, but those things that routinely
consume resources are more efficiently managed with thorough documentation and process
controls.

Revisiting the Importance of Knowing Your Environment


In the information economy, our business intelligence is a corporate asset. For many enterprises,
the information we have about our own business processes, workflows, and information resource
tools may be the most valuable asset the company owns.
The more we know about the environment, the better armed we are to manage the business. The
more we know about our integrated services, how they’re used, when they’re used, and who uses
them, the better prepared we are to support the enterprise mission.
Documenting processes provides sustainability. Too often, especially in small and midsized
businesses, there is a corporate culture of reliance on key individuals who know what to do.
These people who carry corporate information in their heads are perceived as being high-value
resources, when in fact they present great risk. The problem is depth of coverage. When a single
individual is the only resource with expertise in a particular subject area, there is a risk that an
event changing the status of a single person can create problems. If a single IT person handles
cyber security incidents, for example, what happens if that person is out ill or leaves the
company. It’s vital that business processes not rely on “one deep” personnel. Documentation
eases cross-training of other staff, and helps ensure long-term viability of business processes.
Thorough documentation of business intelligence information—whether it’s about the integrated
services network or a customer account—protects the organization from the danger that critical
institutional knowledge exists solely in the hands, or brains, of a single individual.

214
Chapter 10

Be Prepared: Protecting Against Future Business Issues


There is another reason to document IT service delivery processes and workflows. It has to do
with budgeting. Information services networks tend to change more rapidly than legacy
telephone networks. In the past, an enterprise telephone system might have a 10-year lifetime.
The LAN and WAN environment has proven to be more dynamic. In many enterprises, the IT
services network undergoes extensive redesign and re-architecture every 3-to-4 years, with
continual upgrades taking place along the way.
As our integrated voice, data, and video services become more widespread, they become more
tightly coupled with daily business operations. Enhancements and upgrades to the service
network become, for many, key business drivers.
The budgeting process of calculating costs, building a business case, proving return on
investment (ROI), and preparing acquisitions all require documentation. In order to be prepared
for business contingencies, in the IT services network, one commonly adopted best practice is to
maintain and regularly update the requirements, including replacement costs for high-level
elements of the service:
• Hardware—A comprehensive inventory of the network infrastructure can ensure that no
components are overlooked in the life cycle evolution.
• Software—A library of software used, including version information, not only documents
the current environment but also provides a quick recovery checklist when problems
arise.
• Services—An inventory of every service delivered and the user or customer to which it is
delivered can prove vital during potential business continuity events. Knowing who the
users of every service are can ensure good communications and information sharing
during an incident.
• Staffing—Training, cross-training, and skills retention requirements need to be
documented and continually updated.

215
Chapter 10

Legal and Regulatory Issues to Consider


Although this guide can’t begin to address the spectrum of legal and regulatory issues to which
private and public sector organizations may need to adhere, let’s take a brief look at a few of the
most visible and compelling concerns.

SOX
In 2002, SOX was passed. Entitled the Public Company Accounting Reform and Investor
Protection Act of 2002, it’s been very controversial and raised a lot of debate. This law arose
from the furor over corporate accounting scandals at WorldCom, Enron, Tyco, and others. This
flurry of visible accounting scandals and fraud in business reduced public trust in business
management and in accounting and reporting practices. Named after sponsors Senator Paul
Sarbanes and Representative Michael G. Oxley, SOX passed in the House by a vote of 423-3 and
by the Senate 99-0.
This wide-ranging legislation introduced new standards and bolstered existing ones for the
boards, management, and accounting firms that represent public companies. Although SOX
doesn’t correlate directly to telecommunications or IT services, its impact has been far-reaching,
well beyond publicly owned companies.
Major SOX provisions include:
• A Public Company Accounting Oversight Board (PCAOB) was created
• Public companies are required to both evaluate and disclose the effectiveness of their
internal controls
• Financial reports must be certified reports by both Chief Executive Officers (CEOs) and
Chief Financial Officers (CFOs)
• Companies listed on stock exchanges must have independent committees overseeing
relationships with auditors
• Most personal loans to executive officers and directors have been banned
• Both criminal and civil penalties for securities law violations have been increased
• Willfully misstating financial reporting by corporate executives is punishable by
significantly longer maximum jail sentences and larger fines
Although SOX might not apply to all organizations because of its focus on financial
accountability, the principles contained in the act have been widely embraced and adopted across
the private and public sector. What is for some a legal requirement has quickly become a
generally accepted best practice for many. The fundamental SOX focal areas easily extend down
into IT management and delivery of services within the enterprise.
The impact of SOX on business practices surrounding financial and auditing controls has been
widely felt. But the impact isn’t solely in the financial area of business. IT has felt some impact
as well. Those impacts are mostly felt in five areas: risk assessment, control environment, control
activities, monitoring, and information and communication.

216
Chapter 10

Risk Assessment
Every organization is faced with a variety of internal and external risks that must be taken into
consideration. Establishing objectives is a precondition to risk assessment because we are
actually evaluating the risk posed to achieving business objectives. Before we can determine how
to manage risks, we have to understand what they are in relation to specific business goals.
For example, IT managers have to both assess and understand the risks around financial reports,
particularly in the areas of completeness and validity, before they can implement necessary
control mechanisms. Thus, service delivery managers must understand how services, systems,
and resources are being used, and document accordingly.

Control Environment
The control environment sets the tone for the organization and influences human behavior. This
control environment addresses business factors such as management style, organizational ethics,
and delegation of authority and staff development processes.
IT and service delivery groups thrive in an environment in which employees take accountability
for and receive recognition for the success of their projects. This encourages that issues and
concerns are escalated in a timely manner and dealt with effectively. For many organizations,
this means that service delivery employees need to be cross-trained in design, implementation,
quality assurance, and deployment so that they can engage in the complete life cycle of the
converged services technologies.

Control Activities
These are the corporate policies and procedures that govern how to ensure management
directives are accomplished. They help guarantee that appropriate actions are being taken to
mitigate the risk that might interfere with meeting objectives. Control activities occur throughout
every organization, and include routine tasks at all levels and in all functions. They include a
broad range of activities—approval processes, authorization acquisition, verification, financial
reconciliations, operating performance reviews, and asset security.
In the past, many ERP and CRM systems were used to collect data that fed into spreadsheets for
analysis. Manual spreadsheets are prone to human error. Organizations will need to document
usage rules for the integrated services and create an audit trail for each. These systems support
business services and often contribute financial or other key business intelligence information.
It’s important that corporate policies define the business requirements and other documentation
necessary for each and every IT and telecommunications service project.

217
Chapter 10

Monitoring
Internal control systems within the service environment need to be routinely monitored. This
helps assess the quality of the system’s performance over time. This is best accomplished
through a combination of both ongoing monitoring activities and periodic evaluations. Problems
and anomalies identified during monitoring should be reported and documented and corrective
action should be taken. This ensures a continuous improvement in the delivery of fully integrated
services.
Within the service delivery environment, specialized auditing and review processes may be
appropriate to address high-risk areas. Those services and systems that are mission-critical to the
enterprise often require focused attention. The service delivery team should perform frequent
internal assessments of these vital systems and services. One important consideration may be to
have independent third-party assessments performed on a regularly scheduled basis to augment
the work performed by employees. The management team needs to clearly understand the
outcomes of these audits and assessments. As these services roll up to larger corporate business
and financial reporting, the management team will be held accountable for the outcomes.

Information and Communication


Information systems play a pivotal role in all our internal business controls. They provide reports
that make it possible to run and manage the business. These reports contain operational,
financial, and compliance-related information. Effective communication aids the flow of
information throughout the organization.
Service managers can’t identify and address risks without timely, accurate information.
Timeliness in reacting to issues as they occur is vital to maintaining the integrated services
environment and ongoing business operations.

GLBA
GLBA was designed to open competition among banks, securities companies, and insurance
companies. These companies make up what is known as the financial services industry. GLBA
really governs how these companies can merge and own one another. Because people invest
money differently in differing economic times, GLBA introduces controls so that investment
banking firms, for example, might also participate in the insurance business. Banks haven’t
traditionally participated in insurance underwriting, but consolidation of the financial services
sector led to the need for new regulation.
Although a great deal of consolidation in this industry has taken place since the passage of
GLBA, it hasn’t been as widespread as anticipated. For most businesses outside the financial
services industry, GLBA is of no concern. It represents another set of requirements, similar in
many ways to SOX, for financial services. Again, we see regulatory efforts and compliance
requirements for one industry or set of industries being adopted as common best practices in
other sectors.

218
Chapter 10

HIPAA
Congress enacted HIPAA in 1996. The primary goal of HIPAA Title I is to protect health
insurance coverage for workers and their families in the event of change or loss of their jobs.
Title II of HIPAA had wide-ranging impacts on the health care industry. It set in place
requirements for national standards of health care transactions and identification mechanisms for
health care providers, insurance plans and providers, and employers.
As part of Title II, a section called Administrative Simplification also addressed the security and
privacy of individuals’ health-related information. The intent of this section was to bring about
standardization in systems and encourage electronic data interchange (EDI) within the U.S.
health care sector.
Within HIPAA, there is a Privacy Rule that took effect in 2003. This rule provides regulations
for what is termed Protected Health Information. PHI is comprised of information that can be
linked to any individual about the status of that person’s health, health care being provided, or
payment for that health care. The rule has been broadly interpreted to include any portion of
either a patient’s complete medical record or payment history.
The Privacy Rule dictates what may or may not be disclosed by law and to help ensure treatment.
Health care organizations have to take great care to disclose only what is necessary and take
measures to protect all information from undue disclosure.
Beyond the Privacy Rule, there is also a Final Rule on Security Standards that was issued in
2003, with a compliance requirement date in 2005. This Security Rule is intended to complement
the Privacy Rule. It defines three types of safeguards that healthcare entities must put in to place
to comply:
• Administrative
• Physical
• Technical
Each of these types has specific security standards that must be adhered to for compliance. Some
of the standards are required; others are “addressable” or flexible enough to be uniquely
addressed based on circumstances.
Although the intent of this guide isn’t to provide a detailed evaluation of HIPAA requirements
and impacts, because the act is so broad in scope—even to the point of impacting many
enterprise Human Resources organizations—this information is included to give an appreciation
of these regulatory compliance needs as they may overlap into service delivery requirements.
These descriptions of administrative, physical, and technical safeguards are liberally extracted
from a number of resources widely available online. HIPAA has been very widely documented
and many resources say the same things.

219
Chapter 10

Administrative Safeguards:
• Organizations covered under HIPAA must adopt written privacy procedures and
designate a privacy officer. The privacy officer is responsible for developing and
implementing all required policies and procedures.
• Management oversight and organization buy-in to compliance and security controls must
be documented within the organizational procedures and policies.
• Classes of employees that are permitted access to PHI must be documented in
organizational procedures. Employees should have access only to the subset of PHI
needed to perform their jobs.
• Organizations must demonstrate they have implemented an ongoing training program
regarding the handling of PHI, and that it’s provided to all employees that perform any
administrative health plan functions.
• If an entity outsources some part of their business process to a third party, that entity
must document that their vendors also have the procedures and policies in place to
comply with HIPAA requirements. In most cases, organizations accomplish this through
contractual clauses as part of third-party service agreements. One area of concern is the
possibility of a vendor further outsourcing data handling functions to yet another vendor
farther downstream. At each step of the way, appropriate contracts and controls need to
be established.
• Health care entities are responsible for data backups and disaster recovery procedures. A
documented emergency contingency plan should be established that documents data
priority and failure analysis, testing activities, and change control procedures.
• HIPAA dictates an internal audit process that continually reviews compliance and the
potential for security violations. Enterprise policies and procedures need to document the
frequency and scope of audits, and the specific details of the audit procedures. Under
HIPAA, audits should be both regularly scheduled and event driven.
• Procedures for mitigating security breaches should be well documented.
Physical Safeguards:
• The network must include controls governing the introduction and removal of both
hardware and software. Equipment that has reached end of life must be retired in such a
way that PHI cannot be compromised.
• Access to systems that contain health information should be diligently monitored and
controlled.
• Only authorized individuals may access systems hardware and software.
• Access controls must include security plans for facilities, maintenance records, and
visitor sign-in (including a process for escorts).
• Workstations should not be placed in public or high-traffic areas. Monitor screens should
not be viewable by the public. Policies addressing proper workstation use are required.
• If contracted employees are used, they too must be fully trained on their responsibilities.

220
Chapter 10

Technical Safeguards:
• Systems that house PHI must be protected from intrusion. If data is transmitted over an
open network, some form of encryption is required. If private or closed networks are
used, existing access controls may be sufficient to eliminate the need for encryption.
• Every entity involved in the transmission and handling of PHI data is responsible for
ensuring that the data within its systems has not been modified or deleted.
• Data integrity should be maintained and corroborated using check sum, double-keying,
message authentication, and digital signature techniques.
• Message authentication between parties covered under HIPAA is required. Each party
should corroborate that an entity is who it claims to be. Common examples of this
authentication include: password, two or three-way handshakes, telephone callback, and
token systems (including digital certificates).
• Organizations covered under HIPAA must make their documentation available to the
government to determine and verify compliance.
• IT documentation should include a written record of all configuration settings on the
elements of the service network. This is in addition to other policies and procedures
because these components are complex, configurable, and constantly changing. For many
service delivery organizations, this presents a huge challenge.
• Risk analysis and risk management programs must be implemented and documented.
Whether the regulatory issue is SOX, GLBA, or HIPAA, it’s obvious by now that any enterprise
that is required to comply takes on an added set of documentation requirements that necessitates
diligence and attention to detail. Some regulations apply only to specific areas in the private
sector. Public sector groups may have other requirements at the federal, state, and local level.
Although this guide can’t begin to touch on the privacy issues surrounding the protection of
personally identifying information (PII), as of this writing 46 states have active PII legislation in
place.
For many enterprises, another entity’s regulatory challenge can be used as a diligent best practice
with some adaptation. Many best-in-breed organizations adopt a blend of compliance with the
regulations we’ve described coupled with the practices that follow in the remainder of the
chapter.

221
Chapter 10

Methodologies in Best Practices for Management and Oversight


This section will delve into two widely accepted best practices methodologies: the Information
Technology Information Library (ITIL) and ISO 17799.

ITIL
ITIL is a collection of industry best practices that has grown out of earlier efforts by the Central
Computer and Telecommunications Agency (CCTA) in the UK. It provides a collection of
industry best practices, taken from both private and public sector organizations, for information
technology services across all IT infrastructure and operations.
ITIL is published as a series of books, hence the library approach. At one point, there were 31
volumes in incorporated in ITIL. ITIL v3, termed a refresh of the library, was due to be released
in May of 2007 but has been slightly delayed as of this writing. ITIL v3 will contain five
volumes:
• Service Strategy
• Service Design
• Service Transition
• Service Operation
• Continual Service Improvement
The framework in ITIL has been widely adopted by many enterprise organizations. ITIL contains
an area that addresses IT Service Management (ITSM), which garners a great deal of attention
from service delivery organizations. Although ITSM and ITIL aren’t directly related, they’re
used in tandem to provide a customer-focused set of service delivery best practices. Their shared
heritage dates back to the IBM Yellow books and W. Edwards Demings’ Total Quality
Management (TQM) principles. Today, these principles are found in a number of widely
embraced approaches including Six Sigma, Capability Maturity Model Integration (CMMI), and
Business Process Management (BPM).
It’s perhaps noteworthy to remember that early work in quality management and process
engineering was oriented toward manufacturing. What we see here is the evolution of how the
lessons learned in the manufacturing sector evolved over time and are now being applied to the
delivery of information services in today’s converged data, voice, and video networks.
Some of the benefits of a systematic, ITIL-based approach to managing IT services include:
• Cost reduction
• IT service improvements based on best practices
• Increased customer and end user satisfaction
• Enhanced auditability through process and documentation
• Productivity improvements through incorporation of repeatable processes

222
Chapter 10

Service Support
ITSM can bring many benefits through the adoption of best practices. It’s driven by both
technology and the many businesses engaged in ITSM, so it’s continually evolving and
changing.
It’s a fact that many service provider organizations have turned to ITIL for guidance to help
deliver and support their services. ITIL provides useful guidelines, yet it does not actually
provide the service management processes. Service providers are still expected to define the
processes themselves. This is often accomplished with help from outside consultants.
The service management processes alone are not enough. Staff who are expected to follow the
processes also require more detailed work instructions behind the processes.
ITIL provide some guidelines for service management applications, yet does not provide the tool
settings. Thus, after a service provider organization has defined its service management
processes, the organization still needs to find or develop the appropriate application to support
these processes.
Today, nearly all service management applications claim to support ITIL and work out-of-the-
box. Reality requires more work. It’s not unusual for it to take somewhere between 2 weeks and
4 months to configure a service management application so that it can support the processes that
a service provider organization has defined. And that’s the easy part, after all the processes have
been defined and documented.
There are several views of how ITIL might be used. One leading company is Alignability. Figure
10.1 shows a widely adopted view called the Alignability Process Model.

Figure 10.1: The Alignability view of ITIL service support and delivery processes.

223
Chapter 10

Service Delivery
Given that ITIL is a library, covered in a series of books developed over many years, this short
mention cannot begin to do justice to the complexity of this very thorough approach. In the
visuals that follow, we see some flows and concepts specific only to incident management that
are widely adopted by service delivery organizations embracing the ITIL model. Each box in
Figure 10.1 has a detailed and specific set of documentation and process flows to support that
facet of operations.
Incident management is a daily part of operations for every service delivery organization. In
Figure 10.2, we see one ITIL view of incident management and response. This is provided as an
example and may not be applicable as shown for every service delivery organization. The
example provides, as does the entire ITIL collection, a framework that an organization can use
for modeling its own specific process documentation.

Figure 10.2: The ITIL incident management and response flow.

In Figure 10.3, we see one tool for a small part of the incident management and response flow.
Every incident has an impact, and associated with that impact is some sense of urgency to
achieve resolution. For most organizations, this simple three-by-three matrix will provide enough
granularity to prioritize every possible incident and contingency from a flu pandemic to natural
disaster to minor power failures or service disruptions.

224
Chapter 10

IMPACT
Extreme (Critical) High Medium Low
Cannot Restricts Does not
Major Incident significantly
conduct core ability to
And, Multiple agencies business conduct impede
cannot conduct core business
business
business
High
Requires
immediate
attention
Medium
Requires
attention in
near future
Low
Does not
require
significant
urgency

Figure 10.3: Incident prioritization matrix.

Figure 10.4 simply provides an example of drilling down another level. Once priorities have
been applied to a given incident, established responses need to be engaged.

Each priority is related to a certain recovery time . . .


Priority 0: Major Incidents - Larger global incidents Immediate Attention
Multiple agencies cannot conduct core business. Other global incidents may include: 24/7 effort until resolved /contained
Fire, natural disasters, such as floods, earthquakes or volcanic eruptions. Human and n hour escalation / communication
animal disease outbreaks. Hazardous materials incidents. Terrorist incidents, including the
use of weapons of mass destruction. Civil unrest, labor strikes, picketing.

Priority 1: Significant damage, must be resolved / Immediate Attention


recovered immediately 24/7 effort until resolved /contained
n hour escalation / communication

Priority 2: Limited damage, should be resolved /


N hours
recovered immediately
24/7 effort at managerial discretion

Priority 3: Significant damage, does not need


to be resolved / recovered immediately N hours / business day

Priority 4: Limited damage, does not need to be


resolved / recovered immediately N business day(s)

Priority 5: Non-urgent – no impact on the


customer’s ability to work. Could become N/A
a project oriented issue

Figure 10.4: Prioritization response matrix.

225
Chapter 10

The Business Perspective


What we see in this quick series of examples is that the ITIL approach to incident management
provides a consistent methodology for prioritizing and responding to incidents. Coupled with
established process flows, the integrated services delivery organization possesses a complete,
methodical tool set for managing and delivering consistent and predictable integrated network
services.
This consistency and set of repeatable processes ensures that enterprise operations continue on
an even keel no matter what happens. They provide sustainability. For service providers, these
tools ensure business survival, as these integrated services represent the customer revenue
stream. For other enterprises, appropriate modification allows them to support the enterprise
mission and sustain the work that goes into meeting corporate objectives.

ISO 17799
ISO 17799 is actually entitled Information technology - Security techniques - Code of practice
for information security management. It was revised in June 2005 by the International
Organization for Standardization (ISO) and the International Electrotechnical Commission
(IEC). It offers best practice recommendations on information security management. Many
organizations are today embracing ISO 17999 as a broader part of service management and
business operations, folding it into the larger set of corporate compliance efforts. Although
focused primarily on the protection of confidentiality, integrity, and availability of network
services and resources, ISO 17799 reaches far beyond security.

Business Continuity Planning


Business Continuity Planning (BCP) has grown and evolved from what in the past were disaster
recovery initiatives in the enterprise. Like every other facet of managing a unified
communications services network, BCP strategies continually mature as processes are
developed. BCP focuses on organization recovery and service restoration when a disaster or
extended disruption occurs. For many organizations, BCP digs deeper than network services into
what is frequently called an “all hazards plan.”
BCP is simply how an enterprise plans and prepares for future events that could jeopardize
business mission and overall operational health. It’s often a path to discovery of weak processes
and tightly coupled with efforts toward improving information security and risk management
practices. Typically, BCP yields some organizational documentation for use before, during, and
after a business disruption.
BCP methodologies today scale to fit both large and small businesses but initially grew primarily
from the efforts of regulated industries. Common sense indicates that every organization should
have a plan to guarantee sustainable business.

226
Chapter 10

One view that is widely represented in a number of scholarly articles on business continuity
management process consists of the following six procedures.
• Disaster Notification Handling—At this stage, an on-duty manager becomes aware of
either an existing or impending disaster. This procedure is used to assess the situation
using established checklists and determine what protective and recovery actions are
warranted.
• Service Recovery—In the recovery phase, management and recovery teams follow
predefined plans for service recovery.
• Return to Normal Operations—Once service recovery has been completed, the continuity
manager or incident commander implements a predefined return to normal operations.
This collaborative effort shifts the focus of operational control from the recovery team
back into the hands of the normal operations team.
• Service Recovery Testing—The continuity manager and service recovery team remain
engaged to ensure that recovery is complete and the process, as followed, provides for
reliable service recovery and restoration to normal operations.
• Continuity Manual Maintenance—The business continuity team must complete an after-
action analysis and update business continuity process manuals to ensure that action is
taken on any lessons learned from the event.
• Continuity Plan Maintenance—Business continuity planners need to continually revise
the continuity plans for the service infrastructure of the integrated network that provides
mission-critical enterprise services.
There is a great deal of evidence indicating that many organizations don’t invest enough time or
money in BCP. Statistically, building fires permanently close 44 percent of the businesses they
affect. In the World Trade Center bombing in 1993, more than 42 percent, 150 of 350, businesses
didn’t survive. Conversely, in the attacks of September 11, 2001, many businesses had extensive
business continuity plans and were back in operation within days of the devastation.
For many organizations, the plan can simply be a binder stored somewhere away from the main
work location that contains contact information for management and general company staff, a list
of clients and customers, and vendor partners. It’s wise to include any other work process
documentation including the location of the offsite data backups, copies of insurance documents,
and other materials vital business survival.
For large enterprises, with more complex needs, the plan may well include:
• A failover, disaster recovery, or business continuity work site
• A process to routinely test technical requirements and readiness for disaster
• Backup systems for regulatory reporting requirements
• A mechanism to reestablish physical records, perhaps from tape, drive, or microfiche
• A means to establish a new supply chain when vendor partners are also impacted
• For manufacturing operations, the means to establish new production centers may be
required

227
Chapter 10

Every organization, large or small, needs to ensure that their business continuity plans are
realistic. It’s vital that they be easy to use during a crisis. BCP is a crucial peer to crisis
management and disaster recovery planning. It is part of every organization’s overall risk
management strategy.
Development of a business continuity plan can be reduced to five simple steps:
• Analysis
• Solution design
• Implementation
• Testing and organization acceptance
• Maintenance

Analysis
Maintenance

Testing &
Acceptance

Solution
Design

Implementation

Figure 10.5: Five-step BCP life cycle.

A great deal of BCP material is available on the Internet, sponsored by consulting firms who
provide fee-based services. It’s worthwhile to note that there are many freely available basic
tutorials that can help those organizations that are motivated to prepare but financially
constrained from hiring consultants. The simple act of establishing the role of a business
continuity planner can greatly improve organizational readiness for events that cannot be
predicted.

228
Chapter 10

System Access Control


System access control includes several techniques addressed earlier in this guide. Chapter 9
looked at the “Golden Rules” of authentication, authorization, and audit. Rather than dig into the
details of each, we’ll just revisit the basics.
Authentication is the act of ensuring that the user is who they claim to be. Best practices also
drive server or system authentication in the opposite direction, giving the user authenticated
proof that the system or server is indeed the one it purports to be and not a spoofed system
somehow inserted in the traffic flow. Authentication can range from simple passwords to more
complex token-based and/or biometric verification tools.
Factors for Authentication
Simple passwords are often viewed as a very weak form of authentication. Two-factor authentication is
generally considered more acceptable and appropriate for many business applications today. There are
four generally accepted authentication factors that might be used in combination to provide strong user
authentication:
● Something you known—Usually a password or PIN; it’s assumed that only the owner of the
account would know this information
● Something you have—Often a smart card or token; again, assumes only the owner of an account
will have the associated token
● Something you are—Most often a biometric solution: a fingerprint, palm print, voice or retinal scan
● Somewhere you are—May be a simple as inside or outside the corporate firewall for matters of
trust; it might also be a more extensive ring-back system that allows a user to dial in, and then
calls back at a predefined telephone number to grant access; RFID devices may play a role in
future use of location-based authentication.

Authorization is the process of matching the authenticated user or system to a profile that
identifies what that use has permission to do. It’s a combination of services the user is allowed to
use and service network resources to which the user has been granted access. These might
include permission to read-only access or to write files and the ability to execute programs.
Read, write, and execute are the most common authorization focal points.
Audits provide the proof details that may be required later in some form of incident or event
investigation. This includes system logs and history files that provide information about what the
user actually did during a session. Audits provide the documentation for who performed what
function and when. Audits provide accountability, associating users with actions.

229
Chapter 10

Physical and Environmental Security


Beyond the systematic controls of authentication, authorization, and auditing at the system level,
we have physical controls to protect systems. Door locks, controlled work areas, card key
scanners, and automated systems all aid in securing the physical location where systems
delivering key integrated services reside. Physical security may be aa simple as a locked door.
For many small organizations, this simple precaution is adequate. Other enterprises may require
guards, video monitoring, metal detectors, and complex personnel tracking systems to eliminate
any chance of unauthorized access.
Hiding resources by simply not advertising their availability provides a very minimal and weak
approach often termed “security by obscurity.” Proponents of discretion will point out that,
conversely, putting a big sign on the data center door advertising what is inside is probably
unnecessary and inappropriate.

Compliance
ISO 17799 isn’t really a regulation or code that organizations comply with directly. It’s really a
code of practice that helps shape the corporate culture and behaviors. Businesses use ISO 1799 to
help avoid breaking any laws or regulations. For many companies contracts, SLAs and security
policies become part of ISO 17799 adoption. The intent is to ensure that all internal systems
comply with documented corporate policies and standards.

Asset Classifications and Control


IT asset management (ITAM) is another set of business best practices. ITAM brings financial,
contractual, and inventory functions together to manage the IT environment life cycle. This
information is used to make decisions about architectural and service changes in the enterprise
based on knowledge of what is known about the operational environment. It includes both
hardware and software management.
Software management commonly includes licensing to ensure businesses aren’t running on
unlicensed copies of commercial software. Beyond this, software asset management ties closely
to configuration management of corporate systems. Lastly, software asset management is
frequently used as a tool to document compliance with SOX, GLBA, HIPAA, copyright laws,
and so on. There are a number of commercial tools available to assist in managing the enterprise
software inventory.
Hardware asset management focuses on the physical components. In the integrated services
network, this includes all the network elements, nodes, and even cables that deliver mission-
critical business services. Under the umbrella of ISO 17799, many organization document
practices surrounding procurement, administration, re-use, and retirement of hardware assets.

Security Policy
At a high level, the enterprise security policy simply provides a definition of what it means to be
secure. The security policy describes acceptable behavior and approved uses of resources and
often delineates specific security elements to impede unauthorized access. The security policy
speaks to the flow of information by elaborating on what information may enter or leave the
corporate network or zones inside that network.

230
Chapter 10

Managing and Protecting the Network


Managing and protecting the enterprise network is a daily, ongoing exercise in managing risk. As
we’ve begun to learn in this brief overview of regulatory and compliance concerns, we find that
to some extent we must embrace risk. We recognize risks exist. We assess the impact it may
have on business and we develop strategies to either counter, manage, or mitigate it.
In some cases, we can avoid risk entirely by taking some alternative safe approach. In some
cases, we’ll defer the risk to a third party, like an insurance carrier. Sometimes we can reduce the
negative impact, but other times we simply accept the full risk consequences. Each approach
needs to be the result of a conscious business decision.
Chapter 9 introduced Jacobson’s Window for simple risk modeling. This is a good point to refer
back and bring our thoughts together as we work toward the closing of this guide. We’ll do a
quick review here and revisit some key points.

Inconsequential Risk Classes


Inconsequential risks fall into two distinct categories. One inconsequential risk is so unlikely to
occur that is doesn’t warrant undo concern. This risk has a low occurrence rate but a high impact
should it ever occur. We used the example of a meteor striking our data center in Chapter 9, but
we could just as likely take into account a continental power failure or any other very rare event.
When they happen, they are completely outside our control and so devastating to business
operations that spending time brainstorming contingency plans is simply a wasted effort.
At the opposite end of the spectrum is the risk event that has a high rate of occurrence but a low
impact. Spam email provides a perfect example. We get spam email in our inboxes every day.
For most of us, the impact is minimal. We click delete, and it’s gone. Although we may have to
do this several times in the course of a day, the impact to the company is low. In both these
cases, we’re simply accepting the risk rather than investing efforts to mitigate and respond to the
consequences.

Significant Risk Classes


The more significant risk classes are those that have either low-low or high-high occurrence and
impact rates. This is where the majority of all risk analysis and management efforts fall. We
build remediation and mitigation strategies to reduce the risk, spread it across the organization to
reduce the impact, or transfer it through outsourcing, insurance, or some other means. In each
case, we’re striving to minimize disruption to services in the network.

231
Chapter 10

Reducing Risk for Single Occurrence Losses


Chapter 9 used the following illustration to show how risks spread in the real world. This view is
helpful in placing the risks in different quadrants as we determine what our plan or response
might be to a particular type of event.
LOW – Frequency of Occurrence - HIGH

Happens too
rarely to
Accept address
Risks
Max Tolerable
Single
Occurrence
Loss

Mitigate
Risks

Don’t Transfer
Care Risks

LOW - Consequences of Occurrence - HIGH

Figure 10.6: Modeling risks.

This spectrum of risk is very identifiable in real-world planning for business operations. Over
time, losses across this spectrum of risk tend to have similar organizational impacts. From an
actuarial standpoint, the notion of annualized loss expectancy (ALE) is used to quantify risk.
This is often expressed in terms of occurrences per year and the loss resulting from a single
occurrence.
Although ALE is useful for comparing risks, it’s difficult to achieve credible estimates of
occurrence rate for risk events that rarely occur. It’s fairly easy to estimate the consequences, but
not as easy to estimate occurrence rate. Computer network risks frequently stem from human
actions such as simple configuration errors, but fraud, sabotage, viruses, and malicious attacks
may come into play unexpectedly. These occurrence rates are very difficult to quantify.
Despite this difficulty, it’s often wise to define a set of risks that might have disastrous effects
should they occur as a one-time event. Some one-time events might be planned for, while others
might result in circumstance for which there can be no pre-planned business strategy. The
exercise of documenting a brainstorming session for these types of events often leads to creative,
workable strategies for other similar events.

232
Chapter 10

Addressing the Risks


Typical risk management focuses on risks that come from known and identifiable causes—
natural disasters, fires, accidents, death, and so on. Risk is anything that threatens the successful,
ongoing operations of the business. The main objective of risk management is to reduce risks to
tolerable levels. What is tolerable will vary from organization to organization and is based on a
number of factors including technology, people, organizations, and sometimes politics.
Ideally, risk management follows a prioritization scheme like that described in Chapter 9.
Jacobson’s Window offers just one way of looking at risk prioritization. One simple view of the
risk management process is offered from the Nonprofit Risk Management Center at
http://www.nonprofitrisk.org/.

Risk Management Life Cycle


Like every facet of network services we’ve touched on throughout this guide, risk management
has a life cycle of its own. We see in Figure 10.7, the risk management life cycle, from
identification, assessment, response development, to control and closure.

2 - Assessment

1 - Identification

3 – Response
development

5 - Closure
4 - Control

Figure 10.7: The risk management life cycle.

233
Chapter 10

In simple thinking, we assess the risks to the business and determine what the core needs are to
protect the corporate mission. This assessment typically identifies gaps between the current
operational state and what poses acceptable risk. This gap analysis feeds the development of new
policies, procedures, and controls.
As policies and procedures adapt and change to mitigate risk and meet business needs, it’s vital
that they be widely disseminated across the organization. This may present the greatest
challenge, but a policy isn’t effective if employees aren’t aware it exists or what purpose it
serves. Continuous monitoring of policies and controls will lead to ongoing improvements—a
life of risk management within the life cycle of the enterprise itself.

Summary
For much of this guide, we focused on the FCAPS model as a tool for managing faults,
configurations, accounting, performance, and security. FCAPS is a model from industry-driven
standards groups and has been widely adopted in the telecommunications industry as a method
for delivering voice and data services.
It’s important to openly acknowledge that FCAPS isn’t the only model and may not be the best
approach for every organization. Today, many service providers take excerpts from the FCAPS
approach and fold them into the ITIL model, creating an enterprise-specific framework from the
array of best practices tools and techniques available in the marketplace.
For most organizations, there is no one model that fits all needs. The process of compiling the
best facets of each approach is a learning process, maturation. As part of ITIL adoptions, most
organizations perform a maturity self-assessment exercise to identify where they feel they fit
along the curve of maturity as an IT services organization. Very few companies starting the
process place themselves toward the mature end of the scale. The act of beginning, of assessing
where your organization fits and documenting each step of the way, is a recognized best practice
to delivering sustainable, reliable data, voice, and video services in a converged network. At the
end of the day, the highly successful service delivery organization doesn’t have just the ITIL
library. They have the enterprise library of policies, procedures, and practices that, honed over
time, document the best practices for managing the needs of that enterprise.

Download Additional eBooks from Realtime Nexus!


Realtime Nexus—The Digital Library provides world-class expert resources that IT
professionals depend on to learn about the newest technologies. If you found this eBook to be
informative, we encourage you to download more of our industry-leading technology eBooks
and video guides at Realtime Nexus. Please visit http://nexus.realtimepublishers.com.

234

You might also like