You are on page 1of 85

Data Center

Networking
Architecture

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 1
Before We Get Started:
ƒ Put cell phones into silent mode
ƒ Intermediate level session focused on data center front end
architecture
ƒ This session is based upon the Data Center Infrastructure Design
Guide 2.5.
http://www.cisco.com/application/pdf/en/us/guest/netsol/ns107/c649/cc
migration_09186a008073377d.pdf
ƒ Additional Cisco Validated Designs (CVD) can be found at;
http://www.cisco.com/go/cvd
Enterprise Data Center:
http://www.cisco.com/en/US/solutions/ns340/ns414/ns742/ns741/netw
orking_solutions_products_genericcontent0900aecd80601e1d.html#da
tacenter

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 2
Agenda
Data Center Infrastructure

ƒ Core Layer Design


ƒ Aggregation Layer Design
ƒ Access Layer Design
ƒ Density and Scalability Implications
ƒ Scaling Bandwidth and Density
ƒ Spanning Tree Design and Scalability
ƒ Increasing HA in the DC

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 3
Data Center Evolution

Data Center 1.0 Data Center 2.0 Data Center 3.0

Client-Server and Service Oriented and


Mainframe
Distributed Computing Web 2.0 Based
IT Relevance and Control

Consolidate

Virtualize

Automate

Centralized Decentralized Virtualized


Application Architecture Evolution

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 4
Application Centric Architecture
Two Sides Of the Same Coin

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 5
Data Center Architecture Overview
Layers of the Enterprise Multi-Tier Model

ƒ Multi-tier application architecture Enterprise Core


logically overlaid on network
ƒ Layer 2 and layer 3 access DC Core
topologies
ƒ Dual and single attached DC Aggregation/
Distribution
Servers, Mainframe and Blade
Chassis
ƒ Multiple aggregation modules
DC Access
ƒ L2 adjacency requirements
ƒ Stateful services for security and Blade Chassis w/
load balancing Integrated Switch

Blade Chassis w/ Mainframe


Pass Thru w/ OSA
L2 w/ Clustering L3 Access
and NIC Teaming

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 6
Core Layer Design

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 7
Core Layer Design Scaling
Requirements
Campus Access Layer
Campus
Distribution
ƒ Is a separate DC Core
Layer required?
ƒ Consider:
10GigE port density Campus Core
Administrative domains
Anticipate future
requirements
ƒ Key core characteristics DC Core
10GE scalability
Aggregation
Distributed forwarding
architecture
Advanced link load balancing
Scalable IP multicast support
Server Farm Access Layer

Scaling

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 8
Core Layer Design
L2/L3 Characteristics
ƒ Layer 3 Core
Equal cost multi-path (ECMP) load
balancing Campus Core
EIGRP/OSPF for fast convergence
L2 extension through core is
not recommended
DC Core
CEF Hash
ƒ CEF* hashing algorithm Applied to
Default hash is on L3 IP Packets on
addresses only Equal Cost
L3 + L4 port hash will improve Routes
L3
load distribution
L2
CORE1(config)#mls ip cef load full simple
Aggregation
Leverages automatic source port
randomization in client TCP stack

Access

Web Application Database


Servers Servers Servers
*CEF = Cisco Express Forwarding
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 9
Core Layer Design
Routing Protocol Design: OSPF
ƒ Isolate the DC network with a
dedicated OSPF Area.
ƒ Not So Stubby Area (NSSA) helps to Campus Core
limit LSA propagation, but permits DC Subnets
route redistribution (RHI) (Summarized)
Area 0
ƒ Advertise default into NSSA, NSSA DC Core
summarize routes out L0=10.10.1.1 L0=10.10.2.2
Default Default
ƒ Use “auto-cost reference-bandwidth”
to support 10G links
ƒ Loopback interfaces simplify L3 vlan-ospf

troubleshooting (neighbor ID)


Aggregation
ƒ Use passive-network default: enable
on L3 links to allow peering
L0=10.10.3.3 L0=10.10.4.4
ƒ Use authentication: more secure and
avoids undesired adjacencies Access
ƒ Interface hello-dead = 1/3

Web Application Database


Servers Servers Servers

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 10
Core Layer Design
Routing Protocol Design: EIGRP

ƒ Use summarization and default to


isolate the DC from network events Campus Core
ƒ Advertise default into DC with
interface command on core:
ip summary-address eigrp 20 0.0.0.0 0.0.0.0 200 DC Core
Cost of 200 required to prefer Internet default route
over the local NULL0 default route installed by Default Default
summary-address DC Subnets
(Summarized)
ƒ If other default routes exist (from L3 vlan-eigrp

internet edge for example), may need Aggregation


to use distribute lists to filter out
ƒ Use passive-network default
ƒ Summarize DC subnets to core with Access
interface command on agg:
ip summary-address eigrp 20 10.20.0.0 255.255.0.0

Web Application Database


Servers Servers Servers

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 11
Aggregation Layer Design

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 12
Aggregation Layer Design
Spanning Tree Design
ƒ Rapid-PVST+ (802.1w) or
MST (802.1s), Core

ƒ Choice of .1w/.1s based on scale


of logical+virtual ports required Root Primary Root Secondary
HSRP Primary HSRP Secondary
ƒ R-PVST+ is recommended as Active Context Standby Context
best replacement for 802.1d
Fast converging: inherits Cisco
enhancements (Uplink-fast,
Backbone-fast)
Combined with RootGuard,
BPDUGuard, LoopGuard ,and UDLD
achieves for STP stability
Access layer uplink failures:
~300ms – 2sec
Most flexible design options
UDLD global only enables on Fiber
ports, must enable manually on
copper ports Rootguard
LoopGuard
BPDU Guard
UDLD Global

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 13
Aggregation Layer Design
Integrated Services
Services: Firewall, Load Balancing, SSL
Encryption/Decryption

+
ƒ L4-L7 services integrated in Cisco Catalyst® 6500
ƒ Server load balancing, firewall and SSL services may be
deployed in:
Active-standby pairs (ACE, FWSM 2.X)
Active-active pairs (ACE, FWSM 3.1)
ƒ Integrated blades optimize rack space, cabling, mgmt,
providing flexibility and economies of scale
ƒ Influences many aspects of overall design
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 14
Aggregation Layer Design
Active-Standby Service Design

ƒ Active-standby services
Application Control Engine
Core
Firewall Service Module
SSL Module Root Primary Root Secondary
HSRP Primary HSRP Secondary
ƒ Under utilizes: Active Context Standby Context

Access layer uplinks


Service modules
Aggregation switch fabrics
ƒ Advantages:
Widely deployed
Predictable traffic patterns
Easier to configure and manage
Consistent performance under
failure conditions.

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 15
Aggregation Layer Design
Active-Active Service Design

ƒ Active-Active Service Design


Application Control Engine
Core
(ACE)
Active-standby distribution VLAN5: VLAN6:
per context Root Primary Root Primary
HSRP Primary HSRP Primary
Active Context Active Context
Firewall Service Module (3.x)
Two active-standby groups
permit distribution of
contexts across two FWSM’s VLAN 6: VLAN 5:
Root Secondary Root Secondary
ƒ Permits uplink load balancing HSRP Secondary
Standby Context
HSRP Secondary
Standby Context
while having services applied
ƒ Increases overall service
performance vlan6 vlan6

vlan5 vlan6 vlan6 vlan5

Tech Tip: Virtual Context are key to Active/Active designs


Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 16
Aggregation Layer Design
Establishing Inbound Path Preference
Establish Route Preference for
Service Enabled Applications
ƒ Use Route Health Injection Core
feature of ACE
3. EIGRP/OSFP will
ƒ Aligns advertised route propagate RHI VIP to
Core to attract traffic 4. If Context Failover

of VIP with active context on Occurs, RHI and Route


Preference Follow

ACE, FWSM and SSL service


modules
ƒ Avoids unnecessary use 2. If Healthy, Installs
of inter-switch link and Host Route to VIP
on Local MSFC

asymmetrical flows
ƒ Introduce route-map to
RHI injected route to set
vlan6 vlan6
desired metric
vlan5 vlan6 vlan6 vlan5
1. ACE Probes to
Real Servers in VIP
to Determine Health

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 17
Aggregation Layer Design
Scaling the Aggregation Layer

ƒ Aggregation modules provide:


Spanning tree scaling
HSRP Scaling Campus Core
Access layer density
Application services scaling DC Core
SLB/firewall
Fault domain sizing
ƒ Core layer provides inter-agg
module transport:
Provides inter-agg module Aggregation Aggregation Aggregation
transport in multi-tier model Module 1 Module 2 Module 3
Low latency distributed forwarding Server Farm Access Layer
(use DFC’s)
100,000’s PPS forwarding rate Scaling

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 18
Aggregation Layer Design
Using Virtual Route Forwarding (VRF)

MPLS or
VRF-Green other Core
ƒ Enables virtualization/
partitioning of network VRF-Blue DC Core
resources (MSFC, VRF-Red
ACE, FWSM)
Agg1 Agg2
ƒ Permits use of application
services with multiple access
Alternate Primary
topologies Firewall and SLB Contexts on Agg1
Contexts for Green, and 2 to Achieve
ƒ Maps well to path isolation Blue, and Red Active-Active Design
MAN/WAN designs such as VLANs Isolate
with MPLS or Multi-VRF Contexts on Access

(VRF-Lite) 802.1Q
Trunks

ƒ Security policy management


and deployment by user
group/vrf

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 19
Access Layer Design

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 20
Access Layer Design
Defining Layer 2 Access

ƒ Access Layer connects servers &


hosts to the network
DC Core
ƒ Proprietary server protocols
require layer 2 adjacency
L3 Agg1 Agg2
ƒ L2 topologies consist of looped Inter-Switch Link

and loop-free models L2


Primary Root Secondary Root
Looped - VLANs are extended across Primary HSRP Secondary HSRP
inter-switch link trunk Active Services Standby Services

nks
Loop-free - VLANs are not extended

T ru
across inter-switch link trunk

.1q
ƒ L3 routing is typically performed in

802
the aggregation layer
ƒ Stateful services at Agg can be
provided across the L2 access
(FW, SLB, SSL, etc.)

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 21
Access Layer Design
Establish a Deterministic Model
ƒ Align active components in traffic
path on common Agg switch:
DC Core
Primary STP root L3+L4 Hash
Agg-1(config)#spanning-tree vlan 1-10 root primary Path
Pref

Primary HSRP (outbound) L3 Agg1 Agg2


standby 1 priority X L2
Primary Root
Active Service modules/contexts Primary HSRP
Active Services

nks
Def gwy

Path Preference (inbound)

T ru
.1q
Use RHI & Route map

802
Route-map preferred-path
match ip address x.x.x.x
set metric -30
(also see slide 17)

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 22
Access Layer Design
Balancing VLANs on Uplinks: L2 Looped Access
ƒ Distributes load across uplinks
ƒ STP blocks uplink path for
VLANs to secondary root switch DC Core
L3+L4 Hash
ƒ If active/standby service modules
Agg1 Agg2
are used; consider inter-switch L3
link utilization L2
Multiple service modules may be Blue-Primary Root Red-Primary Root
Blue-Primary HSRP Red-Primary HSRP
distributed on both agg switches to Red-Sec Root Blue-Sec Root
achieve balance Red-Sec HSRP Def gwy Def gwy Blue-Sec HSRP

Consider b/w of inter-switch link in a


failure situation
802.1q Trunks
ƒ If active/active service modules
are used;
(ACE and FWSM3.1)
Balance contexts+hsrp across agg
switches
Consider establishing inbound path
preference

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 23
Access Layer Design
Looped Design Model

ƒ VLANs are extended between Primary STP Root Secondary STP Root
aggregation switches, creating Primary HSRP Secondary HSRP
the looped topology Active Services Standby Services
ƒ Spanning Tree is used to L3 Inter-Switch Link
prevent actual loops (Rapid L2
PVST+, MST) F F
.1Q Trunk
ƒ Redundant path exists through a
F F F F F F F
second path that is blocking
ƒ Two looped topology designs:
Triangle and square
FB FB F F
ƒ VLANs may be load balanced B
across access layer uplinks
ƒ Inter-switch link utilization must
be considered as this may be
used to reach active services Looped Triangle Looped Square

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 24
Access Layer Design
L2 Looped Topologies
L2 Triangle L2 Square
L3 L3
L2 L2

VLAN VLAN
10 10

VLAN VLAN VLAN VLAN


10 20 10 20

Row 2, Cabinet 3 Row 9, Cabinet 8 Row 2, Cabinet 3 Row 9, Cabinet 8

Looped Triangle Access Looped Square Access


ƒ Supports VLAN extension/L2 adjacency across ƒ Supports VLAN extension/L2 adjacency across
access layer access layer
ƒ Resiliency achieved with dual homing and STP ƒ Resiliency achieved with dual homing and STP
ƒ Quick convergence with 802.1W/S ƒ Quick convergence with 802.1W/S
ƒ Supports stateful services at aggregation layer ƒ Supports stateful services at aggregation layer
ƒ Proven and widely used ƒ Active-active uplinks align well to ACE/FWSM
ƒ Achieves higher density access layer,
optimizing 10GE aggregation layer density
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 25
Access Layer Design
Benefits of L2 Looped Design
ƒ Services like firewall and load
balancing can easily be deployed at
the aggregation layer and shared DC Core
across multiple access layer switches
L3
ƒ VLANs are primarily
L2
contained between pairs
of access switches but─
ƒ VLANs may be extended
to different access
switches to support
NIC teaming
Clustering L2 adjacency
Administrative reasons
Geographical challenges

Row A, Cab 7 Row K, Cab 3

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 26
Access Layer Design
Loop-Free Design
ƒ Alternative to looped design
ƒ 2 LoopFree Models:
U and Inverted U DC Core

ƒ Benefit: Spanning Tree L3


is enabled but no port is blocking L2
so all links are forwarding L3

L2
ƒ Benefit: less chance of loop condition
due to configuration errors or other
anomalies
ƒ L2-L3 boundary varies by
loop-free model used:
U or Inverted-U
ƒ Implication considerations with
service modules, L2 adjacency,
and single attached servers
LoopFree U LoopFree
Inverted U

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 27
Access Layer Design
Drawbacks of Layer 2 Looped Design

ƒ Main drawback: if frame looping occurs, the


network may become unmanageable due to
the infinite replication of frames
ƒ 802.1w Rapid PVST+ combined with STP
0000.0000.3333
related features and best practices improve
stability and help to prevent loop conditions
UDLD
DST MAC 0000.0000.4444
Loopguard
Rootguard 3/2 3/2
BPDUguard
Limit domain size 3/1 3/1
Stay under STP watermarks for Switch 1 Switch 2
logical and virtual ports DST MAC 0000.0000.4444

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 28
Access Layer Design
Loop-Free Topologies
L2 Loop-Free U L2 Loop-Free Inverted U
L3
L2
L3
L2

VLAN
10

VLAN VLAN VLAN VLAN


10 20 10 20

Loop-Free U Access Loop-Free Inverted U Access


ƒ VLANs contained in switch pairs ƒ Supports VLAN extension
(no extension outside of switch pairs) ƒ No STP blocking; all uplinks active
ƒ No STP blocking; all uplinks active ƒ Access switch uplink failure black holes
ƒ Autostate implications for certain service single attached servers
modules (CSM) ƒ Supports all service module implementations
ƒ ACE supports autostate and per context ƒ Single attached servers prone to isolation
failover during failures
ƒ Potential for split subnets during failure
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 29
Access Layer Design
Loop-Free U Design and Service Modules (1)
Agg1, Active HSRP Agg2, Standby
Services Services
HSRP Secondary on
L3 Agg2 Takes Over as def-
L2 gwy, But How to Reach
Active Service Modules?

MSFC on Agg1

SVI for Interface vlan 10


goes to down state if
uplink is the only
interface in vlan 10

VLAN 10 VLAN 11
VLAN 20, 21

ƒ If the uplink connecting access and aggregation goes down, the VLAN interface on the MSFC
goes down as well due to the way autostate works
ƒ CSM and FWSM has implications as autostate is not conveyed (leaving black hole)
ƒ Tracking and monitoring features may be used to allow failover of service modules based on
uplink failure but would you want a service module failover for one access switch uplink failure?
ƒ Not recommended to use loop-free L2 access with active-standby service module
implementations. (See slide on ACE next)

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 30
Access Layer Design
Loop-Free U Design and Service Modules (2)
ACE Context 1 ACE Context 1
HSRP

L3
L2 HSRP Secondary on Agg2
Takes over as def-gwy

VLAN 10, 11 VLAN 20, 21


With ACE:
ƒ Per context failover with autostate
ƒ If uplink fails to Agg1, ACE can switchover to Agg2 (under 1sec)
ƒ Requires ACE on access trunk side for autostate failover
ƒ May be combined with FWSM3.1 for active-active design
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 31
Access Layer Design
Drawbacks of Loop-Free Inverted-U Design

ƒ Single attached servers are


black-holed if access switch DC Core
uplink fails
L3
ƒ Distributed EtherChannel® can L2
reduce chance of black holing
ƒ NIC teaming improves
resiliency in this design
ƒ Inter-switch link scaling needs
to be considered when using
active-standby service
modules

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 32
Virtual Switching System 1440
Network System Virtualization
Aggregation/Access Server Access

Si
Si Si
Si Si
Si Si
Si

Features Benefits of VSS


Network System Virtualization Increased Operational Efficiency
via Simplified Network
Inter-Chassis Stateful Switch Boost Non-stop Communication
Over (SSO)
Multi-Chassis EtherChannel Scale the System Bandwidth
(MEC) Capacity to 1.4 Tbps
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 33
Virtual Switching System
Data Center
A Virtual Switching in the Data Center increases bandwidth scalability, but still provides a
Layer 2 hierarchical architecture without relying on Spanning Tree…

Data Center VSS design guide Summer 2008

Single router node, Fast


convergence, Scalable
architecture
L2/L3 Core

Dual Active Uplinks,


Fast L2 convergence,
minimized L2 Control L2
Plane, Scalable Aggregation*

Dual-Homed Servers,
Single active uplink per
VLAN (PVST), Fast L2
convergence L2 Access

*Service Module support in August 2008.


Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 34
Access Layer Design
Comparing Looped, Loop-Free and VSS

VLAN Service Single


Uplink vlans
Extension Module Black- Attached Access Switch Must Consider
on Agg Switch
Supported Holing on Server Black- Density per Inter-Switch
in Blocking or
Across Uplink Failure Holing on Agg Module (3) Scaling
Link
Standby State
Access Layer (5) Uplink Failure

Looped
Triangle
- + + + - (3) +

Looped
Square
+ + + + + -
(4)
Loop-Free U + - (4) - + + +
(1,2)
Loop-Free
Inverted U
+ + + (1,2) +/- + -

VSS (6) + + + + + +

1. Use of Distributed EtherChannel Greatly Reduces Chances of Black Holing Condition


2. NIC Teaming Can Eliminate Black Holing Condition
3. When Service Modules Are Used and Active Service Modules Are Aligned to Agg1
4. ACE Module Permits L2 Loopfree Access with per Context Switchover on Uplink failure
5. Applies to when using CSM or FWSM in active/standby arrangement
6. DCVSS design guide to be released in Summer’08

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 35
Access Layer Design:
L3 Design Model

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 36
Access Layer Design
Defining Layer 3 Access

ƒ L3 access switches connect to


aggregation with routed interfaces
DC Core DC Core
ƒ All uplinks are active (ECMP), no
spanning tree blocking occurs
ƒ .1Q trunks between pairs of L3 DC Aggregation
access switches to support L2
adjacency server requirements
ƒ Convergence time is usually better
than Spanning Tree DC Access
L3
ƒ Provides isolation/shelter for L2
hosts affected by broadcasts

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 37
Access Layer Design
Need L3 for Multicast Sources?

ƒ Multicast sources on L2 access


works well with IGMP snooping
ƒ IGMP snooping at access switch DC Core DC Core
automatically limits multicast flow
to interfaces with registered
clients in VLAN
DC Aggregation
ƒ Use L3 when IGMP snooping is
not available or when particular
L3 administrative functions are
required
DC Access
L3
L2

L3 Access with L3 Access with


Multicast Sources Multicast Sources

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 38
Access Layer Design
Benefits of L3 Access

ƒ Minimizes broadcast domains


attaining high level of stability
DC Core DC Core
ƒ Meet server stability requirements
or isolate particular application
environments
ƒ Creates smaller failure domains, DC Aggregation
increasing stability
ƒ All uplinks are active paths, no
blocking (up to ECMP maximum) DC Access
ƒ Fast uplink convergence: L3
L2
failover and fallback, no arp
table to rebuild for aggregation
switches

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 39
Access Layer Design
Drawbacks of Layer 3 Design

ƒ L2 adjacency is limited to access


pairs (clustering and NIC teaming
DC Core DC Core
limited)
ƒ IP address space management is
more difficult, small subnets. DC Aggregation
ƒ If migrating to L3 access, IP
address changes may be difficult
on servers (may break apps)
DC Access
ƒ Normally require services to L3
L2
be deployed at access layer
pair to maintain L2 adjacency
with server and provide stateful
failover

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 40
Access Layer Design
L2 or L3? What Are My Requirements?
Aggregation Aggregation
L3

OSPF, EIGRP
L2

Rapid PVST+ or MST


L3
Access L2 Access

Layer 2 Layer 3
The Choice of One Design Versus the Other:
ƒ Difficulties in managing loops ƒ Ability to extend VLANs
ƒ Staff skillset; time to resolution ƒ Specific application requirements
ƒ Broadcast domain sizing
ƒ Convergence properties
ƒ Oversubscription requirements
ƒ NIC teaming; adjacency
ƒ Link utilization on uplinks
ƒ HA clustering; L2 adjacency ƒ Service module support/placement

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 41
Access Layer Design:
BladeServers

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 42
Blade Server Requirements
Connectivity Options

Using Pass-Through Modules Using Integrated Ethernet Switches

Aggregation
Layer

External L2
Switches

Integrated L2
Switches

Interface 1
Interface 2
Blade Server Chassis Blade Server Chassis

Tech Tip – Avoid hierarchy of L2 switches. Traffic patterns


are difficult to predict during failure conditions.
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 43
Blade Server Requirements
Trunk Failover Feature
ƒ Switch takes down server Aggregation
interfaces if corresponding uplink
fails, forcing NIC teaming failover
ƒ Solves NIC teaming limitations;
prevents black-holing of traffic
ƒ Achieves maximum bandwidth
utilization:
No blocking by STP, but STP is Integrated L2
enabled for loop protection switches

Can distribute trunk failover groups


across switches

ƒ Dependent upon the NIC feature


set for NIC Teaming/failover Blade Server Chassis
Interface 1
Interface 2

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 44
Cisco Virtual Blade Switch (VBS)
Overview of Concept and Benefits
NEW

ƒ Management Simplification
– Operational simplification
• Single switch per rack to manage
• True Plug-n-Play of switches
– Design Simplification:
• Sharing Uplinks helps reduce cables
• Reduction in # of logical nodes in L2/L3
network helps improve network convergence
– Operational Consistency
• Familiar IOS CLI, MIBs and management tools
like CiscoWorks
• Consistent End-to-end features and functionality

ƒ Performance & Scalability


Traditional Virtual Blade – Up to 160G configurable bandwidth out of rack
Blade Switch Switch
– Rack switch allows server to double bandwidth
with no additional cost

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 45
Cisco Catalyst Virtual Blade Switch
Topology Highlighting Key Benefits
Access Layer (Virtual Blade Switch) Aggregation Layer

Mix-n-match
GE & 10GE
switches

Local Traffic
doesn’t go to
distribution
switch

Higher
Resiliency With Catalsyt
with 6500 VSS, all
Etherchannel links utilized

Single Switch / Node Greater Server BW –


(for Spanning Tree or via Active-Active
Layer 3 or Server Connectivity
Management)
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 46
Cisco Catalyst Virtual Blade Switch
Multiple Deployment Options

Common Scenario Separate Rings 4 NIC server Scenario


ƒ Single Virtual ƒ Separate VBS ƒ More Server Bandwidth
– VMware
Blade switch ƒ More resilient
ƒ Creates smaller Rings
* Design Guide recommendations have not been established yet.
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 47
Density and Scalability
Implications in the
Data Center

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 48
Density and Scalability Implications
Modular, Top of Rack, Blade Access Switching Models

ƒ Where are the issues?


Cabling
Power
Cooling
Spanning Tree Scalability
Management
Oversubscription
Sparing
Redundancy
ƒ The right solution is usually
based on business requirements
ƒ Hybrid implementations can and
do exist

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 49
49
Density and Scalability Implications
Server Farm Cabinet Layout
~30-40 1RU Servers per Rack

…… N- Racks per Row


……
Considerations:
How Many Interfaces per Server?
Top of Rack, Blade Switches or End of Row?
Separate Switch for Management Network?
Cabling overhead, under floor, patch systems?
Cooling capacity?
Power distribution/ Redundancy?

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 50
Density and Scalability Implications
Density: How Many NICs to Plan For?
Front End Interface
ƒ Three to four NICs per server
OOB
are common Management
Front end or public interface Backup Network
Storage HBA or
Storage interface (GE, FC) GE NIC
Backup interface
Back end or private interface Back End Interface

integrated Lights Out (iLO)


for OOB mgmt

Cabling remains in cabinets


ƒ May require more than two
TOR switches per rack
30 servers@ 4 ports = 120 ports
required in a single cabinet (3x48 port
1RU switches)
May need hard limits on
cabling capacity
Avoid cross cabinet and other cabling
nightmares
Single Rack-2 Switches Dual Rack-2 Switches

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 51
Density and Scalability Implications
Cabinet Design with Top of Rack Switching
Servers Connect to a Top of Rack (TOR) Switch

ƒ Minimizes the number of cables to


run from each cabinet/rack
ƒ If NIC teaming support: two - TOR
switches are required
ƒ Will two TOR switches provide
enough port density?

Cabling remains in cabinets


ƒ Cooling requirements usually do not
permit a full rack of servers
ƒ Redundant switch power supply are
option
ƒ No Redundant switch processors
ƒ GEC or 10GE Uplink Considerations

Single Rack-2 Switches Dual Rack-2 Switches

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 52
Density and Scalability Implications
Top of Rack (TOR) Switching Model

Pro: Efficient Cabling DC Core 4 Uplinks per Cabinet = 100


Pro: Improved Cooling Uplinks Total to Agg Layer
Con: Number of Devices/Mgmt
Con: Spanning Tree Load

Aggregation
GEC or 10GE Uplinks?

Access

Cabinet 1 Cabinet 2
… Cabinet 25

With ~1,000 Servers/25 Cabinets = 50 Switches

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 53
Density and Scalability Implications
Cabinet Design with Modular Access Switches
Servers Connect Directly to a Modular Switch

ƒ Cable bulk can be difficult to


manage and block cool air flow
ƒ Switch placement at end of row
or within row
ƒ Minimizes cabling to Aggregation
ƒ Reduces number of
uplinks/aggregation ports
ƒ Redundant switch power and
processors are options
ƒ GEC or 10GE Uplink
Considerations
ƒ NEBS Considerations

Cables route under raised floor or in overhead trays

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 54
Density and Scalability Implications
Access Network Topology w Modular Switches

Pro: Fewer Devices/Mgmt DC Core 2 Uplinks per 6509 = 16


Con: Cabling Challenges Uplinks to Agg Layer
Con: Cooling Challenges

Aggregation

GEC or 10GE Uplinks?

Access

With ~1,000 Servers/9 Slot Access Switches= 8 Switches


~8 Access Switches to Manage
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 55
Density and Scalability Implications
Modular, Top of Rack and Blade Comparison

Lower STP Proc

More I/O Cables

Modular Fewer Uplinks


Access

Aggregation

Top of Rack & Higher STP Proc


Blade Switch
Access
Fewer I/O Cables

More Uplinks

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 56
Scaling Bandwidth and Density

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 57
Scaling B/W with GEC and 10GE
Optimizing EtherChannel Utilization

ƒ Ideal is graph on top right


ƒ Bottom left graph more typical
ƒ Analyze the traffic flows in and out of
the server farm:
IP addresses (how many?)
L4 port numbers (randomized?)

ƒ Default L3 hash may not be optimal


for GEC: L4 hash may improve
agg(config)# port-channel load balance
src-dst-port

ƒ 10 GigE gives you effectively the full


bandwidth without hash implications

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 58
Scaling B/W with GEC and 10GE
Migrating Access Layer Uplinks to 10GE

ƒ How do I migrate from GEC to 10GE uplinks?


ƒ How do I increase the 10GE port density at the agg layer?
ƒ Is there a way to regain slots used by service modules?

DC Core

Aggregation

Access Pair 1 … …
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 59
Scaling B/W with GEC and 10GE
Consolidate to ACE

DC Core

ƒ Consider consolidating multiple


service modules onto ACE Module
–SLB
–Firewall
Aggregation
–SSL
ƒ 4/8/16G Fabric Connected
ƒ Active-Active Designs
ƒ Higher CPS + Concurrent CPS
ƒ Single TCP termination, lower latency
ƒ Firewall feature gap needs Access
consideration

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 60
Scaling B/W with GEC and 10GE
Service Layer Switch

ƒ Move certain services out of


aggregation layer DC Core
ƒ Ideal for CSM, SSL modules
ƒ Opens slots in agg layer for
10GE ports
Service Service
ƒ Use separate links for FT Switch1 Aggregation Switch2
paths Si Si
(Redundant)
ƒ Extend only necessary L2
VLANs to service switches
via .1Q trunks (GEC/TenG)

Access

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 61
Increasing Throughput with
Virtual Switching System 1440

ƒ VSS combines a pair of DC Core


Catalyst 6500s into a single
logical switch
ƒ Eliminates Spanning-Tree
blocking ports. Service Service
ƒ Etherchannel load balancing Switch1 Switch2
(Redundant)
ƒ Service module support in
12.2(33)SXI Summer08
(FWSM, ACE, NAM)
ƒ Spanning-tree is a protection
mechanism against rogue
switches and cabling errors Access

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 62
Scaling with 10GE Density
Nexus 7000

ƒ Nexus 7000 provides high


density 10GE scalability. DC Core
ƒ Virtual Device Contexts
(VDC) to logically partition a
single chassis.
ƒ Multi-Chassis Etherchannel Service Service
Switch1 Aggregation Switch2
support (future)
ƒ NX-OS purpose built data Nexus Nexus
center operating system 7000 7000

ƒ Unified Fabric capable

Access

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 63
Spanning Tree Scalability

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 64
Spanning Tree Scalability
Common Questions
ƒ How many VLANs can I
support in a single
aggregation module? DC Core
ƒ Can a “VLAN Anywhere”
model be supported?
Aggregation
ƒ How many access switches
can I support in each
aggregation module?
ƒ What are the maximum
number of logical ports?
ƒ Are there STP hardware
restrictions?

Access Pair 1 … …
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 65
Spanning Tree Scalability
Spanning Tree Protocols Used in the DC

ƒ Rapid PVST+ (802.1w)


Most common in data center today
Scales to large size ~10,000 logical ports
Coupled with UDLD, Loopguard, RootGuard and BPDU Guard,
provides a strong-stable L2 design solution
Easy to implement, proven, scales
ƒ MST (802.1s)
Permits very large scale STP deployments ~30,000 logical ports
Not as flexible as Rapid PVST+
Service module implications (FWSM transparent mode)
More common in service providers and ASPs

This CVD focuses on the Use of Rapid PVST+


Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 66
Spanning Tree Scalability
Spanning Tree Protocol Scaling

MST RPVST+ PVST+

Total Active STP


50,000 Total 10,000 Total 13,000 Total
Logical Interfaces
1,8001 per
Switching
Total Virtual Ports 6,0001 per Module(6700) 1,8001 per
per LineCard Switching Module Switching Module
1200 for Earlier
Modules

1 10 Mbps, 10/100 Mbps, and 100 Mbps Switching Modules Support a Maximum of 1,200
Logical Interfaces per Module
http://www.cisco.com/univercd/cc/td/doc/product/lan/cat6000/122sx/ol_4164.htm#wp26366

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 67
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Number of Total STP Active
Logical Interfaces= DC Core
ƒ Trunks on the switch * active
VLANs on the trunks + number
of non-trunking interfaces on the switch
30 VLANs
In this example, aggregation 1 will have:
10 + 20 + 30 = 60 STP active logical interfaces
Te7/4
AGG1#sh spann summ tot
Switch is in rapid-pvst mode Te7/3
Root bridge for: VLAN0010, VLAN0020, VLAN0030
EtherChannel misconfig guard is enabled
Extended system ID is enabled
Portfast Default is disabled
PortFast BPDU Guard Default is disabled
Portfast BPDU Filter Default is disabled
Loopguard Default is enabled
UplinkFast is disabled 10 VLANs 20 VLANs
BackboneFast is disabled
Pathcost method used is long

Name Blocking Listening Learning Forwarding STP Active


---------------------- -------- --------- -------- ---------- ----------
30 VLANs 0 0 0 60 60
AGG1#

STP Active Column = STP Total Active Logical Interfaces

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 68
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Example: Calculating Total Active
Logical Ports
ƒ 120 VLANs system wide
ƒ No manual pruning performed on Core

trunks
ƒ 1RU access layer environment Layer 3
Layer 2
ƒ 45 access switches each Aggregation 1 Aggregation 2
connected with 4GEC Primary Root Loop Secondary Root

ƒ Dual homed, loop topology


(120 * 45 access links)+120
instances on link to
agg2=5400+120=5520 Access 1 ..… Access 45
ƒ This is under the maximum
recommendation of 10,000 when
using Rapid PVST+

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 69
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Number of Virtual Ports per
Line Card= DC Core
ƒ For line card x: sum of all trunks * VLANs *
(the number of ports in a port-channel if used)
10 + 20 + (30*2) 30 VLANs
=90 Virtual Port’s on line card 7 Te7/1
Te7/2

Te7/4
AGG1#sh vlan virtual-port slot 7
Slot 7 Te7/3
Port Virtual-ports
-------------------------
Te7/1 30 EtherChannel
Te7/2 30 10 VLANs 20 VLANs
Te7/3 10
Te7/4 20
Total virtual ports:90
AGG1#
NOTE: VPs Are Calculated per Port in Channel Groups

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 70
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Example: Calculating Virtual Ports per
Line Card
Core

ƒ 120 VLANs system wide


ƒ no manual pruning performed on
Layer 3
trunks Layer 2
Aggregation 2
ƒ 12 access switches, each Aggregation 1
Primary Root loop Secondary Root

connected with 4GEC across


6700 line card 6748 Line Card

Acc1Acc2 …. Acc12
Access 1 ..… Access 12
EtherChannels to Acces Layer

(120 * 48 access links)


=5,760 virtual ports
ƒ This is above the recommended Maximum number VLANs per port = 37
1800/48=37.5 per port
watermark

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 71
Spanning Tree Scalability
Why STP Watermarks Are Important

Watermarks Are Not Hard Limits But─


ƒ If exceeded, performance is unpredictable
ƒ Larger impact when interface flaps, or shut/no_shut
ƒ Small networks may not see a problem
ƒ Large networks will usually see problems
ƒ Convergence time will be affected
ƒ Pruning off unneeded VLANs will reduce active
logical and virtual port instances

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 72
Spanning Tree Scalability
Design Guidelines

ƒ Add aggregation modules to


scale, dividing up the STP DC Core
domain
ƒ Maximum five hundred HSRP
instances on Sup720 (depends
on other cpu driven processes)
ƒ If logical/virtual ports near upper Aggregation
Module 1
limits perform:
–Manual pruning on trunks
–Add aggregation modules
–Use MST if necessary

Access

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 73
Increasing HA in the Data Center

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 74
Increasing HA in the Data Center
Server High Availability
Common Points of Failure
1. Server network adapter
2. Port on a multi-port
server adapter
L3
L2 3. Network media
(server access)
4. Network media (uplink)
5. Access switch port
6. Access switch module
7. Access switch

These Network Failure Issues Can Be


Without Data Center HA With Data Center HA Addressed by Deployment of Dual
Recommendations Recommendations Attached Servers Using Network Adapter
Teaming Software
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 75
Increasing HA in the Data Center
Common NIC Teaming Configurations
AFT—Adapter Fault Tolerance SFT—Switch Fault Tolerance ALB—Adaptive Load Balancing

Default GW Default GW Default GW


10.2.1.1 10.2.1.1 10.2.1.1
HSRP HSRP HSRP

Heartbeats
Heartbeats
Heartbeats

Eth0: Active Eth1: Standby Eth0: Active Eth1: Standby Eth0: Active Eth1-X: Active

IP=10.2.1.14 IP=10.2.1.14
IP=10.2.1.14 IP=10.2.1.14
MAC =0007.e910.ce0f MAC =0007.e910.ce0e
MAC =0007.e910.ce0f MAC =0007.e910.ce0f
One Port Receives, All Ports Transmit
On Failover, Src MAC Eth1 = Src MAC Eth0 On Failover, Src MAC Eth1 = Src MAC Eth0
Incorporates Fault Tolerance
IP Address Eth1 = IP Address Eth0 IP Address Eth1 = IP Address Eth0
One IP Address and Multiple MAC Addresses

Note: NIC manufacturer drivers are changing and may


operate differently. Also, server OS have started integrating
NIC teaming drivers which may operate differently.

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 76
Increasing HA in the Data Center
Server Attachment: Multiple NICs
You Can Bundle Multiple Links to Allow
Generating Higher Throughputs Between
Servers and Clients
L3
L2
EtherChannel Hash May
Not Permit Full Utilization
for Certain Applications
(Backup Example)

Only One Link Active:


Fault Tolerant Mode

All Links Active: EtherChannels


Load Balancing

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 77
Increasing HA in the Data Center
Failover: What Is the Time to Beat?

ƒ The overall failover time is the


combination of convergence
at L2, L3, + L4 components
Stateful devices can replicate connection
information and typically failover within 3-5sec
EtherChannels < 1sec
STP converges in ~1 sec (802.1w)
HSRP can be tuned to <1s

ƒ Where does TCP break? Microsoft,


Linux, AIX, etc.
Linux and
Microsoft XP Others Tolerate
2003 Server a Longer
Failover Time

TCP Stack Outage

L4 Convergence Tolerance
~ 5s ~ 9s
L3 Convergence L2 Convergence

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 78
Increasing HA in the Data Center
Failover Time Comparison

ƒ STP-802.1w—One sec
ƒ OSPF-EIGRP—Sub sec
ƒ ACE Module with Autostate
ƒ HSRP—Three sec (using 1/3)
ƒ FWSM Module—Three sec
ƒ CSM Module—Five sec
ƒ WinXP/2003 ServerTCP Stack—Nine sec
Failover Time

TCP Stack
Tolerance
Content ~ 9s
Service
Module
HSRP FireWall ~ 5s
~ 3s Service
(may be tuned Module
Spanning Tree ACE to less) ~ 3s
OSPF/EIGRP ~ 1s
~1sec
Sub-second

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 79
Increasing HA in the Data Center
Non-Stop Forwarding/Stateful Switch-Over

ƒ NSF/SSO is a redundancy NSF-Aware NSF-Aware


mechanism for supervisor failover
ƒ SSO synchronizes layer 2 protocol
state, hardware L2/L3 tables (MAC,
FIB, adjacency table), ACL and QoS
tables
ƒ SSO synchronizes state for:
trunks, interfaces, EtherChannels,
port security, SPAN/RSPAN, STP,
UDLD, VTP
ƒ NSF with EIGRP, OSPF, IS-IS, BGP
makes it possible to have no route
flapping during the recovery
ƒ Aggressive EIGRP/OSPF timers do
not work in NSF/SSO environment

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 80
Increasing HA in the Data Center
NSF/SSO in the Data Center
ƒ Redundant Supervisors with SSO in the
access layer:
Improves availability for single attached
servers
ƒ Redundant Supervisors with SSO in the
aggregation layer:
Consider in primary agg layer switch
Prevents service module switchover
(up to ~5sec depending on module)
SSO switchover time less than two sec
12.2.18SXD3 or higher
ƒ Possible implications
HSRP state between Agg switches is
not tracked and will show switchover until
control plane recovers
IGP Timers cannot be aggressive (tradeoff)

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 81
Increasing HA in the Data Center
Best Practices: STP, HSRP, Other
Rapid PVST+
UDLD Global
Spanning Tree Pathcost Method=Long

LACP+L4 Port Hash


L3+ L4 CEF Hash Dist EtherChannel for FT
and Data VLANs
Agg1: FT
Agg2:
STP Primary Root STP Secondary Root
HSRP Primary Data
HSRP Secondary
HSRP Preempt and Delay HSRP Preempt and Delay
Dual Sup with NSF+SSO Single Sup
LACP+L4 Hash
Dist EtherChannel
Min-Links

Rootguard
LoopGuard Blade Chassis
Portfast + with Integrated
BPDUguard Switch

UDLD Global Rapid PVST+: Maximum Number of STP Active Logical


Ports- 8000 and Virtual Ports Per Linecard-1500

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 82
Q and A

Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 83
Recommended Reading
ƒ Continue your Networkers at
Cisco Live learning
experience with further
reading from Cisco Press
ƒ Check the Recommended
Reading flyer for suggested
books
ƒ Visit www.cisco.com/go/cvd
for Design Guides and
Whitepapers about all of
these subjects and more

Available Onsite at the Cisco Company Store


Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 84
Presentation_ID © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 85

You might also like