Sie sind auf Seite 1von 52

The NGN Carrier Ethernet System:

Technologies, Architecture and Deployment Models

Biren Mehta
Sr. Marketing Manager
Agenda
 Market Trends and its Impact on Network Infrastructure

 Cisco Carrier Ethernet Transport Architecture

 Cisco Carrier Ethernet Portfolio

 TCO Leadership

 Closing Remarks
Market Dynamics & Trends
Affecting Next Generation Internet

Traffic Device Emergence of


Growth Proliferation Cloud

4X volume growth (2010-2015) 12 Billion IP Video capable devices by 2015 $43 Billion cloud services revenue by 2013
80.5 Exabyte per month in 2015 ~ 1 mobile device per capita (7.2B) by 2015 Midmarket firms are more likely to migrate
90%+ will be video 3 Billion IPv6 ready mobile devices by 2014 to the cloud
Shifting Revenue & Traffic
Affecting Next Generation Internet
90+%
2011 2013 2016
IP Traffic
Private Line Private Line Private Line
TDM/OTN TDM/OTN TDM/OTN
Traffic Traffic Traffic

~50-70%* 20-30% 0─10%


Circuit Packet
Private/Public Private/Public Private/Public
IP Traffic IP Traffic IP Traffic
Packet
Legacy TDM
~30-50% 70-80% 90+% Traffic

• SP revenue shifting from circuits to packet services**


5 yrs  ~80% revenue derived from packet services
• Packet traffic increasing at 34% CAGR***, Mobile traffic at 78% CAGR****

• Massive change in SP traffic make-up in next 5 years*


*ACG Research 2011, ** Cisco Research 2010, ***Cisco VNI 2011, ****Cisco Mobile VNI 2012
Traditional Carrier Ethernet Architecture
Business Ethernet Services – Overlay on IP Service Network

SP Services Third Party Content Single-Service

Ethernet Transport

Core
Single-Path

Business – Business
Edge

Single-Access
Access

Fiber Access

Business Business
Next Generation Carrier Ethernet Architecture
Converged Infrastructure for Any Service to Any Access

SP Third-Party Any Service


Content Content
Consumer, Business, Wholesale,
Mobile, Cloud

National
IP Core
National Any Path
Data Center/ Data Center/
Cloud/VHO Cloud/VHO
Client to Client
Business to Business
Cloud to Cloud
Regional Regional
Data Edge Data
Center/VSO Center/VSO
Any Access
Unified Ethernet , PON, Cable, DSL,
Access Wireless

Business
Home
Trends Posing Network Challenges
SP Services/ Third-Party Services/ Content
Content

The
Challenges?
Challenges?

National
Data Center/ Core
Increasing Complexity
National
Data Center/
How do I cost efficiently scale to meet
Cloud/VHO Cloud/VHO
current and future network demands

Regional
Data Edge
Regional
Rising Costs
Data Center/VSOHowcan I cost economically offer
Center/VSO business, residential, mobile services over
a converged infrastructure

Metro
(Access/Agg)
Limited Flexibility
How can I simplify the network while
Multi-point improving network economics?

Business

Requiring Carrier Ethernet Networks to Evolve


Packet Vs. TDM Transport
Chart from Infonetics, Text from DT

TDM transport of packets is no longer economically


viable, lacks statistical multiplexing which makes it very
expensive
Full transformation to NGN needs
to occur from core to customer
Long term vision is critical, this will be the network for the
next decade
What is the most effective
technology choice that will:
–Minimize CapEx and OpEx?
–Provide carrier class service delivery?
–Maximize service agility?

Carriers want the deterministic attributes of transport


networks with the flexibility of the internet
NGN Carrier Ethernet Transport Direction
Characteristic SONET Optical Electrical PBB-TE MPLS-TP IP/MPLS
OTN OTN
/
(ROADMs)
SDH
Eline (10GE)
Eline (sub 10GE)
Ethernet E-Tree
E-LAN Cisco focuses on IP/MPLS for the
F/R Carrier Ethernet Transport
Legacy ATM architecture.
TDM
L3VPN Cisco targets MPLS-TP for the
L3 Unicast POTS and Access Networks while
IP L3 Multicast
supporting already Ethernet Bridged
Content
Traffic Engineering
Access
50ms restoration
Cisco also addresses MPLS to the
Multiplexing Technology Time Wave Division Time Division Statistical Statistical Statistical
General
Division access with Unified MPLS
UNI processing Limited None None Typically rich Typically rich Typically rich

Granularity VC-4 Lambda ODU Variable Variable Variable

Technology Maturity
L2 – Transport
Technology Review

• Two technologies for L2 transport over MPLS:


– Ethernet over MPLS (EoMPLS)
• Used for L2 point-to-point link over MPLS cloud
• No MAC learning involved
– Virtual Private LAN Services (VPLS)
• Used for multipoint L2 connections
• Collection of pseudowires tied together by a Virtual Forwarding
Interface (VFI)
• MAC addresses learned on VFI
• Traffic forwarding based on destination MAC addresses
• H-VPLS, an extension of VPLS

11
EoMPLS
Technology Review

Tunnel Ethernet
VC label
label PDU

Pseudowire
Ethernet
Ethernet PDU
PDU

MPLS
Attachment Circuit Attachment Circuit
LDP
P LDP
Aggregation P
Aggregation
Node
Node
Access Node
Access Node Targeted LDP
FTTB CPE
FTTB CPE

• MPLS in the aggregation network and core


• Targeted LDP session between PEs to exchange VC label
• Tunnel label is used to forward packet from PE to P to PE
• VC label is used to identify L2VPN circuit
• Attachment Circuit (AC) can be port-based or VLAN-based (or Ethernet Flow Point
based, see later)

12
VPLS
Technology Review

MPLS
Aggregation Aggregation Access Node
Access Node Node Core Node

VFI VFI

Attachment
Circuit Aggregation Virtual Forwarding
VFI Instance
Ethernet Port Node
or VLAN

Eompls Virtual Circuit


(Pseudowire)

• Attachment Circuit (AC)—Connection to Aggregation using an Ethernet VLAN


• Virtual Circuit (Pseudowire)—EoMPLS tunnel between PEs using a full mesh
• Virtual Forwarding Instance (VFI)—A virtual L2 bridge instance that connects ACs to VCs (PWs);
VFI=VLAN=broadcast domain
• Enhanced with BGP based Autodiscovery (RFC607)
• Scalability issues almost solved via H-VPLS and state-of-the-art NPU technology (2M MAC address/chip)

13
What is MPLS-TP?
• MPLS-TP – MPLS Transport Profile
– Subset and extension of current MPLS functions
– Connection-oriented transport based on MPLS protocols combined with transport style OAM and protection
mechanisms

Data Plane Control Plane


– Standard MPLS Forwarding – NMS provisioning option

– Bidirectional P2P and P2MP LSPs – GMPLS control plane option

• No LSP merging – PW control plane option


• PHP optional

– Standard PW (SS-PW, MS-PW)


OAM Resiliency
– In-band OAM channel (GACH) – Sub-50ms protection switch over
– Connectivity Check (CC): proactive (BFD) – 1:1, 1+1, 1:N path protection
– Connectivity Verification (CV): reactive (BFD) – Linear protection
– Alarm Suppression and Fault Indication with AIS (new tool), RDI – Ring protection
(BFD), and Client Fault Indication (CFI)
– Performance monitoring, proactive and reactive (new tools)
Business Services Architecture
Optimal Edge and Access Independence
Centralised Service Edge
MPLS/Multicast VPN

EoMPLS Pseudowire
EoMPLS PW EVPN or VPLS

Ethernet, TDM, ATM UNI


FR, HDLC, PPP, ATM IP interworking AToM

Distributed Service Edge MPLS/Multicast VPN

MPLS PWE3 EVPN or VPLS

Ethernet, TDM, ATM UNI


FR, HDLC, PPP, ATM IP interworking AToM

Efficient Large Scale Multiservice


Access Network Aggregation Network Core Network

TR101 MLS Service Edge Node*


Access Node
VPWS, VPLS, VPLS LSM
MPLS/IP Service Edge Node
Core Node
Access Node Aggregation Node

xDSL, xPON, Ethernet MPLS/IP over DWDM PPP, IP, MPLS MPLS
Residential Services Architecture
Optimal Service Edge Placement
Centralised Service Edge,
HSI, VoIP, Video unicast Transport
EoMPLS
EoMPLS PW Pseudowire MPLS/IP (PIM or mLDP)
MPLS/Multicast VPN (mLDP)
MVR VPLS LSM: P2MP PW, mLDP LSP IP, PPPoE Sessions
VFI VFI with IGMP snooping, MVR, IGMP Admission Control
IPTV Transport
Access Node UNI:
Non Trunk, N:1 or 1:1 VLAN; IP PIM, MLDP or mLPD VPN may be used if no wholesale

Distributed Service Edge


3play Unicast
PWE3
MPLS/IP (PIM or mLDP)
PWE3 IP, PPPoE Sessions MPLS/Multicast VPN (mLDP)
VFI
IP TV
Efficient Large Scale Multiservice
Access Network Aggregation Network Core Network

TR101 MLS Service Edge Node*


Access Node
VPWS, VPLS, VPLS LSM
MPLS/IP Service Edge Node
Core Node
Access Node Aggregation Node

xDSL, xPON, Ethernet MPLS/IP over DWDM PPP, IP, MPLS MPLS
Mobile Transport Services Architecture
Simplified, Scalable, and Optimized

BSC
ATM RNC
TDM BTS, ATM NodeB ATM or TDM

S1-U
SAE Gateway
IP eNB Mobile Transport Gateway MME

V4 or v6 MPLS VPN
S1-C
X2-C, X2-U
Mobile Transport Gateway
SAE Gateway

Access Network Aggregation Network Core Network

Mobile Transport Gateway


ASR9000
IP/MPLS Transport IP/MPLS Transport IP/MPLS Transport

Cell Site Gateway Aggregation Node Aggregation Node Core Node Core Node
Fiber or uWave Link, Ring DWDM, Fiber Rings, H&S, Hierarchical Topology DWDM, Fiber Rings, Mesh Topology
Architecture Comparisons
Which one to choose?

The architectures options can be evaluated against the following criteria


• Capital Expenditures
• Scalability (Bandwidth / Subscriber, Transport, Policy Control)
• Operational Complexity (Troubleshooting, QoS)
• Reuse of existing Operations procedures
• Availability
• Traffic Patterns
• Economically serving areas of differing subscriber density
• Service Flexibility
• Operational Flexibility

18
Cisco Carrier Ethernet Portfolio
Intelligent, Scalable, Reliable, and Lowest TCO

Cell Router Aggregation Mobile Edge IP Core

• Flexible scalability: network virtualized (nV), any service, any transport


• Proven performance and reliability: Superior voice & video quality with service assurance
• Operational excellence: Unified management, lower OpEx, and lower power consumption

1 0 0 Ti m e s t he C a pa c i t y f or a Fr a c t i on of C ur r e nt C os t
ASR 9000
Future Proof Edge Portfolio
ASR 9922

Cisco Prime IP NGN 48Tb Per Chassis


20 linecard slots
44 RU
N:1 Switch Fabric Redundancy
11+1 DC Power Module Option
8+8 AC Power Module Option
A to Z management 7 1
Integrated Services Module nV Technology
Integrates multiple platforms
6 2 into a single virtual system
Offer virtualized applications

24x10GE ASR 9000v


5 3 80g Capacity
4x10GE Uplinks
44x1GE Downlinks
4 Redundant -24v/48v DC
Single AC power feed
Linerate 10GE 2x100GE

Industry’s First 2x100GE for Edge


ASR 9000 System
nV Technology

ASR 9001 ASR 9006 ASR 9010 ASR 9922


2 RU 6 slots (¼ rack) 10 slots (½ rack) 22 slots (full rack)
LC / Chassis Up to 12x10GE 4 LC + 2 RSP 8 LC + 2 RSP 20 LC
BW / Chassis 120 Gb 240 Gb 3.2 Tb 6.4 Tb 6.4 Tb 12.8 Tb 48 Tb 96 Tb
nV Technology Availability CY‖12 CY‖12 CY‖12 CY‖12

Double your system capacity by upgrading any ASR 9000 product to an ASR 9000 System
ASR 9000 System Benefits
Powered by nV technology

Multi-Dimensional
Scale
96 Tb capacity
184,320 GigE ports
1,920 10 GigE ports
480 100 GigE ports

Simplify
Service Operations
Single management entity: Edge to Access
Velocity Zero touch configuration
Network virtualization: Edge to Access
Integrated Traffic analytics Single click upgrade
Expanding the ASR Family 2G,3G,4G
Ready
Unified
MPLS
Unified
Ethernet

Extending nV to the Access Architectu


re
Access

Pre-Aggregation 3RU, 360G Switching Capacity


(Unified Ethernet Access) Fully Redundant (RSP,PSU, FANs)
SyncE, 1588
300mm, Environmentally Hardened

ASR 903
SP Edge Cell Site Router
(Small Deployments)
(2G, 3G, 4G Ready)

2RU, 120G Switching Capacity 1RU, 16G System Capacity


nV ASR 901
4 Fixed 10GE ASR 54W GE+TDM, 38W GE
MPA’s: 20x1GE, 2x10GE 9001 Enabled SyncE, 1588
SyncE, 1588 300mm, Environmentally
Hardened

Simplify, Unify, Virtualize Access/Aggregation Infrastructure


ASR 9000 nV Technology Overview
SP Services/ Third-Party
Content Services/

Before: nV Technology
Content
After: nV Technology
Cisco
Prime IP NGN

Each device managed Edge and aggregation


separately. Core managed as one virtual
system through Cisco Prime
IP NGN.
Inconsistent features
between edge and Edge
nV Cluster Single release vehicle
aggregation. offering feature consistency.

Siloed service
domains.
Residential Converged Business
nV Offers up to 71% reduction
in OPEX over 6 years vs
competitors.
Aggregation

Inconsistent service nV Satellite Reduced protocol


outages upon device complexity between edge
and aggregation
failure. Access
Up to 84,480 GE ports
Port scale limited to managed through a single
chassis. virtual system
Network Dual-Homing
Today’s Solution: Protocol based approach

Cellsite MLP
Router Bundle

DACS
CR dual-
IP/Service Edge
homing (MR-
APS)
IP/MPLS
L2 Ethernet Ring
(MST/REP-AG,
G.8032)

Ethernet
spoke-and-hub
(MC-LAG)
L2/L3 service resiliency protocols
HSRP/VRRP, 1-way & 2-way PW
redundancy, BGP PIC
Access dual-homing
protocols
L3 Router dual- MST/REP/G.8032/MS Service state sync
homing (L3 T-AG between two nodes:
ECMP) MC-LAG
DHCP, IGMP, IGMP
MR-APS snooping, ANCP,
L3 IGP/BGP ARP, etc state sync
Network Dual-Homing
Tomorrow’s Solution: Self-Protected Service

Cellsite MLP
Replace two nodes with one single virtual node
Router Bundle  simplify dual-homing to be single-homing

DACS ASR 9000 Cluster


CR dual-
IP/Service Edge
homing (MR-
APS)
IP/MPLS
L2 Ethernet Ring
(MST/REP-AG,
G.8032)

Ethernet
spoke-and-hub
(MC-LAG) No need for L2/L3 service resiliency
protocols:
L2/L3 service resiliency protocols
It is a single Virtual Node.
NO need! It’s SINGLE virtual node
Access single-homing
L3 Router dual- Regular LAG
homing (L3 Single Router APS No need to sync Service state between
two nodes:
ECMP) Single routing Adjacency
All L2 and L3 state are sync’d naturally via
control plane extension
Network Virtualization (nV)
Deployment Scenarios

• L2VPN
– SP 3Play and L2 Business VPN
– DCI (data center inter-connect) (both enterprise and SP DCI)
– Ethernet exchange
• Wireline Aggregation
– L3 termination, no IP session
• BNG (distributed or centralized)
• Wireless Back haul
• L3 CPE aggregation
Network Virtualization (nV)
Deployment Example – L2VPN Service

 Active/standby MC-LAG
A A  bandwidth inefficiency
Active PW  4 PWs with 3 standby
 control plane overhead
Active Active
 PW failover time depends on
the number of PWs
Standby PW  slow convergence
LACP
S
S
LACP
 Require additional state
sync (for example, IGMP
Snooping table) to speed up
Standby Standby
service convergence
Solution1: MC-LAG + 2-way PW redundancy  complex
(Currently the best solution in the market)

 Active/active regular LAG


 Single PW
 Link/Node failure is
protected by LAG, PW is
even not aware  super
fast convergence
 State sync naturally
Solution 2: ASR 9000 Cluster  Simple, fast solution
Network Virtualization (nV)
Deployment Example – L3 Service
 CE dual homing to two PE routers. It has 2
separated L3 interface, and run separated
IGP/BGP session with two PE routers
 Traffic load balance over the two ECMP
paths
 When link or node failure, IGP/BGP
Two Routing adjacency goes down. Protocol re-converge.
Adjacency BGP PIC edge feature is used for fast BGP
convergence
 No state sync between two PE routers

 CE dual homing to one virtual PE. Single


routing adjacency over the link bundle
 Traffic load balance over the link bundle
 When link or node failure, bundle remains
up, so upper layer protocol is even not aware
 super fast convergence, and simple
Single Routing
Adjacency  State sync naturally
What’s nV Satellite ?

Satellite Discovery and Control Protocol


Satellite access Satellite access port is
port represented by the virtual
―nvEthernet‖ interface on
the HOST
Satellite Fabric links

One ASR 9000 nV System ASR 9000 Host

• Install special satellite image on the selected access device to make it ASR 9000 satellite
• Running satellite auto discovery and control protocol to make satellite as ―virtual line card‖ of the ASR 9000 Host
• From end user point of view, it’s single virtual system – ASR 9000 nV System. All management, configuration are
done on the Host chassis
• Satellite and Host could co-locate or in different location. There is no distance limit between satellite and Host
Satellite have zero touch configuration*

* If satellite is connected to Host via L1 link


First Satellite Hardware
ASR 9000v
Field Replaceable Fan Tray
Power Feeds • Redundant Fans
• Redundant -48vDC
1 RU ANSI & ETSI
• ToD/PSS Output
Power Feeds Compliant
• Bits Out
• Single AC power
feed
LEDs

4x10G SFP+
44x10/100/1000 Mbps
Pluggables • Initially used as Fabric Ports ONLY
(could be used as access port in the
• Full Line Rate Packet Processing future)
and Traffic Management
• Plug-n-Play In-Band Management
• Automatic Discovery and Provisioning
• Co-Located or Remote Distribution
• Environmentally Hardened
Satellite Connection Models
L1 Connection – Hub & Spoke

Satellite Single home

Single home with uplink


Satellite bundle

ASR 9000
Cluster

Dual home to cluster (or


Satellite two HOSTs)

ASR 9000
Cluster

Satellite Dual home to cluster (or two


HOSTs) with uplink bundle

3
FTTB Case Study
Cluster + Satellite Deployment Models

• 17 individual devices (12 sites) to manage • 5 ASR9000 nV Systems to manage


• Different platform and OS in small and big POP sites • Common SW feature set
• Same operation complexity in small or big POP sites • Satellite is configured and managed on the nV Host.
Minimal operation on the small POP site  rapid
service deployment

5 ASR 9000 nV Systems

Small POP site Big POP site


(10-80Gbps) nV satellite for the small POP
(>80G) site or for small box in the big
site
Mobile Aggregation Case Study
Cluster + Satellite Deployment Models

• ~nx1000 GE ports
• GE for cell site routers aggregation
• ―unlimited‖ backhaul capacity for growth and for local
• GE ports for local devices devices
• Limited GE density per box

9000v
9000v
CO
LTE Core
9000v
MME
VRF Voice
SGW
GE port per
cell site router VRF RAN
MSC

RNC

VRF MGMT Mgmt


CDMA Core
Cell Site Routers 9000v
9000v
9000v

Cisco Confidential 34
FTTH Case Aggregation/Access
Cluster + Satellite Deployment Models
• One nV system to manage, with ~nx1000 GE ports fan
• Dozens of pre-aggregation/access boxes to out
manage • Simplify the access/aggregation dual-homing by link
• Complex network resiliency solutions bundle: active/active forwarding

cluster

Satellite Satellite Satellite Satellite


Traditional wireline GE
aggregation via L2 switch One ASR 9000 nV System
Managing Cell Site Routers
Cluster + Satellite Deployment Models
• Cell site router become ASR 9000 satellite
• nx1000 cell site Routers to manage
• Single ASR 9000 nV system for management,
• Complex L3 routing, BFD, even L3VPN or L2VPN
configuration on the cell site Routers
configuration and image upgrade
• Zero (or minimal) touch on ASR 9000 satellite. Minimal
feature on satellite

CO
satellite
LTE Core
satellite
MME
satellite VRF Voice
satellite SGW

satellite
satellite VRF RAN
MSC
satellite
satellite RNC

VRF MGMT Mgmt


Cell Site Routers CDMA Core
Satellite*
One ASR 9000 nV System

* It could use different hardware as ASR 9000 satellite for the cell site
Cisco Confidential 36
router instead of the existing ASR 9000v
Virtualized Transport
Value Propositions

 Virtual router is always on


 Towards 50msec failure protection
with very high service scale
 Simplify network protocol based High
resiliency to be internal system control Resiliency
plane based

 Leverage ASR9K HOST


ultra-high MD control plane Low
scale and feature set, remove Cost
complex feature from satellite
 low cost satellite hardware

Operational
 One network element to manage a network cloud Savings
 simple service provisioning, image upgrading,
configuration, etc
 Rapid service deployment  plug-and-play, self-
managed access
Cisco Packet Transport
Value Propositions

Mobile CPT 600


Backhaul
CPT 200
MPLS-TP Ethernet

Ethernet
Services
6 slots (480G) 2 slots (160G) OTN DWDM
Up to 352 ports CPT 50 Up to 176 ports
CPT 50
CPT 50
FTTX & CPT 50
TDM
Fixed config satellite
44xGE, 4x10GE
Hardware Software
Unique satellite architecture MPLS-TP, 802.1ad, H-QoS, E-OAM,
HA: SSO, ISSU, MDR MPLS OAM, Sync-E, 1588, LAG, REP,
Active-Standby Control Plane MVR,IGMPv3
Industry’s first, standards-based, unifying packet transport
Cisco Confidential 38
Cisco Packet Transport
Exceptional Power Savings

Ethernet + TDM
Switching

29.75 10.5
Transponder
inches inches
CPT
ROADM

Over 60% Reduction in Rack Space


Over 65% Reduction in Power Consumption

Calculations based on 480G capacity

Powerful Yet Green and Optimized


POT-S in Metro
Deployment Scenario

IP & MPLS
Routers IP & MPLS
Point-to-point
& multipoint
IPoDWDM
Carrier Ethernet Carrier Ethernet
Private Lines Private Lines
Eth & TDM
Eth & TDM
Point-to-point OTN DWDM Switching

POT-S TDM
TDM

POT-S and IPoDWDM complementary


MPLS from Core to Access
Reducing OpEx, CapEx via simplification
National
Data Center/
Cloud/VHO
IP/MPLS
CRS Single
Core
Forwarding
Rich set of connections
Mechanism
(mesh), P2P / P2MP /
Regional
Data Center/ VSO IP/MPLS MP2MP
Single Control
Edge ASR Plane
XaaS Content 9000
Cache MPLS-TP
Single
Mobile Management
Aggregation MPLS-TP System
Connection-
oriented, P2P /
P2MP
In-band OAM
ME CPT
Access
Satellite

Interop Tested
Business

Transport Trust + Packet Efficiency = 20% OpEx Savings


Scaling MPLS Services
With Converged Infrastructure
• Scale - Interconnect 100k Access nodes through an MPLS domain
• Resilience - < 50msec convergence as often as possible
• Simplicity - Operation of big MPLS networks is often considered difficult

Reference Model

PE11 PE2
ABR11 ABR21 Distribution 1
Distribution / /
Core and Edge Aggregation
Aggregation
DSLAM1 DSLAM2

PE12 PE2
ABR12 ABR22 2

1k Nodes / Core

10k Nodes / Aggregation

100k Nodes / Access


Solution - Unified MPLS
Carrier Ethernet Transport Architecture

Cloud
Virtualized
Functions
Consumer
Apps
Video
Processing
Billing
Compute Core
Svc Delivery
Origin
Server
Encryption
Data
Transcode
Center
Device Mgmt
Storage
Switch Edge Unified
MPLS
 Converged: Any Service, Any Aggregation
Path, Any Access
 Operationally Simple: Single
Control Plane Client
 Carrier Class: Fast Reroute and Devices

Network Convergence
Consumer Business Mobile
Assess Solution Business Value
Carrier Ethernet Transport Architecture
ASR 9000 System

Single Management
Entity

How does the total solution


Translate to business value?

Zero Touch
(ASR 9001, ASR 903, ASR 901) (nV ) Configuration

Integrated
Traffic Analytics

Multi- Simplify Service


Competitive TCO Analysis
Input Various TCO Parameters

• Average Sales Price (ASP)


CAPEX
• Engineering, Furnishing and Installation (EF&I)
CAPEX
Mobile
• Power
Res TCO OPEX
Biz • Cooling
• Floor Space
• Network Care (provisioning, fault management,
3 Year OPEX performance management)
Period • Software upgrades
• Vendor maintenance
Requirements & Assumptions
Comparable Architectures

Infrastructure Convergence Network Assumptions over


(Converging Network Silo’ed Domains in the access)
3 Years

1 • 3,500,000 residences in metro area


= with Video, VoIP and Internet.

• 50,000 business establishments in


Operation Simplification metro area with L2 VPN, L3 VPN and
(Integrated Traffic Analytics, Zero Touch Configuration, Optimized Power)
Internet.

2 • 7,000,000 mobile customers in metro


area with Voice, Data and SMS.

Infrastructure Redundancy
(Route Processors, Power Supplies & Fans)

3
Competing Architectures
Breaking the Backhaul OpEX Barriers
Lowest Power Consumption in the Industry

Only 2.375 W/Gbps at cell, 2.5W/Gbps at pre-Agg


(ASR 901-12C-F-D 38W, 16 Gbps)

5 year cell-site power savings NPV is $20+ Million


1 year savings by removing dedicated T1 timing is $48 Million
i.e., Verizon 40,000, Bharti 40,000 cell sites
$100/month for E1/T1
Cisco Carrier Ethernet Architecture
Business Differentiators

Solutions Payback from Cisco® Converged ASR9000


5 System with nV

Cost reduction from


68 Cisco® Ethernet Energy Savings

TCO reduction from


70 Cisco® ASR9000 System with nV
Q&A

#CiscoPlusCA
We value your feedback.
Please be sure to complete the Evaluation Form for this session.

Access today’s presentations at cisco.com/ca/plus

Follow @CiscoCanada and join the #CiscoPlusCA conversation


NMS for Network Management
or Dynamic Control Plane
Working LSP

Client node PE PE Client node


Protect LSP
MPLS-TP LSP (Static or Dynamic)
Pseudowire with e2e and
Section Section segment OAM
Client Signal

Connection Oriented, pre-determined working path and protect path


Transport Tunnel 1:1 protection, switching triggered by in-band OAM,
Options
51 with NMS for static provisioning, or dynamic control plane for routing and signaling
51
ASR 9000 Virtual Chassis Overview
Control Plane EOBC Extension (L1 or L2 connection) Special external EOBC 1G/10G • Single control and
One or two 10G/1G from each RSP port s on RSP (new RSP) management plane,
distributed data
plane  one virtual
chassis
0 1 • Control plane EOBC
Active
RSP
Standby
RSP
Active
RSP
Standby
RSP Internal extension is through
EOBC special RSP
onboard 1G or 10G
ports
• Data plane
LC LC LC LC LC LC LC LC

extension is through
regular LC ports (it
can even mix regular
data ports and
Inter-chassis data link (L1 Regular 10G or 100G data
virtual chassis data
connection) ports (Current or future line plane ports on the
10G or 100 G bundle (up to 32 ports) card) same LC), doesnt
require fabric
chassis  flexible
deployment
5
Scale, Simplify, Virtualizes
Extending Cisco ASR 9000 System to Access & Mobile Networks

ASR 9000 System SP Benefits


ASR 9922

ASR 9010
Multi- Simplify Service
ASR 9006 Dimensional Operations Velocity
ASR 9001 Scale

ASR 903
ASR
9000v
Single 96 Tb IPv6 System
ASR 901

36x More Capacity than the Closest


Competitive Platform

Das könnte Ihnen auch gefallen