• Share
  • Email
  • Embed
  • Like
  • Save
  • Private Content
The NGN Carrier Ethernet System: Technologies, Architecture and Deployment Models
 

The NGN Carrier Ethernet System: Technologies, Architecture and Deployment Models

on

  • 1,817 views

This presentation discusses market trends and its impact on Network infrastructure, Cisco carrier Ethernet Transport Architecture, Cisco carrier Ethernet portfolio and TCO Leadership.

This presentation discusses market trends and its impact on Network infrastructure, Cisco carrier Ethernet Transport Architecture, Cisco carrier Ethernet portfolio and TCO Leadership.

Statistics

Views

Total Views
1,817
Views on SlideShare
1,817
Embed Views
0

Actions

Likes
3
Downloads
200
Comments
0

0 Embeds 0

No embeds

Accessibility

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

    The NGN Carrier Ethernet System: Technologies, Architecture and Deployment Models The NGN Carrier Ethernet System: Technologies, Architecture and Deployment Models Presentation Transcript

    • The NGN Carrier Ethernet System:Technologies, Architecture and Deployment Models Biren Mehta Sr. Marketing Manager
    • Agenda Market Trends and its Impact on Network Infrastructure Cisco Carrier Ethernet Transport Architecture Cisco Carrier Ethernet Portfolio TCO Leadership Closing Remarks
    • Market Dynamics & TrendsAffecting Next Generation Internet Traffic Device Emergence of Growth Proliferation Cloud 4X volume growth (2010-2015) 12 Billion IP Video capable devices by 2015 $43 Billion cloud services revenue by 2013 80.5 Exabyte per month in 2015 ~ 1 mobile device per capita (7.2B) by 2015 Midmarket firms are more likely to migrate 90%+ will be video 3 Billion IPv6 ready mobile devices by 2014 to the cloud
    • Shifting Revenue & Traffic Affecting Next Generation Internet 90+% 2011 2013 2016 IP Traffic Private Line Private Line Private Line TDM/OTN TDM/OTN TDM/OTN Traffic Traffic Traffic ~50-70%* 20-30% 0─10% Circuit Packet Private/Public Private/Public Private/Public IP Traffic IP Traffic IP Traffic Packet Legacy TDM ~30-50% 70-80% 90+% Traffic • SP revenue shifting from circuits to packet services** 5 yrs  ~80% revenue derived from packet services • Packet traffic increasing at 34% CAGR***, Mobile traffic at 78% CAGR**** • Massive change in SP traffic make-up in next 5 years**ACG Research 2011, ** Cisco Research 2010, ***Cisco VNI 2011, ****Cisco Mobile VNI 2012
    • Traditional Carrier Ethernet ArchitectureBusiness Ethernet Services – Overlay on IP Service Network SP Services Third Party Content Single-Service Ethernet Transport Core Single-Path Business – Business Edge Single-Access Access Fiber Access Business Business
    • Next Generation Carrier Ethernet ArchitectureConverged Infrastructure for Any Service to Any Access SP Third-Party Any Service Content Content Consumer, Business, Wholesale, Mobile, Cloud National IP Core National Any Path Data Center/ Data Center/ Cloud/VHO Cloud/VHO Client to Client Business to Business Cloud to Cloud Regional Regional Data Edge Data Center/VSO Center/VSO Any Access Unified Ethernet , PON, Cable, DSL, Access Wireless Business Home
    • Trends Posing Network Challenges SP Services/ Third-Party Services/ Content Content The Challenges? Challenges? National Data Center/ Core Increasing Complexity National Data Center/ do I cost efficiently scale to meet How Cloud/VHO Cloud/VHO current and future network demands Regional Data Edge Regional Rising Costs Data Center/VSOHowcan I cost economically offer Center/VSO business, residential, mobile services over a converged infrastructure Metro (Access/Agg) Limited Flexibility How can I simplify the network while Multi-point improving network economics? Business Requiring Carrier Ethernet Networks to Evolve
    • Packet Vs. TDM TransportChart from Infonetics, Text from DT TDM transport of packets is no longer economically viable, lacks statistical multiplexing which makes it very expensive Full transformation to NGN needs to occur from core to customer Long term vision is critical, this will be the network for the next decade What is the most effective technology choice that will: –Minimize CapEx and OpEx? –Provide carrier class service delivery? –Maximize service agility? Carriers want the deterministic attributes of transport networks with the flexibility of the internet
    • NGN Carrier Ethernet Transport Direction Characteristic SONET Optical Electrical PBB-TE MPLS-TP IP/MPLS OTN OTN / (ROADMs) SDH Eline (10GE) Eline (sub 10GE) Ethernet E-Tree E-LAN Cisco focuses on IP/MPLS for the F/R Carrier Ethernet Transport Legacy ATM architecture. TDM L3VPN Cisco targets MPLS-TP for the L3 Unicast POTS and Access Networks while IP L3 Multicast supporting already Ethernet Bridged Content Traffic Engineering Access 50ms restoration Cisco also addresses MPLS to the Multiplexing Technology Time Wave Division Time Division Statistical Statistical Statistical General Division access with Unified MPLS UNI processing Limited None None Typically rich Typically rich Typically rich Granularity VC-4 Lambda ODU Variable Variable Variable Technology Maturity
    • L2 – TransportTechnology Review • Two technologies for L2 transport over MPLS: – Ethernet over MPLS (EoMPLS) • Used for L2 point-to-point link over MPLS cloud • No MAC learning involved – Virtual Private LAN Services (VPLS) • Used for multipoint L2 connections • Collection of pseudowires tied together by a Virtual Forwarding Interface (VFI) • MAC addresses learned on VFI • Traffic forwarding based on destination MAC addresses • H-VPLS, an extension of VPLS 11
    • EoMPLSTechnology Review Tunnel Ethernet VC label label PDU Pseudowire Ethernet Ethernet PDU PDU MPLS Attachment Circuit Attachment Circuit LDP P LDP Aggregation P Aggregation Node Node Access Node Access Node Targeted LDP FTTB CPE FTTB CPE • MPLS in the aggregation network and core • Targeted LDP session between PEs to exchange VC label • Tunnel label is used to forward packet from PE to P to PE • VC label is used to identify L2VPN circuit • Attachment Circuit (AC) can be port-based or VLAN-based (or Ethernet Flow Point based, see later) 12
    • VPLSTechnology Review MPLS Aggregation Aggregation Access Node Access Node Node Core Node VFI VFI Attachment Circuit Aggregation Virtual Forwarding VFI Instance Ethernet Port Node or VLAN Eompls Virtual Circuit (Pseudowire) • Attachment Circuit (AC)—Connection to Aggregation using an Ethernet VLAN • Virtual Circuit (Pseudowire)—EoMPLS tunnel between PEs using a full mesh • Virtual Forwarding Instance (VFI)—A virtual L2 bridge instance that connects ACs to VCs (PWs); VFI=VLAN=broadcast domain • Enhanced with BGP based Autodiscovery (RFC607) • Scalability issues almost solved via H-VPLS and state-of-the-art NPU technology (2M MAC address/chip) 13
    • What is MPLS-TP? • MPLS-TP – MPLS Transport Profile – Subset and extension of current MPLS functions – Connection-oriented transport based on MPLS protocols combined with transport style OAM and protection mechanismsData Plane Control Plane – Standard MPLS Forwarding – NMS provisioning option – Bidirectional P2P and P2MP LSPs – GMPLS control plane option • No LSP merging – PW control plane option • PHP optional – Standard PW (SS-PW, MS-PW)OAM Resiliency – In-band OAM channel (GACH) – Sub-50ms protection switch over – Connectivity Check (CC): proactive (BFD) – 1:1, 1+1, 1:N path protection – Connectivity Verification (CV): reactive (BFD) – Linear protection – Alarm Suppression and Fault Indication with AIS (new tool), RDI – Ring protection (BFD), and Client Fault Indication (CFI) – Performance monitoring, proactive and reactive (new tools)
    • Business Services ArchitectureOptimal Edge and Access Independence Centralised Service Edge MPLS/Multicast VPN EoMPLSPW EoMPLS Pseudowire EVPN or VPLS Ethernet, TDM, ATM UNI FR, HDLC, PPP, ATM IP interworking AToM Distributed Service Edge MPLS/Multicast VPN MPLS PWE3 EVPN or VPLS Ethernet, TDM, ATM UNI FR, HDLC, PPP, ATM IP interworking AToM Efficient Large Scale Multiservice Access Network Aggregation Network Core Network TR101 MLS Service Edge Node* Access Node VPWS, VPLS, VPLS LSM MPLS/IP Service Edge Node Core Node Access Node Aggregation Node xDSL, xPON, Ethernet MPLS/IP over DWDM PPP, IP, MPLS MPLS
    • Residential Services ArchitectureOptimal Service Edge PlacementCentralised Service Edge, HSI, VoIP, Video unicast Transport EoMPLS EoMPLS PW Pseudowire MPLS/IP (PIM or mLDP) MPLS/Multicast VPN (mLDP) MVR VPLS LSM: P2MP PW, mLDP LSP IP, PPPoE Sessions VFI VFI with IGMP snooping, MVR, IGMP Admission Control IPTV Transport Access Node UNI: Non Trunk, N:1 or 1:1 VLAN; IP PIM, MLDP or mLPD VPN may be used if no wholesaleDistributed Service Edge3play Unicast PWE3 MPLS/IP (PIM or mLDP) PWE3 IP, PPPoE Sessions MPLS/Multicast VPN (mLDP) VFI IP TV Efficient Large Scale Multiservice Access Network Aggregation Network Core Network TR101 MLS Service Edge Node* Access Node VPWS, VPLS, VPLS LSM MPLS/IP Service Edge Node Core Node Access Node Aggregation Node xDSL, xPON, Ethernet MPLS/IP over DWDM PPP, IP, MPLS MPLS
    • Mobile Transport Services ArchitectureSimplified, Scalable, and Optimized BSC ATM RNC TDM BTS, ATM NodeB ATM or TDM S1-U SAE Gateway IP eNB Mobile Transport Gateway MME V4 or v6 MPLS VPN S1-C X2-C, X2-U Mobile Transport Gateway SAE Gateway Access Network Aggregation Network Core Network Mobile Transport Gateway ASR9000 IP/MPLS Transport IP/MPLS Transport IP/MPLS Transport Cell Site Gateway Aggregation Node Aggregation Node Core Node Core Node Fiber or uWave Link, Ring DWDM, Fiber Rings, H&S, Hierarchical Topology DWDM, Fiber Rings, Mesh Topology
    • Architecture ComparisonsWhich one to choose? The architectures options can be evaluated against the following criteria • Capital Expenditures • Scalability (Bandwidth / Subscriber, Transport, Policy Control) • Operational Complexity (Troubleshooting, QoS) • Reuse of existing Operations procedures • Availability • Traffic Patterns • Economically serving areas of differing subscriber density • Service Flexibility • Operational Flexibility 18
    • Cisco Carrier Ethernet PortfolioIntelligent, Scalable, Reliable, and Lowest TCO Cell Router Aggregation Mobile Edge IP Core • Flexible scalability: network virtualized (nV), any service, any transport • Proven performance and reliability: Superior voice & video quality with service assurance • Operational excellence: Unified management, lower OpEx, and lower power consumption 1 0 0 Ti m e s t he C a pa c i t y f or a Fr a c t i on of C ur r e nt C os t
    • ASR 9000Future Proof Edge Portfolio ASR 9922 Cisco Prime IP NGN 48Tb Per Chassis 20 linecard slots 44 RU N:1 Switch Fabric Redundancy 11+1 DC Power Module Option 8+8 AC Power Module Option A to Z management 7 1 Integrated Services Module nV Technology Integrates multiple platforms 6 2 into a single virtual system Offer virtualized applications 24x10GE ASR 9000v 5 3 80g Capacity 4x10GE Uplinks 44x1GE Downlinks 4 Redundant -24v/48v DC Single AC power feed Linerate 10GE 2x100GE Industry’s First 2x100GE for Edge
    • ASR 9000 System nV Technology ASR 9001 ASR 9006 ASR 9010 ASR 9922 2 RU 6 slots (¼ rack) 10 slots (½ rack) 22 slots (full rack)LC / Chassis Up to 12x10GE 4 LC + 2 RSP 8 LC + 2 RSP 20 LCBW / Chassis 120 Gb 240 Gb 3.2 Tb 6.4 Tb 6.4 Tb 12.8 Tb 48 Tb 96 TbnV Technology Availability CY‖12 CY‖12 CY‖12 CY‖12 Double your system capacity by upgrading any ASR 9000 product to an ASR 9000 System
    • ASR 9000 System BenefitsPowered by nV technology Multi-Dimensional Scale 96 Tb capacity 184,320 GigE ports 1,920 10 GigE ports 480 100 GigE ports Simplify Service Operations Single management entity: Edge to Access Velocity Zero touch configuration Network virtualization: Edge to Access Integrated Traffic analytics Single click upgrade
    • Expanding the ASR Family 2G,3G,4G Ready Unified MPLS Unified EthernetExtending nV to the Access Architectu re Access Pre-Aggregation 3RU, 360G Switching Capacity (Unified Ethernet Access) Fully Redundant (RSP,PSU, FANs) SyncE, 1588 300mm, Environmentally Hardened ASR 903 SP Edge Cell Site Router (Small Deployments) (2G, 3G, 4G Ready) 2RU, 120G Switching Capacity 1RU, 16G System Capacity nV ASR 901 4 Fixed 10GE ASR 54W GE+TDM, 38W GE MPA’s: 20x1GE, 2x10GE 9001 Enabled SyncE, 1588 SyncE, 1588 300mm, Environmentally Hardened Simplify, Unify, Virtualize Access/Aggregation Infrastructure
    • ASR 9000 nV Technology Overview SP Services/ Third-Party Content Services/ Before: nV Technology Content After: nV Technology Cisco Prime IP NGN Each device managed Edge and aggregation separately. Core managed as one virtual system through Cisco Prime IP NGN. Inconsistent features between edge and Edge nV Cluster Single release vehicle aggregation. offering feature consistency. Siloed service domains. Residential Converged Business nV Offers up to 71% reduction in OPEX over 6 years vs competitors. Aggregation Inconsistent service nV Satellite Reduced protocol outages upon device complexity between edge and aggregation failure. Access Up to 84,480 GE ports Port scale limited to managed through a single chassis. virtual system
    • Network Dual-HomingToday’s Solution: Protocol based approach Cellsite MLP Router Bundle DACS CR dual- IP/Service Edge homing (MR- APS) IP/MPLS L2 Ethernet Ring (MST/REP-AG, G.8032) Ethernet spoke-and-hub (MC-LAG) L2/L3 service resiliency protocols HSRP/VRRP, 1-way & 2-way PW redundancy, BGP PIC Access dual-homing protocols L3 Router dual- MST/REP/G.8032/MS Service state sync homing (L3 T-AG between two nodes: ECMP) MC-LAG DHCP, IGMP, IGMP MR-APS snooping, ANCP, L3 IGP/BGP ARP, etc state sync
    • Network Dual-HomingTomorrow’s Solution: Self-Protected Service Cellsite MLP Replace two nodes with one single virtual node Router Bundle  simplify dual-homing to be single-homing DACS ASR 9000 Cluster CR dual- IP/Service Edge homing (MR- APS) IP/MPLS L2 Ethernet Ring (MST/REP-AG, G.8032) Ethernet spoke-and-hub (MC-LAG) No need for L2/L3 service resiliency protocols: L2/L3 service resiliency protocols It is a single Virtual Node. NO need! It’s SINGLE virtual node Access single-homing L3 Router dual- Regular LAG homing (L3 Single Router APS No need to sync Service state between two nodes: ECMP) Single routing Adjacency All L2 and L3 state are sync’d naturally via control plane extension
    • Network Virtualization (nV)Deployment Scenarios • L2VPN – SP 3Play and L2 Business VPN – DCI (data center inter-connect) (both enterprise and SP DCI) – Ethernet exchange • Wireline Aggregation – L3 termination, no IP session • BNG (distributed or centralized) • Wireless Back haul • L3 CPE aggregation
    • Network Virtualization (nV)Deployment Example – L2VPN Service  Active/standby MC-LAG A A  bandwidth inefficiency Active PW  4 PWs with 3 standby  control plane overhead Active Active  PW failover time depends on the number of PWs Standby PW  slow convergence LACP S S LACP  Require additional state sync (for example, IGMP Snooping table) to speed up Standby Standby service convergence Solution1: MC-LAG + 2-way PW redundancy  complex (Currently the best solution in the market)  Active/active regular LAG  Single PW  Link/Node failure is protected by LAG, PW is even not aware  super fast convergence  State sync naturally Solution 2: ASR 9000 Cluster  Simple, fast solution
    • Network Virtualization (nV)Deployment Example – L3 Service  CE dual homing to two PE routers. It has 2 separated L3 interface, and run separated IGP/BGP session with two PE routers  Traffic load balance over the two ECMP paths  When link or node failure, IGP/BGP Two Routing adjacency goes down. Protocol re-converge. Adjacency BGP PIC edge feature is used for fast BGP convergence  No state sync between two PE routers  CE dual homing to one virtual PE. Single routing adjacency over the link bundle  Traffic load balance over the link bundle  When link or node failure, bundle remains up, so upper layer protocol is even not aware  super fast convergence, and simple Single Routing Adjacency  State sync naturally
    • What’s nV Satellite ? Satellite Discovery and Control ProtocolSatellite access Satellite access port isport represented by the virtual ―nvEthernet‖ interface on the HOST Satellite Fabric links One ASR 9000 nV System ASR 9000 Host • Install special satellite image on the selected access device to make it ASR 9000 satellite • Running satellite auto discovery and control protocol to make satellite as ―virtual line card‖ of the ASR 9000 Host • From end user point of view, it’s single virtual system – ASR 9000 nV System. All management, configuration are done on the Host chassis • Satellite and Host could co-locate or in different location. There is no distance limit between satellite and Host Satellite have zero touch configuration* * If satellite is connected to Host via L1 link
    • First Satellite HardwareASR 9000v Field Replaceable Fan Tray Power Feeds • Redundant Fans • Redundant -48vDC 1 RU ANSI & ETSI • ToD/PSS Output Power Feeds Compliant • Bits Out • Single AC power feed LEDs 4x10G SFP+ 44x10/100/1000 Mbps Pluggables • Initially used as Fabric Ports ONLY (could be used as access port in the • Full Line Rate Packet Processing future) and Traffic Management • Plug-n-Play In-Band Management • Automatic Discovery and Provisioning • Co-Located or Remote Distribution • Environmentally Hardened
    • Satellite Connection ModelsL1 Connection – Hub & Spoke Satellite Single home Single home with uplink Satellite bundle ASR 9000 Cluster Dual home to cluster (or Satellite two HOSTs) ASR 9000 Cluster Satellite Dual home to cluster (or two HOSTs) with uplink bundle 3
    • FTTB Case StudyCluster + Satellite Deployment Models • 17 individual devices (12 sites) to manage • 5 ASR9000 nV Systems to manage • Different platform and OS in small and big POP sites • Common SW feature set • Same operation complexity in small or big POP sites • Satellite is configured and managed on the nV Host. Minimal operation on the small POP site  rapid service deployment 5 ASR 9000 nV Systems Small POP site Big POP site (10-80Gbps) nV satellite for the small POP (>80G) site or for small box in the big site
    • Mobile Aggregation Case StudyCluster + Satellite Deployment Models • ~nx1000 GE ports • GE for cell site routers aggregation • ―unlimited‖ backhaul capacity for growth and for local • GE ports for local devices devices • Limited GE density per box 9000v 9000v CO LTE Core 9000v MME VRF Voice SGW GE port per cell site router VRF RAN MSC RNC VRF MGMT Mgmt CDMA Core Cell Site Routers 9000v 9000v 9000v Cisco Confidential 34
    • FTTH Case Aggregation/AccessCluster + Satellite Deployment Models • One nV system to manage, with ~nx1000 GE ports fan • Dozens of pre-aggregation/access boxes to out manage • Simplify the access/aggregation dual-homing by link • Complex network resiliency solutions bundle: active/active forwarding cluster Satellite Satellite Satellite Satellite Traditional wireline GE aggregation via L2 switch One ASR 9000 nV System
    • Managing Cell Site RoutersCluster + Satellite Deployment Models • Cell site router become ASR 9000 satellite • nx1000 cell site Routers to manage • Single ASR 9000 nV system for management, • Complex L3 routing, BFD, even L3VPN or L2VPN configuration on the cell site Routers configuration and image upgrade • Zero (or minimal) touch on ASR 9000 satellite. Minimal feature on satellite CO satellite LTE Core satellite MME satellite VRF Voice satellite SGW satellite satellite VRF RAN MSC satellite satellite RNC VRF MGMT Mgmt Cell Site Routers CDMA Core Satellite* One ASR 9000 nV System * It could use different hardware as ASR 9000 satellite for the cell site Cisco Confidential 36 router instead of the existing ASR 9000v
    • Virtualized TransportValue Propositions  Virtual router is always on  Towards 50msec failure protection with very high service scale  Simplify network protocol based High resiliency to be internal system control Resiliency plane based  Leverage ASR9K HOST ultra-high MD control plane Low scale and feature set, remove Cost complex feature from satellite  low cost satellite hardware Operational  One network element to manage a network cloud Savings  simple service provisioning, image upgrading, configuration, etc  Rapid service deployment  plug-and-play, self- managed access
    • Cisco Packet TransportValue Propositions Mobile CPT 600 Backhaul CPT 200 MPLS-TP Ethernet Ethernet Services 6 slots (480G) 2 slots (160G) OTN DWDM Up to 352 ports CPT 50 Up to 176 ports CPT 50 CPT 50 FTTX & CPT 50 TDM Fixed config satellite 44xGE, 4x10GE Hardware Software Unique satellite architecture MPLS-TP, 802.1ad, H-QoS, E-OAM, HA: SSO, ISSU, MDR MPLS OAM, Sync-E, 1588, LAG, REP, Active-Standby Control Plane MVR,IGMPv3 Industry’s first, standards-based, unifying packet transport Cisco Confidential 38
    • Cisco Packet TransportExceptional Power Savings Ethernet + TDM Switching 29.75 10.5 Transponder inches inches CPT ROADM Over 60% Reduction in Rack Space Over 65% Reduction in Power ConsumptionCalculations based on 480G capacity Powerful Yet Green and Optimized
    • POT-S in MetroDeployment Scenario IP & MPLS Routers IP & MPLS Point-to-point & multipoint IPoDWDMCarrier Ethernet Carrier EthernetPrivate Lines Private LinesEth & TDM Eth & TDM Point-to-point OTN DWDM Switching POT-S TDM TDM POT-S and IPoDWDM complementary
    • MPLS from Core to AccessReducing OpEx, CapEx via simplification National Data Center/ Cloud/VHO IP/MPLS CRS SingleCore Forwarding Rich set of connections Mechanism (mesh), P2P / P2MP / Regional Data Center/ VSO IP/MPLS MP2MP Single ControlEdge ASR Plane XaaS Content 9000 Cache MPLS-TP Single Mobile ManagementAggregation MPLS-TP System Connection- oriented, P2P / P2MP In-band OAM ME CPTAccess Satellite Interop Tested Business Transport Trust + Packet Efficiency = 20% OpEx Savings
    • Scaling MPLS ServicesWith Converged Infrastructure • Scale - Interconnect 100k Access nodes through an MPLS domain • Resilience - < 50msec convergence as often as possible • Simplicity - Operation of big MPLS networks is often considered difficult Reference Model PE11 PE2 ABR11 ABR21 Distribution 1 Distribution / / Core and Edge Aggregation Aggregation DSLAM1 DSLAM2 PE12 PE2 ABR12 ABR22 2 1k Nodes / Core 10k Nodes / Aggregation 100k Nodes / Access
    • Solution - Unified MPLSCarrier Ethernet Transport Architecture Cloud Virtualized Functions Consumer Apps Video Processing Billing Compute Core Svc Delivery Origin Server Encryption Data Transcode Center Device Mgmt Storage Switch Edge Unified MPLS  Converged: Any Service, Any Aggregation Path, Any Access  Operationally Simple: Single Control Plane Client  Carrier Class: Fast Reroute and Devices Network Convergence Consumer Business Mobile
    • Assess Solution Business ValueCarrier Ethernet Transport Architecture ASR 9000 System Single Management Entity How does the total solution Translate to business value? Zero Touch (ASR 9001, ASR 903, ASR 901) (nV ) Configuration Integrated Traffic Analytics Multi- Simplify Service
    • Competitive TCO AnalysisInput Various TCO Parameters • Average Sales Price (ASP) CAPEX • Engineering, Furnishing and Installation (EF&I) CAPEX Mobile • Power Res TCO OPEX Biz • Cooling • Floor Space • Network Care (provisioning, fault management, 3 Year OPEX performance management) Period • Software upgrades • Vendor maintenance
    • Requirements & AssumptionsComparable Architectures Infrastructure Convergence Network Assumptions over (Converging Network Silo’ed Domains in the access) 3 Years 1 • 3,500,000 residences in metro area = with Video, VoIP and Internet. • 50,000 business establishments in Operation Simplification metro area with L2 VPN, L3 VPN and (Integrated Traffic Analytics, Zero Touch Configuration, Optimized Power) Internet.2 • 7,000,000 mobile customers in metro area with Voice, Data and SMS. Infrastructure Redundancy (Route Processors, Power Supplies & Fans) 3 Competing Architectures
    • Breaking the Backhaul OpEX BarriersLowest Power Consumption in the Industry Only 2.375 W/Gbps at cell, 2.5W/Gbps at pre-Agg (ASR 901-12C-F-D 38W, 16 Gbps) 5 year cell-site power savings NPV is $20+ Million 1 year savings by removing dedicated T1 timing is $48 Million i.e., Verizon 40,000, Bharti 40,000 cell sites $100/month for E1/T1
    • Cisco Carrier Ethernet ArchitectureBusiness Differentiators Solutions Payback from Cisco® Converged ASR9000 5 System with nV Cost reduction from 68 Cisco® Ethernet Energy Savings TCO reduction from 70 Cisco® ASR9000 System with nV
    • Q&A #CiscoPlusCA
    • We value your feedback.Please be sure to complete the Evaluation Form for this session. Access today’s presentations at cisco.com/ca/plus Follow @CiscoCanada and join the #CiscoPlusCA conversation
    • NMS for Network Management or Dynamic Control Plane Working LSP Client node PE PE Client node Protect LSP MPLS-TP LSP (Static or Dynamic) Pseudowire with e2e and Section Section segment OAM Client SignalConnection Oriented, pre-determined working path and protect pathTransport Tunnel 1:1 protection, switching triggered by in-band OAM,Options with NMS for static provisioning, or dynamic control plane for routing and signaling51 51
    • ASR 9000 Virtual Chassis Overview Control Plane EOBC Extension (L1 or L2 connection) Special external EOBC 1G/10G • Single control and One or two 10G/1G from each RSP port s on RSP (new RSP) management plane, distributed data plane  one virtual chassis 0 1 • Control plane EOBC Active RSP Standby RSP Active RSP Standby RSP Internal extension is through EOBC special RSP onboard 1G or 10G ports • Data plane LC LC LC LC LC LC LC LC extension is through regular LC ports (it can even mix regular data ports and Inter-chassis data link (L1 Regular 10G or 100G data virtual chassis data connection) ports (Current or future line plane ports on the 10G or 100 G bundle (up to 32 ports) card) same LC), doesnt require fabric chassis  flexible deployment 5
    • Scale, Simplify, VirtualizesExtending Cisco ASR 9000 System to Access & Mobile Networks ASR 9000 System SP Benefits ASR 9922 ASR 9010 Multi- Simplify Service ASR 9006 Dimensional Operations Velocity ASR 9001 Scale ASR 903 ASR 9000v Single 96 Tb IPv6 System ASR 901 36x More Capacity than the Closest Competitive Platform