Mellanox Approach to NFV & SDN
 

Mellanox Approach to NFV & SDN

on

  • 1,291 views

Presented by Eran Bello at the "NFV & SDN Summit" held March 2014 in Paris, France ...

Presented by Eran Bello at the "NFV & SDN Summit" held March 2014 in Paris, France

Ideal for Cloud DataCenter, Data Processing Platforms and Network Functions Virtualization
Leading SerDes Technology: High Bandwidth – Advanced Process
10/40/56Gb VPI with PCIe 3.0 Interface
10/40/56Gb High Bandwidth Switch: 36 ports of 10/40/56Gb or 64 ports of 10Gb
RDMA/RoCE technology: Ultra Low Latency Data Transfer
Software Defined Networking: SDN Switch and Control End to End Solution
Cloud Management: OpenStack integration

Paving the way to 100Gb/s Interconnect
End to End Network Interconnect for Compute/Processing and Switching
Software Defined Networking

High Bandwidth, Low Latency and Lower TCO: $/Port/Gb

Statistics

Views

Total Views
1,291
Views on SlideShare
1,260
Embed Views
31

Actions

Likes
1
Downloads
69
Comments
0

4 Embeds 31

https://twitter.com 21
http://www.slideee.com 6
http://www.pinterest.com 2
http://www.linkedin.com 2

Accessibility

Categories

Upload Details

Uploaded via as Microsoft PowerPoint

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

Mellanox Approach to NFV & SDN Mellanox Approach to NFV & SDN Presentation Transcript

  • Eran Bello, Director of Business Development March 2014 | NFV&SDN Summit | Paris, France Mellanox Approach to NFV & SDN
  • © 2014 Mellanox Technologies 2 Leading Supplier of End-to-End Interconnect Solutions Virtual Protocol Interconnect Storage Front / Back-End Server / Compute Switch / Gateway 56G IB & FCoIB 56G InfiniBand 10/40/56GbE & FCoE 10/40/56GbE Virtual Protocol Interconnect Host/Fabric SoftwareICs Switches/GatewaysAdapter Cards Cables/Modules Comprehensive End-to-End InfiniBand and Ethernet Portfolio Metro / WAN
  • © 2014 Mellanox Technologies 3 Virtual Protocol Interconnect (VPI) Technology 64 ports 10GbE 36 ports 40/56GbE 48 10GbE + 12 40/56GbE 36 ports IB up to 56Gb/s 8 VPI subnets Switch OS Layer Mezzanine Card VPI Adapter VPI Switch Ethernet: 10/40/56 Gb/s InfiniBand:10/20/40/56 Gb/s Unified Fabric Manager Networking Storage Clustering Management Applications Acceleration Engines LOM Adapter Card 3.0 From data center to campus and metro connectivity
  • © 2014 Mellanox Technologies 4  Highest Capacity in 1RU • From 12 QSFP to 36 QSFP 40/56Gb ports  4.03Tb/s • 64 x 10GbE • 48x10GbE plus 12x40/56Gbps  Unique Value Proposition • VPI 10/40/56Gbps • End to end solution Ethernet Switch Portfolio SX1036 The Ideal 40GbE ToR/Aggregation SX1024 Non-blocking 10GbE 40GbE ToR SX1016 Highest density 10GbE ToR  Latency • 220ns L2 latency • 330ns L3 latency  Power (SX1036) • SX1036 – 83w • SX1024 – 70w • SX1016 – 64w • 1W per 10Gb port, 2.3W per 40Gb port SX1012 Ideal storage/Database 10/40GbE Switch
  • © 2014 Mellanox Technologies 5 Independent Software Vendors BRAS Firewall DPI CDN Tester/QoE monitor WAN Acceleration Message Router Radio Network Controller Carrier Grade NAT Session Border Controller Classical Network Appliance Approach PE Router SGSN/GGSN Generic High Volume Ethernet Switches Generic High Volume Servers Generic High Volume Storage Orchestrated, automatic remote install Network Functions Virtualisation Approach  Ideal platform for ETSI NFV: Network Functions Virtualization • Consolidate Network Equipment onto standard servers, switches and storage • Leverage Software Defined Networking • Driven within the ETSI: European Telecom Standard Institution  The migration to x86 based platforms is the enabler • 3G/4G Network Core • Load Balancing / Traffic and Policy Enforcement • Internet Security Gateways • Network Monitoring / VPN • CDN / Video Processing and Optimization / IPTV Telecom and Security Network Functions Virtualization ATCA Platforms Compute and Storage Platforms
  • © 2014 Mellanox Technologies 6 HP c7000 with Mellanox 40GbE Interconnect Mezz Adapter HPPN 644161-B22 2P Blade NFF Cables Switch Blade SX1018HP VPI Ready: Same HW for both ETH and IB Highest Capacity: 2P 40GbE PCIe 3.0 x 8 lanes Lowest Latency RoCE (App to App): 1.3us Lowest Power (40GbE) Typ 2-port 40GbE: 5.1W •56Gbps FDR IB / 40GbE QSFP •QSA: QSFP to SFP+ Adapter VPI Ready: Same HW for both ETH and IB Highest Capacity: 2.72 Tb/s bandwidth 16 Internal 40/10GbE ports 18 External 40/10GbE ports (QSFP+) Lowest Latency: 220 nsec latency 40GbE 270 nsec latency 10GbE Lowest Power: 82W (Typical Power with Passive Cables) C-Class double wide form factor Up to two SX1018HP switches per enclosure
  • © 2014 Mellanox Technologies 7- Mellanox Confidential - Mellanox ConnectX-3 Dual-Port 40GbE NIC and Switch release in Q3/2013  14 Compute Blades each using Single EN6132 Dual Port 40GbE NIC  2 Switch Blades each using EN6131 SwitchX-2 32 Ports 40GbE  Compute I/O 1.12 Tbps @40 GbE  Uplink I/O up to 1.44Tbps @40 GbE Dual Star Architecture Dual-Dual Star Architecture 10GbE 10GbE 40GbE 40GbE 22 x 10GbE 18 x 40GbE 40GbE 40GbE 40GbE 40GbE 18 x 40GbE 18 x 40GbE 22 x 10GbE 18 x 40GbE 18 x 40GbE 18 x 40GbE EN4093R EN4093R EN6131 EN6131 EN6131 EN6131 EN6131 EN6131 ITE SWITCH ITE SWITCH IBM PureFlex System  14 Compute Blades each using Dual EN6132 Dual Port 40GbE NIC  4 Switch Blades each using EN6131 SwitchX-2 32 Ports 40GbE  Compute I/O 2.24 Tbps @40 GbE  Uplink I/O up to 2.24 Tbps @40 GbE IBM PureFlex with Mellanox 40GbE Interconnect Single wide chassis: 14x ITEs / Blade Servers support 2 adapters per server Double-wide chassis: 7x ITEs / Blade Servers support 4 adapters per server
  • © 2014 Mellanox Technologies 8 CloudNFV: The 1st ETSI ISG NFV Approved PoC Demo Console Optimization, Active Virtualization and Demo console: NFV Orchestration and Metro Ethernet Switches: Demo Virtual Function: Traffic Telemetry and DPI as a service: Servers, Data Center Switching, Lab Facilities. Systems Integration: Data Path Acceleration: Demo Virtual Function Overall Architecture: High Performance Server and Storage Interconnect Active Data Center
  • © 2014 Mellanox Technologies 9 Mellanox and 6WIND ConnectX-3 NIC Driver for Intel® DPDK 6WIND or Intel® DPDK • Data Plane libraries • Optimized NIC drivers Client’s Application Software High-performance packet processing solutions for • Gateways • Security appliances • UTMs • Virtual appliances • etc. Multicore Processor …… librte_pmd_mlx4 librte_pmd driver provided as an addon into the DPDK (no need to patch the DPDK) Based on the generic librte_eal and librte_ether API of the DPDK. Best design since it co-works with the ibverb framework. librte_crypto_nitrox 6WIND addons VMware …
  • © 2014 Mellanox Technologies 10 Neutron Plug-in OpenStack integration High performance 10/40/56Gbps SR-IOV enabled OpenFlow enabled eSwitch OpenStack Neutron Plug-in PMD for DPDK: VM OS bypass Multi cores and RSS support Delivering bare-metal performance Mellanox NIC SR-IOV with PMD for Intel® DPDK in the Guest VM OS VM OS VM Hypervisor Legacy Software vSwitches SR-IOV eSwitch Hardware Offload OpenFlow enabled VM 6WIND or Intel® DPDK • Data Plane libraries • Optimized NIC drivers Client’s Application Software High-performance packet processing solutions for • Gateways • Security appliances • UTMs • Virtual appliances • etc. Multicore Processor …… librte_pmd_mlx4 librte_crypto_nitrox 6WIND addonsVMware … 10/40/56Gbps
  • © 2014 Mellanox Technologies 11  Slow Application Performance • 1/10GbE • 50us latency for VM to VM connectivity • Slow VM Migration • Slow Storage I/O  Expensive and inefficient • High CPU overhead for I/O processing • Multiple adapters needed  Limited isolation • Minimal QoS and Security in software Mellanox NIC Based I/O Virtualization Advantages  Fastest Application Performance • 10/40GbE with RDMA, 56Gb InfiniBand • Only 2us for VM to VM connectivity • >3.5x faster VM Migration • >6x faster storage access  Superb efficiency • Offload hypervisor CPU, more VMs • I/O consolidation  Best isolation • Hardware-enforced security and QoS OS VM OS VM OS VM Hypervisor Software based vSwitches OS VM OS VM OS VM Hypervisor Legacy Software vSwitches Hardware Offload + vSwitches Legacy NICs
  • © 2014 Mellanox Technologies 12 I/O Virtualization Future – NIC Based Switching OS VM OS VM OS VM Hypervisor OS VM eSwitches (embedded switches) Physical Ports (pPort) NIC/HCA Hardware “LAG” vPorts with multi-level QoS and hardware based congestion control Virtual NICs (vNICs) vPort Security Filters, ACLs, and Tunneling (EoIB/VXLAN/NVGRE) HW Based teaming HW Based VM Switching pPort QoS and DCB vPort Priority tagging Controlled via SDN/OpenFlow  eSwitch supported match fields : • Destination MAC address • VLAN ID • Ether Type • Source/Destination IP address • Source/Destination UDP/TCP port  eSwitch supported actions: • Drop • Allow • Count • Trap/Mirror • Set priority (VLAN priority, egress queue & policer)
  • © 2014 Mellanox Technologies 13 Mellanox and Radware Defense Pro : SDN demo The Traditional Way: Bump in the wire Appliances A Better Way: SDN and OpenFlow with Flow Based Routing
  • © 2014 Mellanox Technologies 14 ConnectX-3 Pro NVGRE and VXLAN Performance 0 2 4 6 8 10 12 2 4 8 16 BandwidthGb/s VM Pairs NvGRE Throughput ConnectX-3 Pro 10GbE NVGRE Offload Disabled 0.00 1.00 2.00 3.00 4.00 5.00 1 VM 2 VMs 3 VMs VxLAN in software 3.50 3.33 4.29 VxLAN HW Offload 0.90 0.89 1.19 CPU%/Bandwidth (Gbit/sce) CPU Usage Per Gbit/sec with VxLAN 0 5 10 15 20 25 1 VM 2 VMs 3 VMs VxLAN in software 2 3 3.5 VxLAN HW Offload 10 19 21 Bandwidth[Gb/s] Total VM Bandwidth when using VxLAN The Foundation of Cloud 2.0 The World’s First NVGRE / VXLAN Offloaded NIC
  • © 2014 Mellanox Technologies 15 6WIND demonstration of 195 Gbps Accelerated Virtual Switch iproute2iptables Fast Path IP IPsec OVS Acceleration TCP VLAN GRE MPLS ACL LAG Custom GTPu NAT Intel® DPDK Shared Memory Statistics Protocol Tables Linux Kernel 6WINDGate fast path statistics Linux Networking Stack 6WINDGate Sync Daemons NIC(s) Multicore Processor Platform Quagga 6WINDGate includes the Mellanox poll mode driver (PMD) - Provide Direct access to the networking hardware: Linux OS Bypass The demo include 5 Mellanox ConnectX®-3 Pro cards with dual 40G Ports.
  • © 2014 Mellanox Technologies 16 Managing the VM Networking Via OpenFlow / SDN Neutron Plug-In SDN Applications SDN Applications Cloud Management OpenStack Manager SDN Controller OS VM Para- virtual OS VM OS VM OS VM SR-IOV to the VM 10/40GbE or InfiniBand ports Embedded Switch OpenFlow Agent Neutron Agent Create/delete, configure policy per VM vNIC Servers tap tap  OpenFlow control over switch and NIC  Adapter hardware acceleration for OpenFlow and overlay functions  Native integration to OpenStack and SDN controllers The Benefits of VM Provision & Fabric Policy in Hardware Isolation, Performance & Offload, Simpler SDN
  • © 2014 Mellanox Technologies 17  Allow Service Orchestration over the Telecom WAN Network.  Leverage OpenStack for the Telecom Datacenter  Leverage Mellanox Neutron Plug-in allow SR-IOV • Near bear metal performance to the VMs CYAN Blue Planet: Carrier Grade SDN Orchestration Platform
  • © 2014 Mellanox Technologies 18  Using CloudBand, service providers can create cloud services that offer virtually limitless growth and that capitalize on their broad range of distributed data center and network resources. By building their own carrier clouds, service providers can meet stringent service level agreements (SLAs) and deliver the performance, access and security that enterprises and consumers demand.  “Network Function Virtualization can provide service providers with significant gains in automation and reductions in costs. Working in conjunction with the Alcatel-Lucent CloudBand Ecosystem, Mellanox’s industry-leading, end-to-end InfiniBand and Ethernet interconnect products with support for NFV provides cloud and telecommunications networks with best-in-class virtualization features, performance and efficiency.” Alcatel-Lucent CloudBand: Mellanox Solution Partner
  • © 2014 Mellanox Technologies 19 Calsoft Labs: Virtual B-RAS solution  High Performance Virtual B-RAS solution  Addresses Broadband service requirements  Intel® DPDK optimized solution  Powered by highly optimized data plane processing software from 6WIND  Performance & capabilities accelerated by Mellanox ConnectX-3 NIC in DELL servers  Delivers 256K PPPoE tunnels on a 2U rack DELL server with Intel Sandy Bridge  Can be integrated with Calsoft Labs Cloud NOC™ orchestration framework or third party NFV management systems.  PPPoX termination with VRF support for Multi- tenants  DHCP support: o DHCP Relay o DHCP Server for IPv4/IPv6  Tunneling: o L2TP and GRE with VRF support o IPsec/PPP interworking per VRF  AAA (Authentication, Authorization, Accounting) – RADIUS  Security: o IP address tracking o Centralized Firewall  QoS: o QoS per service o QoS per subscriber, Hierarchical QoS o Dynamic Bandwidth management Key Features POWERED BY
  • © 2014 Mellanox Technologies 20 RDMA/RoCE I/O Offload RDMA over InfiniBand or Ethernet KERNELHARDWAREUSER RACK 1 OS NIC Buffer 1 Application 1 Application 2 OS Buffer 1 NICBuffer 1 TCP/IP RACK 2 HCA HCA Buffer 1Buffer 1 Buffer 1 Buffer 1 Buffer 1
  • © 2014 Mellanox Technologies 21 6200 1200 800 0 2000 4000 6000 8000 I/O Size - 64 [KB]Bandwidth [MB/s] SCSI Write Example, Linux KVM iSER 16 VMs Write 10GbE Fiber Channel - 8Gb Accelerating Cloud Performance 38 10 0 10 20 30 40 50 Time[s] Migration of Active VM 10GE-A 40GE-A Storage Migration Virtualization 3.5X 20X 6X 40 20 10 20 30 40 50 Message Size - 256 [bytes] Latency[us] VM-to-VM Latency Performance TCP ParaVirtualization RDMA Direct Access 10 GbE Fibre Channel 8Gb 40 GbE iSER 40GbE VMs Write
  • © 2014 Mellanox Technologies 22- Mellanox Confidential - “To make storage cheaper we use lots more network! How do we make Azure Storage scale? RoCE (RDMA over Ethernet) enabled at 40GbE for Windows Azure Storage, achieving massive COGS savings” Microsoft Keynote at Open Networking Summit 2014 on RDMA RDMA at 40GbE Enables Massive Cloud Saving For Microsoft Azure Keynote Albert Greenberg, Microsoft SDN Azure Infrastructure
  • © 2014 Mellanox Technologies 23  Using OpenStack Built-in components and management (Open-iSCSI, tgt target, Cinder), no additional software is required, RDMA is already inbox and used by our OpenStack customers !  Mellanox enable faster performance, with much lower CPU%  Next step is to bypass Hypervisor layers, and add NAS & Object storage Faster Cloud Storage Access Hypervisor (KVM) OS VM OS VM OS VM Adapter Open-iSCSI w iSER Compute Servers Switching Fabric iSCSI/iSER Target (tgt) Adapter Local Disks RDMA Cache Storage Servers OpenStack (Cinder) Using RDMA to accelerate iSCSI storage 0 1000 2000 3000 4000 5000 6000 7000 1 2 4 8 16 32 64 128 256 Bandwidth[MB/s] I/O Size [KB] iSER 4 VMs Write iSER 8 VMs Write iSER 16 VMs Write iSCSI Write 8 vms iSCSI Write 16 VMs PCIe Limit 6X
  • © 2014 Mellanox Technologies 24 Mellanox CloudX OpenCloud Any Software Open NIC, Open Switch, Open Server, Open Rack
  • © 2014 Mellanox Technologies 25 Fat-tree SDN Switch Network 40GbE 56Gbps IB FDR Fabric 40Gbps Fabric Platform 1 Platform 2 40Gbps Fabric Platform X 40Gbps Server Attached and or Network Attached HWA DPI BRAS SGSN GGSN PE Router Firewall CG-NAT SBC STB Ethernet Ethernet Ethernet A A A A A A A A A A A A B B B B B B B B B B B B C C C C C C C C C C C C A A A A B B B B C C C C A A A A B B B B C C C C AA AA BB BB CC CC AA BB CC AA BB CC AA BB CC AA BB CC AA BB CC AA BB CC AA BB CC Server Attached and or Network Attached HWA are Non-Scalable and lead back to the custom appliance based model
  • © 2014 Mellanox Technologies 26 Fat-tree SDN Switch Network 40GbE 56Gbps IB FDR SX1024 Ethernet Switch HWA / Signal Processing Fabric 40Gbps SX1024 Ethernet Switch HWA / Signal Processing Fabric Platform 1 Platform 2 40Gbps SX1024 Ethernet Switch HWA / Signal Processing Fabric Platform X Nx40Gbps Nx40Gbps Nx40Gbps 40Gbps Remote HWA as a Service in NFV Cloud Model DPI BRAS SGSN GGSN PE Router Firewall CG-NAT SBC STB Ethernet Ethernet Ethernet RDMA/RoCE RDMA/RoCE RDMA/RoCE AA BB CC AA BB CC AA BB CC AA BB CC AA BB CC AA BB CC
  • © 2014 Mellanox Technologies 27 Fat-tree SDN Switch Network 10/40/100Gbps ToR Aggregation Ethernet Switch SAN/NAS Storage Compute Storage 10/40/100Gbps 10/40/100Gbps Ethernet Switch SAN/NAS Storage Compute Storage Rack 1 Rack 2 10/40/100Gbps 10/40/100Gbps Ethernet Switch SAN/NAS Storage Compute Storage Rack n 12x10/40/100Gbps 12x10/40/100Gbps 12x10/40/100Gbps 10/40/100Gbps iSCSI SAN/NAS Storage Architecture in an NFV Cloud model iSCSI SAN/NAS Storage over Standard Ethernet Network: Shared Resource RDMA/RoCE RDMA/RoCE RDMA/RoCE
  • © 2014 Mellanox Technologies 28 The GPU as a Service Implementation  GPUs as a network-resident service • Little to no overhead when using FDR InfiniBand  Virtualize and decouple GPU services from CPU services • A new paradigm in cluster flexibility • Lower cost, lower power and ease of use with shared GPU resources • Remove difficult physical requirements of the GPU for standard compute servers GPU CPU GPU CPUGPU CPU GPU CPU GPU CPU GPUs in every server GPUs as a Service CPU VGPU CPU VGPUv CPU VGPU GPUGPUGPUGPUGPUGPUGPUGPUGPUGPUGPU
  • © 2014 Mellanox Technologies 29 Local and Remote GPU HWA Solutions Application/GPU servers GPU as a Service with Mellanox GPUDirect™ 1.0 rCUDA daemon Network Interface CUDA Driver + runtime Network Interface rCUDA library Application Application Server Side Remote GPU Side Application CUDA Driver + runtime CUDA Application Mellanox GPUDirect™ 1.0 enables remote access from every node to any GPU in the system with a single copy Data path is copied through CPU Memory to or from Network Interface and GPU HWA Device GPU as a Service with Mellanox PeerDirect™ Network Interface CUDA Driver + runtime Network Interface rCUDA library Application Application Server Side Remote GPU Side P2P Plugin HCA Driver Peer Driver EXPORT Peer Device Memory Functions ib_umem_* functions are “tunneled” thru the p2p plugin module Mellanox PeerDirect™ enables remote access from every node to any GPU in the system with a zero copy Data path is directly from Network Interface to GPU HWA Device
  • © 2014 Mellanox Technologies 30  Ideal for Cloud Datacenter, Data Processing Platforms and Network Functions Virtualization • Leading SerDes Technology: High Bandwidth – Advanced Process • 10/40/56Gb VPI with PCIe 3.0 Interface • 10/40/56Gb High Bandwidth Switch: 36 ports of 10/40/56Gb or 64 ports of 10Gb • RDMA/RoCE technology: Ultra Low Latency Data Transfer • Software Defined Networking: SDN Switch and Control End to End Solution • Cloud Management: OpenStack integration  Paving the way to 100Gb/s Interconnect • End to End Network Interconnect for Compute/Processing and Switching • Software Defined Networking  High Bandwidth, Low Latency and Lower TCO: $/Port/Gb Mellanox Interconnect Solutions Mellanox Interconnect is Your competitive Advantage!
  • Thank You