Virtualization Acceleration 
Motti Beck, Director Enterprise Market Development 
VMworld 2014 | San Francisco
Virtualization Acceleration is Next 
Management 
Efficiency 
Acceleration 
 Server Virtualization 
 Network Virtualization 
 Storage Virtualization 
 1GbE  10GbE 40GbE  100GbE 
 VMDirectPath, SR-IOV 
 Offloads: Network Protocols, VXLAN 
 On Dashboard for Compute and Storage 
 Software Defined Network 
 Software Defined Storage 
1st Generation 2nd Generation 3rd Generation 
Functionality 
© 2014 Mellanox Technologies 2
IO Acceleration 
TCP/IP Remote Direct Memory Access (RDMA) 
© 2014 Mellanox Technologies 3
RDMA over Converged Ethernet - RoCE 
 RDMA transport over Ethernet 
• Efficient, light-weight transport, layered directly over Ethernet 
• Takes advantage of PFC (Priority Flow Control) in DCB Ethernet 
• IBTA standard 
• Supported in OFED 1.5.1, RHEL 6.X, Windows Server 2012 R2 
 Lowest latency in the Ethernet industry 
• 1.3μs end-to-end RDMA latency 
- Faster application completion 
- Better server utilization 
- Higher scalability 
 Tremendous support momentum by ecosystem 
• Cloud service providers, DB Vendors, Financial ISVs, Server & 
Storage OEMs 
• Entire Ethernet management ecosystem is available 
© 2014 Mellanox Technologies 4
vMotion over RoCE Accelerates vMotion 
Total vMotion Time (seconds) CPU Utilization (Destination) 
70.6272862 
36% Faster 
45.3119161 
80 
70 
60 
50 
40 
30 
20 
10 
0 
TCP/IP RDMA 
90% Less 
0:00 0:05 0:11 0:16 0:21 0:26 0:31 0:36 0:41 0:46 0:51 0:56 1:01 1:06 1:11 1:16 
60 
50 
40 
30 
20 
10 
0 
TCP/IP 1.37 21.5 47.8 47.5 48 43.1 47.8 45.1 43.9 44 40.1 34.9 35.3 34.9 41.1 36.5 45.6 
RDMA 0.74 6.32 6.66 6.98 6.85 6.47 6.45 6.7 4.66 4.57 
% CPU Utilization 
 Destination CPU utilization 92% lower 
 Source CPU utilization 84% lower 
*Source: VMware’s CTO office VMworld 2012 
http://cto.vmware.com/wp-content/uploads/2012/09/RDMAonvSphere.pdf 
© 2014 Mellanox Technologies 5
10X Boost of Live Migration over Hyper-V with SMB Direct 
TCP/IP Compression SMB/RDMA 
Normalized 
Send Bandwidth 
Higher is better 
% CPU #1 
Lower is better 
% CPU #2 
Lower is better 
150Gb/sec 
across 3 links 
Source: TechED’13 Opening Satya Nadella Keynote Session with Jose Baretto 
70 
60 
50 
40 
30 
20 
10 
0 
Seconds 
Live Migration Times* 
* Lower is better 
Click To 
Watch Video 
© 2014 Mellanox Technologies 6
Virtualized Storage Acceleration Running iSER* over ESXi 
Test Setup: ESXi 5.0, 2 VMs, 2 LUNS per VM 
 RDMA Superior Across the Board 
• Throughput & IOP’s 
• Efficiency & CPU Utilization 
• Scalability 
Higher is 
Better 
10x Bandwidth Performance Advantage vs TCP/IP 
2.5x IOPS Performance With iSER Initiator 
Higher is 
Better 
* iSCSI over RDMA 
© 2014 Mellanox Technologies 7
vSphere Storage Access Acceleration over RoCE running iSER 
Dell Fluid Cache for SAN 
Source: https://www.youtube.com/watch?v=uw7UHWWAtig 
Boost Performance with Server-based Caching over RoCE 
© 2014 Mellanox Technologies 8
Maximize VDI Efficiency over RDMA 
 RDMA eliminates storage bottlenecks in VDI deployments 
• Mellanox ConnectX®-3 with RoCE accelerates the access to cache over RDMA 
• 150 Virtual desktops over RoCE vs. 60 virtual desktops over TCP/IP 
ConnectX 
iSCSI 
using 
RDMA 
Active Active 
Nytro 
MegaRAID 
Flash 
Cache 
ConnectX 
iSCSI 
using 
RDMA 
Nytro 
MegaRAID 
Flash 
Cache 
0 20 40 60 80 100 120 140 160 
Intel 10GbE, iSCSI/TCP 
ConnectX3 10GbE, iSCSI/RDMA (iSER) 
ConnectX3 40GbE, iSCSI/RDMA (iSER) 
Number of Virtual Desktop VMs 
http://www.mellanox.com/related-docs/whitepapers/SB_Virtual_Desktop_Infrastructure_Storage_Acceleration_Final.pdf 
iSER Enables 2X More Virtual Desktops 
© 2014 Mellanox Technologies 9
Proven Deployment Over Azure 
Microsoft Keynote 
Albert Greenberg 
Click to SDN in Azure Infrastructure 
Watch Video 
“To make storage cheaper we use lots more network! 
How do we make Azure Storage scale? RoCE (RDMA 
over Ethernet) enabled at 40GbE for Windows Azure 
Storage, achieving massive COGS savings” 
Microsoft Keynote at Open Networking Summit 2014 on RDMA 
RDMA at 40GbE Enables Massive Cloud Saving for Microsoft Azure 
© 2014 Mellanox Technologies 10
Mellanox Accelerates OpenStack Storage 
RDMA Accelerates iSCSI Storage 
Compute Servers 
VM 
OS 
VM 
OS 
VM 
OS 
Hypervisor (KVM) 
Open-iSCSI w iSER 
Adapter 
Storage Servers 
OpenStack (Cinder) 
iSCSI/iSER Target (tgt) 
RDMA Cache 
Adapter Local Disks 
Switching Fabric 
OpenStack Storage Performance * 
1.3 
5.5 
6 
5 
4 
3 
2 
1 
0 
iSCSI over TCP iSER 
GBytes/s 
* iSER patches are available on OpenStack 
branch: https://github.com/mellanox/openstack 
Built-in OpenStack Components/Management & Cinder/iSER to Accelerate Storage Access 
© 2014 Mellanox Technologies 11
Leadership in Strategic Markets with an End to End Portfolio 
Big Data DB/Enterprise Cloud Storage Web 2.0 
End-to-End - Cloud, Enterprise, and Storage Optimized - InfiniBand and Ethernet Portfolio 
ICs Adapter Cards Switches/Gateways Host/Fabric Software Metro / WAN 
Cables/Modules 
© 2014 Mellanox Technologies 12
Thank You

Virtualization Acceleration

  • 1.
    Virtualization Acceleration MottiBeck, Director Enterprise Market Development VMworld 2014 | San Francisco
  • 2.
    Virtualization Acceleration isNext Management Efficiency Acceleration  Server Virtualization  Network Virtualization  Storage Virtualization  1GbE  10GbE 40GbE  100GbE  VMDirectPath, SR-IOV  Offloads: Network Protocols, VXLAN  On Dashboard for Compute and Storage  Software Defined Network  Software Defined Storage 1st Generation 2nd Generation 3rd Generation Functionality © 2014 Mellanox Technologies 2
  • 3.
    IO Acceleration TCP/IPRemote Direct Memory Access (RDMA) © 2014 Mellanox Technologies 3
  • 4.
    RDMA over ConvergedEthernet - RoCE  RDMA transport over Ethernet • Efficient, light-weight transport, layered directly over Ethernet • Takes advantage of PFC (Priority Flow Control) in DCB Ethernet • IBTA standard • Supported in OFED 1.5.1, RHEL 6.X, Windows Server 2012 R2  Lowest latency in the Ethernet industry • 1.3μs end-to-end RDMA latency - Faster application completion - Better server utilization - Higher scalability  Tremendous support momentum by ecosystem • Cloud service providers, DB Vendors, Financial ISVs, Server & Storage OEMs • Entire Ethernet management ecosystem is available © 2014 Mellanox Technologies 4
  • 5.
    vMotion over RoCEAccelerates vMotion Total vMotion Time (seconds) CPU Utilization (Destination) 70.6272862 36% Faster 45.3119161 80 70 60 50 40 30 20 10 0 TCP/IP RDMA 90% Less 0:00 0:05 0:11 0:16 0:21 0:26 0:31 0:36 0:41 0:46 0:51 0:56 1:01 1:06 1:11 1:16 60 50 40 30 20 10 0 TCP/IP 1.37 21.5 47.8 47.5 48 43.1 47.8 45.1 43.9 44 40.1 34.9 35.3 34.9 41.1 36.5 45.6 RDMA 0.74 6.32 6.66 6.98 6.85 6.47 6.45 6.7 4.66 4.57 % CPU Utilization  Destination CPU utilization 92% lower  Source CPU utilization 84% lower *Source: VMware’s CTO office VMworld 2012 http://cto.vmware.com/wp-content/uploads/2012/09/RDMAonvSphere.pdf © 2014 Mellanox Technologies 5
  • 6.
    10X Boost ofLive Migration over Hyper-V with SMB Direct TCP/IP Compression SMB/RDMA Normalized Send Bandwidth Higher is better % CPU #1 Lower is better % CPU #2 Lower is better 150Gb/sec across 3 links Source: TechED’13 Opening Satya Nadella Keynote Session with Jose Baretto 70 60 50 40 30 20 10 0 Seconds Live Migration Times* * Lower is better Click To Watch Video © 2014 Mellanox Technologies 6
  • 7.
    Virtualized Storage AccelerationRunning iSER* over ESXi Test Setup: ESXi 5.0, 2 VMs, 2 LUNS per VM  RDMA Superior Across the Board • Throughput & IOP’s • Efficiency & CPU Utilization • Scalability Higher is Better 10x Bandwidth Performance Advantage vs TCP/IP 2.5x IOPS Performance With iSER Initiator Higher is Better * iSCSI over RDMA © 2014 Mellanox Technologies 7
  • 8.
    vSphere Storage AccessAcceleration over RoCE running iSER Dell Fluid Cache for SAN Source: https://www.youtube.com/watch?v=uw7UHWWAtig Boost Performance with Server-based Caching over RoCE © 2014 Mellanox Technologies 8
  • 9.
    Maximize VDI Efficiencyover RDMA  RDMA eliminates storage bottlenecks in VDI deployments • Mellanox ConnectX®-3 with RoCE accelerates the access to cache over RDMA • 150 Virtual desktops over RoCE vs. 60 virtual desktops over TCP/IP ConnectX iSCSI using RDMA Active Active Nytro MegaRAID Flash Cache ConnectX iSCSI using RDMA Nytro MegaRAID Flash Cache 0 20 40 60 80 100 120 140 160 Intel 10GbE, iSCSI/TCP ConnectX3 10GbE, iSCSI/RDMA (iSER) ConnectX3 40GbE, iSCSI/RDMA (iSER) Number of Virtual Desktop VMs http://www.mellanox.com/related-docs/whitepapers/SB_Virtual_Desktop_Infrastructure_Storage_Acceleration_Final.pdf iSER Enables 2X More Virtual Desktops © 2014 Mellanox Technologies 9
  • 10.
    Proven Deployment OverAzure Microsoft Keynote Albert Greenberg Click to SDN in Azure Infrastructure Watch Video “To make storage cheaper we use lots more network! How do we make Azure Storage scale? RoCE (RDMA over Ethernet) enabled at 40GbE for Windows Azure Storage, achieving massive COGS savings” Microsoft Keynote at Open Networking Summit 2014 on RDMA RDMA at 40GbE Enables Massive Cloud Saving for Microsoft Azure © 2014 Mellanox Technologies 10
  • 11.
    Mellanox Accelerates OpenStackStorage RDMA Accelerates iSCSI Storage Compute Servers VM OS VM OS VM OS Hypervisor (KVM) Open-iSCSI w iSER Adapter Storage Servers OpenStack (Cinder) iSCSI/iSER Target (tgt) RDMA Cache Adapter Local Disks Switching Fabric OpenStack Storage Performance * 1.3 5.5 6 5 4 3 2 1 0 iSCSI over TCP iSER GBytes/s * iSER patches are available on OpenStack branch: https://github.com/mellanox/openstack Built-in OpenStack Components/Management & Cinder/iSER to Accelerate Storage Access © 2014 Mellanox Technologies 11
  • 12.
    Leadership in StrategicMarkets with an End to End Portfolio Big Data DB/Enterprise Cloud Storage Web 2.0 End-to-End - Cloud, Enterprise, and Storage Optimized - InfiniBand and Ethernet Portfolio ICs Adapter Cards Switches/Gateways Host/Fabric Software Metro / WAN Cables/Modules © 2014 Mellanox Technologies 12
  • 13.