SlideShare a Scribd company logo
1 of 23
Download to read offline
Interconnect Your Future With Mellanox
High-Performance Computing
March 2014
Mellanox Performance Advantage (Source TopCrunch)
2014 Results

Higher Performance
with Half of the System Size!

LS-DYNA Applications, Car2Car Benchmark (Seconds)
1500
1400
1300

LS-DYNA is an advanced multiphysics
simulation software (CAE), developed by LSTC
Used in automotive, aerospace, military,
manufacturing, and bioengineering industries

1200
1100
1000
CRAY XC30/Aries
2000 Cores

CRAY XC30/Aries
4000 Cores

FDR InfiniBand (SGI)
2000 Cores

InfiniBand Delivers Highest System Performance, Efficiency and Scalability
All platforms use same CPUs of Intel® Xeon® E5-2690 v2 @3.00GHz, Cray platform is connected with Cray Aries interconnect, SGI platform is connected with Mellanox FDR InfiniBand
© 2014 Mellanox Technologies

2
Mellanox Performance Advantage (Source HPC Advisory Council)

More than
2X Performance!
HOOMD-blue is an highly optimized object-oriented
many-particle dynamics applications that performs
general purpose particle dynamics simulations
Developed by the University of Michigan

InfiniBand Delivers Highest System Performance, Efficiency and Scalability
© 2014 Mellanox Technologies

3
InfiniBand Leadership in TOP500 Petascale-Capable Systems

 Mellanox InfiniBand is the interconnect of choice for Petascale computing
• Accelerates 48% of the sustained Petaflop systems (19 systems out of 40)
© 2014 Mellanox Technologies

4
Mellanox InfiniBand Connected Petascale Systems

Connecting Half of the World’s Petascale Systems
Mellanox Connected Petascale System Examples
© 2014 Mellanox Technologies

5
InfiniBand’s Unsurpassed System Efficiency

Average Efficiency
• InfiniBand: 86%
• Cray: 80%
• 10GbE: 65%
• GigE: 44%

 TOP500 systems listed according to their efficiency
 InfiniBand is the key element responsible for the highest system efficiency
 Mellanox delivers efficiencies of more than 97% with InfiniBand
© 2014 Mellanox Technologies

6
Mellanox in the TOP500 Supercomputing List (Nov’13)
 Mellanox FDR InfiniBand is the fastest interconnect solution on the TOP500
•
•
•
•

More than 12GB/s throughput, less than 0.7usec latency
Being used in 80 systems on the TOP500 list – 1.8X increase from the Nov’12 list
Connects the fastest InfiniBand-based supercomputers – TACC (#7), LRZ (#10)
Enables the two most efficient systems in the TOP200

 Mellanox InfiniBand is the fastest interconnect technology on the list
• Enables the highest system utilization on the TOP500 – more than 97% system efficiency
• Enables the top seven highest utilized systems on the TOP500 list

 Mellanox InfiniBand is the only Petascale-proven, standard interconnect solution
• Connects 19 out of the 40 Petaflop capable systems on the list
• Connects 4X the number of Cray based systems in the Top100, 6.5X in TOP500

 Mellanox’s end-to-end scalable solutions accelerate GPU-based systems
• GPUDirect RDMA technology enables faster communications and higher performance
© 2014 Mellanox Technologies

7
System Example: NASA Ames Research Center Pleiades
 20K InfiniBand nodes
 Mellanox end-to-end FDR and QDR InfiniBand
 Supports variety of scientific and engineering projects
• Coupled atmosphere-ocean models
• Future space vehicle design
• Large-scale dark matter halos and galaxy evolution

Asian Monsoon Water Cycle
High-Resolution Climate Simulations

© 2014 Mellanox Technologies

8
Leading Supplier of End-to-End Interconnect Solutions
Comprehensive End-to-End Software Accelerators and Managment

Management
MXM

FCA

Mellanox Messaging
Acceleration

Fabric Collectives
Acceleration

UFM
Unified Fabric Management

Storage and Data
VSA

UDA

Storage Accelerator
(iSCSI)

Unstructured Data
Accelerator

Comprehensive End-to-End InfiniBand and Ethernet Portfolio
ICs

Adapter Cards

© 2014 Mellanox Technologies

Switches/Gateways

Host/Fabric Software

Metro / WAN

Cables/Modules

9
Converged Interconnect Solutions to Deliver Highest ROI for all Applications

Accelerating Half of the World’s Petascale Systems
Mellanox Connected Petascale System Examples

InfiniBand Enables Lowest Application Cost in the Cloud
(Examples)
© 2014 Mellanox Technologies

Businesses Success Depends on Mellanox

Dominant in Storage Interconnects
10
Mellanox Solutions

© 2014 Mellanox Technologies

11
Virtual Protocol Interconnect (VPI) Technology
VPI Adapter

VPI Switch
Unified Fabric Manager
Switch OS Layer

Applications
Storage

Networking

Clustering

Management

Acceleration Engines
Ethernet: 10/40/56 Gb/s

3.0

64 ports 10GbE
36 ports 40/56GbE
48 10GbE + 12 40/56GbE
36 ports IB up to 56Gb/s
8 VPI subnets

InfiniBand:10/20/40/56 Gb/s

From data center to
campus and metro
connectivity

LOM

Adapter Card

Mezzanine Card

Standard Protocols of InfiniBand and Ethernet on the Same Wire!
© 2014 Mellanox Technologies

12
Mellanox ScalableHPC Communication Library to Accelerate Applications
MPI

OpenSHMEM / PGAS

MXM

FCA

•
•
•
•

•
•
•
•

Berkeley UPC

Reliable Messaging
Hybrid Transport Mechanism
Efficient Memory Registration
Receive Side Tag Matching

Topology Aware Collective Optimization
Hardware Multicast
Separate Virtual Fabric for Collectives
CORE-Direct Hardware Offload

Reduce Collective Latency

100.0
80.0
60.0
40.0
20.0
0.0

Latency (us)

Latency (us)

Barrier Collective Latency

0

500

1000

1500

2000

Processes (PPN=8)

Without FCA

© 2014 Mellanox Technologies

With FCA

2500

3000
2500
2000
1500
1000
500
0
0

500

1000

1500

2000

2500

Processes (PPN=8)

Without FCA

With FCA

13
Mellanox Connect-IB The World’s Fastest Adapter
 The 7th generation of Mellanox interconnect adapters
 World’s first 100Gb/s interconnect adapter (dual-port FDR 56Gb/s InfiniBand)
 Delivers 137 million messages per second – 4X higher than competition
 World leading scalable transport – no dependency on system size

© 2014 Mellanox Technologies

14
Smart Offloads for MPI/SHMEM/PGAS/UPC Collective Operations

Ideal







System noise

CORE-Direct (Offload)

CORE-Direct - Asynchronous

CORE-Direct Technology
US Department of Energy (DOE) funded project – ORNL and Mellanox
Adapter-based hardware offloading for collectives operations
Includes floating-point capability on the adapter for data reductions
CORE-Direct API is exposed through the Mellanox drivers

© 2014 Mellanox Technologies

15
1

System
Memory

CPU

CPU

1

GPUDirect RDMA for Highest GPU Performance

GPU

Chip
set

Chip
set

System
Memory

GPU

InfiniBand
GPU
Memory

InfiniBand

GPUDirect RDMA

GPU
Memory

Source: Prof. DK Panda

67% Lower Latency
© 2014 Mellanox Technologies

5X Increase in Throughput
16
Remote GPU Access through rCUDA
GPU servers
CUDA Application

Application

GPU as a Service
Client Side

Server Side

Application

rCUDA daemon

rCUDA library
CUDA
Driver + runtime

Network Interface

Network Interface

CPU
VGPU
CPU
VGPU

CUDA
Driver + runtime

GPU
GPU
GPU
GPU
GPU
GPU
GPU
GPU
GPU
GPU
GPU

CPU
VGPU

rCUDA provides remote access from
every node to any GPU in the system
© 2014 Mellanox Technologies

17
Campus and Metro RDMA Long Reach Solutions
Example:
4 MetroX TX6100 systems over 6 km

Example:
4 MetroX TX6100 systems
• Connect IB over 2-4km

“A common problem is the time cost of moving data
between datacenters, which can slow computations
and delay results. Mellanox's MetroX lets us unify
systems across campus, and maintain the high-speed
access our researchers need, regardless of the
physical location of their work.”

• Replace Obsidian SDR

Mike Shuey, Purdue University

Example:
© 2014 Mellanox Technologies

2 MetroX TX6100 systems over 8 km

18
Variety of Clustering Topologies
CLOS (Fat Tree)

Typically enables best performance, lowest latency
Non-Blocking Network
Alleviates bandwidth bottleneck closer to the root.
Most common topology in many supercomputers

Hypercube

Supported by SGI

© 2014 Mellanox Technologies

Mesh / 3D Torus

Blocking network, good for applications with locality
Support for dedicate sub-networks
Simple expansion for future growth
Not limited to storage connection only at cube edges

DragonFly+

Concept of connecting “groups” together in a full-graph
Flexible definition of intra-group interconnection

19
The Mellanox Advatage
 Connect-IB delivers superior performance: 100Gb/s, 0.7usec latency, 137 million messages/sec
 ScalableHPC software library provides leading performance for MPI, OpenSHMEM/PGAS and UPC
 Superiors applications offloads: RDMA, Collectives, scalable transport (Dynamically Connected)
 Flexible topologies: Fat Tree, mesh, 3D Torus, Dragonfly+
 Standard based solution, Open source support, large eco-system, one solution for all applications
 Converged I/O – compute, storage, management on single fabric
 Long term roadmap

© 2014 Mellanox Technologies

20
Technology Roadmap – One-Generation Lead over the Competition

Mellanox

56Gbs

40Gbs

20Gbs

Terascale

Petascale

3rd

“Roadrunner”

Virginia Tech (Apple)

Mellanox Connected

© 2014 Mellanox Technologies

Exascale

1st

TOP500 2003

2000

200Gbs

100Gbs

2005

Mega Supercomputers

2010

2015

2020
21
The Only Provider of End-to-End 40/56Gb/s Solutions
Comprehensive End-to-End InfiniBand and Ethernet Portfolio
ICs

Adapter Cards

Switches/Gateways

Host/Fabric Software

Metro / WAN

Cables/Modules

From Data Center to Metro and WAN

X86, ARM and Power based Compute and Storage Platforms
The Interconnect Provider For 10Gb/s and Beyond
© 2014 Mellanox Technologies

22
For more information: HPC@mellanox.com

More Related Content

What's hot

InfiniBand Growth Trends - TOP500 (July 2015)
InfiniBand Growth Trends - TOP500 (July 2015)InfiniBand Growth Trends - TOP500 (July 2015)
InfiniBand Growth Trends - TOP500 (July 2015)Mellanox Technologies
 
Interop Tokyo 2014 -- Mellanox Demonstrations
Interop Tokyo 2014 -- Mellanox DemonstrationsInterop Tokyo 2014 -- Mellanox Demonstrations
Interop Tokyo 2014 -- Mellanox DemonstrationsMellanox Technologies
 
Advancing Applications Performance With InfiniBand
Advancing Applications Performance With InfiniBandAdvancing Applications Performance With InfiniBand
Advancing Applications Performance With InfiniBandMellanox Technologies
 
Announcing the Mellanox ConnectX-5 100G InfiniBand Adapter
Announcing the Mellanox ConnectX-5 100G InfiniBand AdapterAnnouncing the Mellanox ConnectX-5 100G InfiniBand Adapter
Announcing the Mellanox ConnectX-5 100G InfiniBand Adapterinside-BigData.com
 
The Top 10 Business Reasons for 10GbE iSCSI
The Top 10 Business Reasons for 10GbE iSCSIThe Top 10 Business Reasons for 10GbE iSCSI
The Top 10 Business Reasons for 10GbE iSCSIEmulex Corporation
 
Proven and Emerging Use Cases of Software Defined Network
Proven and Emerging Use Cases of Software Defined NetworkProven and Emerging Use Cases of Software Defined Network
Proven and Emerging Use Cases of Software Defined NetworkOpen Networking Summits
 
SDN Network virtualization, NFV & MPLS synergies
SDN Network virtualization, NFV & MPLS synergiesSDN Network virtualization, NFV & MPLS synergies
SDN Network virtualization, NFV & MPLS synergiesHector.Avalos
 
SDN Service Provider Use Cases
SDN Service Provider Use CasesSDN Service Provider Use Cases
SDN Service Provider Use CasesSDxCentral
 
latencyin fiber optic networks
latencyin fiber optic networkslatencyin fiber optic networks
latencyin fiber optic networksMapYourTech
 
Transport SDN - Use Cases and Lessons Learnt
Transport SDN - Use Cases and Lessons LearntTransport SDN - Use Cases and Lessons Learnt
Transport SDN - Use Cases and Lessons LearntADVA
 

What's hot (20)

Mellanox's Sales Strategy
Mellanox's Sales StrategyMellanox's Sales Strategy
Mellanox's Sales Strategy
 
InfiniBand Growth Trends - TOP500 (July 2015)
InfiniBand Growth Trends - TOP500 (July 2015)InfiniBand Growth Trends - TOP500 (July 2015)
InfiniBand Growth Trends - TOP500 (July 2015)
 
Mellanox IBM
Mellanox IBMMellanox IBM
Mellanox IBM
 
Mellanox's Technological Advantage
Mellanox's Technological AdvantageMellanox's Technological Advantage
Mellanox's Technological Advantage
 
Interop Tokyo 2014 -- Mellanox Demonstrations
Interop Tokyo 2014 -- Mellanox DemonstrationsInterop Tokyo 2014 -- Mellanox Demonstrations
Interop Tokyo 2014 -- Mellanox Demonstrations
 
Advancing Applications Performance With InfiniBand
Advancing Applications Performance With InfiniBandAdvancing Applications Performance With InfiniBand
Advancing Applications Performance With InfiniBand
 
Interconnect Product Portfolio
Interconnect Product PortfolioInterconnect Product Portfolio
Interconnect Product Portfolio
 
Announcing the Mellanox ConnectX-5 100G InfiniBand Adapter
Announcing the Mellanox ConnectX-5 100G InfiniBand AdapterAnnouncing the Mellanox ConnectX-5 100G InfiniBand Adapter
Announcing the Mellanox ConnectX-5 100G InfiniBand Adapter
 
Mellanox Approach to NFV & SDN
Mellanox Approach to NFV & SDNMellanox Approach to NFV & SDN
Mellanox Approach to NFV & SDN
 
The Great IT Migration
The Great IT MigrationThe Great IT Migration
The Great IT Migration
 
Interop: The 10GbE Top 10
Interop: The 10GbE Top 10Interop: The 10GbE Top 10
Interop: The 10GbE Top 10
 
The Top 10 Business Reasons for 10GbE iSCSI
The Top 10 Business Reasons for 10GbE iSCSIThe Top 10 Business Reasons for 10GbE iSCSI
The Top 10 Business Reasons for 10GbE iSCSI
 
Mellanox's Operational Excellence
Mellanox's Operational ExcellenceMellanox's Operational Excellence
Mellanox's Operational Excellence
 
SDN use cases_2014
SDN use cases_2014SDN use cases_2014
SDN use cases_2014
 
Proven and Emerging Use Cases of Software Defined Network
Proven and Emerging Use Cases of Software Defined NetworkProven and Emerging Use Cases of Software Defined Network
Proven and Emerging Use Cases of Software Defined Network
 
SDN Network virtualization, NFV & MPLS synergies
SDN Network virtualization, NFV & MPLS synergiesSDN Network virtualization, NFV & MPLS synergies
SDN Network virtualization, NFV & MPLS synergies
 
Mellanox Market Leading Solutions
Mellanox Market Leading SolutionsMellanox Market Leading Solutions
Mellanox Market Leading Solutions
 
SDN Service Provider Use Cases
SDN Service Provider Use CasesSDN Service Provider Use Cases
SDN Service Provider Use Cases
 
latencyin fiber optic networks
latencyin fiber optic networkslatencyin fiber optic networks
latencyin fiber optic networks
 
Transport SDN - Use Cases and Lessons Learnt
Transport SDN - Use Cases and Lessons LearntTransport SDN - Use Cases and Lessons Learnt
Transport SDN - Use Cases and Lessons Learnt
 

Similar to Interconnect Your Future With Mellanox

Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand SolutionsMellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand Solutionsinside-BigData.com
 
InfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and RoadmapInfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and Roadmapinside-BigData.com
 
Co-Design Architecture for Exascale
Co-Design Architecture for ExascaleCo-Design Architecture for Exascale
Co-Design Architecture for Exascaleinside-BigData.com
 
Mellnox Interconnect presentation in OpenPOWER Brazil workshop
Mellnox Interconnect presentation in OpenPOWER Brazil workshopMellnox Interconnect presentation in OpenPOWER Brazil workshop
Mellnox Interconnect presentation in OpenPOWER Brazil workshopGanesan Narayanasamy
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task ComputingEric Van Hensbergen
 
Interconnect Your Future: Paving the Road to Exascale
Interconnect Your Future: Paving the Road to ExascaleInterconnect Your Future: Paving the Road to Exascale
Interconnect Your Future: Paving the Road to Exascaleinside-BigData.com
 
InfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and RoadmapInfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and Roadmapinside-BigData.com
 
HPC DAY 2017 | The network part in accelerating Machine-Learning and Big-Data
HPC DAY 2017 | The network part in accelerating Machine-Learning and Big-DataHPC DAY 2017 | The network part in accelerating Machine-Learning and Big-Data
HPC DAY 2017 | The network part in accelerating Machine-Learning and Big-DataHPC DAY
 
IBM 40Gb Ethernet - A competitive alternative to Infiniband
IBM 40Gb Ethernet - A competitive alternative to InfinibandIBM 40Gb Ethernet - A competitive alternative to Infiniband
IBM 40Gb Ethernet - A competitive alternative to InfinibandAngel Villar Garea
 
The evolution of data center network fabrics
The evolution of data center network fabricsThe evolution of data center network fabrics
The evolution of data center network fabricsCisco Canada
 
InfiniBand in the Enterprise Data Center.pdf
InfiniBand in the Enterprise Data Center.pdfInfiniBand in the Enterprise Data Center.pdf
InfiniBand in the Enterprise Data Center.pdfbui thequan
 
Ceph Day London 2014 - Ceph Over High-Performance Networks
Ceph Day London 2014 - Ceph Over High-Performance Networks Ceph Day London 2014 - Ceph Over High-Performance Networks
Ceph Day London 2014 - Ceph Over High-Performance Networks Ceph Community
 
Ceph Day New York 2014: Ceph over High Performance Networks
Ceph Day New York 2014: Ceph over High Performance NetworksCeph Day New York 2014: Ceph over High Performance Networks
Ceph Day New York 2014: Ceph over High Performance NetworksCeph Community
 
Advanced Networking: The Critical Path for HPC, Cloud, Machine Learning and more
Advanced Networking: The Critical Path for HPC, Cloud, Machine Learning and moreAdvanced Networking: The Critical Path for HPC, Cloud, Machine Learning and more
Advanced Networking: The Critical Path for HPC, Cloud, Machine Learning and moreinside-BigData.com
 
Mellanox Announcements at SC15
Mellanox Announcements at SC15Mellanox Announcements at SC15
Mellanox Announcements at SC15inside-BigData.com
 
Mellanox for OpenStack - OpenStack最新情報セミナー 2014年10月
Mellanox for OpenStack  - OpenStack最新情報セミナー 2014年10月Mellanox for OpenStack  - OpenStack最新情報セミナー 2014年10月
Mellanox for OpenStack - OpenStack最新情報セミナー 2014年10月VirtualTech Japan Inc.
 
Cloud Camp Milan 2K9 Telecom Italia: Where P2P?
Cloud Camp Milan 2K9 Telecom Italia: Where P2P?Cloud Camp Milan 2K9 Telecom Italia: Where P2P?
Cloud Camp Milan 2K9 Telecom Italia: Where P2P?Gabriele Bozzi
 
CloudCamp Milan 2009: Telecom Italia
CloudCamp Milan 2009: Telecom ItaliaCloudCamp Milan 2009: Telecom Italia
CloudCamp Milan 2009: Telecom ItaliaGabriele Bozzi
 
76924356 synopsis-network
76924356 synopsis-network76924356 synopsis-network
76924356 synopsis-networklklokesh
 

Similar to Interconnect Your Future With Mellanox (20)

Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand SolutionsMellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
 
InfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and RoadmapInfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and Roadmap
 
Co-Design Architecture for Exascale
Co-Design Architecture for ExascaleCo-Design Architecture for Exascale
Co-Design Architecture for Exascale
 
Mellnox Interconnect presentation in OpenPOWER Brazil workshop
Mellnox Interconnect presentation in OpenPOWER Brazil workshopMellnox Interconnect presentation in OpenPOWER Brazil workshop
Mellnox Interconnect presentation in OpenPOWER Brazil workshop
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task Computing
 
Interconnect Your Future: Paving the Road to Exascale
Interconnect Your Future: Paving the Road to ExascaleInterconnect Your Future: Paving the Road to Exascale
Interconnect Your Future: Paving the Road to Exascale
 
InfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and RoadmapInfiniBand In-Network Computing Technology and Roadmap
InfiniBand In-Network Computing Technology and Roadmap
 
HPC DAY 2017 | The network part in accelerating Machine-Learning and Big-Data
HPC DAY 2017 | The network part in accelerating Machine-Learning and Big-DataHPC DAY 2017 | The network part in accelerating Machine-Learning and Big-Data
HPC DAY 2017 | The network part in accelerating Machine-Learning and Big-Data
 
IBM 40Gb Ethernet - A competitive alternative to Infiniband
IBM 40Gb Ethernet - A competitive alternative to InfinibandIBM 40Gb Ethernet - A competitive alternative to Infiniband
IBM 40Gb Ethernet - A competitive alternative to Infiniband
 
The evolution of data center network fabrics
The evolution of data center network fabricsThe evolution of data center network fabrics
The evolution of data center network fabrics
 
What is 3d torus
What is 3d torusWhat is 3d torus
What is 3d torus
 
InfiniBand in the Enterprise Data Center.pdf
InfiniBand in the Enterprise Data Center.pdfInfiniBand in the Enterprise Data Center.pdf
InfiniBand in the Enterprise Data Center.pdf
 
Ceph Day London 2014 - Ceph Over High-Performance Networks
Ceph Day London 2014 - Ceph Over High-Performance Networks Ceph Day London 2014 - Ceph Over High-Performance Networks
Ceph Day London 2014 - Ceph Over High-Performance Networks
 
Ceph Day New York 2014: Ceph over High Performance Networks
Ceph Day New York 2014: Ceph over High Performance NetworksCeph Day New York 2014: Ceph over High Performance Networks
Ceph Day New York 2014: Ceph over High Performance Networks
 
Advanced Networking: The Critical Path for HPC, Cloud, Machine Learning and more
Advanced Networking: The Critical Path for HPC, Cloud, Machine Learning and moreAdvanced Networking: The Critical Path for HPC, Cloud, Machine Learning and more
Advanced Networking: The Critical Path for HPC, Cloud, Machine Learning and more
 
Mellanox Announcements at SC15
Mellanox Announcements at SC15Mellanox Announcements at SC15
Mellanox Announcements at SC15
 
Mellanox for OpenStack - OpenStack最新情報セミナー 2014年10月
Mellanox for OpenStack  - OpenStack最新情報セミナー 2014年10月Mellanox for OpenStack  - OpenStack最新情報セミナー 2014年10月
Mellanox for OpenStack - OpenStack最新情報セミナー 2014年10月
 
Cloud Camp Milan 2K9 Telecom Italia: Where P2P?
Cloud Camp Milan 2K9 Telecom Italia: Where P2P?Cloud Camp Milan 2K9 Telecom Italia: Where P2P?
Cloud Camp Milan 2K9 Telecom Italia: Where P2P?
 
CloudCamp Milan 2009: Telecom Italia
CloudCamp Milan 2009: Telecom ItaliaCloudCamp Milan 2009: Telecom Italia
CloudCamp Milan 2009: Telecom Italia
 
76924356 synopsis-network
76924356 synopsis-network76924356 synopsis-network
76924356 synopsis-network
 

More from Mellanox Technologies

More from Mellanox Technologies (14)

InfiniBand FAQ
InfiniBand FAQInfiniBand FAQ
InfiniBand FAQ
 
CloudX – Expand Your Cloud into the Future
CloudX – Expand Your Cloud into the FutureCloudX – Expand Your Cloud into the Future
CloudX – Expand Your Cloud into the Future
 
Mellanox VXLAN Acceleration
Mellanox VXLAN AccelerationMellanox VXLAN Acceleration
Mellanox VXLAN Acceleration
 
Mellanox High Performance Networks for Ceph
Mellanox High Performance Networks for CephMellanox High Performance Networks for Ceph
Mellanox High Performance Networks for Ceph
 
InfiniBand Essentials Every HPC Expert Must Know
InfiniBand Essentials Every HPC Expert Must KnowInfiniBand Essentials Every HPC Expert Must Know
InfiniBand Essentials Every HPC Expert Must Know
 
Become a Supercomputer Hero
Become a Supercomputer HeroBecome a Supercomputer Hero
Become a Supercomputer Hero
 
Unified Fabric Manager - HP Insight CMU Connector
Unified Fabric Manager - HP Insight CMU ConnectorUnified Fabric Manager - HP Insight CMU Connector
Unified Fabric Manager - HP Insight CMU Connector
 
Print 'N Fly - SC13
Print 'N Fly - SC13Print 'N Fly - SC13
Print 'N Fly - SC13
 
Storage, Cloud, Web 2.0, Big Data Driving Growth
Storage, Cloud, Web 2.0, Big Data Driving GrowthStorage, Cloud, Web 2.0, Big Data Driving Growth
Storage, Cloud, Web 2.0, Big Data Driving Growth
 
Mellanox Financial Overview
Mellanox Financial OverviewMellanox Financial Overview
Mellanox Financial Overview
 
Mellanox Storage Solutions
Mellanox Storage SolutionsMellanox Storage Solutions
Mellanox Storage Solutions
 
I/O virtualization with InfiniBand and 40 Gigabit Ethernet
I/O virtualization with InfiniBand and 40 Gigabit EthernetI/O virtualization with InfiniBand and 40 Gigabit Ethernet
I/O virtualization with InfiniBand and 40 Gigabit Ethernet
 
Big Data Benchmarking with RDMA solutions
Big Data Benchmarking with RDMA solutions Big Data Benchmarking with RDMA solutions
Big Data Benchmarking with RDMA solutions
 
Scale Out Database Solution
Scale Out Database SolutionScale Out Database Solution
Scale Out Database Solution
 

Recently uploaded

My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clashcharlottematthew16
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsMiki Katsuragi
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 

Recently uploaded (20)

My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clash
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering Tips
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 

Interconnect Your Future With Mellanox

  • 1. Interconnect Your Future With Mellanox High-Performance Computing March 2014
  • 2. Mellanox Performance Advantage (Source TopCrunch) 2014 Results Higher Performance with Half of the System Size! LS-DYNA Applications, Car2Car Benchmark (Seconds) 1500 1400 1300 LS-DYNA is an advanced multiphysics simulation software (CAE), developed by LSTC Used in automotive, aerospace, military, manufacturing, and bioengineering industries 1200 1100 1000 CRAY XC30/Aries 2000 Cores CRAY XC30/Aries 4000 Cores FDR InfiniBand (SGI) 2000 Cores InfiniBand Delivers Highest System Performance, Efficiency and Scalability All platforms use same CPUs of Intel® Xeon® E5-2690 v2 @3.00GHz, Cray platform is connected with Cray Aries interconnect, SGI platform is connected with Mellanox FDR InfiniBand © 2014 Mellanox Technologies 2
  • 3. Mellanox Performance Advantage (Source HPC Advisory Council) More than 2X Performance! HOOMD-blue is an highly optimized object-oriented many-particle dynamics applications that performs general purpose particle dynamics simulations Developed by the University of Michigan InfiniBand Delivers Highest System Performance, Efficiency and Scalability © 2014 Mellanox Technologies 3
  • 4. InfiniBand Leadership in TOP500 Petascale-Capable Systems  Mellanox InfiniBand is the interconnect of choice for Petascale computing • Accelerates 48% of the sustained Petaflop systems (19 systems out of 40) © 2014 Mellanox Technologies 4
  • 5. Mellanox InfiniBand Connected Petascale Systems Connecting Half of the World’s Petascale Systems Mellanox Connected Petascale System Examples © 2014 Mellanox Technologies 5
  • 6. InfiniBand’s Unsurpassed System Efficiency Average Efficiency • InfiniBand: 86% • Cray: 80% • 10GbE: 65% • GigE: 44%  TOP500 systems listed according to their efficiency  InfiniBand is the key element responsible for the highest system efficiency  Mellanox delivers efficiencies of more than 97% with InfiniBand © 2014 Mellanox Technologies 6
  • 7. Mellanox in the TOP500 Supercomputing List (Nov’13)  Mellanox FDR InfiniBand is the fastest interconnect solution on the TOP500 • • • • More than 12GB/s throughput, less than 0.7usec latency Being used in 80 systems on the TOP500 list – 1.8X increase from the Nov’12 list Connects the fastest InfiniBand-based supercomputers – TACC (#7), LRZ (#10) Enables the two most efficient systems in the TOP200  Mellanox InfiniBand is the fastest interconnect technology on the list • Enables the highest system utilization on the TOP500 – more than 97% system efficiency • Enables the top seven highest utilized systems on the TOP500 list  Mellanox InfiniBand is the only Petascale-proven, standard interconnect solution • Connects 19 out of the 40 Petaflop capable systems on the list • Connects 4X the number of Cray based systems in the Top100, 6.5X in TOP500  Mellanox’s end-to-end scalable solutions accelerate GPU-based systems • GPUDirect RDMA technology enables faster communications and higher performance © 2014 Mellanox Technologies 7
  • 8. System Example: NASA Ames Research Center Pleiades  20K InfiniBand nodes  Mellanox end-to-end FDR and QDR InfiniBand  Supports variety of scientific and engineering projects • Coupled atmosphere-ocean models • Future space vehicle design • Large-scale dark matter halos and galaxy evolution Asian Monsoon Water Cycle High-Resolution Climate Simulations © 2014 Mellanox Technologies 8
  • 9. Leading Supplier of End-to-End Interconnect Solutions Comprehensive End-to-End Software Accelerators and Managment Management MXM FCA Mellanox Messaging Acceleration Fabric Collectives Acceleration UFM Unified Fabric Management Storage and Data VSA UDA Storage Accelerator (iSCSI) Unstructured Data Accelerator Comprehensive End-to-End InfiniBand and Ethernet Portfolio ICs Adapter Cards © 2014 Mellanox Technologies Switches/Gateways Host/Fabric Software Metro / WAN Cables/Modules 9
  • 10. Converged Interconnect Solutions to Deliver Highest ROI for all Applications Accelerating Half of the World’s Petascale Systems Mellanox Connected Petascale System Examples InfiniBand Enables Lowest Application Cost in the Cloud (Examples) © 2014 Mellanox Technologies Businesses Success Depends on Mellanox Dominant in Storage Interconnects 10
  • 11. Mellanox Solutions © 2014 Mellanox Technologies 11
  • 12. Virtual Protocol Interconnect (VPI) Technology VPI Adapter VPI Switch Unified Fabric Manager Switch OS Layer Applications Storage Networking Clustering Management Acceleration Engines Ethernet: 10/40/56 Gb/s 3.0 64 ports 10GbE 36 ports 40/56GbE 48 10GbE + 12 40/56GbE 36 ports IB up to 56Gb/s 8 VPI subnets InfiniBand:10/20/40/56 Gb/s From data center to campus and metro connectivity LOM Adapter Card Mezzanine Card Standard Protocols of InfiniBand and Ethernet on the Same Wire! © 2014 Mellanox Technologies 12
  • 13. Mellanox ScalableHPC Communication Library to Accelerate Applications MPI OpenSHMEM / PGAS MXM FCA • • • • • • • • Berkeley UPC Reliable Messaging Hybrid Transport Mechanism Efficient Memory Registration Receive Side Tag Matching Topology Aware Collective Optimization Hardware Multicast Separate Virtual Fabric for Collectives CORE-Direct Hardware Offload Reduce Collective Latency 100.0 80.0 60.0 40.0 20.0 0.0 Latency (us) Latency (us) Barrier Collective Latency 0 500 1000 1500 2000 Processes (PPN=8) Without FCA © 2014 Mellanox Technologies With FCA 2500 3000 2500 2000 1500 1000 500 0 0 500 1000 1500 2000 2500 Processes (PPN=8) Without FCA With FCA 13
  • 14. Mellanox Connect-IB The World’s Fastest Adapter  The 7th generation of Mellanox interconnect adapters  World’s first 100Gb/s interconnect adapter (dual-port FDR 56Gb/s InfiniBand)  Delivers 137 million messages per second – 4X higher than competition  World leading scalable transport – no dependency on system size © 2014 Mellanox Technologies 14
  • 15. Smart Offloads for MPI/SHMEM/PGAS/UPC Collective Operations Ideal      System noise CORE-Direct (Offload) CORE-Direct - Asynchronous CORE-Direct Technology US Department of Energy (DOE) funded project – ORNL and Mellanox Adapter-based hardware offloading for collectives operations Includes floating-point capability on the adapter for data reductions CORE-Direct API is exposed through the Mellanox drivers © 2014 Mellanox Technologies 15
  • 16. 1 System Memory CPU CPU 1 GPUDirect RDMA for Highest GPU Performance GPU Chip set Chip set System Memory GPU InfiniBand GPU Memory InfiniBand GPUDirect RDMA GPU Memory Source: Prof. DK Panda 67% Lower Latency © 2014 Mellanox Technologies 5X Increase in Throughput 16
  • 17. Remote GPU Access through rCUDA GPU servers CUDA Application Application GPU as a Service Client Side Server Side Application rCUDA daemon rCUDA library CUDA Driver + runtime Network Interface Network Interface CPU VGPU CPU VGPU CUDA Driver + runtime GPU GPU GPU GPU GPU GPU GPU GPU GPU GPU GPU CPU VGPU rCUDA provides remote access from every node to any GPU in the system © 2014 Mellanox Technologies 17
  • 18. Campus and Metro RDMA Long Reach Solutions Example: 4 MetroX TX6100 systems over 6 km Example: 4 MetroX TX6100 systems • Connect IB over 2-4km “A common problem is the time cost of moving data between datacenters, which can slow computations and delay results. Mellanox's MetroX lets us unify systems across campus, and maintain the high-speed access our researchers need, regardless of the physical location of their work.” • Replace Obsidian SDR Mike Shuey, Purdue University Example: © 2014 Mellanox Technologies 2 MetroX TX6100 systems over 8 km 18
  • 19. Variety of Clustering Topologies CLOS (Fat Tree) Typically enables best performance, lowest latency Non-Blocking Network Alleviates bandwidth bottleneck closer to the root. Most common topology in many supercomputers Hypercube Supported by SGI © 2014 Mellanox Technologies Mesh / 3D Torus Blocking network, good for applications with locality Support for dedicate sub-networks Simple expansion for future growth Not limited to storage connection only at cube edges DragonFly+ Concept of connecting “groups” together in a full-graph Flexible definition of intra-group interconnection 19
  • 20. The Mellanox Advatage  Connect-IB delivers superior performance: 100Gb/s, 0.7usec latency, 137 million messages/sec  ScalableHPC software library provides leading performance for MPI, OpenSHMEM/PGAS and UPC  Superiors applications offloads: RDMA, Collectives, scalable transport (Dynamically Connected)  Flexible topologies: Fat Tree, mesh, 3D Torus, Dragonfly+  Standard based solution, Open source support, large eco-system, one solution for all applications  Converged I/O – compute, storage, management on single fabric  Long term roadmap © 2014 Mellanox Technologies 20
  • 21. Technology Roadmap – One-Generation Lead over the Competition Mellanox 56Gbs 40Gbs 20Gbs Terascale Petascale 3rd “Roadrunner” Virginia Tech (Apple) Mellanox Connected © 2014 Mellanox Technologies Exascale 1st TOP500 2003 2000 200Gbs 100Gbs 2005 Mega Supercomputers 2010 2015 2020 21
  • 22. The Only Provider of End-to-End 40/56Gb/s Solutions Comprehensive End-to-End InfiniBand and Ethernet Portfolio ICs Adapter Cards Switches/Gateways Host/Fabric Software Metro / WAN Cables/Modules From Data Center to Metro and WAN X86, ARM and Power based Compute and Storage Platforms The Interconnect Provider For 10Gb/s and Beyond © 2014 Mellanox Technologies 22
  • 23. For more information: HPC@mellanox.com