SlideShare a Scribd company logo
1 of 27
Download to read offline
AdamRoe
HPCSolutionsArchitect
LAD’17:Lustre*GenerationalPerformance
Improvements&NewFeatures
HighLevelAbstract
2
Lustre* has had a number of compelling new features added in
recent releases; this talk will look at those features in detail and see
how well they all work together from both a performance and
functionality perspective. Comparing some of the numbers from last
year we will see how far the Lustre* filesystem has come in such a
short period of time (LAD’16 to LAD’17), comparing the same use
cases observing the generational improvements in the technology.
Agenda
• Hero Numbers: Generational Performance March 2016 – today
• Generational Metadata performance improvements
• Small file performance on OpenZFS (no Data-on-MDT)
• Has LDISKFS changed since last year, how does performance look today
• Scaling with DNE Phase 2
• How does PFL effect Performance
3
SummaryofLastyearsTalk
4
• No LDISKFS Numbers with DNE:
• Stability issues I observed have been resolved, see some new DNE 2 numbers with LDISKFS
in this talk
• DNE Phase 2 Scalability:
• Scalability was reasonable before, do new Lustre release demonstrate better scalability
• Using DNE 2 In Production:
• Yes, I am still using DNE2 in production successfully for 18 months
Copyright ® Intel Corporation 2016. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries.
* Other names and brands may be claimed as the property of others.
TestbedArchitecture
Bw-1-00
Bw-1-01
Bw-1-15
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Intel® HPC
Orchestrator
Headnode
GenericLustre1
GenericLustre2
.
.
.
.
.
.
.
GenericLustre10
Intel® Omni-Path Switching
(100Gbps)
10G Base-T Ethernet Switching
Intel® Omni Path Cabling
10G Base-T RJ45
Copyright ® Intel Corporation 2016. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries.
* Other names and brands may be claimed as the property of others.
TestbedArchitecture(Cont.)
Server
§ 10x Generic Lustre servers with two slightly different configurations
– Each System comprises of:
– 2x Intel® Xeon E5-2697v3 (Haswell) CPU’s
– 1x Intel® Omni-Path x16 HFI
– 128GB DDR4 2133MHz Memory
– Eight of the nodes contain - 4x Intel P3600 2.0TB 2.5” (U.2) NVMe devices, while the other two have 4x Intel® P3700 800GB 2.5” (U.2)
NVMe devices
– One node equipped with 2x Intel® S3700 400GB’s for MGT
§ 16x 2S Intel® Xeon E5v4 (Broadwell) Compute nodes
– 1x Intel® HPC Orchestrator (Beta 2) Headnode
– Hardware Components:
– 2x Intel® Xeon E5-2697v4 (Broadwell) CPU’s
– 1x Intel® Omni-Path x16 HFI
– 128GB DDR4 2400MHz Memory
– Local boot SSD
§ 100Gbps Intel® Omni-Path Fabric
– None-blocking fabric with single switch design.
– Server side optimisations: “options hfi1 sge_copy_mode=2 krcvqs=4 wss_threshold=70”
– Improve generic RDMA performance, only recommended on Lustre server side that do physically do any MPI
Why?
7
GenerationalPerformanceImprovements
Lustre 2.9EA to Lustre 2.10.1
8
MetadataPerformance
• LDISKFS quite close to ZFS for
file create about 75%
• Removal still a way to go
• When testing on slower storage
difference are marginal
• Demonstrates a good level of
improvement.
9
Lustre Metadata performance 2.9EA (Mid 2016) vs. 2.10.1 (today), MDTEST: Single MDT Performance.
0
10000
20000
30000
40000
50000
60000
70000
80000
90000
100000
b2.9 + zfs-0.6.5.7-1 b2.9 + zfs-0.7RC1 b2.10 + zfs 0.7RC5 b2.10.1 + zfs 0.7.1 b2.10 + ldiskfs
File Create/Remove 1 MDT: Generational
Improvements
File Create File Remove
/mnt/zlfs2/mdtest -i 3 -I 10000 -F -C -T -r -u -d /mnt/zlfs2/test1.out
SmallFilePerformance(4k)
• Single MDT operation up 4x
compared to this time last year
• Scaling of DNE2 still not linear,
but better
• Create performance trails off due
to lack of clients
• Clear benefit versus the previous
release
10
Lustre Small file performance 2.9EA (Mid 2016) vs. 2.10.1 (today), MDTEST: Single MDT Performance. No
Data-on-MDT used leveraging DNE Phase 2 up to 8 MDT’s on Separate servers.
0
50000
100000
150000
200000
250000
300000
350000
1x MDT 2x MDT 3x MDT 4x MDT 5x MDT 6x MDT 7x MDT 8x MDT
Small File Performance:
Lustre 2.9EA (ZFS 0.6.5-7) vs. Lustre 2.10.1 (ZFS 0.7.1)
LAD'17 Create LAD'17 Removal LAD'16 Create LAD'16 Removal
/mnt/zlfs2/mdtest -i 3 -I 10000 -z 1 -b 1 -L -u -F –w 4096 -d /mnt/zlfs2/*DNE-DIR*/test1.out
LDISKFS:GenerationalMetadataImprovements
11
LDISKFSPerformance
• Some performance boost was
expected, but not this much
• Shows positive trend version
to version
• 2.10 to 2.10.1 - LU-7899
12
Lustre 2.9EA vs. Lustre 2.10.0 vs. Lustre 2.10.1
56887.869
74418.092
79191.709
94895.636
125084.373
107051.123
0
20000
40000
60000
80000
100000
120000
140000
File	Create File	Removal
LDISKFS:	Lustre	2.9EA	vs.	Lustre	2.10.0	vs.	Lustre	2.10.1EA
LDISKFS	2.9EA LDISKFS	2.10.0 LDISKFS	2.10.1EA
DNEPhaseIIScaling
Lustre 2.9EAvs.2.10.1
13
Normalised:ScalingLAD’16toLAD’17:DNEPhase2
• Neither are linear
• Overall scalability dropped a
little, but ultimate number is
much higher
• Some work to do to get this close
to OST scalability
14
Generational Performance and scalability of DNE Phase 2 on OpenZFS
1x	MDT 2x	MDT	 3x	MDT 4x	MDT 5x	MDT 6x	MDT 7x	MDT 8x	MDT
Normalised:	DNE	Phase	2	Lustre	2.9EA	vs.	Lustre	2.10.1
2.9EA	ZFS	0.6.5 2.10.1	ZFS	0.7.1 Linear
DNEPhaseIIonLDISKFS
15
LDISKFSDNEPhase2Scaling&Functionality
• Totally stable, versus pervious
testing
• Scaling stops after 2 MDT’s
• Clients unable to push that
much I/O
• You can see from the previous
slide 200 – 250k is my HW limit
16
Following up from last year where I couldn’t give DNE2 on LDISKFS.
0
200000
400000
600000
800000
1000000
1200000
1x	MDT 2x	MDT	 3x	MDT 4x	MDT 5x	MDT 6x	MDT 7x	MDT 8x	MDT
DNE	Phase	2	Scaling:	Lustre	2.10.1	LDISKFS
Create	LDISKFS Linear
ProgressiveFileLayout
17
ProgressiveFileLayout
18
Stripe
=1
Stripe
=2
Stripe
=4
Stripe
=8
Stripe
=16
4MB 64MB 256MB 1GB 10GB
• Example Layout
• lfs setstripe -E 4M -c 1 -E 64M -c 2 -E 256M
-c 4 -E 1G -c 8 -E 10G -c 16 -E -1 -c -
1 /mnt/zlfs2/pfl_test01/
19
/mnt/zlfs2/pfl_test01/
lcm_layout_gen: 0
lcm_entry_count: 6
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 0
lcme_extent.e_end: 4194304
stripe_count: 1 stripe_size: 1048576 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 4194304
lcme_extent.e_end: 67108864
stripe_count: 2 stripe_size: 1048576 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 67108864
lcme_extent.e_end: 268435456
stripe_count: 4 stripe_size: 1048576 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 268435456
lcme_extent.e_end: 1073741824
stripe_count: 8 stripe_size: 1048576 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 1073741824
lcme_extent.e_end: 10737418240
stripe_count: 16 stripe_size: 1048576 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 10737418240
lcme_extent.e_end: EOF
stripe_count: -1 stripe_size: 1048576 stripe_offset: -1
PFLPerformanceWhenusingIOR
• Each files stripe dynamically grows
based on file size
• Write performance up 4.6%, read
within margin of error
• Certainly not detrimental to
performance
20
Lustre 2.10.1; IOR Performance, file per process (256 files, 16GB per file) mean performance MB/s. PFL as
described before versus traditional -1 stripe.
0
5000
10000
15000
20000
25000
30000
35000
40000
45000
Read Write
PFL	vs.	No	PFL:	IOR	MB/s
PFL No	PFL
/mnt/zlfs2/IOR -wr -C -F -i 3 -t 1m -b 1m -s 16384 -a MPIIO -o /mnt/zlfs2/pfl_new/testme1.file
21
/mnt/zlfs2/pfl_new1/
lcm_layout_gen: 0
lcm_entry_count: 6
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 0
lcme_extent.e_end: 4194304
stripe_count: 1 stripe_size: 1048576 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 4194304
lcme_extent.e_end: 67108864
stripe_count: 2 stripe_size: 2097152 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 67108864
lcme_extent.e_end: 268435456
stripe_count: 4 stripe_size: 16777216 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 268435456
lcme_extent.e_end: 1073741824
stripe_count: 8 stripe_size: 33554432 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 1073741824
lcme_extent.e_end: 10737418240
stripe_count: 16 stripe_size: 134217728 stripe_offset: -1
lcme_id: N/A
lcme_flags: 0
lcme_extent.e_start: 10737418240
lcme_extent.e_end: EOF
stripe_count: -1 stripe_size: 268435456 stripe_offset: -1
PFLPerformancewithusingIOR(cont.)
• PFL is giving us the opportunity to
optimise the stripe relative to data
type
• Write up 8.7% on base results and
and 4.1% relative to test one
• Reads get a 9.2% boost with larger
stripe sizes
22
Lets optimise the stripe size this time, assuming the a larger stripe size for the higher stripe counts
lfs setstripe -E 4M -S 1M -c 1 -E 64M -S 2M -c 2 -E 256M -S 16M -c 4 -E 1G -S 32M -c 64 -E 10G -S 128M -c
16 -E -1 -S 256M -c -1 /mnt/zlfs2/pfl_new/
0
5000
10000
15000
20000
25000
30000
35000
40000
45000
50000
Read Write
PFL	vs.	No	PFL:	IOR	MB/s
PFL No	PFL PFL	Optimised
/mnt/zlfs2/IOR -wr -C -F -i 3 -t 1m -b 1m -s 16384 -a MPIIO -o /mnt/zlfs2/pfl_new/testme1.file
GenerationalSummary
Lad’16toLAD’17
23
KeyTakeaways
• Small file / metadata performance across the board is simply just better
• Amazing work done on OpenZFS to get 0.7.1 to where it is today
• Performance is comparable to LDISKFS of previous releases
• Overall DNE Phase 2 scalability is very similar to what we have seen before
• Overall usage and stability feels better, but was good before
• Optimising Striping layouts with PFL is essential, striping is done for you and can be
configured for best performance
24
Ifyoutakeanythingawayfromthistalk…
25
2.2x
7.7x
0
20000
40000
60000
80000
100000
120000
140000
LDISKFS	Create OpenZFS	Create
LAD'16	to	LAD'17:	Metadata	Performance
LAD'16 LAD'17
Copyright ® Intel Corporation 2016. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries.
* Other names and brands may be claimed as the property of others.
LegalInformation
All information provided here is subject to change without notice. Contact your Intel representative to obtain the latest Intel product specifications and roadmaps
Tests document performance of components on a particular test, in specific systems. Differences in hardware, software, or configuration will affect actual performance. Consult other sources of information to
evaluate performance as you consider your purchase. For more complete information about performance and benchmark results, visit http://www.intel.com/performance.
Intel technologies’ features and benefits depend on system configuration and may require enabled hardware, software or service activation. Performance varies depending on system configuration. No
computer system can be absolutely secure. Check with your system manufacturer or retailer or learn more at http://www.intel.com/content/www/us/en/software/intel-solutions-for-lustre-software.html.
Intel technologies may require enabled hardware, specific software, or services activation. Check with your system manufacturer or retailer.
You may not use or facilitate the use of this document in connection with any infringement or other legal analysis concerning Intel products described herein. You agree to grant Intel a non-exclusive, royalty-
free license to any patent claim thereafter drafted which includes subject matter disclosed herein.
No license (express or implied, by estoppel or otherwise) to any intellectual property rights is granted by this document.
The products described may contain design defects or errors known as errata which may cause the product to deviate from published specifications. Current characterized errata are available on request.
Intel disclaims all express and implied warranties, including without limitation, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement, as well as any warranty arising
from course of performance, course of dealing, or usage in trade.
This document contains information on products, services and/or processes in development. All information provided here is subject to change without notice. Contact your Intel representative to obtain the
latest forecast, schedule, specifications and roadmaps.
A "Mission Critical Application" is any application in which failure of the Intel Product could result, directly or indirectly, in personal injury or death. SHOULD YOU PURCHASE OR USE INTEL'S PRODUCTS FOR
ANY SUCH MISSION CRITICAL APPLICATION, YOU SHALL INDEMNIFY AND HOLD INTEL AND ITS SUBSIDIARIES, SUBCONTRACTORS AND AFFILIATES, AND THE DIRECTORS, OFFICERS, AND EMPLOYEES OF
EACH, HARMLESS AGAINST ALL CLAIMS COSTS, DAMAGES, AND EXPENSES AND REASONABLE ATTORNEYS' FEES ARISING OUT OF, DIRECTLY OR INDIRECTLY, ANY CLAIM OF PRODUCT LIABILITY,
PERSONAL INJURY, OR DEATH ARISING IN ANY WAY OUT OF SUCH MISSION CRITICAL APPLICATION, WHETHER OR NOT INTEL OR ITS SUBCONTRACTOR WAS NEGLIGENT IN THE DESIGN, MANUFACTURE,
OR WARNING OF THE INTEL PRODUCT OR ANY OF ITS PARTS.
Intel may make changes to specifications and product descriptions at any time, without notice. Designers must not rely on the absence or characteristics of any features or instructions marked "reserved" or
"undefined". Intel reserves these for future definition and shall have no responsibility whatsoever for conflicts or incompatibilities arising from future changes to them. The information here is subject to change
without notice. Do not finalize a design with this information.
Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries.
* Other names and brands may be claimed as the property of others.
© 2016 Intel Corporation
Copyright © 2017 Intel Corporation.
Intel, the Intel logo are trademarks of Intel Corporation.
Other names and brands may be claimed as the property of others.

More Related Content

What's hot

Shak larry-jeder-perf-and-tuning-summit14-part1-final
Shak larry-jeder-perf-and-tuning-summit14-part1-finalShak larry-jeder-perf-and-tuning-summit14-part1-final
Shak larry-jeder-perf-and-tuning-summit14-part1-finalTommy Lee
 
クラウド環境におけるキャッシュメモリQoS制御の評価
クラウド環境におけるキャッシュメモリQoS制御の評価クラウド環境におけるキャッシュメモリQoS制御の評価
クラウド環境におけるキャッシュメモリQoS制御の評価Ryousei Takano
 
MapReduce: A useful parallel tool that still has room for improvement
MapReduce: A useful parallel tool that still has room for improvementMapReduce: A useful parallel tool that still has room for improvement
MapReduce: A useful parallel tool that still has room for improvementKyong-Ha Lee
 
Netmap presentation
Netmap presentationNetmap presentation
Netmap presentationAmir Razmjou
 
Leveraging memory in sql server
Leveraging memory in sql serverLeveraging memory in sql server
Leveraging memory in sql serverChris Adkin
 
Building Apache Cassandra clusters for massive scale
Building Apache Cassandra clusters for massive scaleBuilding Apache Cassandra clusters for massive scale
Building Apache Cassandra clusters for massive scaleAlex Thompson
 
PostgreSQL performance archaeology
PostgreSQL performance archaeologyPostgreSQL performance archaeology
PostgreSQL performance archaeologyTomas Vondra
 
Shak larry-jeder-perf-and-tuning-summit14-part2-final
Shak larry-jeder-perf-and-tuning-summit14-part2-finalShak larry-jeder-perf-and-tuning-summit14-part2-final
Shak larry-jeder-perf-and-tuning-summit14-part2-finalTommy Lee
 
Column store indexes and batch processing mode (nx power lite)
Column store indexes and batch processing mode (nx power lite)Column store indexes and batch processing mode (nx power lite)
Column store indexes and batch processing mode (nx power lite)Chris Adkin
 
PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016
PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016
PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016Tomas Vondra
 
PostgreSQL + ZFS best practices
PostgreSQL + ZFS best practicesPostgreSQL + ZFS best practices
PostgreSQL + ZFS best practicesSean Chittenden
 
Collaborate nfs kyle_final
Collaborate nfs kyle_finalCollaborate nfs kyle_final
Collaborate nfs kyle_finalKyle Hailey
 
Collaborate vdb performance
Collaborate vdb performanceCollaborate vdb performance
Collaborate vdb performanceKyle Hailey
 
Replicate Oracle to Oracle, Oracle to MySQL, and Oracle to Analytics
Replicate Oracle to Oracle, Oracle to MySQL, and Oracle to AnalyticsReplicate Oracle to Oracle, Oracle to MySQL, and Oracle to Analytics
Replicate Oracle to Oracle, Oracle to MySQL, and Oracle to AnalyticsLinas Virbalas
 
Xdp and ebpf_maps
Xdp and ebpf_mapsXdp and ebpf_maps
Xdp and ebpf_mapslcplcp1
 
Adaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and EigensolversAdaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and Eigensolversinside-BigData.com
 

What's hot (20)

Shak larry-jeder-perf-and-tuning-summit14-part1-final
Shak larry-jeder-perf-and-tuning-summit14-part1-finalShak larry-jeder-perf-and-tuning-summit14-part1-final
Shak larry-jeder-perf-and-tuning-summit14-part1-final
 
クラウド環境におけるキャッシュメモリQoS制御の評価
クラウド環境におけるキャッシュメモリQoS制御の評価クラウド環境におけるキャッシュメモリQoS制御の評価
クラウド環境におけるキャッシュメモリQoS制御の評価
 
MapReduce: A useful parallel tool that still has room for improvement
MapReduce: A useful parallel tool that still has room for improvementMapReduce: A useful parallel tool that still has room for improvement
MapReduce: A useful parallel tool that still has room for improvement
 
Netmap presentation
Netmap presentationNetmap presentation
Netmap presentation
 
Leveraging memory in sql server
Leveraging memory in sql serverLeveraging memory in sql server
Leveraging memory in sql server
 
Building Apache Cassandra clusters for massive scale
Building Apache Cassandra clusters for massive scaleBuilding Apache Cassandra clusters for massive scale
Building Apache Cassandra clusters for massive scale
 
PostgreSQL performance archaeology
PostgreSQL performance archaeologyPostgreSQL performance archaeology
PostgreSQL performance archaeology
 
Shak larry-jeder-perf-and-tuning-summit14-part2-final
Shak larry-jeder-perf-and-tuning-summit14-part2-finalShak larry-jeder-perf-and-tuning-summit14-part2-final
Shak larry-jeder-perf-and-tuning-summit14-part2-final
 
NFS and Oracle
NFS and OracleNFS and Oracle
NFS and Oracle
 
Column store indexes and batch processing mode (nx power lite)
Column store indexes and batch processing mode (nx power lite)Column store indexes and batch processing mode (nx power lite)
Column store indexes and batch processing mode (nx power lite)
 
PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016
PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016
PostgreSQL na EXT4, XFS, BTRFS a ZFS / FOSDEM PgDay 2016
 
Upgrade & ndmp
Upgrade & ndmpUpgrade & ndmp
Upgrade & ndmp
 
PostgreSQL + ZFS best practices
PostgreSQL + ZFS best practicesPostgreSQL + ZFS best practices
PostgreSQL + ZFS best practices
 
Speeding up ps and top
Speeding up ps and topSpeeding up ps and top
Speeding up ps and top
 
Collaborate nfs kyle_final
Collaborate nfs kyle_finalCollaborate nfs kyle_final
Collaborate nfs kyle_final
 
Collaborate vdb performance
Collaborate vdb performanceCollaborate vdb performance
Collaborate vdb performance
 
Userspace networking
Userspace networkingUserspace networking
Userspace networking
 
Replicate Oracle to Oracle, Oracle to MySQL, and Oracle to Analytics
Replicate Oracle to Oracle, Oracle to MySQL, and Oracle to AnalyticsReplicate Oracle to Oracle, Oracle to MySQL, and Oracle to Analytics
Replicate Oracle to Oracle, Oracle to MySQL, and Oracle to Analytics
 
Xdp and ebpf_maps
Xdp and ebpf_mapsXdp and ebpf_maps
Xdp and ebpf_maps
 
Adaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and EigensolversAdaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and Eigensolvers
 

Similar to Lustre Generational Performance Improvements & New Features

Q1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptx
Q1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptxQ1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptx
Q1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptxMemory Fabric Forum
 
Threading Successes 06 Allegorithmic
Threading Successes 06   AllegorithmicThreading Successes 06   Allegorithmic
Threading Successes 06 Allegorithmicguest40fc7cd
 
Improving DragonFly's performance with PostgreSQL by Francois Tigeot
Improving DragonFly's performance with PostgreSQL by Francois TigeotImproving DragonFly's performance with PostgreSQL by Francois Tigeot
Improving DragonFly's performance with PostgreSQL by Francois Tigeoteurobsdcon
 
Fast datastacks - fast and flexible nfv solution stacks leveraging fd.io
Fast datastacks - fast and flexible nfv solution stacks leveraging fd.ioFast datastacks - fast and flexible nfv solution stacks leveraging fd.io
Fast datastacks - fast and flexible nfv solution stacks leveraging fd.ioOPNFV
 
NVIDIA GPUs Power HPC & AI Workloads in Cloud with Univa
NVIDIA GPUs Power HPC & AI Workloads in Cloud with UnivaNVIDIA GPUs Power HPC & AI Workloads in Cloud with Univa
NVIDIA GPUs Power HPC & AI Workloads in Cloud with Univainside-BigData.com
 
Big Lab Problems Solved with Spectrum Scale: Innovations for the Coral Program
Big Lab Problems Solved with Spectrum Scale: Innovations for the Coral ProgramBig Lab Problems Solved with Spectrum Scale: Innovations for the Coral Program
Big Lab Problems Solved with Spectrum Scale: Innovations for the Coral Programinside-BigData.com
 
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...In-Memory Computing Summit
 
CPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCoburn Watson
 
SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...
SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...
SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...Chester Chen
 
Van jaconson netchannels
Van jaconson netchannelsVan jaconson netchannels
Van jaconson netchannelsSusant Sahani
 
Ceph Day Netherlands - Ceph @ BIT
Ceph Day Netherlands - Ceph @ BIT Ceph Day Netherlands - Ceph @ BIT
Ceph Day Netherlands - Ceph @ BIT Ceph Community
 
Experiences with Power 9 at A*STAR CRC
Experiences with Power 9 at A*STAR CRCExperiences with Power 9 at A*STAR CRC
Experiences with Power 9 at A*STAR CRCGanesan Narayanasamy
 
What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014
What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014
What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014Philippe Fierens
 
Introduction to DPDK
Introduction to DPDKIntroduction to DPDK
Introduction to DPDKKernel TLV
 
Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...
Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...
Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...Akihiro Hayashi
 
Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...
Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...
Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...MLconf
 
Ceph Day Berlin: Ceph on All Flash Storage - Breaking Performance Barriers
Ceph Day Berlin: Ceph on All Flash Storage - Breaking Performance BarriersCeph Day Berlin: Ceph on All Flash Storage - Breaking Performance Barriers
Ceph Day Berlin: Ceph on All Flash Storage - Breaking Performance BarriersCeph Community
 
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Odinot Stanislas
 

Similar to Lustre Generational Performance Improvements & New Features (20)

Q1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptx
Q1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptxQ1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptx
Q1 Memory Fabric Forum: Using CXL with AI Applications - Steve Scargall.pptx
 
Threading Successes 06 Allegorithmic
Threading Successes 06   AllegorithmicThreading Successes 06   Allegorithmic
Threading Successes 06 Allegorithmic
 
Improving DragonFly's performance with PostgreSQL by Francois Tigeot
Improving DragonFly's performance with PostgreSQL by Francois TigeotImproving DragonFly's performance with PostgreSQL by Francois Tigeot
Improving DragonFly's performance with PostgreSQL by Francois Tigeot
 
Fast datastacks - fast and flexible nfv solution stacks leveraging fd.io
Fast datastacks - fast and flexible nfv solution stacks leveraging fd.ioFast datastacks - fast and flexible nfv solution stacks leveraging fd.io
Fast datastacks - fast and flexible nfv solution stacks leveraging fd.io
 
NVIDIA GPUs Power HPC & AI Workloads in Cloud with Univa
NVIDIA GPUs Power HPC & AI Workloads in Cloud with UnivaNVIDIA GPUs Power HPC & AI Workloads in Cloud with Univa
NVIDIA GPUs Power HPC & AI Workloads in Cloud with Univa
 
Latest HPC News from NVIDIA
Latest HPC News from NVIDIALatest HPC News from NVIDIA
Latest HPC News from NVIDIA
 
Big Lab Problems Solved with Spectrum Scale: Innovations for the Coral Program
Big Lab Problems Solved with Spectrum Scale: Innovations for the Coral ProgramBig Lab Problems Solved with Spectrum Scale: Innovations for the Coral Program
Big Lab Problems Solved with Spectrum Scale: Innovations for the Coral Program
 
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
IMCSummit 2015 - Day 1 Developer Track - Evolution of non-volatile memory exp...
 
CPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performance
 
SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...
SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...
SF Big Analytics & SF Machine Learning Meetup: Machine Learning at the Limit ...
 
Van jaconson netchannels
Van jaconson netchannelsVan jaconson netchannels
Van jaconson netchannels
 
Ceph Day Netherlands - Ceph @ BIT
Ceph Day Netherlands - Ceph @ BIT Ceph Day Netherlands - Ceph @ BIT
Ceph Day Netherlands - Ceph @ BIT
 
Experiences with Power 9 at A*STAR CRC
Experiences with Power 9 at A*STAR CRCExperiences with Power 9 at A*STAR CRC
Experiences with Power 9 at A*STAR CRC
 
What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014
What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014
What we unlearned_and_learned_by_moving_from_m9000_to_ssc_ukoug2014
 
Introduction to DPDK
Introduction to DPDKIntroduction to DPDK
Introduction to DPDK
 
Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...
Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...
Exploring Compiler Optimization Opportunities for the OpenMP 4.x Accelerator...
 
Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...
Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...
Alex Smola, Professor in the Machine Learning Department, Carnegie Mellon Uni...
 
Ceph Day Berlin: Ceph on All Flash Storage - Breaking Performance Barriers
Ceph Day Berlin: Ceph on All Flash Storage - Breaking Performance BarriersCeph Day Berlin: Ceph on All Flash Storage - Breaking Performance Barriers
Ceph Day Berlin: Ceph on All Flash Storage - Breaking Performance Barriers
 
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
 
Lrz kurs: big data analysis
Lrz kurs: big data analysisLrz kurs: big data analysis
Lrz kurs: big data analysis
 

More from inside-BigData.com

Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...inside-BigData.com
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networksinside-BigData.com
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
 
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...inside-BigData.com
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networksinside-BigData.com
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoringinside-BigData.com
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecastsinside-BigData.com
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Updateinside-BigData.com
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19inside-BigData.com
 
Energy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic TuningEnergy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic Tuninginside-BigData.com
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODinside-BigData.com
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Accelerationinside-BigData.com
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficientlyinside-BigData.com
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Erainside-BigData.com
 
CUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computingCUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Clusterinside-BigData.com
 

More from inside-BigData.com (20)

Major Market Shifts in IT
Major Market Shifts in ITMajor Market Shifts in IT
Major Market Shifts in IT
 
Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networks
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
 
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networks
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecasts
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Update
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19
 
Energy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic TuningEnergy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic Tuning
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
 
State of ARM-based HPC
State of ARM-based HPCState of ARM-based HPC
State of ARM-based HPC
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Acceleration
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Era
 
CUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computingCUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computing
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Cluster
 
Overview of HPC Interconnects
Overview of HPC InterconnectsOverview of HPC Interconnects
Overview of HPC Interconnects
 

Recently uploaded

Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsHyundai Motor Group
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksSoftradix Technologies
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptxLBM Solutions
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphNeo4j
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure servicePooja Nehwal
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAndikSusilo4
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 

Recently uploaded (20)

Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other Frameworks
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptx
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & Application
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 

Lustre Generational Performance Improvements & New Features

  • 2. HighLevelAbstract 2 Lustre* has had a number of compelling new features added in recent releases; this talk will look at those features in detail and see how well they all work together from both a performance and functionality perspective. Comparing some of the numbers from last year we will see how far the Lustre* filesystem has come in such a short period of time (LAD’16 to LAD’17), comparing the same use cases observing the generational improvements in the technology.
  • 3. Agenda • Hero Numbers: Generational Performance March 2016 – today • Generational Metadata performance improvements • Small file performance on OpenZFS (no Data-on-MDT) • Has LDISKFS changed since last year, how does performance look today • Scaling with DNE Phase 2 • How does PFL effect Performance 3
  • 4. SummaryofLastyearsTalk 4 • No LDISKFS Numbers with DNE: • Stability issues I observed have been resolved, see some new DNE 2 numbers with LDISKFS in this talk • DNE Phase 2 Scalability: • Scalability was reasonable before, do new Lustre release demonstrate better scalability • Using DNE 2 In Production: • Yes, I am still using DNE2 in production successfully for 18 months
  • 5. Copyright ® Intel Corporation 2016. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. * Other names and brands may be claimed as the property of others. TestbedArchitecture Bw-1-00 Bw-1-01 Bw-1-15 . . . . . . . . . . . . . . . Intel® HPC Orchestrator Headnode GenericLustre1 GenericLustre2 . . . . . . . GenericLustre10 Intel® Omni-Path Switching (100Gbps) 10G Base-T Ethernet Switching Intel® Omni Path Cabling 10G Base-T RJ45
  • 6. Copyright ® Intel Corporation 2016. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. * Other names and brands may be claimed as the property of others. TestbedArchitecture(Cont.) Server § 10x Generic Lustre servers with two slightly different configurations – Each System comprises of: – 2x Intel® Xeon E5-2697v3 (Haswell) CPU’s – 1x Intel® Omni-Path x16 HFI – 128GB DDR4 2133MHz Memory – Eight of the nodes contain - 4x Intel P3600 2.0TB 2.5” (U.2) NVMe devices, while the other two have 4x Intel® P3700 800GB 2.5” (U.2) NVMe devices – One node equipped with 2x Intel® S3700 400GB’s for MGT § 16x 2S Intel® Xeon E5v4 (Broadwell) Compute nodes – 1x Intel® HPC Orchestrator (Beta 2) Headnode – Hardware Components: – 2x Intel® Xeon E5-2697v4 (Broadwell) CPU’s – 1x Intel® Omni-Path x16 HFI – 128GB DDR4 2400MHz Memory – Local boot SSD § 100Gbps Intel® Omni-Path Fabric – None-blocking fabric with single switch design. – Server side optimisations: “options hfi1 sge_copy_mode=2 krcvqs=4 wss_threshold=70” – Improve generic RDMA performance, only recommended on Lustre server side that do physically do any MPI
  • 9. MetadataPerformance • LDISKFS quite close to ZFS for file create about 75% • Removal still a way to go • When testing on slower storage difference are marginal • Demonstrates a good level of improvement. 9 Lustre Metadata performance 2.9EA (Mid 2016) vs. 2.10.1 (today), MDTEST: Single MDT Performance. 0 10000 20000 30000 40000 50000 60000 70000 80000 90000 100000 b2.9 + zfs-0.6.5.7-1 b2.9 + zfs-0.7RC1 b2.10 + zfs 0.7RC5 b2.10.1 + zfs 0.7.1 b2.10 + ldiskfs File Create/Remove 1 MDT: Generational Improvements File Create File Remove /mnt/zlfs2/mdtest -i 3 -I 10000 -F -C -T -r -u -d /mnt/zlfs2/test1.out
  • 10. SmallFilePerformance(4k) • Single MDT operation up 4x compared to this time last year • Scaling of DNE2 still not linear, but better • Create performance trails off due to lack of clients • Clear benefit versus the previous release 10 Lustre Small file performance 2.9EA (Mid 2016) vs. 2.10.1 (today), MDTEST: Single MDT Performance. No Data-on-MDT used leveraging DNE Phase 2 up to 8 MDT’s on Separate servers. 0 50000 100000 150000 200000 250000 300000 350000 1x MDT 2x MDT 3x MDT 4x MDT 5x MDT 6x MDT 7x MDT 8x MDT Small File Performance: Lustre 2.9EA (ZFS 0.6.5-7) vs. Lustre 2.10.1 (ZFS 0.7.1) LAD'17 Create LAD'17 Removal LAD'16 Create LAD'16 Removal /mnt/zlfs2/mdtest -i 3 -I 10000 -z 1 -b 1 -L -u -F –w 4096 -d /mnt/zlfs2/*DNE-DIR*/test1.out
  • 12. LDISKFSPerformance • Some performance boost was expected, but not this much • Shows positive trend version to version • 2.10 to 2.10.1 - LU-7899 12 Lustre 2.9EA vs. Lustre 2.10.0 vs. Lustre 2.10.1 56887.869 74418.092 79191.709 94895.636 125084.373 107051.123 0 20000 40000 60000 80000 100000 120000 140000 File Create File Removal LDISKFS: Lustre 2.9EA vs. Lustre 2.10.0 vs. Lustre 2.10.1EA LDISKFS 2.9EA LDISKFS 2.10.0 LDISKFS 2.10.1EA
  • 14. Normalised:ScalingLAD’16toLAD’17:DNEPhase2 • Neither are linear • Overall scalability dropped a little, but ultimate number is much higher • Some work to do to get this close to OST scalability 14 Generational Performance and scalability of DNE Phase 2 on OpenZFS 1x MDT 2x MDT 3x MDT 4x MDT 5x MDT 6x MDT 7x MDT 8x MDT Normalised: DNE Phase 2 Lustre 2.9EA vs. Lustre 2.10.1 2.9EA ZFS 0.6.5 2.10.1 ZFS 0.7.1 Linear
  • 16. LDISKFSDNEPhase2Scaling&Functionality • Totally stable, versus pervious testing • Scaling stops after 2 MDT’s • Clients unable to push that much I/O • You can see from the previous slide 200 – 250k is my HW limit 16 Following up from last year where I couldn’t give DNE2 on LDISKFS. 0 200000 400000 600000 800000 1000000 1200000 1x MDT 2x MDT 3x MDT 4x MDT 5x MDT 6x MDT 7x MDT 8x MDT DNE Phase 2 Scaling: Lustre 2.10.1 LDISKFS Create LDISKFS Linear
  • 18. ProgressiveFileLayout 18 Stripe =1 Stripe =2 Stripe =4 Stripe =8 Stripe =16 4MB 64MB 256MB 1GB 10GB • Example Layout • lfs setstripe -E 4M -c 1 -E 64M -c 2 -E 256M -c 4 -E 1G -c 8 -E 10G -c 16 -E -1 -c - 1 /mnt/zlfs2/pfl_test01/
  • 19. 19 /mnt/zlfs2/pfl_test01/ lcm_layout_gen: 0 lcm_entry_count: 6 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 0 lcme_extent.e_end: 4194304 stripe_count: 1 stripe_size: 1048576 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 4194304 lcme_extent.e_end: 67108864 stripe_count: 2 stripe_size: 1048576 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 67108864 lcme_extent.e_end: 268435456 stripe_count: 4 stripe_size: 1048576 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 268435456 lcme_extent.e_end: 1073741824 stripe_count: 8 stripe_size: 1048576 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 1073741824 lcme_extent.e_end: 10737418240 stripe_count: 16 stripe_size: 1048576 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 10737418240 lcme_extent.e_end: EOF stripe_count: -1 stripe_size: 1048576 stripe_offset: -1
  • 20. PFLPerformanceWhenusingIOR • Each files stripe dynamically grows based on file size • Write performance up 4.6%, read within margin of error • Certainly not detrimental to performance 20 Lustre 2.10.1; IOR Performance, file per process (256 files, 16GB per file) mean performance MB/s. PFL as described before versus traditional -1 stripe. 0 5000 10000 15000 20000 25000 30000 35000 40000 45000 Read Write PFL vs. No PFL: IOR MB/s PFL No PFL /mnt/zlfs2/IOR -wr -C -F -i 3 -t 1m -b 1m -s 16384 -a MPIIO -o /mnt/zlfs2/pfl_new/testme1.file
  • 21. 21 /mnt/zlfs2/pfl_new1/ lcm_layout_gen: 0 lcm_entry_count: 6 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 0 lcme_extent.e_end: 4194304 stripe_count: 1 stripe_size: 1048576 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 4194304 lcme_extent.e_end: 67108864 stripe_count: 2 stripe_size: 2097152 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 67108864 lcme_extent.e_end: 268435456 stripe_count: 4 stripe_size: 16777216 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 268435456 lcme_extent.e_end: 1073741824 stripe_count: 8 stripe_size: 33554432 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 1073741824 lcme_extent.e_end: 10737418240 stripe_count: 16 stripe_size: 134217728 stripe_offset: -1 lcme_id: N/A lcme_flags: 0 lcme_extent.e_start: 10737418240 lcme_extent.e_end: EOF stripe_count: -1 stripe_size: 268435456 stripe_offset: -1
  • 22. PFLPerformancewithusingIOR(cont.) • PFL is giving us the opportunity to optimise the stripe relative to data type • Write up 8.7% on base results and and 4.1% relative to test one • Reads get a 9.2% boost with larger stripe sizes 22 Lets optimise the stripe size this time, assuming the a larger stripe size for the higher stripe counts lfs setstripe -E 4M -S 1M -c 1 -E 64M -S 2M -c 2 -E 256M -S 16M -c 4 -E 1G -S 32M -c 64 -E 10G -S 128M -c 16 -E -1 -S 256M -c -1 /mnt/zlfs2/pfl_new/ 0 5000 10000 15000 20000 25000 30000 35000 40000 45000 50000 Read Write PFL vs. No PFL: IOR MB/s PFL No PFL PFL Optimised /mnt/zlfs2/IOR -wr -C -F -i 3 -t 1m -b 1m -s 16384 -a MPIIO -o /mnt/zlfs2/pfl_new/testme1.file
  • 24. KeyTakeaways • Small file / metadata performance across the board is simply just better • Amazing work done on OpenZFS to get 0.7.1 to where it is today • Performance is comparable to LDISKFS of previous releases • Overall DNE Phase 2 scalability is very similar to what we have seen before • Overall usage and stability feels better, but was good before • Optimising Striping layouts with PFL is essential, striping is done for you and can be configured for best performance 24
  • 26. Copyright ® Intel Corporation 2016. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. * Other names and brands may be claimed as the property of others. LegalInformation All information provided here is subject to change without notice. Contact your Intel representative to obtain the latest Intel product specifications and roadmaps Tests document performance of components on a particular test, in specific systems. Differences in hardware, software, or configuration will affect actual performance. Consult other sources of information to evaluate performance as you consider your purchase. For more complete information about performance and benchmark results, visit http://www.intel.com/performance. Intel technologies’ features and benefits depend on system configuration and may require enabled hardware, software or service activation. Performance varies depending on system configuration. No computer system can be absolutely secure. Check with your system manufacturer or retailer or learn more at http://www.intel.com/content/www/us/en/software/intel-solutions-for-lustre-software.html. Intel technologies may require enabled hardware, specific software, or services activation. Check with your system manufacturer or retailer. You may not use or facilitate the use of this document in connection with any infringement or other legal analysis concerning Intel products described herein. You agree to grant Intel a non-exclusive, royalty- free license to any patent claim thereafter drafted which includes subject matter disclosed herein. No license (express or implied, by estoppel or otherwise) to any intellectual property rights is granted by this document. The products described may contain design defects or errors known as errata which may cause the product to deviate from published specifications. Current characterized errata are available on request. Intel disclaims all express and implied warranties, including without limitation, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement, as well as any warranty arising from course of performance, course of dealing, or usage in trade. This document contains information on products, services and/or processes in development. All information provided here is subject to change without notice. Contact your Intel representative to obtain the latest forecast, schedule, specifications and roadmaps. A "Mission Critical Application" is any application in which failure of the Intel Product could result, directly or indirectly, in personal injury or death. SHOULD YOU PURCHASE OR USE INTEL'S PRODUCTS FOR ANY SUCH MISSION CRITICAL APPLICATION, YOU SHALL INDEMNIFY AND HOLD INTEL AND ITS SUBSIDIARIES, SUBCONTRACTORS AND AFFILIATES, AND THE DIRECTORS, OFFICERS, AND EMPLOYEES OF EACH, HARMLESS AGAINST ALL CLAIMS COSTS, DAMAGES, AND EXPENSES AND REASONABLE ATTORNEYS' FEES ARISING OUT OF, DIRECTLY OR INDIRECTLY, ANY CLAIM OF PRODUCT LIABILITY, PERSONAL INJURY, OR DEATH ARISING IN ANY WAY OUT OF SUCH MISSION CRITICAL APPLICATION, WHETHER OR NOT INTEL OR ITS SUBCONTRACTOR WAS NEGLIGENT IN THE DESIGN, MANUFACTURE, OR WARNING OF THE INTEL PRODUCT OR ANY OF ITS PARTS. Intel may make changes to specifications and product descriptions at any time, without notice. Designers must not rely on the absence or characteristics of any features or instructions marked "reserved" or "undefined". Intel reserves these for future definition and shall have no responsibility whatsoever for conflicts or incompatibilities arising from future changes to them. The information here is subject to change without notice. Do not finalize a design with this information. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. * Other names and brands may be claimed as the property of others. © 2016 Intel Corporation
  • 27. Copyright © 2017 Intel Corporation. Intel, the Intel logo are trademarks of Intel Corporation. Other names and brands may be claimed as the property of others.