This document summarizes a presentation on introducing the Persistent Memory Development Kit (PMDK) into PostgreSQL to utilize persistent memory (PMEM). The presentation covers: (1) hacking the PostgreSQL write-ahead log (WAL) and relation files to directly memory copy to PMEM, (2) evaluating the hacks which showed a 3% improvement to transactions and 30% reduction to checkpoint time, and (3) tips for PMEM programming like cache flushing and avoiding volatile layers.
Slide at OpenStack Summit 2018 Vancouver
Session Info and Video: https://www.openstack.org/videos/vancouver-2018/can-we-boost-more-hpc-performance-integrate-ibm-power-servers-with-gpus-to-openstack-environment
Slides at OpenStack Summit 2017 Sydney
Session Info and Video: https://www.openstack.org/videos/sydney-2017/100gbps-openstack-for-providing-high-performance-nfv
Slide at OpenStack Summit 2018 Vancouver
Session Info and Video: https://www.openstack.org/videos/vancouver-2018/can-we-boost-more-hpc-performance-integrate-ibm-power-servers-with-gpus-to-openstack-environment
Slides at OpenStack Summit 2017 Sydney
Session Info and Video: https://www.openstack.org/videos/sydney-2017/100gbps-openstack-for-providing-high-performance-nfv
HKG18-411 - Introduction to OpenAMP which is an open source solution for hete...Linaro
Session ID: HKG18-411
Session Name: HKG18-411 - Introduction to OpenAMP which is an open source solution for heterogeneous system orchestration and communication
Speaker: Wendy Liang
Track: IoT, Embedded
★ Session Summary ★
Introduction to OpenAMP which is an open source solution for heterogeneous system orchestration and communication
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/hkg18/hkg18-411/
Presentation: http://connect.linaro.org.s3.amazonaws.com/hkg18/presentations/hkg18-411.pdf
Video: http://connect.linaro.org.s3.amazonaws.com/hkg18/videos/hkg18-411.mp4
---------------------------------------------------
★ Event Details ★
Linaro Connect Hong Kong 2018 (HKG18)
19-23 March 2018
Regal Airport Hotel Hong Kong
---------------------------------------------------
Keyword: IoT, Embedded
'http://www.linaro.org'
'http://connect.linaro.org'
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
Clear Containers is an Open Containers Initiative (OCI) “runtime” that launches an Intel VT-x secured hypervisor rather than a standard Linux container. An introduction of Clear Containers will be provided, followed by an overview of CNM networking plugins which have been created to enhance network connectivity using Clear Containers. More specifically, we will show demonstrations of using VPP with DPDK and SRIO-v based networks to connect Clear Containers. Pending time we will provide and walk through a hands on example of using VPP with Clear Containers.
About the speaker: Manohar Castelino is a Principal Engineer for Intel’s Open Source Technology Center. Manohar has worked on networking, network management, network processors and virtualization for over 15 years. Manohar is currently an architect and developer with the ciao (clearlinux.org/ciao) and the clear containers (https://github.com/01org/cc-oci-runtime) projects focused on networking. Manohar has spoken at many Container Meetups and internal conferences.
Session ID: SFO17-509
Session Name: Deep Learning on ARM Platforms
- SFO17-509
Speaker: Jammy Zhou
Track:
★ Session Summary ★
A new era of deep learning is coming with algorithm evolvement, powerful computing platforms and large dataset availability. This session will focus on existing and potential heterogeneous accelerator solutions (GPU, FPGA, DSP, and etc) for ARM platforms and the work ahead from platform perspective.
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/sfo17/sfo17-509/
Presentation:
Video:
---------------------------------------------------
★ Event Details ★
Linaro Connect San Francisco 2017 (SFO17)
25-29 September 2017
Hyatt Regency San Francisco Airport
---------------------------------------------------
Keyword:
http://www.linaro.org
http://connect.linaro.org
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://twitter.com/linaroorg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
In this deck from the ECSS Symposium, Abe Stern from NVIDIA presents: CUDA-Python and RAPIDS for blazing fast scientific computing.
"We will introduce Numba and RAPIDS for GPU programming in Python. Numba allows us to write just-in-time compiled CUDA code in Python, giving us easy access to the power of GPUs from a powerful high-level language. RAPIDS is a suite of tools with a Python interface for machine learning and dataframe operations. Together, Numba and RAPIDS represent a potent set of tools for rapid prototyping, development, and analysis for scientific computing. We will cover the basics of each library and go over simple examples to get users started. Finally, we will briefly highlight several other relevant libraries for GPU programming."
Watch the video: https://wp.me/p3RLHQ-lvu
Learn more: https://developer.nvidia.com/rapids
and
https://www.xsede.org/for-users/ecss/ecss-symposium
Sign up for our insideHPC Newsletter: http://insidehp.com/newsletter
Review of the CLIP (Cloud Infrastructure Project) at the Vienna BioCenter. CLIP is an OpenStack deployment for research computing, set out to replace on-site legacy HPC systems. We talk about how we have setup our continuous deployment process, our infrastructure as code approach, continuous testing and verification, monitoring and the pitfalls and surprises we encountered along the way.
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...Linaro
Event: Arm Architecture HPC Workshop by Linaro and HiSilicon
Location: Santa Clara, CA
Speaker: Andrew J Younge
Talk Title: Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Supercomputing
Talk Desc: The Vanguard program looks to expand the potential technology choices for leadership-class High Performance Computing (HPC) platforms, not only for the National Nuclear Security Administration (NNSA) but for the Department of Energy (DOE) and wider HPC community. Specifically, there is a need to expand the supercomputing ecosystem by investing and developing emerging, yet-to-be-proven technologies and address both hardware and software challenges together, as well as to prove-out the viability of such novel platforms for production HPC workloads.
The first deployment of the Vanguard program will be Astra, a prototype Petascale Arm supercomputer to be sited at Sandia National Laboratories during 2018. This talk will focus on the arthictecural details of Astra and the significant investments being made towards the maturing the Arm software ecosystem. Furthermore, we will share initial performance results based on our pre-general availability testbed system and outline several planned research activities for the machine.
Bio: Andrew Younge is a R&D Computer Scientist at Sandia National Laboratories with the Scalable System Software group. His research interests include Cloud Computing, Virtualization, Distributed Systems, and energy efficient computing. Andrew has a Ph.D in Computer Science from Indiana University, where he was the Persistent Systems fellow and a member of the FutureGrid project, an NSF-funded experimental cyberinfrastructure test-bed. Over the years, Andrew has held visiting positions at the MITRE Corporation, the University of Southern California / Information Sciences Institute, and the University of Maryland, College Park. He received his Bachelors and Masters of Science from the Computer Science Department at Rochester Institute of Technology (RIT) in 2008 and 2010, respectively.
In this deck, Jean-Pierre Panziera from Atos presents: BXI - Bull eXascale Interconnect.
"Exascale entails an explosion of performance, of the number of nodes/cores, of data volume and data movement. At such a scale, optimizing the network that is the backbone of the system becomes a major contributor to global performance. The interconnect is going to be a key enabling technology for exascale systems. This is why one of the cornerstones of Bull’s exascale program is the development of our own new-generation interconnect. The Bull eXascale Interconnect or BXI introduces a paradigm shift in terms of performance, scalability, efficiency, reliability and quality of service for extreme workloads."
Watch the video: http://wp.me/p3RLHQ-gJa
Learn more: https://bull.com/bull-exascale-interconnect/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Ronald P. Luijten from IBM Research in Zurich presents: DOME 64-bit μDataCenter.
I like to call it a datacenter in a shoebox. With the combination of power and energy efficiency, we believe the microserver will be of interest beyond the DOME project, particularly for cloud data centers and Big Data analytics applications."
The microserver’s team has designed and demonstrated a prototype 64-bit microserver using a PowerPC based chip from Freescale Semiconductor running Linux Fedora and IBM DB2. At 133 × 55 mm2 the microserver contains all of the essential functions of today’s servers, which are 4 to 10 times larger in size. Not only is the microserver compact, it is also very energy-efficient.
Watch the video: http://wp.me/p3RLHQ-gJM
Learn more: https://www.zurich.ibm.com/microserver/
Sign up for our insideHPC Newsletter: http://insideHPC/newsletter
Utilizing AMD GPUs: Tuning, programming models, and roadmapGeorge Markomanolis
A presentation at FOSDEM 2022 about AMD GPUs, tuning, programming models and software roadmap. This is continuation from the previous talk (FOSDEM 2021)
Nagios Conference 2012 - Dan Wittenberg - Case Study: Scaling Nagios Core at ...Nagios
Dan Wittenberg's presentation on using Nagios at a Fortune 50 Company
The presentation was given during the Nagios World Conference North America held Sept 25-28th, 2012 in Saint Paul, MN. For more information on the conference (including photos and videos), visit: http://go.nagios.com/nwcna
HKG18-411 - Introduction to OpenAMP which is an open source solution for hete...Linaro
Session ID: HKG18-411
Session Name: HKG18-411 - Introduction to OpenAMP which is an open source solution for heterogeneous system orchestration and communication
Speaker: Wendy Liang
Track: IoT, Embedded
★ Session Summary ★
Introduction to OpenAMP which is an open source solution for heterogeneous system orchestration and communication
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/hkg18/hkg18-411/
Presentation: http://connect.linaro.org.s3.amazonaws.com/hkg18/presentations/hkg18-411.pdf
Video: http://connect.linaro.org.s3.amazonaws.com/hkg18/videos/hkg18-411.mp4
---------------------------------------------------
★ Event Details ★
Linaro Connect Hong Kong 2018 (HKG18)
19-23 March 2018
Regal Airport Hotel Hong Kong
---------------------------------------------------
Keyword: IoT, Embedded
'http://www.linaro.org'
'http://connect.linaro.org'
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
Clear Containers is an Open Containers Initiative (OCI) “runtime” that launches an Intel VT-x secured hypervisor rather than a standard Linux container. An introduction of Clear Containers will be provided, followed by an overview of CNM networking plugins which have been created to enhance network connectivity using Clear Containers. More specifically, we will show demonstrations of using VPP with DPDK and SRIO-v based networks to connect Clear Containers. Pending time we will provide and walk through a hands on example of using VPP with Clear Containers.
About the speaker: Manohar Castelino is a Principal Engineer for Intel’s Open Source Technology Center. Manohar has worked on networking, network management, network processors and virtualization for over 15 years. Manohar is currently an architect and developer with the ciao (clearlinux.org/ciao) and the clear containers (https://github.com/01org/cc-oci-runtime) projects focused on networking. Manohar has spoken at many Container Meetups and internal conferences.
Session ID: SFO17-509
Session Name: Deep Learning on ARM Platforms
- SFO17-509
Speaker: Jammy Zhou
Track:
★ Session Summary ★
A new era of deep learning is coming with algorithm evolvement, powerful computing platforms and large dataset availability. This session will focus on existing and potential heterogeneous accelerator solutions (GPU, FPGA, DSP, and etc) for ARM platforms and the work ahead from platform perspective.
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/sfo17/sfo17-509/
Presentation:
Video:
---------------------------------------------------
★ Event Details ★
Linaro Connect San Francisco 2017 (SFO17)
25-29 September 2017
Hyatt Regency San Francisco Airport
---------------------------------------------------
Keyword:
http://www.linaro.org
http://connect.linaro.org
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://twitter.com/linaroorg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
In this deck from the ECSS Symposium, Abe Stern from NVIDIA presents: CUDA-Python and RAPIDS for blazing fast scientific computing.
"We will introduce Numba and RAPIDS for GPU programming in Python. Numba allows us to write just-in-time compiled CUDA code in Python, giving us easy access to the power of GPUs from a powerful high-level language. RAPIDS is a suite of tools with a Python interface for machine learning and dataframe operations. Together, Numba and RAPIDS represent a potent set of tools for rapid prototyping, development, and analysis for scientific computing. We will cover the basics of each library and go over simple examples to get users started. Finally, we will briefly highlight several other relevant libraries for GPU programming."
Watch the video: https://wp.me/p3RLHQ-lvu
Learn more: https://developer.nvidia.com/rapids
and
https://www.xsede.org/for-users/ecss/ecss-symposium
Sign up for our insideHPC Newsletter: http://insidehp.com/newsletter
Review of the CLIP (Cloud Infrastructure Project) at the Vienna BioCenter. CLIP is an OpenStack deployment for research computing, set out to replace on-site legacy HPC systems. We talk about how we have setup our continuous deployment process, our infrastructure as code approach, continuous testing and verification, monitoring and the pitfalls and surprises we encountered along the way.
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...Linaro
Event: Arm Architecture HPC Workshop by Linaro and HiSilicon
Location: Santa Clara, CA
Speaker: Andrew J Younge
Talk Title: Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Supercomputing
Talk Desc: The Vanguard program looks to expand the potential technology choices for leadership-class High Performance Computing (HPC) platforms, not only for the National Nuclear Security Administration (NNSA) but for the Department of Energy (DOE) and wider HPC community. Specifically, there is a need to expand the supercomputing ecosystem by investing and developing emerging, yet-to-be-proven technologies and address both hardware and software challenges together, as well as to prove-out the viability of such novel platforms for production HPC workloads.
The first deployment of the Vanguard program will be Astra, a prototype Petascale Arm supercomputer to be sited at Sandia National Laboratories during 2018. This talk will focus on the arthictecural details of Astra and the significant investments being made towards the maturing the Arm software ecosystem. Furthermore, we will share initial performance results based on our pre-general availability testbed system and outline several planned research activities for the machine.
Bio: Andrew Younge is a R&D Computer Scientist at Sandia National Laboratories with the Scalable System Software group. His research interests include Cloud Computing, Virtualization, Distributed Systems, and energy efficient computing. Andrew has a Ph.D in Computer Science from Indiana University, where he was the Persistent Systems fellow and a member of the FutureGrid project, an NSF-funded experimental cyberinfrastructure test-bed. Over the years, Andrew has held visiting positions at the MITRE Corporation, the University of Southern California / Information Sciences Institute, and the University of Maryland, College Park. He received his Bachelors and Masters of Science from the Computer Science Department at Rochester Institute of Technology (RIT) in 2008 and 2010, respectively.
In this deck, Jean-Pierre Panziera from Atos presents: BXI - Bull eXascale Interconnect.
"Exascale entails an explosion of performance, of the number of nodes/cores, of data volume and data movement. At such a scale, optimizing the network that is the backbone of the system becomes a major contributor to global performance. The interconnect is going to be a key enabling technology for exascale systems. This is why one of the cornerstones of Bull’s exascale program is the development of our own new-generation interconnect. The Bull eXascale Interconnect or BXI introduces a paradigm shift in terms of performance, scalability, efficiency, reliability and quality of service for extreme workloads."
Watch the video: http://wp.me/p3RLHQ-gJa
Learn more: https://bull.com/bull-exascale-interconnect/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Ronald P. Luijten from IBM Research in Zurich presents: DOME 64-bit μDataCenter.
I like to call it a datacenter in a shoebox. With the combination of power and energy efficiency, we believe the microserver will be of interest beyond the DOME project, particularly for cloud data centers and Big Data analytics applications."
The microserver’s team has designed and demonstrated a prototype 64-bit microserver using a PowerPC based chip from Freescale Semiconductor running Linux Fedora and IBM DB2. At 133 × 55 mm2 the microserver contains all of the essential functions of today’s servers, which are 4 to 10 times larger in size. Not only is the microserver compact, it is also very energy-efficient.
Watch the video: http://wp.me/p3RLHQ-gJM
Learn more: https://www.zurich.ibm.com/microserver/
Sign up for our insideHPC Newsletter: http://insideHPC/newsletter
Utilizing AMD GPUs: Tuning, programming models, and roadmapGeorge Markomanolis
A presentation at FOSDEM 2022 about AMD GPUs, tuning, programming models and software roadmap. This is continuation from the previous talk (FOSDEM 2021)
Nagios Conference 2012 - Dan Wittenberg - Case Study: Scaling Nagios Core at ...Nagios
Dan Wittenberg's presentation on using Nagios at a Fortune 50 Company
The presentation was given during the Nagios World Conference North America held Sept 25-28th, 2012 in Saint Paul, MN. For more information on the conference (including photos and videos), visit: http://go.nagios.com/nwcna
Keiichi Matsuzawa, Takahiro Shinagawa.
In Proceedings of the 9th IEEE International Conference on Cloud Computing Technology and Science (CloudCom 2017), Dec 2017.
http://dx.doi.org/10.1109/CloudCom.2017.20
Best paper canditate
Presentation in GNOME Asia 2015 which focus in the topic of "In Memory Computing" with the tittle An Overview, Maximize the Ability of The GNU/LINUX Operating System using In Memory Computation for Academic, Business and Government.
This paper are mainly focus in moving the root file system from storage device into memory (RAM) device, minimizing bottleneck and maximizing performance, benchmarking and a little snippet script how to do that.
C++ Programming and the Persistent Memory Developers KitIntel® Software
Topics
Introduction to Persistent Memory
Introduction to Persistent Memory Developers Kit (PMDK)
Working with PMDK
Persistent Memory Programming with PMDK C++ Bindings
Developed for the Denver Art Museum by Ashley Blewer, this slide-deck covers some of the basics of diagnosing issues with Archivematica. Ashley covers everything from the software components involved with Archivematica, to monitoring logs, system monitoring, and upgrading your system. The presentation concludes with some useful links for tech-savvy preservationists, and Archivematica-unfamiliar system's administrators!
Implementing data and databases on K8s within the Dutch governmentDoKC
A small walkthrough of projects within the dutch government running Data(bases) on OpenShift. This talk shares success stories, provides a proven recipe to `get it done` and debunks some of the FUD.
About Sebastiaan:
I have always been a weird DBA, trying to combine Databases with out-of-the-box thinking and a DevOps mindset. Around 2016 I fell in love with both Postgres and Kubernetes, and I then committed my life to enabling Dutch organisations with running their Database workloads CloudNative.
Over the last few years I worked as a private contractor for 2 large government agencies doing exactly that, and I want to share my and others (success stories) hoping to enable and inspire Data on Kubernetes adoption.
Backup and restore methods are concepts that everyone knows the importance of. Over the years, open-source tools emerged like MyDumper, Xtrabackup, and Mariabackup. Also, with MySQL 8 new shell, new utils for dump and restore were introduced as well.
In this presentation, we are going to compare the newest backup/restore methods with the most used ones. We will see how parallelization can influence the speed of backup and restore process and also how the compression algorithms can influence the performance.
In this talk, we will compare mysqldump, mydumper/myloader, mysqlpump, MySQL Shell utils, and Xtrabackup.
The Forefront of the Development for NVDIMM on Linux KernelYasunori Goto
This is talk for Open Source Summit Japan 2020
--------------------------
NVDIMM (Non Volatile DIMM) is the most interesting device, because it has not only characteristic of memory but also storage. To support NVDIMM, Linux kernel provides three access methods for users. - Storage (Sector) mode - Filesystem DAX(=Direct Access) mode - Device DAX mode. In the above three methods, Filesystem DAX is the most expected access method, because applications can write data to the NVDIMM area directory, and it is easier to use than Device DAX mode. So, some software already uses it with official support. However, Filesystem-DAX is still "experimental status" in the upstream community due to some difficult issues . In this session, Yasunori Goto will talk to the forefront of the development of NVDIMM, and Ruan Shiyang will talk about his challenge with the latest status from CLK2019.
Performance Optimization of SPH Algorithms for Multi/Many-Core ArchitecturesDr. Fabio Baruffa
In the framework of the Intel Parallel Computing Centre at the Research Campus Garching in Munich, our group at LRZ presents recent results on performance optimization of Gadget-3, a widely used community code for computational astrophysics. We identify and isolate a sample code kernel, which is representative of a typical Smoothed Particle Hydrodynamics (SPH) algorithm and focus on threading parallelism optimization, change of the data layout into Structure of Arrays (SoA), compiler auto-vectorization and algorithmic improvements in the particle sorting. We measure lower execution time and improved threading scalability both on Intel Xeon (2.6× on Ivy Bridge) and Xeon Phi (13.7× on Knights Corner) systems. First tests on second generation Xeon Phi (Knights Landing) demonstrate the portability of the devised optimization solutions to upcoming architectures.
Webinar: Three Reasons Why NAS is No Good for AI and Machine LearningStorage Switzerland
Artificial Intelligence (AI) and Machine Learning (ML) are becoming mainstream initiatives at many organizations. Data is at the heart of AI and ML. Immediate access to large data sets is pivotal to successful ML outcomes. Without data, there is no learning. The goal of AI and ML is to try to simulate human thinking and understanding. AI and ML initiatives cannot however be realized unless the data processing layer has immediate access to, and a constant supply of, data.
The problem is that NAS solutions, often those designed for HPC environments, is what most organizations try to leverage as the AI/ML storage architectures. Legacy storage systems, like NAS, cannot support AI and ML workloads, because they were architected when spinning disk and slower networking technologies were the industry standard.
Join Storage Switzerland and WekaIO for our on demand webinar to learn the three reasons why NAS is no good for AI and ML:
* NAS wasn’t architected to leverage today’s flash technology and can’t keep pace with the I/O demands, leaving GPUs starved for data
* NAS has no or very rudimentary Cloud Integration. Tiering to the cloud can play an integral role in AI and ML workloads
* NAS data protection schemes are expensive given the amount of data required to feed an AI/ML environment
We have published a document, "A Global Data Infrastructure for Data Sharing Between Businesses".
This document introduces the current trends toward the implementation of digital management tools that support cross border data sharing between businesses, which will be indispensable for future business transformations and pandemic responses. Today we find ourselves at the confluence of multiple evolving global trends. These include the emergence of new data driven business models, the expansion of B2B platform business, the accelerating pace of digital transformation, the growing expectations for the fulfillment of Sustainable Development Goals (SDGs) and other social needs, the rise of New Glocalism, the growth of stakeholder capitalism, and the Great Reset. In this article, we discuss the challenges of establishing a global data infrastructure for data sharing between businesses as a key ICT infrastructure for the construction of a next generation society, and the efforts that are being made to address these challenges.
NTT Laboratories
J. Arai, S. Yagi, H. Uchiyama, T. Honjo, T. Inagaki, K. Inaba, T. Ikuta, H. Takesue, K. Horikawa
This material is a poster exhibited at the ITBL community booth in SC19 (The International Conference for High Performance Computing, Networking, Storage, and Analysis 2019).
NTT Software Innovation Center
Hiroki Miura, Kota Tsuyuzaki, Junya Arai, Kohei Yamaguchi, Kengo Okitsu, Shinji Morishita
This material is a poster exhibited at the ITBL community booth in SC19 (The International Conference for High Performance Computing, Networking, Storage, and Analysis 2019).
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Quarkus Hidden and Forbidden ExtensionsMax Andersen
Quarkus has a vast extension ecosystem and is known for its subsonic and subatomic feature set. Some of these features are not as well known, and some extensions are less talked about, but that does not make them less interesting - quite the opposite.
Come join this talk to see some tips and tricks for using Quarkus and some of the lesser known features, extensions and development techniques.
AI Fusion Buddy Review: Brand New, Groundbreaking Gemini-Powered AI AppGoogle
AI Fusion Buddy Review: Brand New, Groundbreaking Gemini-Powered AI App
👉👉 Click Here To Get More Info 👇👇
https://sumonreview.com/ai-fusion-buddy-review
AI Fusion Buddy Review: Key Features
✅Create Stunning AI App Suite Fully Powered By Google's Latest AI technology, Gemini
✅Use Gemini to Build high-converting Converting Sales Video Scripts, ad copies, Trending Articles, blogs, etc.100% unique!
✅Create Ultra-HD graphics with a single keyword or phrase that commands 10x eyeballs!
✅Fully automated AI articles bulk generation!
✅Auto-post or schedule stunning AI content across all your accounts at once—WordPress, Facebook, LinkedIn, Blogger, and more.
✅With one keyword or URL, generate complete websites, landing pages, and more…
✅Automatically create & sell AI content, graphics, websites, landing pages, & all that gets you paid non-stop 24*7.
✅Pre-built High-Converting 100+ website Templates and 2000+ graphic templates logos, banners, and thumbnail images in Trending Niches.
✅Say goodbye to wasting time logging into multiple Chat GPT & AI Apps once & for all!
✅Save over $5000 per year and kick out dependency on third parties completely!
✅Brand New App: Not available anywhere else!
✅ Beginner-friendly!
✅ZERO upfront cost or any extra expenses
✅Risk-Free: 30-Day Money-Back Guarantee!
✅Commercial License included!
See My Other Reviews Article:
(1) AI Genie Review: https://sumonreview.com/ai-genie-review
(2) SocioWave Review: https://sumonreview.com/sociowave-review
(3) AI Partner & Profit Review: https://sumonreview.com/ai-partner-profit-review
(4) AI Ebook Suite Review: https://sumonreview.com/ai-ebook-suite-review
#AIFusionBuddyReview,
#AIFusionBuddyFeatures,
#AIFusionBuddyPricing,
#AIFusionBuddyProsandCons,
#AIFusionBuddyTutorial,
#AIFusionBuddyUserExperience
#AIFusionBuddyforBeginners,
#AIFusionBuddyBenefits,
#AIFusionBuddyComparison,
#AIFusionBuddyInstallation,
#AIFusionBuddyRefundPolicy,
#AIFusionBuddyDemo,
#AIFusionBuddyMaintenanceFees,
#AIFusionBuddyNewbieFriendly,
#WhatIsAIFusionBuddy?,
#HowDoesAIFusionBuddyWorks
AI Pilot Review: The World’s First Virtual Assistant Marketing SuiteGoogle
AI Pilot Review: The World’s First Virtual Assistant Marketing Suite
👉👉 Click Here To Get More Info 👇👇
https://sumonreview.com/ai-pilot-review/
AI Pilot Review: Key Features
✅Deploy AI expert bots in Any Niche With Just A Click
✅With one keyword, generate complete funnels, websites, landing pages, and more.
✅More than 85 AI features are included in the AI pilot.
✅No setup or configuration; use your voice (like Siri) to do whatever you want.
✅You Can Use AI Pilot To Create your version of AI Pilot And Charge People For It…
✅ZERO Manual Work With AI Pilot. Never write, Design, Or Code Again.
✅ZERO Limits On Features Or Usages
✅Use Our AI-powered Traffic To Get Hundreds Of Customers
✅No Complicated Setup: Get Up And Running In 2 Minutes
✅99.99% Up-Time Guaranteed
✅30 Days Money-Back Guarantee
✅ZERO Upfront Cost
See My Other Reviews Article:
(1) TubeTrivia AI Review: https://sumonreview.com/tubetrivia-ai-review
(2) SocioWave Review: https://sumonreview.com/sociowave-review
(3) AI Partner & Profit Review: https://sumonreview.com/ai-partner-profit-review
(4) AI Ebook Suite Review: https://sumonreview.com/ai-ebook-suite-review
In 2015, I used to write extensions for Joomla, WordPress, phpBB3, etc and I ...Juraj Vysvader
In 2015, I used to write extensions for Joomla, WordPress, phpBB3, etc and I didn't get rich from it but it did have 63K downloads (powered possible tens of thousands of websites).
Climate Science Flows: Enabling Petabyte-Scale Climate Analysis with the Eart...Globus
The Earth System Grid Federation (ESGF) is a global network of data servers that archives and distributes the planet’s largest collection of Earth system model output for thousands of climate and environmental scientists worldwide. Many of these petabyte-scale data archives are located in proximity to large high-performance computing (HPC) or cloud computing resources, but the primary workflow for data users consists of transferring data, and applying computations on a different system. As a part of the ESGF 2.0 US project (funded by the United States Department of Energy Office of Science), we developed pre-defined data workflows, which can be run on-demand, capable of applying many data reduction and data analysis to the large ESGF data archives, transferring only the resultant analysis (ex. visualizations, smaller data files). In this talk, we will showcase a few of these workflows, highlighting how Globus Flows can be used for petabyte-scale climate analysis.
Globus Connect Server Deep Dive - GlobusWorld 2024Globus
We explore the Globus Connect Server (GCS) architecture and experiment with advanced configuration options and use cases. This content is targeted at system administrators who are familiar with GCS and currently operate—or are planning to operate—broader deployments at their institution.
Exploring Innovations in Data Repository Solutions - Insights from the U.S. G...Globus
The U.S. Geological Survey (USGS) has made substantial investments in meeting evolving scientific, technical, and policy driven demands on storing, managing, and delivering data. As these demands continue to grow in complexity and scale, the USGS must continue to explore innovative solutions to improve its management, curation, sharing, delivering, and preservation approaches for large-scale research data. Supporting these needs, the USGS has partnered with the University of Chicago-Globus to research and develop advanced repository components and workflows leveraging its current investment in Globus. The primary outcome of this partnership includes the development of a prototype enterprise repository, driven by USGS Data Release requirements, through exploration and implementation of the entire suite of the Globus platform offerings, including Globus Flow, Globus Auth, Globus Transfer, and Globus Search. This presentation will provide insights into this research partnership, introduce the unique requirements and challenges being addressed and provide relevant project progress.
How to Position Your Globus Data Portal for Success Ten Good PracticesGlobus
Science gateways allow science and engineering communities to access shared data, software, computing services, and instruments. Science gateways have gained a lot of traction in the last twenty years, as evidenced by projects such as the Science Gateways Community Institute (SGCI) and the Center of Excellence on Science Gateways (SGX3) in the US, The Australian Research Data Commons (ARDC) and its platforms in Australia, and the projects around Virtual Research Environments in Europe. A few mature frameworks have evolved with their different strengths and foci and have been taken up by a larger community such as the Globus Data Portal, Hubzero, Tapis, and Galaxy. However, even when gateways are built on successful frameworks, they continue to face the challenges of ongoing maintenance costs and how to meet the ever-expanding needs of the community they serve with enhanced features. It is not uncommon that gateways with compelling use cases are nonetheless unable to get past the prototype phase and become a full production service, or if they do, they don't survive more than a couple of years. While there is no guaranteed pathway to success, it seems likely that for any gateway there is a need for a strong community and/or solid funding streams to create and sustain its success. With over twenty years of examples to draw from, this presentation goes into detail for ten factors common to successful and enduring gateways that effectively serve as best practices for any new or developing gateway.
Software Engineering, Software Consulting, Tech Lead.
Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Security,
Spring Transaction, Spring MVC,
Log4j, REST/SOAP WEB-SERVICES.
Developing Distributed High-performance Computing Capabilities of an Open Sci...Globus
COVID-19 had an unprecedented impact on scientific collaboration. The pandemic and its broad response from the scientific community has forged new relationships among public health practitioners, mathematical modelers, and scientific computing specialists, while revealing critical gaps in exploiting advanced computing systems to support urgent decision making. Informed by our team’s work in applying high-performance computing in support of public health decision makers during the COVID-19 pandemic, we present how Globus technologies are enabling the development of an open science platform for robust epidemic analysis, with the goal of collaborative, secure, distributed, on-demand, and fast time-to-solution analyses to support public health.
May Marketo Masterclass, London MUG May 22 2024.pdfAdele Miller
Can't make Adobe Summit in Vegas? No sweat because the EMEA Marketo Engage Champions are coming to London to share their Summit sessions, insights and more!
This is a MUG with a twist you don't want to miss.
Gamify Your Mind; The Secret Sauce to Delivering Success, Continuously Improv...Shahin Sheidaei
Games are powerful teaching tools, fostering hands-on engagement and fun. But they require careful consideration to succeed. Join me to explore factors in running and selecting games, ensuring they serve as effective teaching tools. Learn to maintain focus on learning objectives while playing, and how to measure the ROI of gaming in education. Discover strategies for pitching gaming to leadership. This session offers insights, tips, and examples for coaches, team leads, and enterprise leaders seeking to teach from simple to complex concepts.
Atelier - Innover avec l’IA Générative et les graphes de connaissancesNeo4j
Atelier - Innover avec l’IA Générative et les graphes de connaissances
Allez au-delà du battage médiatique autour de l’IA et découvrez des techniques pratiques pour utiliser l’IA de manière responsable à travers les données de votre organisation. Explorez comment utiliser les graphes de connaissances pour augmenter la précision, la transparence et la capacité d’explication dans les systèmes d’IA générative. Vous partirez avec une expérience pratique combinant les relations entre les données et les LLM pour apporter du contexte spécifique à votre domaine et améliorer votre raisonnement.
Amenez votre ordinateur portable et nous vous guiderons sur la mise en place de votre propre pile d’IA générative, en vous fournissant des exemples pratiques et codés pour démarrer en quelques minutes.
Good afternoon, everyone.
I'm very happy to speak to you today.
My presentation today is about introducing PMDK into PostgreSQL.
First, let me introduce myself.
I have worked on system software, such as distributed block storage or operating system.
This is my first time to dive into PostgreSQL.
I try to refine open-source software by a new storage and a new library, and I choose PostgreSQL.
Today, I will show you our first ideas, and I want to have better ideas.
So any discussions and comments are welcome.
This is an overview of my talk.
Let me begin with introduction.
Persistent Memory, PMEM, is an emerging memory-like storage device.
It’s non-volatile, byte-addressable, and as fast as DRAM.
There are several types of PMEM released or announced.
We use NVDIMM-N in this presentation.
This is a chart telling how PMEM is fast.
This shows read and write latency of each device, so lower is better.
The leftmost is DRAM and the middle three are PMEM.
Solid-state disk and spindle disk are far from DRAM and PMEM.
Zoom into the left four.
Every type of PMEM is close to DRAM.
So it’s fast enough for storage use.
Popular databases are already on the way to PMEM.
Microsoft SQL Server, SAP HANA and MariaDB are getting to PMEM.
And, of cource, PostgreSQL is.
An early work is reported in PGCon two years ago.
So I think it’s time for us to get into PMEM.
What do we need to use PMEM?
We need hardware support and software support such as those written here.
Today, I’d like to focus on performance improvement and talk about DAX and PMDK.
Both of the two are already supported on Linux and Windows.
I’d like to explain DAX, PMDK, and then what we did.
The current I/O stack is as shown in the left.
PostgreSQL uses File I/O APIs such as read and write system calls in POSIX.
Then, go to the middle.
PostgreSQL can run on DAX-enabled filesystem without change because the provided APIs are the same as before.
It can also run faster than before because DAX reads and writes file data directly to PMEM.
It bypasses page cache to remove unnecessary copies.
Note that there is still context switch between user and kernel spaces.
Next, go to the right.
DAX-enabled filesystem also provides memory-mapped file feature.
It maps the file on PMEM directly into user space.
The application can use CPU instructions to access the file data without context switches so it can run much faster.
In this way, PMDK provides primitive memory functions such as CPU cache flush or memory barrier to make sure that the data reaches PMEM.
So we don’t need to make such functions by ourselves.
Finally, what we did is here.
We hacked PostgreSQL to use memory-mapped files and the PMDK library.
Then we compared the performance of the middle and the right to evaluate our hacks.
For details, I’ll show you in later slides.
Finally, what we did is here.
We hacked PostgreSQL to use memory-mapped files and the PMDK library.
Then we compared the performance of the middle and the right to evaluate our hacks.
For details, I’ll show you in later slides.
The memory-mapped file and PMDK can improve the performance of I/O-intensive workload
because it reduces context switches and the overhead of API calls.
To make it sure, we run a micro-benchmark that performs 8-KiB synchronous write.
This emulates the XLOG I/O inside PostgreSQL.
The results shows that ...
We think it seems good to apply PMDK, so we try to introduce it into PostgreSQL.
Next, I’ll talk about our hacks and evaluation.
Before going into detail, I’ll show you our approach.
There are two important points; how to hack and what we hack.
First, how to hack.
We get two ideas and choose the first one.
We think it’s an easier way so it’s reasonable for our first step.
Second, what we hack.
We look into PostgreSQL to find out what kind of files are used in it.
We choose XLOG and Relation segment files because it’s critical for transaction performance or many writes occur during checkpoint.
Roughly speaking, we replace five system calls including read and write by using PMDK.
Blue four functions are provided by libpmem. It’s the base library of PMDK.
Note that the memory-mapping function and unmapping funcion take the length of the mapped file.
Please be careful that the memory-mapped file cannot extend or shrink while being mapped.
Okay, let’s go to XLOG.
An XLOG segment file contains write-ahead log records.
It’s critical for transaction performance.
I know that you are all PostgreSQL experts, so the details are skipped.
I’ll show you how we hack XLOG.
We have PostgreSQL memory-map every segment file and memory-copy to it from the XLOG buffer.
Note that the fixed-length file is highly compatible with memory-mapping because it don’t need to either extend or shrink.
My colleague Yoshimi created a patchset.
It has about 1,000-line changes and it’s available on pgsql-hackers mailing list.
So please search PMDK.
Evaluation setup.
I use one x64 computer with two NUMA nodes.
I dedicate Node 0 to PostgreSQL server and Node 1 to client such as pgbench.
The postgresql.conf is like this.
To evaluate XLOG hacks, I compare transaction throughput by using pgbench on the five conditions below.
The parameters are here. The scale factor is 200 [two-hundred] and the runtime is 30 minutes.
The condition (e) corresponds to our hacks.
On the conditions (c), (d), and (e), the server use PMEM for XLOG segment files, but each of them uses different wal_sync_method.
So we should compare these to evaluate our hacks.
(a) and (b) are for your information.
They use Optane SSD for XLOG segment files, instead of PMEM.
This is the results.
Each bar shows the throughput in kilo TPS, so higher is better.
The condition (e) achieves the best result.
Comparing (c) and (e), the throughput is improved by 3 percent.
This is because of our hacks.
So we achieved the improvement by three percent.
It is roughly the same as Yoshimi reported on the hackers mailing list.
It seems small in the percentage, but I think not-so-small in the absolute value.
For future work, we will do performance profiling to find hotspot.
Also, we will search for a query pattern for which our hack is more effective.
In the abstract of our talk, we said we achieved up to 1.8 [one point eight] as much throughput
by using an INSERT-intensive benchmark of our own.
This is quite artificial, but we think our hack is effective for such workload,
so we search for a more real-life pattern.
That’s all of the XLOG hacks.
Then, let me explain the Relation hacks.
The relation segment file is so-called data file.
I think it’s critical for checkpoint duration.
The details are skipped again because I believe you are experts.
I have PostgreSQL memory-map only every 1-GiB segment file and memory-copy to it from the shared buffer.
The reason of “only 1-GiB” is that it’s difficult for me to handle extending and shrinking in proper PostgreSQL manner.
The relation segment file can extend or shrink up to 1 GiB per file but the memory-mapped file cannot do so while being mapped.
Remapping the file on every extend or shrink could be a solution, but it seemed difficult to implement.
I decided to map only 1-GiB segment file that cannot extend any more, and I ignored shrink.
So my work is still in progress.
I created a patchet.
It has 150-line changes.
My patchset is not published yet because it’s still under test.
Setup is the same as before except postgresql.conf.
I set the checkpoint_timeout to 1day. It’s large enough not to kick a checkpoint automatically.
I also set the checkpoint_completion_target to zero to complete every checkpoint as soon as possible.
I compare the checkpoint duration time as follows.
It's a little tricky but the important point is that, on every condition, the server flushes out the same amount of pages on the shared buffer during checkpoint.
I set the amount to 7.37 (seven point thirty-seven) gibibytes.
Then I initialize the database, restart the server, and have the shared buffer dirty by pgbench.
Next I send the CHECKPOINT query to the server to kick force-immediate checkpoint.
After checkpoint done, I check the administration log to make sure the amount of flushed pages is 7.37-GiB.
Finally I get the total checkpoint time for the result.
I compare the three conditions <here>.
The condition (c) corresponds to our hacks.
On the conditions (b) and (c), the server uses PMEM for PGDATA [pee-gee-data].
The condition (a) is for your information.
It uses Optane SSD for PGDATA, instead of PMEM.
Moreover, I profile the performance of (b) and (c) by Linux perf.
This is the results.
Each bar shows the checkpoint duration, so lower is better.
Comparing (b) and (c), the time is improved by 30 percent.
Then, profile of (b) and (c).
This shows relative CPU cycles, regarding the total cycles of (b) as 100 [one hundred].
The (c) indicates 71 [seventy-one] that almost matches the improvement by 30 [thirty] percent.
First, look at the sky blue portion.
This stands for the overhead of sys_write, the write system call.
Note that the data copy in the write system call is separated as the deep blue portion.
The overhead falls by 20 percentage point. This is the main reason of our improvement.
Second, look at the green portion. It falls by 7 percentage point.
The green stands for the context switches between user and kernel spaces.
Third, the deep blue and the brown portions.
Both of the two stand for memory copy but the blue is that of kernel space while the brown is user space.
The summation of the two falls a little, by 3 percentage point.
The user space function is provided by PMDK.
The function uses instruction set extension such as SSE2 or AVX if possible.
So it could take less time than the general function in kernel space.
Finally, the orange portion, LockBufHdr [lock-buf-header].
This becomes rather longer than before, but I don’t know the reason why.
We shortened checkpoint duration by 30%.
I know that a regular checkpoint is not performed at full speed like as our evaluation.
However, I think the server can give its computing resource to the other purposes.
So it may serve more transactions than before during checkpoint.
We reduce the overhead of system calls and the context switches between user and kernel spaces.
It’s the benefits of using memory-mapped files.
However, there is an open issue.
The time of LockBufHdr became rather longer.
I’ll find the reason for future work.
This is the conclusion of our evaluation.
We improved transaction throughput and shortened checkpoint duration.
I think it’s not so bad in an easier way, but we must bring out more potential from PMEM.
Because our improvement is far from 2.5 times in microbenchmark.
I think another way is to have data structures on DRAM persist on PMEM directly.
So I will continue my work and deep dive into PostgreSQL.
Then I’ll show some tips related to PMEM.
First, CPU cache flush and cache-bypassing store
These methods are important because the data should reach nothing but PMEM,
that is, do not stop at half, volatile middle layer such as CPU caches;
or the data will be lost when the program or system crashes.
For that purpose, the x64 offers two instruction families, CLFLUSH [see-el-flush] and MOVNT [move-en-tee].
The CLFLUSH flushes data out of CPU caches to memory.
The MOVNT stores data to memory, bypassing CPU caches.
PMDK supports both instructions by the functions <here>.
So you don’t need to write assembly codes.
Second, memory-mapped file and Relation extension, extension in file size.
The two are not compatible, because memory-mapped file cannot be extended while it is being mapped.
I find two naive workarounds, but neither is perfect.
The first is remapping a segment file on extend, that is, unmapping, extending, then mapping again.
It’s compatible, but also time-consuming, so it’s not good to remap in a small unit such as 8-KiB.
The second is pre-allocating possible maximum size.
It’s not time-comsuming, but it wastes free space of PMEM.
There will be a middle way, of cource, but what I want to say is that,
we must rethink traditional data structures, especially for DBMS, based on fixed-length page on the disk.
PMEM is memory, so you can use a pointer, for example.
Such properties will drastically change persistent data structures for DBMS.
We should take care of page table for PMEM.
For x64, the regular page size is 4-KiB, but it could be so small that many page faults occur during storing large data.
Its overhead has a significant impact on the performance.
Modern processors and operating systems support hugepage such as 2-MiB or 1-GiB page, for x64.
The hugepage will improve performance of PMEM by reducing page walk and TLB [tee-el-bee] miss.
The good news is that, PMDK on x64 considers hugepage
by aligning the mapping address on hugepage boundary when the file is large enough.
So you can enjoy it with ease.
Also pre-warming the page table for PMEM will also make the performance better
by reducing page fault on main runtime.
We also take care of NUMA effect.
It’s critical for stable performance on multi-processing system.
Consider <this> figure, our evaluation setup.
For the CPU on node 0, accessing to local memory on node 0 is fast, while remote node 1 is slow.
This also applies to PCIe SSD, but PMEM is more sensitive.
The operating system schedules which processes run on which CPU cores.
So the server process can run on either node 0 or 1, if you do not control explicitly,
and it runs slowly because of memory access over NUMA node.
We can control the effect by binding the processes to a certain node.
For that purpose, we can use numactl [numa-control] command as like <this>.
If you assign 0 into X here, your process runs on only the node 0, and never run on the node 1.
And this is the last tip.
I think we should have a new common sense of hotspot.
PMEM is much faster than the traditional disk storage so something other than storage access could be the hotspot of transaction.
When we use disk storage, the time of disk access is usually the hotspot.
But when we use PMEM, the access time decreases and other portions could be the hotspot,
such as described <here>, locking, memory copy, or pre-processing.
Actually, we fell into the SQL parse trap when we run pgbench and we avoided it by prepared statement.
Finally, conclusion.
We applied PMDK into PostgreSQL in an easier way that uses memory-mapped files memory copy.
And we achieved not-so-bad results.
I also showed some tips related to PMEM.
I believe that PMEM will change software design drastically, and we should change software and our mind to bring out PMEM’s potential much more.
I will continue my work, and I hope that you will get interested in.
So let’s try PMEM and PMDK.