In this deck from the Perth HPC Conference, Werner Scholz from XENON Systems presents: Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPC.
"A decade ago, 100 watts per CPU was devastating to thermal design. Today, Intel’s highest performing CPUs (e.g. Intel Cascade Lake-AP 9282 processor) have a thermal design envelope of 400 watts. There really is no end in sight, and accommodating more power is critical to advancing performance. The ability to dissipate the resulting heat is the hard ceiling that systems face in terms of performance – giving greater importance to liquid cooling breakthroughs. With liquid cooling, less energy is expended to cool systems – a significant savings in HPC deployments with arrays of servers drawing energy and generating heat. Electrical current drives the CPU and enables it to function. This electrical power is converted into thermal energy (heat). To maintain a stable temperature, the CPU needs to be cooled by efficiently removing this heat and releasing it. Liquid cooling is the best way to cool a system because liquid transfers heat much more efficiently than air. From an environmental perspective, liquid cooling reduces both those characteristics to create a smarter and more ecological approach on a grand scale. The cascade of value continues, as ambient heat removed from systems can then be used to heat buildings and augment or replace traditional heating systems. It’s an intelligent approach to thermal management, distributing the economic value of reduced energy use and transforming heat into an enterprise asset."
Watch the video: https://wp.me/p3RLHQ-kZa
Learn more: https://www.xenon.com.au/
and
http://hpcadvisorycouncil.com/events/2019/australia-conference/agenda.php
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Reliability, Availability, and Serviceability (RAS) on ARM64 status - SFO17-203Linaro
Session ID: SFO17-203
Session Name: Reliability, Availability, and Serviceability (RAS) on ARM64 status - SFO17-203
Speaker: Fu Wei
Track: LEG
★ Session Summary ★
This presentation gives an updated RAS architecture on ARM64 base on RAS extension (in ARMv8.2), SDEI (Software Delegated Exception Interface), APEI, UEFI PI-SMM. Will talk about all the components of the new RAS architecture on ARM64, gives audience the current status and the next step of development.
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/sfo17/sfo17-203/
Presentation:
Video: https://www.youtube.com/watch?v=NReFBzbeWi0
---------------------------------------------------
★ Event Details ★
Linaro Connect San Francisco 2017 (SFO17)
25-29 September 2017
Hyatt Regency San Francisco Airport
---------------------------------------------------
Keyword:
'http://www.linaro.org'
'http://connect.linaro.org'
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://twitter.com/linaroorg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
In this talk Jiří Pírko discusses the design and evolution of the VLAN implementation in Linux, the challenges and pitfalls as well as hardware acceleration and alternative implementations.
Jiří Pírko is a major contributor to kernel networking and the creator of libteam for link aggregation.
Reliability, Availability, and Serviceability (RAS) on ARM64 status - SFO17-203Linaro
Session ID: SFO17-203
Session Name: Reliability, Availability, and Serviceability (RAS) on ARM64 status - SFO17-203
Speaker: Fu Wei
Track: LEG
★ Session Summary ★
This presentation gives an updated RAS architecture on ARM64 base on RAS extension (in ARMv8.2), SDEI (Software Delegated Exception Interface), APEI, UEFI PI-SMM. Will talk about all the components of the new RAS architecture on ARM64, gives audience the current status and the next step of development.
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/sfo17/sfo17-203/
Presentation:
Video: https://www.youtube.com/watch?v=NReFBzbeWi0
---------------------------------------------------
★ Event Details ★
Linaro Connect San Francisco 2017 (SFO17)
25-29 September 2017
Hyatt Regency San Francisco Airport
---------------------------------------------------
Keyword:
'http://www.linaro.org'
'http://connect.linaro.org'
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://twitter.com/linaroorg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
In this talk Jiří Pírko discusses the design and evolution of the VLAN implementation in Linux, the challenges and pitfalls as well as hardware acceleration and alternative implementations.
Jiří Pírko is a major contributor to kernel networking and the creator of libteam for link aggregation.
BUD17-209: Reliability, Availability, and Serviceability (RAS) on ARM64 Linaro
BUD17-209: Reliability, Availability, and Serviceability (RAS) on ARM64
Speaker: Wei Fu
Track: LEG
★ Session Summary ★
The RAS architecture base on RAS extension, SDEI, APEI.
1. the definition, Function and importance of RAS
2. Hardware requirement: CPU, GIC and RAS Extension
3. Firmware support : SDEI dispatcher and APEI support, and how to take advantage of RAS Extension.
4. OS support: APEI driver and SDEI client, error report mechanism(trace event)
5. Userspace recorder: rasdaemon
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/bud17/bud17-209/
Presentation:
Video:
---------------------------------------------------
★ Event Details ★
Linaro Connect Budapest 2017 (BUD17)
6-10 March 2017
Corinthia Hotel, Budapest,
Erzsébet krt. 43-49,
1073 Hungary
---------------------------------------------------
ARM64, LEG, RAS
http://www.linaro.org
http://connect.linaro.org
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://twitter.com/linaroorg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
"
During the CXL Forum at OCP Global Summit, memory system architect Jungmin Choi of SK hynix talks about the need for memory bandwidth and capacity, and the SK hynix Niagara solution.
Kernel Recipes 2019 - No NMI? No Problem! – Implementing Arm64 Pseudo-NMIAnne Nicolas
As the name would suggest, a Non-Maskable Interrupt (NMI) is an interrupt-like feature that is unaffected by the disabling of classic interrupts. In Linux, NMIs are involved in some features such as performance event monitoring, hard-lockup detector, on demand state dumping, etc… Their potential to fire when least expected can fill the most seasoned kernel hackers with dread.
AArch64 (aka arm64 in the Linux tree) does not provide architected NMIs, a consequence being that features benefiting from NMIs see their use limited on AArch64. However, the Arm Generic Interrupt Controller (GIC) supports interrupt prioritization and masking, which, among other things, provides a way to control whether or not a set of interrupts can be signaled to a CPU.
This talk will cover how, using the GIC interrupt priorities, we provide a way to configure some interrupts to behave in an NMI-like manner on AArch64. We’ll discuss the implementation, some of the complications that ensued and also some of the benefits obtained from it.
Julien Thierry
In this video, Jeff Larkin from NVIDIA presents: Unified Memory on Summit.
"The biggest problems in science require supercomputers of unprecedented capability. That’s why the US Department of Energy’s Oak Ridge National Laboratory (ORNL) launched Summit, a system 8 times more powerful than ORNL’s previous top-ranked system Titan. Summit is providing scientists with incredible computing power to solve challenges in energy, artificial intelligence, human health, and other research areas, that were simply out of reach until now. These discoveries will help shape our understanding of the universe, bolster US economic competitiveness, and contribute to a better future.
The Summit Application Readiness Workshop had the primary objective of providing the detailed technical information and hands-on help required for select application teams to meet the scalability and performance metrics required for Early Science proposals. It is the expectation that at the time of the workshop, Summit Phase I will be available in a configuration that will allow the teams to demonstrate these metrics using representative science runs. This will be an important milestone for the projects, and a requisite for a successful Early Science proposal. Technical representatives from the IBM/NVIDIA Center of Excellence will be delivering a few plenary presentations, but most of the time will be set aside for the extended application teams to carry out hands-on technical work on Summit."
Learn more: https://www.olcf.ornl.gov/summit-application-readiness-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Switzerland HPC Conference, Francis Lam from Huawei presents: A Fresh Look at HPC from Huawei Enterprise.
"High performance computing is rapidly finding new uses in many applications and businesses, enabling the creation of disruptive products and services. Huawei, a global leader in information and communication technologies, brings a broad spectrum of innovative solutions to HPC. This talk examines Huawei's world class HPC solutions and explores creative new ways to solve HPC problems."
Watch the video: http://wp.me/p3RLHQ-gCV
Learn more: http://e.huawei.com/en/solutions/business-needs/data-center/high-performance-computing
and
http://www.hpcadvisorycouncil.com/events/2017/swiss-workshop/agenda.php
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
BUD17-209: Reliability, Availability, and Serviceability (RAS) on ARM64 Linaro
BUD17-209: Reliability, Availability, and Serviceability (RAS) on ARM64
Speaker: Wei Fu
Track: LEG
★ Session Summary ★
The RAS architecture base on RAS extension, SDEI, APEI.
1. the definition, Function and importance of RAS
2. Hardware requirement: CPU, GIC and RAS Extension
3. Firmware support : SDEI dispatcher and APEI support, and how to take advantage of RAS Extension.
4. OS support: APEI driver and SDEI client, error report mechanism(trace event)
5. Userspace recorder: rasdaemon
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/bud17/bud17-209/
Presentation:
Video:
---------------------------------------------------
★ Event Details ★
Linaro Connect Budapest 2017 (BUD17)
6-10 March 2017
Corinthia Hotel, Budapest,
Erzsébet krt. 43-49,
1073 Hungary
---------------------------------------------------
ARM64, LEG, RAS
http://www.linaro.org
http://connect.linaro.org
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://twitter.com/linaroorg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
"
During the CXL Forum at OCP Global Summit, memory system architect Jungmin Choi of SK hynix talks about the need for memory bandwidth and capacity, and the SK hynix Niagara solution.
Kernel Recipes 2019 - No NMI? No Problem! – Implementing Arm64 Pseudo-NMIAnne Nicolas
As the name would suggest, a Non-Maskable Interrupt (NMI) is an interrupt-like feature that is unaffected by the disabling of classic interrupts. In Linux, NMIs are involved in some features such as performance event monitoring, hard-lockup detector, on demand state dumping, etc… Their potential to fire when least expected can fill the most seasoned kernel hackers with dread.
AArch64 (aka arm64 in the Linux tree) does not provide architected NMIs, a consequence being that features benefiting from NMIs see their use limited on AArch64. However, the Arm Generic Interrupt Controller (GIC) supports interrupt prioritization and masking, which, among other things, provides a way to control whether or not a set of interrupts can be signaled to a CPU.
This talk will cover how, using the GIC interrupt priorities, we provide a way to configure some interrupts to behave in an NMI-like manner on AArch64. We’ll discuss the implementation, some of the complications that ensued and also some of the benefits obtained from it.
Julien Thierry
In this video, Jeff Larkin from NVIDIA presents: Unified Memory on Summit.
"The biggest problems in science require supercomputers of unprecedented capability. That’s why the US Department of Energy’s Oak Ridge National Laboratory (ORNL) launched Summit, a system 8 times more powerful than ORNL’s previous top-ranked system Titan. Summit is providing scientists with incredible computing power to solve challenges in energy, artificial intelligence, human health, and other research areas, that were simply out of reach until now. These discoveries will help shape our understanding of the universe, bolster US economic competitiveness, and contribute to a better future.
The Summit Application Readiness Workshop had the primary objective of providing the detailed technical information and hands-on help required for select application teams to meet the scalability and performance metrics required for Early Science proposals. It is the expectation that at the time of the workshop, Summit Phase I will be available in a configuration that will allow the teams to demonstrate these metrics using representative science runs. This will be an important milestone for the projects, and a requisite for a successful Early Science proposal. Technical representatives from the IBM/NVIDIA Center of Excellence will be delivering a few plenary presentations, but most of the time will be set aside for the extended application teams to carry out hands-on technical work on Summit."
Learn more: https://www.olcf.ornl.gov/summit-application-readiness-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Switzerland HPC Conference, Francis Lam from Huawei presents: A Fresh Look at HPC from Huawei Enterprise.
"High performance computing is rapidly finding new uses in many applications and businesses, enabling the creation of disruptive products and services. Huawei, a global leader in information and communication technologies, brings a broad spectrum of innovative solutions to HPC. This talk examines Huawei's world class HPC solutions and explores creative new ways to solve HPC problems."
Watch the video: http://wp.me/p3RLHQ-gCV
Learn more: http://e.huawei.com/en/solutions/business-needs/data-center/high-performance-computing
and
http://www.hpcadvisorycouncil.com/events/2017/swiss-workshop/agenda.php
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
The number of internet-connected devices is growing exponentially, enabling an increasing number of edge applications in environments such as smart cities, retail, and industry 4.0. These intelligent solutions often require processing large amounts of data, running models to enable image recognition, predictive analytics, autonomous systems, and more. Increasing system workloads and data processing capacity at the edge is essential to minimize latency, improve responsiveness, and reduce network traffic back to data centers. Purpose-built systems such as Supermicro’s short-depth, multi-node SuperEdge, powered by 3rd Gen Intel® Xeon® Scalable processors, increase compute and I/O density at the edge and enable businesses to further accelerate innovation.
Join this webinar to discover new insights in edge-to-cloud infrastructures and learn how Supermicro SuperEdge multi-node solutions leverage data center scale, performance, and efficiency for 5G, IoT, and Edge applications.
Data Centers: Compute Faster and More Sustainably with Microconvective Cooling®JetCool Technologies
As the demand for AI, ML, and big data grows, data centers are facing significant energy challenges. It is estimated that 3% of the planet's total energy consumption is used by data centers with 30% of that attributed to cooling equipment.
Inferior in-rack cooling solutions use considerable energy and resources often requiring power- and water-intensive infrastructure like chillers and evaporative coolers. Depending on the local climate, these systems can be used year-round with U.S. data centers alone consuming an estimated 660 billion liters of water in 2020.
Processors rely on optimized cooling solutions to maximize power density and increase compute performance. With suboptimal cooling solutions, heat can cause processor throttling and limited device lifetime. As power density and processor performance increase so do the expectations for cooling technology.
Recognizing the need for a simple, scalable, and sustainable liquid cooling option for data centers, JetCool developed SmartPlate. SmartPlate is a simple water cooling system that outfits processors with a fully-sealed, drop-in replacement cooling solution that lowers chip temperatures by 37%. SmartPlate technology helps accelerate performance by extracting 30% faster processing for TDPs over 1,000 Watts. Sustainability is one of JetCool's primary principles. Our consumers use 50% less electricity and 90% less water than the competing liquid cooling solutions. Reach out to our team to learn more about partnering with JetCool today.
Una reproducción de articulos sobre eficiencia energética para datacenters e instalaciones de telecomunicaciones con tecnología limpia de solid oxide fuel cell (SOFC)
Join us for an exciting and informative preview of the broadest range of next-generation systems optimized for tomorrow’s data center workloads, Powered by 4th Gen Intel® Xeon® Scalable Processors (formerly codenamed Sapphire Rapids).
Experts from Supermicro and Intel will discuss how the upcoming Supermicro X13 systems will enable new performance levels utilizing state-of-the-art technology, including DDR5, PCIe 5.0, Compute Express Link™ 1.1, and Intel® Advanced Matrix Extensions (Intel AMX).
Introducing the new On-Demand Private Cloud. Supermicro and InMotion Hosting joined forces to design a Data Center POD solution that allows data centers to take control of their cloud costs by lowering the total cost per VM. This all-in-one solution consisting of small hyper-converged building blocks, are built for your business to achieve significant on-demand flexibility and scalability. Utilizing a consumption based model eliminates multiple vendors by consolidating hardware, software, networking, management, and administration enabling your data center to grow and shrink based on your business’ needs. Bring your business to the next level with an On-Demand Private Cloud, let it help you achieve greater profitability by eliminating overpriced inflated, high-cost licensing fees. This streamlined model removes overpriced licensing fees, and enables you to increase profits while reducing overhead costs. This On-Demand Private Cloud solution is built with Supermicro’s green, power efficient and high density, compute servers; OpenStack’s open-source software, and Ceph’s object, file, and block storage.
Join this webinar to hear industry experts from InMotion Hosting and Supermicro discuss:
- Building your next-generation data center infrastructure
- Enables private data centers with validated solutions from Supermicro and InMotion
- Hosting to improve operational efficiency
- Reach new segments with Kubernetes, Machine Learning, and Artificial Intelligence
- Achieve on-demand cloud computing, high availability, data redundancy, and flexibility
- Take control of your cloud costs and lower your total cost of ownership
- Hyper-dense hardware enables economies of scale for power, cooling, and physical space
Applying Cloud Techniques to Address Complexity in HPC System Integrationsinside-BigData.com
In this video from the HPC User Forum at Argonne, Arno Kolster from Providentia Worldwide presents: Applying Cloud Techniques to Address Complexity in HPC System Integrations.
"The Oak Ridge Leadership Computing Facility (OLCF) and technology consulting company Providentia Worldwide recently collaborated to develop an intelligence system that combines real-time updates from the IBM AC922 Summit supercomputer with local weather and operational data from its adjacent cooling plant, with the goal of optimizing Summit’s energy efficiency. The OLCF proposed the idea and provided facility data, and Providentia developed a scalable platform to integrate and analyze the data."
Watch the video: https://wp.me/p3RLHQ-kOg
Learn more: http://www.providentiaworldwide.com/
and
http://hpcuserforum.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
T12.Fujitsu World Tour India 2016-Your Datacenter‘s backboneFujitsu India
Growing complexity in business processes have a direct impact on your company's storage , data protection and archival architecture. Understand from our senior Storage expert as how can you manage these challenges while achieving your cost goals
Optimize Content Delivery with Multi-Access Edge ComputingRebekah Rodriguez
Edge applications today require strong networks and connections to core data center products. Supermicro’s Twin systems deliver performance from the edge to the core or cloud data center and leverage our latest content delivery network (CDN) technologies.
Join this webinar to learn about the different Supermicro CDN technologies that support Edge Computing based on 3GPP standards and have proven successful for our customers. Product experts will also discuss the importance of an elastic CDN to optimize the availability of resources for low-latency and high-bandwidth applications and share an introduction to our new GrandTwin® server.
Similar to Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPC (20)
In this deck from the Stanford HPC Conference, Shahin Khan from OrionX describes major market Shifts in IT.
"We will discuss the digital infrastructure of the future enterprise and the state of these trends."
"We work with clients on the impact of Digital Transformation (DX) on them, their customers, and their messages. Generally, they want to track, in one place, trends like IoT, 5G, AI, Blockchain, and Quantum Computing. And they want to know what these trends mean, how they affect each other, and when they demand action, and how to formulate and execute an effective plan. If that describes you, we can help."
Watch the video: https://wp.me/p3RLHQ-lPP
Learn more: http://orionx.net
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Preparing to program Aurora at Exascale - Early experiences and future direct...inside-BigData.com
In this deck from IWOCL / SYCLcon 2020, Hal Finkel from Argonne National Laboratory presents: Preparing to program Aurora at Exascale - Early experiences and future directions.
"Argonne National Laboratory’s Leadership Computing Facility will be home to Aurora, our first exascale supercomputer. Aurora promises to take scientific computing to a whole new level, and scientists and engineers from many different fields will take advantage of Aurora’s unprecedented computational capabilities to push the boundaries of human knowledge. In addition, Aurora’s support for advanced machine-learning and big-data computations will enable scientific workflows incorporating these techniques along with traditional HPC algorithms. Programming the state-of-the-art hardware in Aurora will be accomplished using state-of-the-art programming models. Some of these models, such as OpenMP, are long-established in the HPC ecosystem. Other models, such as Intel’s oneAPI, based on SYCL, are relatively-new models constructed with the benefit of significant experience. Many applications will not use these models directly, but rather, will use C++ abstraction libraries such as Kokkos or RAJA. Python will also be a common entry point to high-performance capabilities. As we look toward the future, features in the C++ standard itself will become increasingly relevant for accessing the extreme parallelism of exascale platforms.
This presentation will summarize the experiences of our team as we prepare for Aurora, exploring how to port applications to Aurora’s architecture and programming models, and distilling the challenges and best practices we’ve developed to date. oneAPI/SYCL and OpenMP are both critical models in these efforts, and while the ecosystem for Aurora has yet to mature, we’ve already had a great deal of success. Importantly, we are not passive recipients of programming models developed by others. Our team works not only with vendor-provided compilers and tools, but also develops improved open-source LLVM-based technologies that feed both open-source and vendor-provided capabilities. In addition, we actively participate in the standardization of OpenMP, SYCL, and C++. To conclude, I’ll share our thoughts on how these models can best develop in the future to support exascale-class systems."
Watch the video: https://wp.me/p3RLHQ-lPT
Learn more: https://www.iwocl.org/iwocl-2020/conference-program/
and
https://www.anl.gov/topic/aurora
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Greg Wahl from Advantech presents: Transforming Private 5G Networks.
Advantech Networks & Communications Group is driving innovation in next-generation network solutions with their High Performance Servers. We provide business critical hardware to the world's leading telecom and networking equipment manufacturers with both standard and customized products. Our High Performance Servers are highly configurable platforms designed to balance the best in x86 server-class processing performance with maximum I/O and offload density. The systems are cost effective, highly available and optimized to meet next generation networking and media processing needs.
“Advantech’s Networks and Communication Group has been both an innovator and trusted enabling partner in the telecommunications and network security markets for over a decade, designing and manufacturing products for OEMs that accelerate their network platform evolution and time to market.” Said Advantech Vice President of Networks & Communications Group, Ween Niu. “In the new IP Infrastructure era, we will be expanding our expertise in Software Defined Networking (SDN) and Network Function Virtualization (NFV), two of the essential conduits to 5G infrastructure agility making networks easier to install, secure, automate and manage in a cloud-based infrastructure.”
In addition to innovation in air interface technologies and architecture extensions, 5G will also need a new generation of network computing platforms to run the emerging software defined infrastructure, one that provides greater topology flexibility, essential to deliver on the promises of high availability, high coverage, low latency and high bandwidth connections. This will open up new parallel industry opportunities through dedicated 5G network slices reserved for specific industries dedicated to video traffic, augmented reality, IoT, connected cars etc. 5G unlocks many new doors and one of the keys to its enablement lies in the elasticity and flexibility of the underlying infrastructure.
Advantech’s corporate vision is to enable an intelligent planet. The company is a global leader in the fields of IoT intelligent systems and embedded platforms. To embrace the trends of IoT, big data, and artificial intelligence, Advantech promotes IoT hardware and software solutions with the Edge Intelligence WISE-PaaS core to assist business partners and clients in connecting their industrial chains. Advantech is also working with business partners to co-create business ecosystems that accelerate the goal of industrial intelligence."
Watch the video: https://wp.me/p3RLHQ-lPQ
* Company website: https://www.advantech.com/
* Solution page: https://www2.advantech.com/nc/newsletter/NCG/SKY/benefits.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
In this deck from the Stanford HPC Conference, Katie Lewis from Lawrence Livermore National Laboratory presents: The Incorporation of Machine Learning into Scientific Simulations at Lawrence Livermore National Laboratory.
"Scientific simulations have driven computing at Lawrence Livermore National Laboratory (LLNL) for decades. During that time, we have seen significant changes in hardware, tools, and algorithms. Today, data science, including machine learning, is one of the fastest growing areas of computing, and LLNL is investing in hardware, applications, and algorithms in this space. While the use of simulations to focus and understand experiments is well accepted in our community, machine learning brings new challenges that need to be addressed. I will explore applications for machine learning in scientific simulations that are showing promising results and further investigation that is needed to better understand its usefulness."
Watch the video: https://youtu.be/NVwmvCWpZ6Y
Learn more: https://computing.llnl.gov/research-area/machine-learning
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...inside-BigData.com
In this deck from the Stanford HPC Conference, DK Panda from Ohio State University presents: How to Achieve High-Performance, Scalable and Distributed DNN Training on Modern HPC Systems?
"This talk will start with an overview of challenges being faced by the AI community to achieve high-performance, scalable and distributed DNN training on Modern HPC systems with both scale-up and scale-out strategies. After that, the talk will focus on a range of solutions being carried out in my group to address these challenges. The solutions will include: 1) MPI-driven Deep Learning, 2) Co-designing Deep Learning Stacks with High-Performance MPI, 3) Out-of- core DNN training, and 4) Hybrid (Data and Model) parallelism. Case studies to accelerate DNN training with popular frameworks like TensorFlow, PyTorch, MXNet and Caffe on modern HPC systems will be presented."
Watch the video: https://youtu.be/LeUNoKZVuwQ
Learn more: http://web.cse.ohio-state.edu/~panda.2/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
In this deck from the Stanford HPC Conference, Nick Nystrom and Paola Buitrago provide an update from the Pittsburgh Supercomputing Center.
Nick Nystrom is Chief Scientist at the Pittsburgh Supercomputing Center (PSC). Nick is architect and PI for Bridges, PSC's flagship system that successfully pioneered the convergence of HPC, AI, and Big Data. He is also PI for the NIH Human Biomolecular Atlas Program’s HIVE Infrastructure Component and co-PI for projects that bring emerging AI technologies to research (Open Compass), apply machine learning to biomedical data for breast and lung cancer (Big Data for Better Health), and identify causal relationships in biomedical big data (the Center for Causal Discovery, an NIH Big Data to Knowledge Center of Excellence). His current research interests include hardware and software architecture, applications of machine learning to multimodal data (particularly for the life sciences) and to enhance simulation, and graph analytics.
Watch the video: https://youtu.be/LWEU1L1o7yY
Learn more: https://www.psc.edu/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Stanford HPC Conference, Ryan Quick from Providentia Worldwide describes how DNNs can be used to improve EDA simulation runs.
"Systems Intelligence relies on a variety of methods for providing insight into the core mechanisms for driving automated behavioral changes in self-healing command and control platforms. This talk reports on initial efforts with leveraging Semiconductor Electronic Design Automation (EDA) telemetry data from cross-domain sources including power, network, storage, nodes, and applications in neural networks as a driving method for insight into SI automation systems."
Watch the video: https://youtu.be/2WbR8tq-XbM
Learn more: http://www.providentiaworldwide.com/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoringinside-BigData.com
In this deck from the Stanford HPC Conference, Nicole Xu from Stanford University describes how she transformed a common jellyfish into a bionic creature that is part animal and part machine.
"Animal locomotion and bioinspiration have the potential to expand the performance capabilities of robots, but current implementations are limited. Mechanical soft robots leverage engineered materials and are highly controllable, but these biomimetic robots consume more power than corresponding animal counterparts. Biological soft robots from a bottom-up approach offer advantages such as speed and controllability but are limited to survival in cell media. Instead, biohybrid robots that comprise live animals and self- contained microelectronic systems leverage the animals’ own metabolism to reduce power constraints and body as an natural scaffold with damage tolerance. We demonstrate that by integrating onboard microelectronics into live jellyfish, we can enhance propulsion up to threefold, using only 10 mW of external power input to the microelectronics and at only a twofold increase in cost of transport to the animal. This robotic system uses 10 to 1000 times less external power per mass than existing swimming robots in literature and can be used in future applications for ocean monitoring to track environmental changes."
Watch the video: https://youtu.be/HrmJFyvInj8
Learn more: https://sanfrancisco.cbslocal.com/2020/02/05/stanford-research-project-common-jellyfish-bionic-sea-creatures/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Stanford HPC Conference, Peter Dueben from the European Centre for Medium-Range Weather Forecasts (ECMWF) presents: Machine Learning for Weather Forecasts.
"I will present recent studies that use deep learning to learn the equations of motion of the atmosphere, to emulate model components of weather forecast models and to enhance usability of weather forecasts. I will than talk about the main challenges for the application of deep learning in cutting-edge weather forecasts and suggest approaches to improve usability in the future."
Peter is contributing to the development and optimization of weather and climate models for modern supercomputers. He is focusing on a better understanding of model error and model uncertainty, on the use of reduced numerical precision that is optimised for a given level of model error, on global cloud- resolving simulations with ECMWF's forecast model, and the use of machine learning, and in particular deep learning, to improve the workflow and predictions. Peter has graduated in Physics and wrote his PhD thesis at the Max Planck Institute for Meteorology in Germany. He worked as Postdoc with Tim Palmer at the University of Oxford and has taken up a position as University Research Fellow of the Royal Society at the European Centre for Medium-Range Weather Forecasts (ECMWF) in 2017.
Watch the video: https://youtu.be/ks3fkRj8Iqc
Learn more: https://www.ecmwf.int/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Gilad Shainer from the HPC AI Advisory Council describes how this organization fosters innovation in the high performance computing community.
"The HPC-AI Advisory Council’s mission is to bridge the gap between high-performance computing (HPC) and Artificial Intelligence (AI) use and its potential, bring the beneficial capabilities of HPC and AI to new users for better research, education, innovation and product manufacturing, bring users the expertise needed to operate HPC and AI systems, provide application designers with the tools needed to enable parallel computing, and to strengthen the qualification and integration of HPC and AI system products."
Watch the video: https://wp.me/p3RLHQ-lNz
Learn more: http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Today RIKEN in Japan announced that the Fugaku supercomputer will be made available for research projects aimed to combat COVID-19.
"Fugaku is currently being installed and is scheduled to be available to the public in 2021. However, faced with the devastating disaster unfolding before our eyes, RIKEN and MEXT decided to make a portion of the computational resources of Fugaku available for COVID-19-related projects ahead of schedule while continuing the installation process.
Fugaku is being developed not only for the progress in science, but also to help build the society dubbed as the “Society 5.0” by the Japanese government, where all people will live safe and comfortable lives. The current initiative to fight against the novel coronavirus is driven by the philosophy behind the development of Fugaku."
Initial Projects
Exploring new drug candidates for COVID-19 by "Fugaku"
Yasushi Okuno, RIKEN / Kyoto University
Prediction of conformational dynamics of proteins on the surface of SARS-Cov-2 using Fugaku
Yuji Sugita, RIKEN
Simulation analysis of pandemic phenomena
Nobuyasu Ito, RIKEN
Fragment molecular orbital calculations for COVID-19 proteins
Yuji Mochizuki, Rikkyo University
In this deck from the Performance Optimisation and Productivity group, Lubomir Riha from IT4Innovations presents: Energy Efficient Computing using Dynamic Tuning.
"We now live in a world of power-constrained architectures and systems and power consumption represents a significant cost factor in the overall HPC system economy. For these reasons, in recent years researchers, supercomputing centers and major vendors have developed new tools and methodologies to measure and optimize the energy consumption of large-scale high performance system installations. Due to the link between energy consumption, power consumption and execution time of an application executed by the final user, it is important for these tools and the methodology used to consider all these aspects, empowering the final user and the system administrator with the capability of finding the best configuration given different high level objectives.
This webinar focused on tools designed to improve the energy-efficiency of HPC applications using a methodology of dynamic tuning of HPC applications, developed under the H2020 READEX project. The READEX methodology has been designed for exploiting the dynamic behaviour of software. At design time, different runtime situations (RTS) are detected and optimized system configurations are determined. RTSs with the same configuration are grouped into scenarios, forming the tuning model. At runtime, the tuning model is used to switch system configurations dynamically.
The MERIC tool, that implements the READEX methodology, is presented. It supports manual or binary instrumentation of the analysed applications to simplify the analysis. This instrumentation is used to identify and annotate the significant regions in the HPC application. Automatic binary instrumentation annotates regions with significant runtime. Manual instrumentation, which can be combined with automatic, allows code developer to annotate regions of particular interest."
Watch the video: https://wp.me/p3RLHQ-lJP
Learn more: https://pop-coe.eu/blog/14th-pop-webinar-energy-efficient-computing-using-dynamic-tuning
and
https://code.it4i.cz/vys0053/meric
Sign up for our insideHPC Newsletter: http://insidehpc.com/newslett
In this deck from GTC Digital, William Beaudin from DDN presents: HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD.
Enabling high performance computing through the use of GPUs requires an incredible amount of IO to sustain application performance. We'll cover architectures that enable extremely scalable applications through the use of NVIDIA’s SuperPOD and DDN’s A3I systems.
The NVIDIA DGX SuperPOD is a first-of-its-kind artificial intelligence (AI) supercomputing infrastructure. DDN A³I with the EXA5 parallel file system is a turnkey, AI data storage infrastructure for rapid deployment, featuring faster performance, effortless scale, and simplified operations through deeper integration. The combined solution delivers groundbreaking performance, deploys in weeks as a fully integrated system, and is designed to solve the world's most challenging AI problems.
Watch the video: https://wp.me/p3RLHQ-lIV
Learn more: https://www.ddn.com/download/nvidia-superpod-ddn-a3i-ai400-appliance-with-the-exa5-filesystem/
and
https://www.nvidia.com/en-us/gtc/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Paul Isaacs from Linaro presents: State of ARM-based HPC. This talk provides an overview of applications and infrastructure services successfully ported to Aarch64 and benefiting from scale.
"With its debut on the TOP500, the 125,000-core Astra supercomputer at New Mexico’s Sandia Labs uses Cavium ThunderX2 chips to mark Arm’s entry into the petascale world. In Japan, the Fujitsu A64FX Arm-based CPU in the pending Fugaku supercomputer has been optimized to achieve high-level, real-world application performance, anticipating up to one hundred times the application execution performance of the K computer. K was the first computer to top 10 petaflops in 2011."
Watch the video: https://wp.me/p3RLHQ-lIT
Learn more: https://www.linaro.org/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Versal Premium ACAP for Network and Cloud Accelerationinside-BigData.com
Today Xilinx announced Versal Premium, the third series in the Versal ACAP portfolio. The Versal Premium series features highly integrated, networked and power-optimized cores and the industry’s highest bandwidth and compute density on an adaptable platform. Versal Premium is designed for the highest bandwidth networks operating in thermally and spatially constrained environments, as well as for cloud providers who need scalable, adaptable application acceleration.
Versal is the industry’s first adaptive compute acceleration platform (ACAP), a revolutionary new category of heterogeneous compute devices with capabilities that far exceed those of conventional silicon architectures. Developed on TSMC’s 7-nanometer process technology, Versal Premium combines software programmability with dynamically configurable hardware acceleration and pre-engineered connectivity and security features to enable a faster time-to- market. The Versal Premium series delivers up to 3X higher throughput compared to current generation FPGAs, with built-in Ethernet, Interlaken, and cryptographic engines that enable fast and secure networks. The series doubles the compute density of currently deployed mainstream FPGAs and provides the adaptability to keep pace with increasingly diverse and evolving cloud and networking workloads.
Learn more: https://insidehpc.com/2020/03/xilinx-announces-versal-premium-acap-for-network-and-cloud-acceleration/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Zettar: Moving Massive Amounts of Data across Any Distance Efficientlyinside-BigData.com
In this video from the Rice Oil & Gas Conference, Chin Fang from Zettar presents: Moving Massive Amounts of Data across Any Distance Efficiently.
The objective of this talk is to present two on-going projects aiming at improving and ensuring highly efficient bulk transferring or streaming of massive amounts of data over digital connections across any distance. It examines the current state of the art, a few very common misconceptions, the differences among the three major type of data movement solutions, a current initiative attempting to improve the data movement efficiency from the ground up, and another multi-stage project that shows how to conduct long distance large scale data movement at speed and scale internationally. Both projects have real world motivations, e.g. the ambitious data transfer requirements of Linac Coherent Light Source II (LCLS-II) [1], a premier preparation project of the U.S. DOE Exascale Computing Initiative (ECI) [2]. Their immediate goals are described and explained, together with the solution used for each. Findings and early results are reported. Possible future works are outlined.
Watch the video: https://wp.me/p3RLHQ-lBX
Learn more: https://www.zettar.com/
and
https://rice2020oghpc.rice.edu/program-2/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Rice Oil & Gas Conference, Bradley McCredie from AMD presents: Scaling TCO in a Post Moore's Law Era.
"While foundries bravely drive forward to overcome the technical and economic challenges posed by scaling to 5nm and beyond, Moore’s law alone can provide only a fraction of the performance / watt and performance / dollar gains needed to satisfy the demands of today’s high performance computing and artificial intelligence applications. To close the gap, multiple strategies are required. First, new levels of innovation and design efficiency will supplement technology gains to continue to deliver meaningful improvements in SoC performance. Second, heterogenous compute architectures will create x-factor increases of performance efficiency for the most critical applications. Finally, open software frameworks, APIs, and toolsets will enable broad ecosystems of application level innovation."
Watch the video:
Learn more: http://amd.com
and
https://rice2020oghpc.rice.edu/program-2/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
In this deck from the ECSS Symposium, Abe Stern from NVIDIA presents: CUDA-Python and RAPIDS for blazing fast scientific computing.
"We will introduce Numba and RAPIDS for GPU programming in Python. Numba allows us to write just-in-time compiled CUDA code in Python, giving us easy access to the power of GPUs from a powerful high-level language. RAPIDS is a suite of tools with a Python interface for machine learning and dataframe operations. Together, Numba and RAPIDS represent a potent set of tools for rapid prototyping, development, and analysis for scientific computing. We will cover the basics of each library and go over simple examples to get users started. Finally, we will briefly highlight several other relevant libraries for GPU programming."
Watch the video: https://wp.me/p3RLHQ-lvu
Learn more: https://developer.nvidia.com/rapids
and
https://www.xsede.org/for-users/ecss/ecss-symposium
Sign up for our insideHPC Newsletter: http://insidehp.com/newsletter
In this deck from FOSDEM 2020, Colin Sauze from Aberystwyth University describes the development of a RaspberryPi cluster for teaching an introduction to HPC.
"The motivation for this was to overcome four key problems faced by new HPC users:
* The availability of a real HPC system and the effect running training courses can have on the real system, conversely the availability of spare resources on the real system can cause problems for the training course.
* A fear of using a large and expensive HPC system for the first time and worries that doing something wrong might damage the system.
* That HPC systems are very abstract systems sitting in data centres that users never see, it is difficult for them to understand exactly what it is they are using.
* That new users fail to understand resource limitations, in part because of the vast resources in modern HPC systems a lot of mistakes can be made before running out of resources. A more resource constrained system makes it easier to understand this.
The talk will also discuss some of the technical challenges in deploying an HPC environment to a Raspberry Pi and attempts to keep that environment as close to a "real" HPC as possible. The issue to trying to automate the installation process will also be covered."
Learn more: https://github.com/colinsauze/pi_cluster
and
https://fosdem.org/2020/schedule/events/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from ATPESC 2019, Ken Raffenetti from Argonne presents an overview of HPC interconnects.
"The Argonne Training Program on Extreme-Scale Computing (ATPESC) provides intensive, two-week training on the key skills, approaches, and tools to design, implement, and execute computational science and engineering applications on current high-end computing systems and the leadership-class computing systems of the future."
Watch the video: https://wp.me/p3RLHQ-luc
Learn more: https://extremecomputingtraining.anl.gov/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPC
1. 1
Dr. Werner Scholz
XENON Systems, CTO and Head of R&D
werners@xenon.com.au
www.xenon.com.au
Liquid Cooling:
Exceeding the Limits of Air Cooling to
Unlock Greater Potential in
High Performance Computing
27 Aug. 2019
2. 2
XENON Systems – Who We Are
Australian company
established in 1996.
Global Onsite
hardware support
and installation
services in over 80
countries
Focused on
innovation and
investment in
Research &
Development
In-house
technical
ability to build
low volume
custom
designed
servers
Direct Relationship
with all major
component
manufacturers to
lower cost and speed
up support
Subsidiaries:
Mediaproxy
Global leader in compliance
logging and transport stream
monitoring for broadcast and TV
industries.
XDT/Catapult
Software for film and post
production industries.
XENOptics
Fibre automation solutions for
SDN in data centres
3. 3
XENON Systems – A History of High Performance Solutions
2018
Todd Energy (NZ)
HPC Cluster for Reservoir
Modelling
Fully managed service
CSL
HPC Cluster based on
Cisco UCS
4. 4
Why do we need to drive Data Centre Efficiency?
Refs:
https://www.theregister.co.uk/2016/07/25/semiconductor_industry_association_international_technology_roadmap_for_semiconductors/
http://www.semiconductors.org/main/2015_international_technology_roadmap_for_semiconductors_itrs/
●
Reduce power consumption
●
Reduce fossil fuel consumption
● Reduce CO2
emissions
●
Reduce cost
●
Save the planet
What is your motivation?
5. 5
Why do we need to drive Data Centre Efficiency?
Refs:
https://www.theregister.co.uk/2016/07/25/semiconductor_industry_association_international_technology_roadmap_for_semiconductors/
http://www.semiconductors.org/main/2015_international_technology_roadmap_for_semiconductors_itrs/
●
Reduce power consumption
●
Reduce fossil fuel consumption
● Reduce CO2
emissions
●
Reduce cost
●
Save the planet
What is your motivation?
Some scenarios predict that computers will
consume more power than the world can
produce by 2040.
(Semiconductor Industry Association Roadmap 2015)
6. 6
Component Trends and Demand are Challenging
Refs:
https://www.karlrupp.net/2018/02/42-years-of-microprocessor-trend-data/
●
Single thread performance increases are slowing
●
Number of logical cores and transistors per socket is
increasing
●
Power efficiency gains per core are slowing
●
Power per socket is increasing
7. 7
Component Trends and Demand are Challenging
Refs:
https://www.datacenterdynamics.com/opinions/its-not-just-about-chip-density-considering-liquid-cooling-for-your-data-center/
https://www.tomshardware.com/reviews/intel-xeon-w-3175x-cpu,5976-3.html
●
GPU: currently up to 300 W per device
●
CPU:
●
Jumping from typ. 150 W to
●
225 W (AMD EPYC 7742 with 64 cores)
●
255 W (Intel® Xeon® W-3175X) and recently
●
400 W (Intel Xeon 9282 “Cascade Lake AP” with 56
cores)
●
even higher for overclocked CPUs
8. 8
Impact on Data Centres
Refs:
11 PUE Data Center for the Masses A Webcast by The Green Grid.pdf
https://www.thegreengrid.org/
Challenges:
●
Power density/power per rack increases
●
Power feeds
●
Power distribution
●
Power cost
●
Hot spots
●
DC/Rack/Component Cooling
9. 9
Setting a Target: PUE = 1
Refs:
https://gigaom.com/2012/03/26/whose-data-centers-are-more-efficient-facebooks-or-googles/
https://googleblog.blogspot.com/2012/03/measuring-to-improve-comprehensive-real.html
https://www.datacenterknowledge.com/archives/2013/04/18/facebook-unveils-live-dashboard-on-pue-water-use
https://www.google.com/about/datacenters/efficiency/internal/index.html#measuring-efficiency
https://www.facebook.com/notes/facebook-engineering/designing-a-very-efficient-data-center/10150148003778920/
https://www.energystar.gov/ia/partners/prod_development/downloads/EPA_Datacenter_Report_Congress_Final1.pdf
*) Live Facebook data centre dashboards
https://www.facebook.com/PrinevilleDataCenter/app/399244020173259/
https://www.facebook.com/ForestCityDataCenter/app/288655784601722/
https://www.facebook.com/LuleaDataCenter/app/115276998849912/
https://www.facebook.com/AltoonaDataCenter/app/602730866540556/
PUE = Power Usage Effectiveness
DCIE = Data Center Infrastructure Efficiency
Total Facility Energy Non IT Facility Energy
IT Equipment Energy IT Equipment Energy
PUE = = 1+
DCIE = 1/PUE
Data Centre PUE
Google (2008) 1.21
Microsoft (2008) 1.22
Google (2010) 1.16
Facebook (2011) 1.08
Typical Data Centre (2011) 1.50
Microsoft (2012) 1.20
Google (2012) 1.14
Switch Supernap 7 (2015) 1.18
Facebook (Prineville 2015) 1.08
Allied Control (Bitfury 2015) (2PIC) 1.02
Green IT Cube (2016) (RDHX) 1.07
Supermicro (2017) 1.06
Facebook Prineville DC (yesterday*) 1.04
Facebook Forest City DC (yesterday*) 1.08
Facebook Lulea DC (yesterday*) 1.16
Facebook Altoona DC (yesterday*) 1.14
10. 10
How to Drive Data Centre Efficiency Gains?
Refs:
https://www.asetek.com/data-center/data-center-oem-partners/intel/
https://www.xenon.com.au/solutions/high-frequency-trading-solutions/
Some ideas:
●
Improve Data Centre design
●
Location selection (more on that later)
●
Improve component selection (for efficiency)
●
Improve component design (offload, BIGLittle)
●
Develop smarter algorithms:
Do more with less (but doing more with more is becoming cheaper faster...)
●
Smarter load distribution: In-Server, In-Rack, In-DC (avoid hot spots)
●
Improve Cooling Efficiency
11. 11
Liquid Cooling Options for Every Situation
Liquid cooling includes a whole family of options
In-server
●
Closed Loop Liquid Cooling
●
Actively pumped
●
Passive solutions
●
Heat Pipe and Solid Conduction
In-rack
●
LAAC (Liquid Assisted Air Cooling)
●
Rear Door Heat Exchangers
●
In-rack CDU
In-Data Centre
●
Immersion Cooling
●
Single phase immersion
●
2-phase immersion
●
“Full DC immersion”
13. 13
Passive In-Server Cooling
Refs
https://www.calyos-tm.com/processor-cooling-for-datacenter :
http://www.lemtech.com/products/heatsinks/#Process
1) Passive closed loop liquid cooling
●
Cover longer distances (to condenser)
●
Larger condenser for better efficiency
●
Enables higher TDP
2) Heat Pipe and Solid Conduction
No active parts
No risk of failure of active parts
Fully sealed systems: no leakage
14. 14
In-Rack Cooling
Refs:
https://www.asetek.com/data-center/technology-for-data-centers/inracklaac/
Rear Door Heat Exchangers
●
“Hot aisle” confinement within rack
●
Eliminates hotspots and “overcooling”
●
Retrofit onto existing racks usually possible
●
Easy installation
●
“Zero U”
●
Small space requirement (additional rack depth)
InRack LAAC (“Liquid Assisted Air Cooling”)
●
Rack-mounted 2U cabinet containing liquid-to-air
(L2A) heat exchanger
●
Capable of rejecting up to 6.4kW of total processor
power
●
Captures 60% to 80% of server heat with Asetek
D2C cooling loops
InRackCDU, OnRackCDU
●
Liquid-to-liquid (L2L) heat exchanger
●
Rejects up to 80kW of heat from the rack
●
Captures 60% to 80% of server heat with Asetek
D2C cooling loops
●
2.5x-5x increases in rack density
15. 15
Recent Project
OpenStack HPC Cluster
●
Intel 0.5U server (2U 4N)
●
Intel 2U servers for Ceph NVMe Storage
●
Supermicro 1U server with 4x NVIDIA V100
●
Water cooled CPUs and GPUs
Air cooled Liquid cooled
CPU LINPACK
CPU1 temp 73 C 52 C
CPU2 temp 76 C 55 C
GPU LINPACK
Performance (TFLOPS) 18.44 19.07
GPU power (typ.) 300 W 300 W
GPU power (max.) 375 W 379 W
GPU temp. (typ.) 50-60 C 40 C
GPU temp. (max.) 65 C 49 C
18. 18
How to Further Drive DC Efficiency Gains
●
Measure, measure, measure (power, PUE, load, temp., humidity)
●
Improve Thresholds and Sensitivities
●
Increase DC temp. as much as possible
●
Increase humidity range as much as possible
●
Improve DC design and airflow management
●
Optimize tile layout and cold air outlets to match IT load
●
Blanking plates and side panels
●
Barriers
●
CRAC unit return air inlet design
●
Hot and Cold aisle containment
●
Use free cooling (as climate permits)
●
Invest in CFD modelling (ROI could be <6 months)
●
Sensor lights to reduce light load
●
Reduce power conversion
●
Use high efficiency UPS
19. 19
From a Google Case Study
Refs:
Joe Kava, "Central Network Room (CNR) PU Efficiency Project - A True Story."
2011 Data Centre Efficiency Summit held on May 24 in Zürich, Switzerland.
https://www.youtube.com/watch?v=APynRrGuZJA
http://static.googleusercontent.com/media/www.google.com/en/us/corporate/datacenter/dc-best-practices-google.pdf
20. 20
Conclusions
Benefits of Liquid Cooling
●
Reduced number/elimination of fans and other air cooling components
●
Data Centre temp. and humidity can increase (to comfortable levels)
●
Data Centre noise level reduction
●
Power reduction
●
Cost reduction
●
Higher power density systems and racks
●
Performance increases (turbo/overclocking)
●
Reliability increases
●
Fewer/no fans / fan failures
●
Immersion cooling reduces corrosion risk, reduces failures, increases component life
Integrated with careful Data Centre Design further gains can be made
21. 21
Contacting XENON Systems
XENON Systems Pty Ltd
10 Westall Road
Clayton South, VIC 3169
P +61 3 9549 1111
F +61 3 9549 1199
info@xenon.com.au
www.xenon.com.au
ABN: 39 074 339 316
Contact Details
Werner Scholz
CTO and Head of R&D
03 9549 1140
werners@xenon.com.au
Michael Shing
Business Development Manager
03 9549 1144
michaels@xenon.com.au
22. 22
Dr. Werner Scholz
XENON Systems, CTO and Head of R&D
werners@xenon.com.au
Thank you!
Questions?
www.xenon.com.au
24. 24
XENON Deep Learning and AI Solutions
Deep Learning Workshops
XENON offers hands-on training for
developers, data scientists, and
researchers looking to solve
challenging problems with Deep Learning (DL).
These workshops are led by certified
instructors and can be held publicly or
privately, throughout Australia and NZ.
For private workshops business managers
can request an on-site highly customised DL
curriculum for their team. The workshops cover
the foundations of deep learning such as Image
Classification and Neural Network Deployment
using popular frameworks.
Outcomes:
• Understand the components of DL systems
intuitively
• Train a Deep Neural Network (DNN) to
correctly classify images it has never seen
before
• Deploy deep neural networks into
applications
Artificial Intelligence Consulting
XENON designs and delivers
Artificial Intelligence (AI)
solutions that include the very
latest high-performance CPU, GPU,
network, and storage hardware for
optimizing performance of deep learning
software and AI container stacks.
By understanding customer expectations
for AI and performing capability analysis
of IT systems, XENON applies a multi-step
approach to implementing AI and deep
learning within organisations including:
• Efficient storage & backup systems for
AI datasets
• Software and hardware matched to
specific AI functions
• AI technical resources and training
• AI Results measurement
Analytics Software
Anticipate future outcomes and
steer your business in the right
direction fast with XENON range
of data analytics software solutions.
KINETICA
Insight engine for the Extreme Data
Economy for both streaming data and
data-at-rest. Harness the power of data
to drive ever-changing business.
Accelerated parallel computing—powered
by NVIDIA® GPUs—tackles the
unpredictability and complexity of
extreme data.
ESI MINESET
Predictive Analytics software with
integrated machine learning enables easy
interactive exploration of data in the
cloud or on premise.
GPU Computing
Graphics processing units (GPUs)
can offload compute-intensive
functions in code from CPUs,
giving scientific computing researchers
and commercial businesses orders-of-
magnitude faster processing times.
XENON empowers you to build a highly
customised High Performance Computing
(HPC) cluster, server or workstation based
on NVIDIA® Tesla® and Quadro® GPUs.
In addition, XENON has partnered closely
with NVIDIA® and is:
• HPC Partner – Elite
• Deep Learning Partner – Elite
• Professional Visualization Partner – Elite
• GPU Virtualization Partner – Preferred
• DGX partner in Australia
25. 25
XENON Infrastructure
High Performance Computing
XENON offers a full range of
industry leading Servers and Storage
technologies designed
for High Performance Computing (HPC) users.
XENON’s HPC portfolio includes:
• HPC solution design
• High Performance-Servers, Clusters, Storage
and Parallel file systems
• High Performance Networking technologies
• A comprehensive Software Stack for HPC
including:
• Performance monitoring tools
• Development tools
• Application libraries
• Job Management and Scheduling
• File Systems(Lustre, GPFS, Panasas, NFS)
• System/Cluster Management (xCAT, Bright
Computing)
• Operating environments (RedHat, CentOS,
SuSE)
HighPerformanceNetworking
XENONdeliversHighPerformance
Computing (HPC)networkingsolutions
basedonInfinibandand
Ethernettechnology.Infinibandisacommunications
standardusedinHPC whichfeaturesveryhigh
throughputand verylowlatency.
Infinibandtechnologysolutionsinclude:
• Mellanox
• IntelOmnipath
Theseproductsareidealforlargescale,lowlatency
datatransfersforMessagePassing Interface(MPI)
trafficusedinHPCclusters.
Ethernettechnologysolutionsinclude:
• Mellanox+Cumulus
• Arista
• ExtremeNetworks
Theseproductsareidealforlowlatency software
definednetworks.
High Frequency Trading
XENON has a strong track record of
providing innovative High Frequency
Trading (HFT) solutions to leading
investment and proprietary trading firms in global
financial markets.
XENON’s HFT product portfolio consists of:
• eXtreme HFT Servers
• Arista Low-latency Switches
• Solarflare Server Adapters and Software
Applications
• Exablaze Network Adapters and Low- latency
Switches
• Spectracom Network Time Server
• Mellanox/Infiniband
XENON also partners with MRV Communications,
a global supplier of packet and optical solutions
that power the world’s largest networks.
ServerSolutions
XENON understands that computing
workloads are becoming more highly
sophisticatedandspecialisedand
these workload require computing systems that are
tuned to the performance demands andcharacteristics
ofeachworkload.
XENON can offer the right combination of systems to
meet any business challenge as well as maximise your
ITinvestment.
XENON offers a comprehensive range of server
solutions dedicated to solving each individual’s
businessproblemthatinclude:
• Single/DualProcessorservers
• Large Memory Systems (Multi-processors/ Large
memory:4-way,8-way, ScaleMP)
• HighDensityserversolutionsincluding Bladesystems
• CloudMicroservers
26. 26
XENON Solutions and Services
Big Data Analytics
All Big Data Solutions are built
and configured by XENON to
achieve the best performance.
Each solution is custom designed by
XENON engineers to cater towards
specific and unique project requirements
to establish optimisation and achieve
maximum scalability and efficiency.
Each solution is designed to:
• Simultaneously store and process large
datasets in a distributed environment
across servers and storage for extensive
structured and unstructured data
mining and analysis
• Tailor and deploy validated reference
architectures
• Lower costs and drive insights from your
data
VirtualDesktopInfrastructure
XENONVirtualDesktopInfrastructure
(VDI) solutionscreateacentralised
desktop experiencethatisaccessible
frommultiple devices(tablets,laptops&
workstations) withabroadbandconnection,
offering improveddatasecurity,reduced
downtime, andanaccelerateddeployment
process thatisquickandpainless.
Multiple VDI appliances can be clustered
together and deployed to support thousands of
users. VDI can help simplify devicemanagement
in a BYOD environment as corporate resources
are not stored on the device but instead remain
onthe premisesorinthecloud.
XENON VDI solutions feature GPU acceleration
utilising NVIDIA GRID technology and are
deliveredvia:
• CitrixXenDesktop
• VMwareHorizon
Professional Services
XENON offers an array of services
to assist users with complex
computing challenges and
problems. Our team of specialist all have
broad industry knowledge, experience and
expertise.
What sets us apart are our long-term
relationships with our customers. The trust
they place in us is founded upon the
personal ongoing service we offer, long
after the initial deployment of their
solution. We forge long-standing
relationships through a strong
commitment to quality, service, and value.
XENON can offer expert services tailored to
your unique requirements to assist you
with your Compute, Storage and
Networking challenges. Services can
include areas such as solution design,
troubleshooting to hardware rentals.
Managed Services
XENON software solutions makes it
easy to deploy and manage High
Performance Computing (HPC)
clusters, Big Data clusters, and OpenStack -
in the datacentre and in the cloud.
XENON can support a range of HPC Cluster
management products including: Rocks,
xCAT, Slurm, PBS and Bright. This enables
customers to focus on critical research and
business rather than struggling to manage
compute environments.
In addition, XENON along with it partners
can deliver tailored training and provide
knowledge transfer in both class-room or
less informal work environments.