In this deck from HPCKP'19, Karl Schultz from TACC presents: OpenHPC: Community Building Blocks for HPC Systems.
"Over the last several years, OpenHPC has emerged as a community-driven stack providing a variety of common, pre-built ingredients to deploy and manage an HPC Linux cluster including provisioning tools, resource management, I/O clients, runtimes, development tools, containers, and a variety of scientific libraries. Formed initially in November 2015 and formalized as a Linux Foundation project in June 2016, OpenHPC has been adding new software components and now supports multiple OSes and architectures. This presentation will present an overview of the project, currently available software, and highlight more recent changes along with general project updates and future plans."
Learn more: https://openhpc.community/
and
https://hpckp.org/
Watch the video: https://wp.me/p3RLHQ-kH2
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Warewulf 4.3.0 is an open source, stateless operating system provisioning system that is lightweight, flexible, and container-based. It uses containers to define the base OS for nodes and overlays to customize configurations on a per-node basis using templates. Nodes boot using PXE and iPXE to load the kernel, container, and system overlay before contacting the warewulfd server to retrieve the runtime overlay and complete booting into the container OS in a stateless manner. Future plans include adding new APIs and additional management features.
This document discusses deploying IPv6 on OpenStack. It provides an overview of IPv6, including that IPv6 addresses the shortage of IPv4 addresses by providing a vastly larger 128-bit address space. It describes IPv6 address types and allocation methods. It also discusses IPv6 configuration modes in OpenStack, including stateless address autoconfiguration (SLAAC) and DHCPv6 stateless and stateful modes. Additionally, it covers deployment options for IPv6 on OpenStack like dual stack, NAT64/DNS64, and network tunnels. It provides details on IPv6 address and router advertisement configuration in OpenStack.
In order to understand HAL layers of Android Framework, having Linux device driver knowledge is important. Hence Day-2 of the workshop focuses on the same.
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime RipardAnne Nicolas
Every modern multimedia-oriented ARM SoC usually has a number of display controllers, to drive a screen or an LCD panel, and a GPU, to provide 3D acceleration. The Linux kernel framework of choice to support these controllers is the DRM subsystem.
This talk will walk through the DRM stack, the architecture of a DRM/KMS driver and the interaction between the display and GPU drivers. The presentation is based on the work we have done to develop a DRM driver for the Allwinner SoCs display controller with multiple outputs, such as parallel display interfaces, HDMI or MIPI-DSI. The work done to make the ARM Mali OpenGL driver work on top of a mainline DRM/KMS driver will also be detailed, as well as the more traditional, Mesa-based, solution used in a variety of other platforms.
Maxime Ripard, Free Electrons
Buildroot is a tool that generates embedded Linux systems by automating the configuration, compilation, and packaging of the system. It produces a root filesystem image ready to deploy on the target architecture. The build process compiles packages and a cross-compilation toolchain, then generates images containing the root filesystem, kernel, and other files needed by the target system. The output of Buildroot is organized into subdirectories containing the built images, toolchain, target and host files, and a staging area simulating the target filesystem.
An unique module combining various previous modules you have learnt by combing Linux administration, Hardware knowledge, Linux as OS, C/Computer programming areas. This is a complete module on Embedded OS, as of now no books are written on this with such practical aspects. Here is a consolidated material to get real hands-on perspective about building custom Embedded Linux distribution in ARM.
The U-Boot is an "Universal Bootloader" ("Das U-Boot") is a monitor program that is under GPL. This production quality boot-loader is used as default boot loader by several board vendors. It is easily portable and easy to port and to debug by supporting PPC, ARM, MIPS, x86,m68k, NIOS, Microblaze architectures. Here is a presentation that introduces U-Boot.
Warewulf 4.3.0 is an open source, stateless operating system provisioning system that is lightweight, flexible, and container-based. It uses containers to define the base OS for nodes and overlays to customize configurations on a per-node basis using templates. Nodes boot using PXE and iPXE to load the kernel, container, and system overlay before contacting the warewulfd server to retrieve the runtime overlay and complete booting into the container OS in a stateless manner. Future plans include adding new APIs and additional management features.
This document discusses deploying IPv6 on OpenStack. It provides an overview of IPv6, including that IPv6 addresses the shortage of IPv4 addresses by providing a vastly larger 128-bit address space. It describes IPv6 address types and allocation methods. It also discusses IPv6 configuration modes in OpenStack, including stateless address autoconfiguration (SLAAC) and DHCPv6 stateless and stateful modes. Additionally, it covers deployment options for IPv6 on OpenStack like dual stack, NAT64/DNS64, and network tunnels. It provides details on IPv6 address and router advertisement configuration in OpenStack.
In order to understand HAL layers of Android Framework, having Linux device driver knowledge is important. Hence Day-2 of the workshop focuses on the same.
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime RipardAnne Nicolas
Every modern multimedia-oriented ARM SoC usually has a number of display controllers, to drive a screen or an LCD panel, and a GPU, to provide 3D acceleration. The Linux kernel framework of choice to support these controllers is the DRM subsystem.
This talk will walk through the DRM stack, the architecture of a DRM/KMS driver and the interaction between the display and GPU drivers. The presentation is based on the work we have done to develop a DRM driver for the Allwinner SoCs display controller with multiple outputs, such as parallel display interfaces, HDMI or MIPI-DSI. The work done to make the ARM Mali OpenGL driver work on top of a mainline DRM/KMS driver will also be detailed, as well as the more traditional, Mesa-based, solution used in a variety of other platforms.
Maxime Ripard, Free Electrons
Buildroot is a tool that generates embedded Linux systems by automating the configuration, compilation, and packaging of the system. It produces a root filesystem image ready to deploy on the target architecture. The build process compiles packages and a cross-compilation toolchain, then generates images containing the root filesystem, kernel, and other files needed by the target system. The output of Buildroot is organized into subdirectories containing the built images, toolchain, target and host files, and a staging area simulating the target filesystem.
An unique module combining various previous modules you have learnt by combing Linux administration, Hardware knowledge, Linux as OS, C/Computer programming areas. This is a complete module on Embedded OS, as of now no books are written on this with such practical aspects. Here is a consolidated material to get real hands-on perspective about building custom Embedded Linux distribution in ARM.
The U-Boot is an "Universal Bootloader" ("Das U-Boot") is a monitor program that is under GPL. This production quality boot-loader is used as default boot loader by several board vendors. It is easily portable and easy to port and to debug by supporting PPC, ARM, MIPS, x86,m68k, NIOS, Microblaze architectures. Here is a presentation that introduces U-Boot.
OpenWrt is a Linux distribution for embedded systems that runs on many routers and networking devices today. In this session we'll talk about OpenWrt's origins, architecture and get down to building apps for the platform.
Along the way we will touch on some basic firmware concepts and at last present the final working OpenWrt router and its capabilities.
Anton Lerner, Architect at Sitaro, computer geek, developer and occasional maker.
Sitaro provides total cyber protection for small business and home networks. Sitaro prevents massive scale IoT cyber attacks.
Find out more information in the meetup event page - https://www.meetup.com/Tel-Aviv-Yafo-Linux-Kernel-Meetup/events/245319189/
Part 01 Linux Kernel Compilation (Ubuntu)Tushar B Kute
Presentation on "Linux Kernel Compilation" (Ubuntu based).
Presented at Army Institute of Technology, Pune for FDP on "Basics of Linux Kernel Programming". by Tushar B Kute (http://tusharkute.com).
Linux Kernel and Driver Development TrainingStephan Cadene
This document provides information about a Linux Kernel and Driver Development training from Free Electrons. It begins with an overview of the course and hardware that will be used. It then discusses Free Electrons as a company and their online resources. The document also provides generic course information and guidelines for participation and the practical labs.
RISC-V Boot Process: One Step at a TimeAtish Patra
- OpenSBI is an open-source implementation of the RISC-V Supervisor Binary Interface (SBI) specifications. It provides runtime services in M-mode to facilitate booting of operating systems.
- OpenSBI supports various RISC-V platforms including SiFive boards, QEMU, and is integrated with projects like U-Boot and EDK2. It provides a standardized way for operating systems to interface with the underlying hardware.
- Future work includes supporting more platforms, implementing the SBI v0.2 specification, and enabling features like sequential CPU boot and hypervisor support. OpenSBI aims to establish a stable boot ecosystem for RISC-V.
Embedded Android system development workshop is focused on integrating new device with Android framework. Our hands-on approach makes Emertxe as the best institute to learn android system development training. This workshop deep dives into Android porting, Android Hardware Abstraction Layer (HAL), Android Services and Linux device driver ecosystem. This workshop based training program will enable you to efficiently integrate new hardware with Android HAL / Framework.
File systems provide an organized way to store and access data on storage devices like hard drives. The Linux file system hierarchy standard defines a common structure across Linux distributions with directories like /bin, /etc, /home, /usr, and /var. Common Linux file system types include ext2, ext3, ext4 for disks, initramfs for RAM, and JFFS2 for flash storage. File systems can also be distributed across a network using NFS or optimized for specific purposes like squashfs for read-only files. Partitions divide available storage space to better manage files, users, and data security.
This document provides an overview of the Red Hat Ansible Automation Platform. It begins with discussing why automation is important, citing industry analysts and research showing that automation is a strategic priority. It then discusses why the Red Hat Ansible platform specifically, noting its leadership position in Forrester evaluations. The rest of the document discusses what makes a platform, covering the key elements of creating, operating, and consuming automation. It details the various components of the Ansible platform that address the automation lifecycle.
This presentation will provide the information about the Linux Root File systems and its hierarchy. So any technocrate who is willing to gain info about root files of Linux can easily understand . preffered for Embedded system design Students who are pursuing diploma courses in various CDAC centers.
X / DRM (Direct Rendering Manager) Architectural OverviewMoriyoshi Koizumi
This document contains diagrams and descriptions related to the X Window System architecture for direct and indirect graphics rendering. It shows how OpenGL applications interact with the X server and Mesa library to perform direct graphics rendering using the kernel's Direct Rendering Infrastructure (DRI) and devices' direct rendering capabilities. It also summarizes the DRM authentication process where an application receives a magic cookie from the kernel to authenticate with the DRM driver.
For new age touch-based embedded devices, Android is becoming a popular OS going beyond mobile phones. With its roots from Embedded Linux, Android framework offers benefits in terms of rich libraries, open-source and multi-device support. Emertxe’s hands-on Embedded Android Training Course is designed to customize, build and deploy custom Embedded OS on ARM target. Rich set of projects will make your learning complete.
Rootless Containers & Unresolved issuesAkihiro Suda
Rootless containers allow users to run containers without root privileges by leveraging user and namespace isolation techniques. While rootless containers mitigate some security risks, there are still unresolved issues around sub-user management, networking, and adoption by runtimes and image builders. Rootless containers also cannot prevent all attacks if a container is broken out of. Container runtimes are working to improve support for rootless containers to further enhance security.
Kvm performance optimization for ubuntuSim Janghoon
This document discusses various techniques for optimizing KVM performance on Linux systems. It covers CPU and memory optimization through techniques like vCPU pinning, NUMA affinity, transparent huge pages, KSM, and virtio_balloon. For networking, it discusses vhost-net, interrupt handling using MSI/MSI-X, and NAPI. It also covers block device optimization through I/O scheduling, cache mode, and asynchronous I/O. The goal is to provide guidance on configuring these techniques for workloads running in KVM virtual machines.
Learning AOSP - Android Linux Device DriverNanik Tolaram
This document discusses Android and Linux device drivers. It provides an overview of Android's core low-level software and hardware drivers, which perform minimum tasks and access the framework layer or libraries. It also describes the interaction between the kernel, framework, and virtual filesystem for key drivers like binder, logger, and USB. Configuration options for Android in the Linux kernel are listed, and resources for learning more about Android open source are provided.
- Ironic is an OpenStack service and Nova driver that provides bare metal provisioning capabilities, allowing deployment of images to physical machines without using virtualization.
- It has a distributed architecture with components like ironic-api, ironic-conductor, and deploy ramdisk for provisioning hardware.
- Ironic supports various technologies for remote access and booting including PXE, IPMI, and iSCSI. Configuration options allow it to work with both homogeneous and heterogeneous hardware environments.
The document discusses Compatibility Test Suite (CTS) which tests Android device compatibility. It describes how to download, build, install and run CTS. It also provides examples of common CTS test failures and their solutions such as adding missing features, modifying device configurations, and using alternative test methods.
This document provides an overview of the data structures and functions used to implement ethernet drivers in the Linux kernel. It discusses the net_device and sk_buff structures that represent network interfaces and packets. It also describes how the driver interacts with the kernel via polling, interrupts, and NAPI to handle reception and transmission of frames. Finally, it provides an example of the key components needed for a simple ethernet driver, including initialization, setup, open/close, transmission, and reception functions.
The document discusses Ext4 journaling and the write barrier feature. It notes that the write barrier forces a flush-to-disk call after writing the journal to ensure consistency. However, this can cause sluggishness when storage is full during OTA updates. Disabling the write barrier allows reordering of cache-to-disk writes, reducing latency and improving performance, though it introduces a small risk of filesystem corruption in the event of a power failure. Tests showed disabling the barrier reduced fsync latency and improved SQLite transactions per second on HDD and EMMC storage.
Presentació a càrrec d'Ismael Fernández i Cristian
Gomollón (tècnics d'Aplicacions al CSUC) duta a terme a la jornada de formació "Com usar el servei de càlcul del CSUC" celebrada el 8 d'octubre de 2019 al CSUC.
Process Scheduler and Balancer in Linux KernelHaifeng Li
The document discusses process scheduling in the Linux kernel. It begins with an introduction to schedulers and outlines Linux scheduler history, including round-robin, O(N), O(1), and Completely Fair schedulers. It then provides more details on the Completely Fair Scheduler's use of virtual runtime to determine process execution order and maintain fairness. The document also briefly discusses the real-time scheduler and key data structures used for scheduling like task_struct and sched_domain. It concludes with an overview of the CFS load balancer's hierarchical design and methods used to balance tasks among CPU cores.
Extending OpenShift Origin: Build Your Own Cartridge with Bill DeCoste of Red...OpenShift Origin
Extending OpenShift Origin: Build Your Own Cartridge
Presenters: Bill DeCoste
Cartridges allow developers to provide services running on top of the Red Hat OpenShift Platform-as-a-Service (PaaS). OpenShift already provides cartridges for numerous web application frameworks and databases. Writing your own cartridges allows you to customize or enhance an existing service, or provide new services. In this session, the presenter will discuss best practices for cartridge development and the latest changes in the OpenShift cartridge support.
* Latest changes made in the platform to ease cartridge development
* OpenShift Cartridges vs. plugins
* Outline for development of a new cartridge
* Customization of existing cartridges
* Quickstarts: leveraging a cartridge or cartridges to provide a complete application
In this deck from the 2017 MVAPICH User Group, Karl Schultz from Intel presents: OpenHPC: Project Overview and Updates.
"There is a growing sense within the HPC community for the need to have an open community effort to more efficiently build, test, and deliver integrated HPC software components and tools. To address this need, OpenHPC launched as a Linux Foundation collaborative project in 2016 with combined participation from academia, national labs, and industry. The project's mission is to provide a reference collection of open-source HPC software components and best practices in order to lower barriers to deployment and advance the use of modern HPC methods and tools. This presentation will present an overview of the project, highlight packaging conventions and currently available software, outline integration testing/validation efforts with bare-metal installation recipes, and discuss recent updates and expected future plans."
Watch the video: https://wp.me/p3RLHQ-hpH
Learn more: http://openhpc.community/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
OpenWrt is a Linux distribution for embedded systems that runs on many routers and networking devices today. In this session we'll talk about OpenWrt's origins, architecture and get down to building apps for the platform.
Along the way we will touch on some basic firmware concepts and at last present the final working OpenWrt router and its capabilities.
Anton Lerner, Architect at Sitaro, computer geek, developer and occasional maker.
Sitaro provides total cyber protection for small business and home networks. Sitaro prevents massive scale IoT cyber attacks.
Find out more information in the meetup event page - https://www.meetup.com/Tel-Aviv-Yafo-Linux-Kernel-Meetup/events/245319189/
Part 01 Linux Kernel Compilation (Ubuntu)Tushar B Kute
Presentation on "Linux Kernel Compilation" (Ubuntu based).
Presented at Army Institute of Technology, Pune for FDP on "Basics of Linux Kernel Programming". by Tushar B Kute (http://tusharkute.com).
Linux Kernel and Driver Development TrainingStephan Cadene
This document provides information about a Linux Kernel and Driver Development training from Free Electrons. It begins with an overview of the course and hardware that will be used. It then discusses Free Electrons as a company and their online resources. The document also provides generic course information and guidelines for participation and the practical labs.
RISC-V Boot Process: One Step at a TimeAtish Patra
- OpenSBI is an open-source implementation of the RISC-V Supervisor Binary Interface (SBI) specifications. It provides runtime services in M-mode to facilitate booting of operating systems.
- OpenSBI supports various RISC-V platforms including SiFive boards, QEMU, and is integrated with projects like U-Boot and EDK2. It provides a standardized way for operating systems to interface with the underlying hardware.
- Future work includes supporting more platforms, implementing the SBI v0.2 specification, and enabling features like sequential CPU boot and hypervisor support. OpenSBI aims to establish a stable boot ecosystem for RISC-V.
Embedded Android system development workshop is focused on integrating new device with Android framework. Our hands-on approach makes Emertxe as the best institute to learn android system development training. This workshop deep dives into Android porting, Android Hardware Abstraction Layer (HAL), Android Services and Linux device driver ecosystem. This workshop based training program will enable you to efficiently integrate new hardware with Android HAL / Framework.
File systems provide an organized way to store and access data on storage devices like hard drives. The Linux file system hierarchy standard defines a common structure across Linux distributions with directories like /bin, /etc, /home, /usr, and /var. Common Linux file system types include ext2, ext3, ext4 for disks, initramfs for RAM, and JFFS2 for flash storage. File systems can also be distributed across a network using NFS or optimized for specific purposes like squashfs for read-only files. Partitions divide available storage space to better manage files, users, and data security.
This document provides an overview of the Red Hat Ansible Automation Platform. It begins with discussing why automation is important, citing industry analysts and research showing that automation is a strategic priority. It then discusses why the Red Hat Ansible platform specifically, noting its leadership position in Forrester evaluations. The rest of the document discusses what makes a platform, covering the key elements of creating, operating, and consuming automation. It details the various components of the Ansible platform that address the automation lifecycle.
This presentation will provide the information about the Linux Root File systems and its hierarchy. So any technocrate who is willing to gain info about root files of Linux can easily understand . preffered for Embedded system design Students who are pursuing diploma courses in various CDAC centers.
X / DRM (Direct Rendering Manager) Architectural OverviewMoriyoshi Koizumi
This document contains diagrams and descriptions related to the X Window System architecture for direct and indirect graphics rendering. It shows how OpenGL applications interact with the X server and Mesa library to perform direct graphics rendering using the kernel's Direct Rendering Infrastructure (DRI) and devices' direct rendering capabilities. It also summarizes the DRM authentication process where an application receives a magic cookie from the kernel to authenticate with the DRM driver.
For new age touch-based embedded devices, Android is becoming a popular OS going beyond mobile phones. With its roots from Embedded Linux, Android framework offers benefits in terms of rich libraries, open-source and multi-device support. Emertxe’s hands-on Embedded Android Training Course is designed to customize, build and deploy custom Embedded OS on ARM target. Rich set of projects will make your learning complete.
Rootless Containers & Unresolved issuesAkihiro Suda
Rootless containers allow users to run containers without root privileges by leveraging user and namespace isolation techniques. While rootless containers mitigate some security risks, there are still unresolved issues around sub-user management, networking, and adoption by runtimes and image builders. Rootless containers also cannot prevent all attacks if a container is broken out of. Container runtimes are working to improve support for rootless containers to further enhance security.
Kvm performance optimization for ubuntuSim Janghoon
This document discusses various techniques for optimizing KVM performance on Linux systems. It covers CPU and memory optimization through techniques like vCPU pinning, NUMA affinity, transparent huge pages, KSM, and virtio_balloon. For networking, it discusses vhost-net, interrupt handling using MSI/MSI-X, and NAPI. It also covers block device optimization through I/O scheduling, cache mode, and asynchronous I/O. The goal is to provide guidance on configuring these techniques for workloads running in KVM virtual machines.
Learning AOSP - Android Linux Device DriverNanik Tolaram
This document discusses Android and Linux device drivers. It provides an overview of Android's core low-level software and hardware drivers, which perform minimum tasks and access the framework layer or libraries. It also describes the interaction between the kernel, framework, and virtual filesystem for key drivers like binder, logger, and USB. Configuration options for Android in the Linux kernel are listed, and resources for learning more about Android open source are provided.
- Ironic is an OpenStack service and Nova driver that provides bare metal provisioning capabilities, allowing deployment of images to physical machines without using virtualization.
- It has a distributed architecture with components like ironic-api, ironic-conductor, and deploy ramdisk for provisioning hardware.
- Ironic supports various technologies for remote access and booting including PXE, IPMI, and iSCSI. Configuration options allow it to work with both homogeneous and heterogeneous hardware environments.
The document discusses Compatibility Test Suite (CTS) which tests Android device compatibility. It describes how to download, build, install and run CTS. It also provides examples of common CTS test failures and their solutions such as adding missing features, modifying device configurations, and using alternative test methods.
This document provides an overview of the data structures and functions used to implement ethernet drivers in the Linux kernel. It discusses the net_device and sk_buff structures that represent network interfaces and packets. It also describes how the driver interacts with the kernel via polling, interrupts, and NAPI to handle reception and transmission of frames. Finally, it provides an example of the key components needed for a simple ethernet driver, including initialization, setup, open/close, transmission, and reception functions.
The document discusses Ext4 journaling and the write barrier feature. It notes that the write barrier forces a flush-to-disk call after writing the journal to ensure consistency. However, this can cause sluggishness when storage is full during OTA updates. Disabling the write barrier allows reordering of cache-to-disk writes, reducing latency and improving performance, though it introduces a small risk of filesystem corruption in the event of a power failure. Tests showed disabling the barrier reduced fsync latency and improved SQLite transactions per second on HDD and EMMC storage.
Presentació a càrrec d'Ismael Fernández i Cristian
Gomollón (tècnics d'Aplicacions al CSUC) duta a terme a la jornada de formació "Com usar el servei de càlcul del CSUC" celebrada el 8 d'octubre de 2019 al CSUC.
Process Scheduler and Balancer in Linux KernelHaifeng Li
The document discusses process scheduling in the Linux kernel. It begins with an introduction to schedulers and outlines Linux scheduler history, including round-robin, O(N), O(1), and Completely Fair schedulers. It then provides more details on the Completely Fair Scheduler's use of virtual runtime to determine process execution order and maintain fairness. The document also briefly discusses the real-time scheduler and key data structures used for scheduling like task_struct and sched_domain. It concludes with an overview of the CFS load balancer's hierarchical design and methods used to balance tasks among CPU cores.
Extending OpenShift Origin: Build Your Own Cartridge with Bill DeCoste of Red...OpenShift Origin
Extending OpenShift Origin: Build Your Own Cartridge
Presenters: Bill DeCoste
Cartridges allow developers to provide services running on top of the Red Hat OpenShift Platform-as-a-Service (PaaS). OpenShift already provides cartridges for numerous web application frameworks and databases. Writing your own cartridges allows you to customize or enhance an existing service, or provide new services. In this session, the presenter will discuss best practices for cartridge development and the latest changes in the OpenShift cartridge support.
* Latest changes made in the platform to ease cartridge development
* OpenShift Cartridges vs. plugins
* Outline for development of a new cartridge
* Customization of existing cartridges
* Quickstarts: leveraging a cartridge or cartridges to provide a complete application
In this deck from the 2017 MVAPICH User Group, Karl Schultz from Intel presents: OpenHPC: Project Overview and Updates.
"There is a growing sense within the HPC community for the need to have an open community effort to more efficiently build, test, and deliver integrated HPC software components and tools. To address this need, OpenHPC launched as a Linux Foundation collaborative project in 2016 with combined participation from academia, national labs, and industry. The project's mission is to provide a reference collection of open-source HPC software components and best practices in order to lower barriers to deployment and advance the use of modern HPC methods and tools. This presentation will present an overview of the project, highlight packaging conventions and currently available software, outline integration testing/validation efforts with bare-metal installation recipes, and discuss recent updates and expected future plans."
Watch the video: https://wp.me/p3RLHQ-hpH
Learn more: http://openhpc.community/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the 2016 HPC Advisory Council Switzerland Conference, Kenneth Hoste from the University Ghent presents an introduction to EasyBuild, an open-source framework for (automatically) getting scientific software installed on HPC systems.
Watch the video presentation: http://wp.me/p3RLHQ-f8J
Learn more: https://github.com/hpcugent/easybuild
See more talks from the Switzerland HPC Conference:
http://insidehpc.com/2016-swiss-hpc-conference/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Workflow story: Theory versus Practice in large enterprises by Marcin PiebiakNETWAYS
Uphill battle against large enterprise it environments and IT corporate culture. How those difficulties turned out opportunities and clever implementations. Interesting modules, integrations and workflow pieces.
The document provides an overview and cheat sheet for using the OpenShift command line interface. It defines what OpenShift is, lists common commands for login/authentication, managing projects and resources, building and deploying applications, and provides an example of creating a new project, adding users, building an application from source code and image, and checking the status of deployed resources.
Hands on with embedded linux using zero hardwareRajesh Sola
This document provides an overview of hands-on embedded Linux development using zero hardware. It discusses setting up a cross-development toolchain, building a Linux kernel and root filesystem for ARM, emulating the system using QEMU, and approaches for booting the system using U-Boot from an SD card or over a network using TFTP. The document also covers cross-compiling applications, creating static and dynamic libraries, and writing simple kernel modules.
openSUSE on ARM provides openSUSE distributions such as Tumbleweed, Leap, and Kubic/MicroOS that run on ARM architectures like ARMv7 and ARMv64. The distributions are built in the Open Build Service and tested using openQA. Recent improvements included adding more ARM build power in OBS, enabling ARM builds on more projects, and adding tests in openQA. The openSUSE wiki was also updated with new information on ARM boards and the latest tests. Future work includes improving build speeds, enabling more ARMv7 tests, and getting feedback on additional systems.
The document provides instructions for installing and configuring OpenERP on Ubuntu, including downloading and installing Ubuntu, installing required packages like PostgreSQL, configuring the PostgreSQL database, downloading and installing the OpenERP server and client, configuring the OpenERP files, and starting the OpenERP server and client.
Lxc – next gen virtualization for cloud intro (cloudexpo)Boden Russell
This document provides an introduction and overview of Linux containers as next-generation virtualization for cloud computing. It discusses how Linux containers provide better performance and flexibility than traditional virtual machines through the use of cgroups and namespaces. It also covers how containerization is gaining industry momentum and provides lower total cost of ownership through integration with modern Linux kernels and open source tooling. Finally, it defines key Linux container technologies, compares containers to hypervisors, and discusses building and securing Linux containers.
Kubernetes and OpenStack at Scale at OpenStack Summit Boston 2017
Imagine being able to stand up thousands of tenants with thousands of apps, running thousands of Docker-formatted container images and routes, all on a self-healing cluster and elastic infrastructure. Now, take that one step further - all of those images being updatable through a single upload to the registry, and with zero downtime. In this session, you will see just that.
In this presentation, we will walk through a recent benchmarking deployment using Kubernetes and OpenStack on the Cloud Native Computing Foundation’s (CNCF's) 1,000 node cluster with OpenStack and Red Hat’s OpenShift Container Platform, the enterprise-ready Kubernetes for developers.
You'll also what's been happening in subsequent rounds of testing in Red Hat's own SCALE lab and the CNCF cluster and how we are working with the relevant open source communities including OpenStack, Kubernetes, and Ansible to continue to raise the bar for horizontal scaling of these platforms via community powered innovation.
This document provides an overview of using R and high performance computers (HPC). It discusses why HPC is useful when data becomes too large for a local machine, and strategies like moving to more powerful hardware, using parallel packages, or rewriting code. It also covers topics like accessing HPC resources through batch jobs, setting up the R environment, profiling code, and using packages like purrr and foreach to parallelize workflows. The overall message is that HPC can scale up R analyses, but developers must adapt their code for parallel and distributed processing.
OpenDaylight SDN Controller - IntroductionEueung Mulyana
This document provides an overview of OpenDaylight (ODL) and instructions for getting started with ODL on a Raspberry Pi 3 using Mininet and the REST interface. It includes the following sections:
1. Introduction to ODL including its architecture, community, and releases.
2. Instructions for installing Java, downloading and running ODL, and installing features using Karaf.
3. Directions for connecting Mininet to the ODL instance and using DLUX and RESTconf to view and interact with the network topology and nodes.
4. Information on the REST interface and how to use curl and Yang UI/Yangman to send requests and view responses regarding the network topology and inventory.
Multi-OS Continuous Packaging with docker and Project-Builder.orgBruno Cornec
This document introduces Project Builder, a tool for continuously packaging software projects. It allows packaging to be done in an operating system agnostic way across different Linux distributions and package formats. Project Builder uses container technologies like Docker to provide isolated build environments for projects. It aims to make packaging easier for upstream projects and developers to integrate with various distribution repositories and deployment systems.
Continuous Integration and DevOps with Open Build Service(OBS)Ralf Dannert
speech/workshop held at 6. Secure Linux Administration Conference(SLAC) in Berlin
content is mixed English/German
* Open Buildservice (OBS)
* typical workflows with OBS
* command-line tool (osc)
* source Services in OBS
* continuous integration with Jenkins (track upstream github)
* Image build(iso) in OBS using KIWI
his workshop will shed light on a modern solution to solve application portability, building, delivery, packaging, and system dependency issues. Containers especially Docker have seen accelerated adoption in the web, cloud and recently the enterprise. HPC environments are seeing something similar to the introduction of HPC containers Singularity and Shifter. They provide a good use case for solving software portability, not to mention ensure repeatability of results. Not to mention their ECO system provides for the better development, delivery, testing workflows that were alien to most of HPC environments. This workshop will cover the Theory and hands-on of containers and Its ecosystem. Introducing Docker and singularity containers; Docker as a general-purpose container for almost any app, Singularity as the particular container technology for HPC. The workshop will go over the foundations of the containers platform, including an overview of the platform system components: images, containers, repositories, clustering, and orchestration. The strategy is to demonstrate through "live demo, and hands-on exercises." The reuse case of containers in building a portable distributed application cluster running a variety of workloads including HPC workload.
Similar to OpenHPC: Community Building Blocks for HPC Systems (20)
The document discusses the top 5 technologies that all organizations must understand: digital transformation, quantum computing, IoT, 5G, and AI/HPC. It provides an overview of each technology including opportunities and threats to organizations. The document emphasizes that understanding these emerging technologies is mandatory as the information revolution changes many aspects of life and business.
Preparing to program Aurora at Exascale - Early experiences and future direct...inside-BigData.com
In this deck from IWOCL / SYCLcon 2020, Hal Finkel from Argonne National Laboratory presents: Preparing to program Aurora at Exascale - Early experiences and future directions.
"Argonne National Laboratory’s Leadership Computing Facility will be home to Aurora, our first exascale supercomputer. Aurora promises to take scientific computing to a whole new level, and scientists and engineers from many different fields will take advantage of Aurora’s unprecedented computational capabilities to push the boundaries of human knowledge. In addition, Aurora’s support for advanced machine-learning and big-data computations will enable scientific workflows incorporating these techniques along with traditional HPC algorithms. Programming the state-of-the-art hardware in Aurora will be accomplished using state-of-the-art programming models. Some of these models, such as OpenMP, are long-established in the HPC ecosystem. Other models, such as Intel’s oneAPI, based on SYCL, are relatively-new models constructed with the benefit of significant experience. Many applications will not use these models directly, but rather, will use C++ abstraction libraries such as Kokkos or RAJA. Python will also be a common entry point to high-performance capabilities. As we look toward the future, features in the C++ standard itself will become increasingly relevant for accessing the extreme parallelism of exascale platforms.
This presentation will summarize the experiences of our team as we prepare for Aurora, exploring how to port applications to Aurora’s architecture and programming models, and distilling the challenges and best practices we’ve developed to date. oneAPI/SYCL and OpenMP are both critical models in these efforts, and while the ecosystem for Aurora has yet to mature, we’ve already had a great deal of success. Importantly, we are not passive recipients of programming models developed by others. Our team works not only with vendor-provided compilers and tools, but also develops improved open-source LLVM-based technologies that feed both open-source and vendor-provided capabilities. In addition, we actively participate in the standardization of OpenMP, SYCL, and C++. To conclude, I’ll share our thoughts on how these models can best develop in the future to support exascale-class systems."
Watch the video: https://wp.me/p3RLHQ-lPT
Learn more: https://www.iwocl.org/iwocl-2020/conference-program/
and
https://www.anl.gov/topic/aurora
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Greg Wahl from Advantech presents: Transforming Private 5G Networks.
Advantech Networks & Communications Group is driving innovation in next-generation network solutions with their High Performance Servers. We provide business critical hardware to the world's leading telecom and networking equipment manufacturers with both standard and customized products. Our High Performance Servers are highly configurable platforms designed to balance the best in x86 server-class processing performance with maximum I/O and offload density. The systems are cost effective, highly available and optimized to meet next generation networking and media processing needs.
“Advantech’s Networks and Communication Group has been both an innovator and trusted enabling partner in the telecommunications and network security markets for over a decade, designing and manufacturing products for OEMs that accelerate their network platform evolution and time to market.” Said Advantech Vice President of Networks & Communications Group, Ween Niu. “In the new IP Infrastructure era, we will be expanding our expertise in Software Defined Networking (SDN) and Network Function Virtualization (NFV), two of the essential conduits to 5G infrastructure agility making networks easier to install, secure, automate and manage in a cloud-based infrastructure.”
In addition to innovation in air interface technologies and architecture extensions, 5G will also need a new generation of network computing platforms to run the emerging software defined infrastructure, one that provides greater topology flexibility, essential to deliver on the promises of high availability, high coverage, low latency and high bandwidth connections. This will open up new parallel industry opportunities through dedicated 5G network slices reserved for specific industries dedicated to video traffic, augmented reality, IoT, connected cars etc. 5G unlocks many new doors and one of the keys to its enablement lies in the elasticity and flexibility of the underlying infrastructure.
Advantech’s corporate vision is to enable an intelligent planet. The company is a global leader in the fields of IoT intelligent systems and embedded platforms. To embrace the trends of IoT, big data, and artificial intelligence, Advantech promotes IoT hardware and software solutions with the Edge Intelligence WISE-PaaS core to assist business partners and clients in connecting their industrial chains. Advantech is also working with business partners to co-create business ecosystems that accelerate the goal of industrial intelligence."
Watch the video: https://wp.me/p3RLHQ-lPQ
* Company website: https://www.advantech.com/
* Solution page: https://www2.advantech.com/nc/newsletter/NCG/SKY/benefits.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
In this deck from the Stanford HPC Conference, Katie Lewis from Lawrence Livermore National Laboratory presents: The Incorporation of Machine Learning into Scientific Simulations at Lawrence Livermore National Laboratory.
"Scientific simulations have driven computing at Lawrence Livermore National Laboratory (LLNL) for decades. During that time, we have seen significant changes in hardware, tools, and algorithms. Today, data science, including machine learning, is one of the fastest growing areas of computing, and LLNL is investing in hardware, applications, and algorithms in this space. While the use of simulations to focus and understand experiments is well accepted in our community, machine learning brings new challenges that need to be addressed. I will explore applications for machine learning in scientific simulations that are showing promising results and further investigation that is needed to better understand its usefulness."
Watch the video: https://youtu.be/NVwmvCWpZ6Y
Learn more: https://computing.llnl.gov/research-area/machine-learning
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...inside-BigData.com
In this deck from the Stanford HPC Conference, DK Panda from Ohio State University presents: How to Achieve High-Performance, Scalable and Distributed DNN Training on Modern HPC Systems?
"This talk will start with an overview of challenges being faced by the AI community to achieve high-performance, scalable and distributed DNN training on Modern HPC systems with both scale-up and scale-out strategies. After that, the talk will focus on a range of solutions being carried out in my group to address these challenges. The solutions will include: 1) MPI-driven Deep Learning, 2) Co-designing Deep Learning Stacks with High-Performance MPI, 3) Out-of- core DNN training, and 4) Hybrid (Data and Model) parallelism. Case studies to accelerate DNN training with popular frameworks like TensorFlow, PyTorch, MXNet and Caffe on modern HPC systems will be presented."
Watch the video: https://youtu.be/LeUNoKZVuwQ
Learn more: http://web.cse.ohio-state.edu/~panda.2/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
In this deck from the Stanford HPC Conference, Nick Nystrom and Paola Buitrago provide an update from the Pittsburgh Supercomputing Center.
Nick Nystrom is Chief Scientist at the Pittsburgh Supercomputing Center (PSC). Nick is architect and PI for Bridges, PSC's flagship system that successfully pioneered the convergence of HPC, AI, and Big Data. He is also PI for the NIH Human Biomolecular Atlas Program’s HIVE Infrastructure Component and co-PI for projects that bring emerging AI technologies to research (Open Compass), apply machine learning to biomedical data for breast and lung cancer (Big Data for Better Health), and identify causal relationships in biomedical big data (the Center for Causal Discovery, an NIH Big Data to Knowledge Center of Excellence). His current research interests include hardware and software architecture, applications of machine learning to multimodal data (particularly for the life sciences) and to enhance simulation, and graph analytics.
Watch the video: https://youtu.be/LWEU1L1o7yY
Learn more: https://www.psc.edu/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
The document discusses using systems intelligence and artificial intelligence/neural networks to enhance semiconductor electronic design automation (EDA) workflows by collecting telemetry data from EDA jobs and infrastructure and analyzing it using complex event processing, machine learning models, and messaging substrates to provide insights that could optimize EDA pipelines and infrastructure. The approach aims to allow both internal and external augmentation of EDA processes and environments through unsupervised and incremental learning.
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoringinside-BigData.com
In this deck from the Stanford HPC Conference, Nicole Xu from Stanford University describes how she transformed a common jellyfish into a bionic creature that is part animal and part machine.
"Animal locomotion and bioinspiration have the potential to expand the performance capabilities of robots, but current implementations are limited. Mechanical soft robots leverage engineered materials and are highly controllable, but these biomimetic robots consume more power than corresponding animal counterparts. Biological soft robots from a bottom-up approach offer advantages such as speed and controllability but are limited to survival in cell media. Instead, biohybrid robots that comprise live animals and self- contained microelectronic systems leverage the animals’ own metabolism to reduce power constraints and body as an natural scaffold with damage tolerance. We demonstrate that by integrating onboard microelectronics into live jellyfish, we can enhance propulsion up to threefold, using only 10 mW of external power input to the microelectronics and at only a twofold increase in cost of transport to the animal. This robotic system uses 10 to 1000 times less external power per mass than existing swimming robots in literature and can be used in future applications for ocean monitoring to track environmental changes."
Watch the video: https://youtu.be/HrmJFyvInj8
Learn more: https://sanfrancisco.cbslocal.com/2020/02/05/stanford-research-project-common-jellyfish-bionic-sea-creatures/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Stanford HPC Conference, Peter Dueben from the European Centre for Medium-Range Weather Forecasts (ECMWF) presents: Machine Learning for Weather Forecasts.
"I will present recent studies that use deep learning to learn the equations of motion of the atmosphere, to emulate model components of weather forecast models and to enhance usability of weather forecasts. I will than talk about the main challenges for the application of deep learning in cutting-edge weather forecasts and suggest approaches to improve usability in the future."
Peter is contributing to the development and optimization of weather and climate models for modern supercomputers. He is focusing on a better understanding of model error and model uncertainty, on the use of reduced numerical precision that is optimised for a given level of model error, on global cloud- resolving simulations with ECMWF's forecast model, and the use of machine learning, and in particular deep learning, to improve the workflow and predictions. Peter has graduated in Physics and wrote his PhD thesis at the Max Planck Institute for Meteorology in Germany. He worked as Postdoc with Tim Palmer at the University of Oxford and has taken up a position as University Research Fellow of the Royal Society at the European Centre for Medium-Range Weather Forecasts (ECMWF) in 2017.
Watch the video: https://youtu.be/ks3fkRj8Iqc
Learn more: https://www.ecmwf.int/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck, Gilad Shainer from the HPC AI Advisory Council describes how this organization fosters innovation in the high performance computing community.
"The HPC-AI Advisory Council’s mission is to bridge the gap between high-performance computing (HPC) and Artificial Intelligence (AI) use and its potential, bring the beneficial capabilities of HPC and AI to new users for better research, education, innovation and product manufacturing, bring users the expertise needed to operate HPC and AI systems, provide application designers with the tools needed to enable parallel computing, and to strengthen the qualification and integration of HPC and AI system products."
Watch the video: https://wp.me/p3RLHQ-lNz
Learn more: http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Today RIKEN in Japan announced that the Fugaku supercomputer will be made available for research projects aimed to combat COVID-19.
"Fugaku is currently being installed and is scheduled to be available to the public in 2021. However, faced with the devastating disaster unfolding before our eyes, RIKEN and MEXT decided to make a portion of the computational resources of Fugaku available for COVID-19-related projects ahead of schedule while continuing the installation process.
Fugaku is being developed not only for the progress in science, but also to help build the society dubbed as the “Society 5.0” by the Japanese government, where all people will live safe and comfortable lives. The current initiative to fight against the novel coronavirus is driven by the philosophy behind the development of Fugaku."
Initial Projects
Exploring new drug candidates for COVID-19 by "Fugaku"
Yasushi Okuno, RIKEN / Kyoto University
Prediction of conformational dynamics of proteins on the surface of SARS-Cov-2 using Fugaku
Yuji Sugita, RIKEN
Simulation analysis of pandemic phenomena
Nobuyasu Ito, RIKEN
Fragment molecular orbital calculations for COVID-19 proteins
Yuji Mochizuki, Rikkyo University
In this deck from the Performance Optimisation and Productivity group, Lubomir Riha from IT4Innovations presents: Energy Efficient Computing using Dynamic Tuning.
"We now live in a world of power-constrained architectures and systems and power consumption represents a significant cost factor in the overall HPC system economy. For these reasons, in recent years researchers, supercomputing centers and major vendors have developed new tools and methodologies to measure and optimize the energy consumption of large-scale high performance system installations. Due to the link between energy consumption, power consumption and execution time of an application executed by the final user, it is important for these tools and the methodology used to consider all these aspects, empowering the final user and the system administrator with the capability of finding the best configuration given different high level objectives.
This webinar focused on tools designed to improve the energy-efficiency of HPC applications using a methodology of dynamic tuning of HPC applications, developed under the H2020 READEX project. The READEX methodology has been designed for exploiting the dynamic behaviour of software. At design time, different runtime situations (RTS) are detected and optimized system configurations are determined. RTSs with the same configuration are grouped into scenarios, forming the tuning model. At runtime, the tuning model is used to switch system configurations dynamically.
The MERIC tool, that implements the READEX methodology, is presented. It supports manual or binary instrumentation of the analysed applications to simplify the analysis. This instrumentation is used to identify and annotate the significant regions in the HPC application. Automatic binary instrumentation annotates regions with significant runtime. Manual instrumentation, which can be combined with automatic, allows code developer to annotate regions of particular interest."
Watch the video: https://wp.me/p3RLHQ-lJP
Learn more: https://pop-coe.eu/blog/14th-pop-webinar-energy-efficient-computing-using-dynamic-tuning
and
https://code.it4i.cz/vys0053/meric
Sign up for our insideHPC Newsletter: http://insidehpc.com/newslett
The document discusses how DDN A3I storage solutions and Nvidia's SuperPOD platform can enable HPC at scale. It provides details on DDN's A3I appliances that are optimized for AI and deep learning workloads and validated for Nvidia's DGX-2 SuperPOD reference architecture. The solutions are said to deliver the fastest performance, effortless scaling, reliability and flexibility for data-intensive workloads.
In this deck, Paul Isaacs from Linaro presents: State of ARM-based HPC. This talk provides an overview of applications and infrastructure services successfully ported to Aarch64 and benefiting from scale.
"With its debut on the TOP500, the 125,000-core Astra supercomputer at New Mexico’s Sandia Labs uses Cavium ThunderX2 chips to mark Arm’s entry into the petascale world. In Japan, the Fujitsu A64FX Arm-based CPU in the pending Fugaku supercomputer has been optimized to achieve high-level, real-world application performance, anticipating up to one hundred times the application execution performance of the K computer. K was the first computer to top 10 petaflops in 2011."
Watch the video: https://wp.me/p3RLHQ-lIT
Learn more: https://www.linaro.org/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Versal Premium ACAP for Network and Cloud Accelerationinside-BigData.com
Today Xilinx announced Versal Premium, the third series in the Versal ACAP portfolio. The Versal Premium series features highly integrated, networked and power-optimized cores and the industry’s highest bandwidth and compute density on an adaptable platform. Versal Premium is designed for the highest bandwidth networks operating in thermally and spatially constrained environments, as well as for cloud providers who need scalable, adaptable application acceleration.
Versal is the industry’s first adaptive compute acceleration platform (ACAP), a revolutionary new category of heterogeneous compute devices with capabilities that far exceed those of conventional silicon architectures. Developed on TSMC’s 7-nanometer process technology, Versal Premium combines software programmability with dynamically configurable hardware acceleration and pre-engineered connectivity and security features to enable a faster time-to- market. The Versal Premium series delivers up to 3X higher throughput compared to current generation FPGAs, with built-in Ethernet, Interlaken, and cryptographic engines that enable fast and secure networks. The series doubles the compute density of currently deployed mainstream FPGAs and provides the adaptability to keep pace with increasingly diverse and evolving cloud and networking workloads.
Learn more: https://insidehpc.com/2020/03/xilinx-announces-versal-premium-acap-for-network-and-cloud-acceleration/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Zettar: Moving Massive Amounts of Data across Any Distance Efficientlyinside-BigData.com
In this video from the Rice Oil & Gas Conference, Chin Fang from Zettar presents: Moving Massive Amounts of Data across Any Distance Efficiently.
The objective of this talk is to present two on-going projects aiming at improving and ensuring highly efficient bulk transferring or streaming of massive amounts of data over digital connections across any distance. It examines the current state of the art, a few very common misconceptions, the differences among the three major type of data movement solutions, a current initiative attempting to improve the data movement efficiency from the ground up, and another multi-stage project that shows how to conduct long distance large scale data movement at speed and scale internationally. Both projects have real world motivations, e.g. the ambitious data transfer requirements of Linac Coherent Light Source II (LCLS-II) [1], a premier preparation project of the U.S. DOE Exascale Computing Initiative (ECI) [2]. Their immediate goals are described and explained, together with the solution used for each. Findings and early results are reported. Possible future works are outlined.
Watch the video: https://wp.me/p3RLHQ-lBX
Learn more: https://www.zettar.com/
and
https://rice2020oghpc.rice.edu/program-2/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from the Rice Oil & Gas Conference, Bradley McCredie from AMD presents: Scaling TCO in a Post Moore's Law Era.
"While foundries bravely drive forward to overcome the technical and economic challenges posed by scaling to 5nm and beyond, Moore’s law alone can provide only a fraction of the performance / watt and performance / dollar gains needed to satisfy the demands of today’s high performance computing and artificial intelligence applications. To close the gap, multiple strategies are required. First, new levels of innovation and design efficiency will supplement technology gains to continue to deliver meaningful improvements in SoC performance. Second, heterogenous compute architectures will create x-factor increases of performance efficiency for the most critical applications. Finally, open software frameworks, APIs, and toolsets will enable broad ecosystems of application level innovation."
Watch the video:
Learn more: http://amd.com
and
https://rice2020oghpc.rice.edu/program-2/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
In this deck from the ECSS Symposium, Abe Stern from NVIDIA presents: CUDA-Python and RAPIDS for blazing fast scientific computing.
"We will introduce Numba and RAPIDS for GPU programming in Python. Numba allows us to write just-in-time compiled CUDA code in Python, giving us easy access to the power of GPUs from a powerful high-level language. RAPIDS is a suite of tools with a Python interface for machine learning and dataframe operations. Together, Numba and RAPIDS represent a potent set of tools for rapid prototyping, development, and analysis for scientific computing. We will cover the basics of each library and go over simple examples to get users started. Finally, we will briefly highlight several other relevant libraries for GPU programming."
Watch the video: https://wp.me/p3RLHQ-lvu
Learn more: https://developer.nvidia.com/rapids
and
https://www.xsede.org/for-users/ecss/ecss-symposium
Sign up for our insideHPC Newsletter: http://insidehp.com/newsletter
In this deck from FOSDEM 2020, Colin Sauze from Aberystwyth University describes the development of a RaspberryPi cluster for teaching an introduction to HPC.
"The motivation for this was to overcome four key problems faced by new HPC users:
* The availability of a real HPC system and the effect running training courses can have on the real system, conversely the availability of spare resources on the real system can cause problems for the training course.
* A fear of using a large and expensive HPC system for the first time and worries that doing something wrong might damage the system.
* That HPC systems are very abstract systems sitting in data centres that users never see, it is difficult for them to understand exactly what it is they are using.
* That new users fail to understand resource limitations, in part because of the vast resources in modern HPC systems a lot of mistakes can be made before running out of resources. A more resource constrained system makes it easier to understand this.
The talk will also discuss some of the technical challenges in deploying an HPC environment to a Raspberry Pi and attempts to keep that environment as close to a "real" HPC as possible. The issue to trying to automate the installation process will also be covered."
Learn more: https://github.com/colinsauze/pi_cluster
and
https://fosdem.org/2020/schedule/events/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
In this deck from ATPESC 2019, Ken Raffenetti from Argonne presents an overview of HPC interconnects.
"The Argonne Training Program on Extreme-Scale Computing (ATPESC) provides intensive, two-week training on the key skills, approaches, and tools to design, implement, and execute computational science and engineering applications on current high-end computing systems and the leadership-class computing systems of the future."
Watch the video: https://wp.me/p3RLHQ-luc
Learn more: https://extremecomputingtraining.anl.gov/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
Nunit vs XUnit vs MSTest Differences Between These Unit Testing Frameworks.pdfflufftailshop
When it comes to unit testing in the .NET ecosystem, developers have a wide range of options available. Among the most popular choices are NUnit, XUnit, and MSTest. These unit testing frameworks provide essential tools and features to help ensure the quality and reliability of code. However, understanding the differences between these frameworks is crucial for selecting the most suitable one for your projects.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Trusted Execution Environment for Decentralized Process MiningLucaBarbaro3
Presentation of the paper "Trusted Execution Environment for Decentralized Process Mining" given during the CAiSE 2024 Conference in Cyprus on June 7, 2024.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
OpenHPC: Community Building Blocks for HPC Systems
1. OpenHPC: Community Building
Blocks for HPC
Karl W. Schulz, Ph.D.
Research Associate Professor
Oden Institute for Computational Engineering and Sciences and Dell Medical School
https://openhpc.community
HPC Knowledge Meeting
hpcpk'19 ⬩ June 13, 2019 ⬩ Barcelona, Spain
2. 2
Outline
• What is OpenHPC: project overview
-mission/vision
-members, governance, adoption
• More technical details
-key takeaways
-packaging conventions
-build/test environment
-highlights from latest release and what's coming
3. What is the project’s mission and vision?
Mission: to provide a reference collection of open-source HPC software
components and best practices, lowering barriers to deployment, advancement,
and use of modern HPC methods and tools.
Vision: OpenHPC components and best practices will enable and accelerate
innovation and discoveries by broadening access to state-of-the-art, open-source
HPC methods and tools in a consistent environment, supported by a
collaborative, worldwide community of HPC users, developers, researchers,
administrators, and vendors.
3
5. How is the community project governed?
5
Governance is dual-pronged with a Governing Board + Technical Steering Committee
OpenHPC
Governing
Board
OpenHPC Technical
Steering Committee (TSC)
Project
Leader
Testing
Coordinators
Component
Development Reps.
End-User / Site
Reps.
Maintainers
Technical Project
Leadership
HPC
Community
Responsible for budgetary
oversight, marketing, long-term
roadmap guidance
TSC terms are for 1 year: call for nominations put
out each summer in May - deadline is June 14!
7. OpenHPC: a building block repository
[ Key takeaway ]
● OpenHPC provides a collection of pre-built ingredients common in HPC
environments; fundamentally it is a software repository
● The repository is published for use with Linux distro
package managers:
○ yum (CentOS/RHEL)
○ zypper (SLES)
● You can pick relevant bits of interest for your site
○ if you prefer a resource manager that is not included, you can build that locally and still
leverage the scientific libraries and development environment
○ similarly, you might prefer to utilize a different provisioning system
○ public package repositories also makes it easy to include desired elements in customized
containers (e.g. Docker, Singularity, Charliecloud)
7
MPI stack(s)
Scientific Libraries
Admin Tools
8. [ Key takeaway ]
● Designation of
compiler/MPI variants
is fundamental to
OpenHPC’s build, test,
and end-user
environment
● OpenHPC packaging is
architected to allow
multiple variants to be
added over time
● example: two gcc
variants are installed
OpenHPC: A Hierarchical HPC Software Environment
9
$ module list
Currently Loaded Modules:
1) autotools 2) prun/1.2 3) gnu7/7.3.0 4) mvapich2/2.2 5) ohpc
$ module avail
--------------------------------- /opt/ohpc/pub/moduledeps/gnu7-mvapich2 -----------------------------------
adios/1.13.1 mpiP/3.4.1 phdf5/1.10.2 py3-scipy/1.1.0 superlu_dist/5.3.0
boost/1.67.0 mumps/5.1.2 pnetcdf/1.9.0 scalapack/2.0.2 tau/2.27.1
fftw/3.3.7 netcdf-cxx/4.3.0 ptscotch/6.0.4 scalasca/2.3.1 trilinos/12.12.1
hypre/2.14.0 netcdf-fortran/4.4.4 py2-mpi4py/3.0.0 scorep/4.0
imb/2018.1 netcdf/4.6.1 py2-scipy/1.1.0 sionlib/1.7.1
mfem/3.3.2 petsc/3.9.1 py3-mpi4py/3.0.0 slepc/3.9.1
------------------------------------- /opt/ohpc/pub/moduledeps/gnu7 ----------------------------------------
R/3.5.0 likwid/4.3.2 mvapich2/2.2 (L) openmpi3/3.1.0 py2-numpy/1.14.3 superlu/5.2.1
gsl/2.4 metis/5.1.0 ocr/1.0.1 pdtoolkit/3.25 py3-numpy/1.14.3
hdf5/1.10.2 mpich/3.2.1 openblas/0.2.20 plasma/2.8.0 scotch/6.0.4
--------------------------------------- /opt/ohpc/pub/modulefiles -------------------------------------------
EasyBuild/3.6.1 clustershell/1.8 gnu7/7.3.0 (L) ohpc (L) prun/1.2 (L)
autotools (L) cmake/3.11.1 hwloc/1.11.10 papi/5.6.0 singularity/2.5.1
charliecloud/0.2.4 gnu/5.4.0 llvm5/5.0.1 pmix/2.1.1 valgrind/3.13.0
[ compiler tier ]
[ MPI tier ]
9. Quick (hierarchical) installation example
● Once an OpenHPC repo is
enabled, package installation
follows standard package
manager semantics
● Package interdependencies
maintained within RPMs
● Consider example install of
PETSc for gnu7/mvapich2 on
bare bones head node
10
# yum install petsc-gnu7-mvapich2-ohpc
…
=============================================================================================
Package Arch Version Repository Size
=============================================================================================
Installing:
petsc-gnu7-mvapich2-ohpc x86_64 3.7.6-104.1 OpenHPC-updates 13 M
Installing for dependencies:
gnu7-compilers-ohpc x86_64 7.1.0-29.3 OpenHPC-updates 51 M
mvapich2-gnu7-ohpc x86_64 2.2-26.2 OpenHPC-updates 8.6 M
openblas-gnu7-ohpc x86_64 0.2.19-25.1 OpenHPC-updates 3.4 M
phdf5-gnu7-mvapich2-ohpc x86_64 1.10.0-83.1 OpenHPC-updates 2.9 M
prun-ohpc noarch 1.1-21.1 OpenHPC 9.5 k
scalapack-gnu7-mvapich2-ohpc x86_64 2.0.2-39.1 OpenHPC-updates 3.4 M
Transaction Summary
=============================================================================================
10. OpenHPC: validated recipes for system install
[ Key takeaway ]
● In addition to being a package
repository, OpenHPC provides
validated recipes for
bare-metal system installs
● Recipes organized by OS,
architecture, and key
administrative components
● 10 different recipes available
with latest release
● the docs-ohpc RPM installs
these recipes (along with shell
scripts encapsuating all
commands)
11
x86_64:
● Install_guide-CentOS7-Warewulf-PBSPro-1.3.8-x86_64.pdf
● Install_guide-CentOS7-Warewulf-SLURM-1.3.8-x86_64.pdf
● Install_guide-CentOS7-xCAT-Stateful-SLURM-1.3.8-x86_64.pdf
● Install_guide-CentOS7-xCAT-Stateless-SLURM-1.3.8-x86_64.pdf
● Install_guide-SLE_12-Warewulf-PBSPro-1.3.8-x86_64.pdf
● Install_guide-SLE_12-Warewulf-SLURM-1.3.8-x86_64.pdf
aarch64:
● Install_guide-CentOS7-Warewulf-PBSPro-1.3.8-aarch64.pdf
● Install_guide-CentOS7-Warewulf-SLURM-1.3.8-aarch64.pdf
● Install_guide-SLE_12-Warewulf-PBSPro-1.3.8-aarch64.pdf
● Install_guide-SLE_12-Warewulf-SLURM-1.3.8-aarch64.pdf
PBS Pro SLURMWarewulf xCAT
excellent place to start if you are new to the project
CentOS 7 SLES12
15. Packaging Conventions
● OpenHPC tries to play nicely with underlying OS’es and other repositories
● We endeavor to avoid namespace conflict in a variety of ways:
○ install paths: end-user oriented components housed under /opt/ohpc/pub (also allow for
multiple versions to coexist)
○ package names: RPMs provided via OpenHPC repos include “-ohpc” as the suffix for the
package name, for example:
○ dependencies: compiler/MPI variant dependencies are explicitly managed
lmod-ohpc-7.4.8-11.1.aarch64.rpm
%if "%{compiler_family}" == "gnu7"
BuildRequires: gnu7-compilers%{PROJ_DELIM} >= 7.2.0
Requires: gnu7-compilers%{PROJ_DELIM} >= 7.2.0
%endif
%if "%{mpi_family}" == "mvapich2"
BuildRequires: mvapich2-%{compiler_family}%{PROJ_DELIM}
Requires: mvapich2-%{compiler_family}%{PROJ_DELIM}
%endif
[ relevant logic from OHPC_macros ]
16
16. Packaging Conventions
● Packaging adopts a consistent environment variable schema in modulefiles
(we rely on Lmod implementation of modules):
○ ${PACKAGE_DIR} - top level install path of package
○ ${PACKAGE_BIN} - path to binaries supplied by package
○ ${PACKAGE_INC} - path to include headers for package
○ ${PACKAGE_LIB} - path to dynamic libraries for package
● Recommend using these vars in build scripts, Makefiles, and interactive
execution
● Let’s consider an end user interaction example with previous PETSc install:
assume we are a user wanting to build a PETSC hello world in C
17
$ module load gnu8 mvapich2
$ module load petsc
$ mpicc -I$PETSC_INC petsc_hello.c -L$PETSC_LIB –lpetsc
17. Packaging conventions: custom ohpc RPM plugin
To avoid namespace collision for resolving dynamic libraries (.so's), we apply an "ohpc" color delimiter
to libraries installed in /opt/ohpc/ path
18
# rpm -q --requires gnu7-compilers-ohpc | egrep “libc.so|libgc”
libgcc_s.so.1()(64bit)(ohpc)
libgcc_s.so.1(GCC_3.0)(64bit)(ohpc)
libgcc_s.so.1(GCC_3.3)(64bit)(ohpc)
libgcc_s.so.1(GCC_4.2.0)(64bit)(ohpc)
libgcc_s.so.1(GCC_4.3.0)(64bit)(ohpc)
libc.so.6()(64bit)
libc.so.6(GLIBC_2.10)(64bit)
libc.so.6(GLIBC_2.11)(64bit)
libc.so.6(GLIBC_2.14)(64bit)
libc.so.6(GLIBC_2.16)(64bit)
libc.so.6(GLIBC_2.17)(64bit)
libc.so.6(GLIBC_2.2.5)(64bit)
libc.so.6(GLIBC_2.3)(64bit)
libc.so.6(GLIBC_2.3.2)(64bit)
libc.so.6(GLIBC_2.3.3)(64bit)
libc.so.6(GLIBC_2.6)(64bit)
libc.so.6(GLIBC_2.7)(64bit)
.so’s self contained
within ohpc gcc build
.so’s required from base
OS supplied packages
Convention introduced
in v1.3.4 release
custom RPM
dependency analysis
using ohpc plugin
18. OpenHPC Development Infrastructure
What are we using to get the job done....?
The usual software engineering stuff:
● GitHub (SCM and issue tracking/planning)
● Continuous Integration (CI) Testing (Jenkins)
● Documentation (Latex)
● Bats (unit testing for Bash)
Capable build/packaging system:
● Require flexible system to manage builds for multiple distros, multiple
compiler/MPI family combinations, and dependencies across packages
● Have engineered a system using Open Build Service (OBS) which is
supported by back-end git
○ git houses .spec files, tarballs, patches, documentation recipes, and
integration tests
○ OBS performs automated builds and dependency analysis
19
https://github.com/openhpc/ohpc
https://build.openhpc.community
19. Reproducible builds using OBS
● We use the Open Build Service
(OBS) to manage build
processes
● OBS can drive builds for
multiple OS’s and architectures
● Repeatable builds carried out
in chroot environment
● Generates binary and src rpms
● Client/server architecture
supports distributed build
slaves and multiple
architectures
● OpenHPC adopts a single
.spec file input for builds
● Dependencies are self-
managed and designed to
avoid namespace conflict with
distro provided variants
https://build.openhpc.community
parent
package
child package(s)
<link project='OpenHPC:1.3:Update6:Factory'
package='boost-gnu-openmpi'>
<patches>
<topadd>%define compiler_family gnu8</topadd>
<topadd>%define mpi_family mpich</topadd>
</patches>
</link>
[ OBS _link file ]
20. 21
Build system - distributed build servers
• We leverage ability in OBS
to have distributed build
servers:
- x86_64: hosted at TACC
(thanks to kind hardware
donations) and on EC2
- aarch64: hosted at Packet
(thanks to Works on Arm
project)
21. Integration/Test/Validation
Post-build testing is another key element for us and the intent is to build upon existing component-level
validation with targeted cluster-validation and scaling initiatives including:
• install recipes
• cross-package interaction
• development environment
• mimic use cases common in HPC deployments
• upgrade mechanism
OS Distribution
Hardware +
Integrated Cluster Testing
+
Dev
Tools
Parallel
Libs
Perf.
Tools
User Env
Mini
Apps
Serial
Libs
System
Tools
Resource
Manager
Provisioner
I/O Libs
Compilers
Software
OpenHPC
Individual Component Validation
22. Post Install Integration Tests - Overview
23
Package version............... : test-suite-1.0.0
Build user.................... : jilluser
Build host.................... : master4-centos71.localdomain
Configure date................ : 2015-10-26 09:23
Build architecture............ : x86_64-unknown-linux-gnu
Test suite configuration...... : long
Submodule Configuration:
User Environment:
RMS test harness.......... : enabled
Munge..................... : enabled
Apps...................... : enabled
Compilers................. : enabled
MPI....................... : enabled
HSN....................... : enabled
Modules................... : enabled
OOM....................... : enabled
Dev Tools:
Valgrind.................. : enabled
R base package............ : enabled
TBB....................... : enabled
CILK...................... : enabled
Performance Tools:
mpiP Profiler........ .... : enabled
Papi...................... : enabled
PETSc..................... : enabled
TAU....................... : enabled
Example ./configure output (non-root)
Global testing harness
includes a number of
embedded subcomponents:
• major components have
configuration options to
enable/disable
• end user tests need to touch all
of the supported compiler and
MPI families
• we abstract this to repeat the
tests with different
compiler/MPI environments:
• gcc/Intel compiler toolchains
• Intel, OpenMPI, MPICH,
MVAPICH2 MPI families
Libraries:
Adios .................... : enabled
Boost .................... : enabled
Boost MPI................. : enabled
FFTW...................... : enabled
GSL....................... : enabled
HDF5...................... : enabled
HYPRE..................... : enabled
IMB....................... : enabled
Metis..................... : enabled
MUMPS..................... : enabled
NetCDF.................... : enabled
Numpy..................... : enabled
OPENBLAS.................. : enabled
PETSc..................... : enabled
PHDF5..................... : enabled
ScaLAPACK................. : enabled
Scipy..................... : enabled
Superlu................... : enabled
Superlu_dist.............. : enabled
Trilinos ................. : enabled
Apps:
MiniFE.................... : enabled
MiniDFT................... : enabled
HPCG...................... : enabled
PRK....................... : enabledinstallable as test-suite-ohpc RPM
23. 24
Community Jenkins Instance for CI Testing
http://test.openhpc.community:8080
All recipes exercised in CI system (start w/ bare-metal installs + integration test suite)
OpenHPC CI Infrastructure
Thanks to the Texas Advanced Computing Center (TACC) for hosting support and to Intel, Cavium, and Dell for hardware donations.
add description
S W Name ↓ Last Success Last Failure Last Duration
(1.2) - (centos7.2,x86_64) - (warewulf+pbspro) - long cycle 1 day 8 hr - #39 N/A 58 min
(1.2) - (centos7.2,x86_64) - (warewulf+pbspro) - short cycle 1 day 4 hr - #155 4 days 10 hr - #109 14 min
(1.2) - (centos7.2,x86_64) - (warewulf+slurm) - long cycle 2 days 4 hr - #244 4 days 4 hr - #219 1 hr 0 min
(1.2) - (centos7.2,x86_64) - (warewulf+slurm) - short cycle 2 hr 46 min - #554 8 days 15 hr - #349 14 min
(1.2) - (centos7.2,x86_64) - (warewulf+slurm+PXSE) - long cycle 1 day 6 hr - #39 4 days 10 hr - #20 2 hr 29 min
(1.2) - (sles12sp1,x86_64) - (warewulf+pbspro) - short cycle 1 day 3 hr - #166 4 days 10 hr - #86 12 min
(1.2) - (sles12sp1,x86_64) - (warewulf+slurm) - short cycle 1 day 2 hr - #259 8 days 20 hr - #72 14 min
(1.2) - (sles12sp1,x86_64) - (warewulf,slurm) - long test cycle 1 day 5 hr - #97 6 days 19 hr - #41 54 min
(1.2) - aarch64 - (centos7.2) - (warewulf+slurm) 2 days 21 hr - #3 N/A 0.41 sec
(1.2) - aarch64 - (sles12sp1) - (warewulf+slurm) 1 day 8 hr - #45 2 days 21 hr - #41 2 hr 13 min
Icon: S M L Legend RSS for all RSS for failures RSS for just latest builds
KARL W. SCHULZ | LOG OUTsearch
1.1.1 All Interactive admin +1.2
ENABLE AUTO REFRESH
New Item
People
Build History
Edit View
Project Relationship
Check File Fingerprint
Manage Jenkins
My Views
Lockable Resources
Credentials
No builds in the queue.
master
1 Idle
2 Idle
3 Idle
OpenHPC CI Infrastructure
Thanks to the Texas Advanced Computing Center (TACC) for hosting support and to Intel, Cavium, and Dell for hardware donations.
S W Name ↓ Last Success Last Failure Last Dura
(1.2) - (centos7.2,x86_64) - (warewulf+pbspro) - long cycle 1 day 8 hr - #39 N/A 58 min
(1.2) - (centos7.2,x86_64) - (warewulf+pbspro) - short cycle 1 day 4 hr - #155 4 days 10 hr - #109 14 min
(1.2) - (centos7.2,x86_64) - (warewulf+slurm) - long cycle 2 days 4 hr - #244 4 days 4 hr - #219 1 hr 0 mi
(1.2) - (centos7.2,x86_64) - (warewulf+slurm) - short cycle 2 hr 46 min - #554 8 days 15 hr - #349 14 min
(1.2) - (centos7.2,x86_64) - (warewulf+slurm+PXSE) - long cycle 1 day 6 hr - #39 4 days 10 hr - #20 2 hr 29 m
(1.2) - (sles12sp1,x86_64) - (warewulf+pbspro) - short cycle 1 day 3 hr - #166 4 days 10 hr - #86 12 min
(1.2) - (sles12sp1,x86_64) - (warewulf+slurm) - short cycle 1 day 2 hr - #259 8 days 20 hr - #72 14 min
(1.2) - (sles12sp1,x86_64) - (warewulf,slurm) - long test cycle 1 day 5 hr - #97 6 days 19 hr - #41 54 min
(1.2) - aarch64 - (centos7.2) - (warewulf+slurm) 2 days 21 hr - #3 N/A 0.41 sec
(1.2) - aarch64 - (sles12sp1) - (warewulf+slurm) 1 day 8 hr - #45 2 days 21 hr - #41 2 hr 13 m
Icon: S M L Legend RSS for all RSS for failures RSS for just
KARL W. SCHULsearch
Build Queue
Build Executor Status
1.1.1 All Interactive admin +1.2
Jenkins EN
our CI"build" is
really a cluster
25. Additions and Upstream Version Changes
26
• Part of the motivation for community effort like OpenHPC is the rapidity of S/W updates
that occurs in our space
• We have been doing releases on a roughly quarterly cadence:
- convention is to go with latest stable release of upstream components
- additional components added over time
v1.0
v1.0.1
v1.1
v1.1.1
v1.2
v1.2.1
v1.3
v1.3.1
v1.3.2
v1.3.3
v1.3.4
v1.3.5
v1.3.6
v1.3.7
v1.3.8
40
50
60
70
80
90
100
Jul-15 Jan-16 Aug-16 Mar-17 Sep-17 Apr-18 Oct-18 May-19 Dec-19
#ofComponetsIncluded
1.8%
33.9%
8.1%
51.6%
4.5%
25.8%
40.9%
31.9%
26.7%
41.6%
43.8%
40.7%
43.2%
36.4%
0%
10%
20%
30%
40%
50%
60%
v1.0.1 v1.1 v1.1.1 v1.2 v1.2.1 v1.3 v1.3.1 v1.3.2 v1.3.3 v1.3.4 v1.3.5 v1.3.6 v1.3.7 v1.3.8
%ComponetsUpdated
26. OpenHPC v1.3.8 - S/W components
27
● 3rd
Party libraries
are built for each
compiler/MPI
family
● Resulting
repositories
currently
comprised of
~700 RPMs
new since v1.3.1
Released June 11, 2019
Functional Areas Components
Base OS CentOS 7.6, SLES12 SP4
Architecture aarch64, x86_64
Administrative Tools
Conman, Ganglia, Lmod, LosF, Nagios, NHC, pdsh, pdsh-mod-slurm, prun, EasyBuild, ClusterShell,
mrsh, Genders, Shine, Spack, test-suite
Provisioning Warewulf, xCAT
Resource Mgmt. SLURM, Munge, PBS Professional, PMIx
Runtimes Charliecloud, OpenMP, OCR, Singularity
I/O Services Lustre client, BeeGFS client*
Numerical/Scientific
Libraries
Boost, GSL, FFTW, Hypre, Metis, MFEM, Mumps, OpenBLAS, OpenCoarrays, PETSc, PLASMA,
Scalapack, Scotch, SLEPc, SuperLU, SuperLU_Dist, Trilinos
I/O Libraries HDF5 (pHDF5), NetCDF/pNetCDF (including C++ and Fortran interfaces), Adios
Compiler Families GNU (gcc, g++, gfortran), Clang/LLVM, Intel Parallel Studio*
MPI Families MVAPICH2, OpenMPI, MPICH, Intel MPI*
Development Tools Autotools, cmake, hwloc, mpi4py, R, SciPy/NumPy, Valgrind
Performance Tools
PAPI, IMB, Likwid, mpiP, pdtoolkit TAU, Scalasca, ScoreP, SIONLib, GeoPM, msr-safe, Dimemas,
Extrae, Paraver, OSU Microbenchmarks
new in v1.3.8 release
27. How to Request Additional Software?
● We have a simple submission site for new
requests:
○ https://github.com/openhpc/submissions
● Example components added via this
mechanism since the v1.2. release (Nov’ 16)
○ BeeGFS client
○ xCAT recipe
○ hwloc
○ Singularity
○ LLVM/clang
○ PLASMA
28
○ pNetCDF
○ SCOTCH
○ SLEPc
○ PMIx
○ MPI4py
○ Likwid
○ MFEM
○ NHC
○ Charliecloud
○ GeoPM
○ Dimemas/Extrae,
Paraver
○ OpenCoarrays
○ OSU Benchmarks
28. What's coming for next big release?
29
● Will be migrating from 1.3.x series to 2.0 release (targeting SC'19)
● Target major new distro versions:
○ CentOS 8
○ SUSE Leap 15.1 (switching to Leap from SLES)
● Once we release a new major branch (e.g. v2.x) which supports a newer major
distro version at time t, we will restrict updates to previous branch (e.g. v1.3.x) to
include:
○ security patches (e.g. address known CVEs)
○ significant bugs affecting functionality
○ older branch stays in maintenance mode for ∆𝒕 = 1 year
1.3.x
2.x
CentOS7, SLE12
CentOS8, Leap 15.1
! + ∆!!
29. 30
Summary
• Provided an overview of the Linux Foundation project along with key takeaways
and packaging details
• Community project efforts provides package repository for use with native OS
package managers
- for use on bare-metal or cloud/containerized environments
- significant companion test effort with documentation recipes for deployment and
execution in HPC- runtime environment
- RHEL8/CentOS8 and SUSE Leap 15.1 coming later this year
30. Thanks for your time….
● OpenHPC Home: http://www.openhpc.community/
● Primary GitHub Site: https://github.com/openhpc/ohpc
● Package Repositories: http://build.openhpc.community/OpenHPC:/
● Component Submission: https://github.com/openhpc/submissions
● System Registry: System Registration Form
● CI Infrastructure: http://test.openhpc.community:8080
● OpenHPC Wiki: https://github.com/openhpc/ohpc/wiki
● Mailing Lists: http://www.openhpc.community/support/mail-lists/
○ openhpc-announce
○ openhpc-users
○ openhpc-devel