This document provides an overview of high performance computing technologies including cluster management, workload management, storage, networking, and GPU computing. It discusses Bright Cluster Manager for unified cluster management, Moab HPC Suite for job scheduling, Intel Cluster Ready certification, Panasas storage, InfiniBand networking, NVIDIA GPUs, and other parallel computing technologies. The document is an introduction to key HPC solutions from transtec, a company with over 30 years of experience in scientific computing.
The HP sx2000 chipset is Hewlett-Packard's new enterprise systems chipset designed to provide scalability, reliability, manageability and performance to meet demanding server needs. The chipset supports the new dual-core Intel Itanium 2 and Intel Itanium 2 9M processors, and allows HP Itanium 2 processors and PA-8900 RISC processors to concurrently operate in HP Integrity Superdome partitions. This provides flexibility and maximizes return on investment. The sx2000 chipset delivers key benefits for HP Integrity and HP 9000 servers in areas like performance, price/performance, reliability and availability, virtualization and management capabilities, and investment protection.
This document provides an overview of Data ONTAP storage management. It describes Data ONTAP storage architecture, how disks and aggregates work, RAID protection levels, and managing volumes, snapshots, clones, and other storage objects. It is a technical guide intended for storage administrators on configuring and maintaining Data ONTAP storage.
The document provides information about:
1) Configuring and managing the AppServer runtime environment including components, operation modes, and fault tolerance.
2) Using the OpenEdge administration framework including OpenEdge Management, NameServers, and agents to configure and start AppServer instances.
3) Setting up the environment for AppServer execution including environment variables, work directories, and log files.
The document discusses Juniper Networks' ISSU (In-Service Software Upgrade) tool. It describes the key design principles for ISSU, including maintaining control plane and forwarding plane functionality during upgrades. It then provides an overview of Juniper's Unified ISSU, which allows complete operating system upgrades with no control plane disruption and minimal traffic interruption. Unified ISSU was designed according to best practices and delivers upgrades in the shortest time with minimal risk.
This document provides instructions on installing, getting started with, and configuring ManageEngine ServiceDesk Plus. It outlines the system requirements, describes how to install the software on Windows and Linux, and explains how to start and access the application. It also provides an overview of the key modules in ServiceDesk Plus for request management, asset management, purchasing, contracts, and more. Finally, it points the user to the Configurations section for setting up the application settings.
Aa plc-xrassa coal handling and preparation plant design instructionSerena Fu
This document provides a design instruction for the Xrassa Coal Handling and Preparation Plant project. It includes:
- An overview of the project including the coal processing capacity of 14 million tons annually to supply the nearby 2340MW Xrassa Super Thermal Power Station.
- Details of the coal washing process which will use heavy media vessels, heavy media cyclones, spiral separators, and other equipment to process raw coal from nearby mines into a washed mixed coal product with ash content ≤30% for power generation.
- Information on the plant location, coal source and quality, equipment selection, product structure, and other key design aspects and specifications for the project.
The design aims to utilize proven
The document provides an overview and administrator guide for OnTime 2008 project management software. It discusses the OnTime product line including OnTime for Windows, Web Edition, VS.NET, Customer Portal, and Remote Server. It also covers installing and setting up OnTime, including planning deployment, installing OnTime and its components, setting up the OnTime database, configuring system options and email notifications.
The HP sx2000 chipset is Hewlett-Packard's new enterprise systems chipset designed to provide scalability, reliability, manageability and performance to meet demanding server needs. The chipset supports the new dual-core Intel Itanium 2 and Intel Itanium 2 9M processors, and allows HP Itanium 2 processors and PA-8900 RISC processors to concurrently operate in HP Integrity Superdome partitions. This provides flexibility and maximizes return on investment. The sx2000 chipset delivers key benefits for HP Integrity and HP 9000 servers in areas like performance, price/performance, reliability and availability, virtualization and management capabilities, and investment protection.
This document provides an overview of Data ONTAP storage management. It describes Data ONTAP storage architecture, how disks and aggregates work, RAID protection levels, and managing volumes, snapshots, clones, and other storage objects. It is a technical guide intended for storage administrators on configuring and maintaining Data ONTAP storage.
The document provides information about:
1) Configuring and managing the AppServer runtime environment including components, operation modes, and fault tolerance.
2) Using the OpenEdge administration framework including OpenEdge Management, NameServers, and agents to configure and start AppServer instances.
3) Setting up the environment for AppServer execution including environment variables, work directories, and log files.
The document discusses Juniper Networks' ISSU (In-Service Software Upgrade) tool. It describes the key design principles for ISSU, including maintaining control plane and forwarding plane functionality during upgrades. It then provides an overview of Juniper's Unified ISSU, which allows complete operating system upgrades with no control plane disruption and minimal traffic interruption. Unified ISSU was designed according to best practices and delivers upgrades in the shortest time with minimal risk.
This document provides instructions on installing, getting started with, and configuring ManageEngine ServiceDesk Plus. It outlines the system requirements, describes how to install the software on Windows and Linux, and explains how to start and access the application. It also provides an overview of the key modules in ServiceDesk Plus for request management, asset management, purchasing, contracts, and more. Finally, it points the user to the Configurations section for setting up the application settings.
Aa plc-xrassa coal handling and preparation plant design instructionSerena Fu
This document provides a design instruction for the Xrassa Coal Handling and Preparation Plant project. It includes:
- An overview of the project including the coal processing capacity of 14 million tons annually to supply the nearby 2340MW Xrassa Super Thermal Power Station.
- Details of the coal washing process which will use heavy media vessels, heavy media cyclones, spiral separators, and other equipment to process raw coal from nearby mines into a washed mixed coal product with ash content ≤30% for power generation.
- Information on the plant location, coal source and quality, equipment selection, product structure, and other key design aspects and specifications for the project.
The design aims to utilize proven
The document provides an overview and administrator guide for OnTime 2008 project management software. It discusses the OnTime product line including OnTime for Windows, Web Edition, VS.NET, Customer Portal, and Remote Server. It also covers installing and setting up OnTime, including planning deployment, installing OnTime and its components, setting up the OnTime database, configuring system options and email notifications.
Advent Net Manage Engine Service Desk Plus Help Admin Guideguestf80501
The document provides an overview of the key modules and functionality of ManageEngine ServiceDesk Plus. It describes the main modules for managing requests, problems, changes, assets, purchases, and contracts. The request module functions as the help desk where technicians resolve user issues. Problem management seeks to identify the root cause of incidents to prevent recurrences. Change management ensures standardized processes for handling changes. Additional modules allow for asset, purchase, and contract management.
The document provides an overview of the key modules and functionality of ManageEngine ServiceDesk Plus. It describes the main modules for managing requests, problems, changes, assets, purchases, and contracts. The request module functions as the help desk where technicians resolve user issues. Problem management seeks to identify the root cause of incidents to prevent recurrences. Change management ensures standardized processes for handling changes to minimize disruption.
This document provides guidelines for achieving the 100% Pure Java standard. It discusses the purpose of the standard which is to ensure portability and common behavior across Java platforms. The document outlines rules for developing pure Java programs, including avoiding the use of native code, file paths, or platform-specific behavior. It also provides hints and explanations of potential portability problems to help developers create fully portable Java applications.
This document discusses considerations for business managers regarding the total cost of ownership of SOA gateways. It covers factors like cost of implementation, which can be impacted by a gateway's deployability across hardware, software, and virtual form factors, as well as its extensibility through SDKs and standards support. The cost of ongoing operation is also discussed, including manageability, scalability, reliability, and costs associated with updating and upgrading gateways over time. Layer 7 Technologies is highlighted as offering multiple deployment options and a focus on avoiding vendor lock-in.
This document is a book about expert Oracle database architecture for versions 9i and 10g. It discusses programming techniques and solutions. The book covers topics such as developing successful Oracle applications, database architecture overview, database files, memory structures, SQL and PL/SQL, concurrency and locking, performance, backup and recovery, and more. It aims to help readers understand Oracle architecture and how to write efficient and optimized database applications.
This document outlines the machine requirements, prerequisites, and expected outcomes for courses on Linux fundamentals, system administration, networking, shell scripting, and internals.
For the fundamentals course, a Pentium 2 500MHz computer with 32MB RAM is recommended. The system administration course requires completion of fundamentals first. After completing fundamentals and system administration, the equivalent of a junior administrator would be achieved. Additional courses build on this knowledge towards senior roles.
This document is the user manual for SQL Developer version 2.2.0. It provides information on getting started with SQL Developer, including license registration and configuring database drivers. It describes the main features of the SQL Developer desktop interface, including the main menu, tool bar, window areas, and output window. It provides details on the database navigator, connection dialog, SQL editors, bookmarks, diagram editor, database info, settings, and extensions available in SQL Developer.
This technical report describes a virtualized Microsoft Exchange, SQL Server, and SharePoint solution using VMware vSphere and NetApp storage. It evaluates three storage protocols: Fibre Channel (FC), iSCSI, and NFS. For the FC solution, the report details the hardware requirements, including NetApp FAS storage arrays, disk shelves, Ethernet and FC networking components. It then describes the FC storage architecture and layout, including aggregates, volumes, and VM datastores. Performance tests validate the storage efficiency, workload performance, and backup/restore capabilities of the FC-based virtualized Microsoft application environment.
Aaxeon Technologies is changing its name to Antaira to reflect its expansion into new industries and geographic regions. The company will continue operating as normal and customers will see no changes to service, support, or warranties. The president expresses appreciation for customers and looks forward to continued growth together under the new name, Antaira.
Ammonia Plant Selection Sizing and Troubleshooting Karl Kolmetz
Ammonia is produced basically from water, air, and energy. The energy source is
usually hydrocarbons, thus providing hydrogen as well, but may also be coal or
electricity. Steam reforming of light hydrocarbons is the most efficient route, with
about 77% of world ammonia capacity being based on natural gas.
The total energy consumption for the production of ammonia in a modern steam
reforming plant is 40-50% above the thermodynamic minimum. More than half of the
excess consumption is due to compression losses. The practical minimum
consumption is assumed to be 130% of the theoretical minimum.
This guideline covers items in making an ammonia plant. Starting from the raw
material, equipment, manufacturing process and the economics of ammonia plant.
This document provides an introduction to using Actix Analyzer software for analyzing GSM network performance. It covers loading and viewing drive test and other radio network data, performing queries and filters on the data, configuring cell sites and networks, and generating reports. Key features discussed include mapping cells and drive test data, binning and aggregating data, exploring data on charts and tables, and using preconfigured applications and reports for common analysis tasks.
This EMC Engineering TechBook provides insight and understanding of some options available for managing your data center connectivity, including information on some new software management tools developed to bridge the gap in the I/O consolidation environment.
Product description vital qip next generation v7 2_en_feb09(1)Roy Muy Golfo
VitalQIP is a leading IP management software that provides centralized management of IP addresses, DNS, DHCP, and ENUM. It offers automated IP address assignment, accurate network inventory, flexible user interfaces, redundancy of DNS and DHCP servers, and integration with existing DNS and DHCP environments. The document provides an overview of VitalQIP's capabilities, including its architecture with enterprise server, database, and client components. It describes the user interfaces of the core product, such as the new next-generation web GUI, command line interface, and web services API. The document also outlines many features of VitalQIP related to DNS, DHCP, reporting, user permissions, and high availability.
LoadRunner enables you to test your system under controlled and peak load conditions. To generate load, LoadRunner runs thousands of Virtual Users that are distributed over a network.
This document provides an overview and implementation guide for Juniper's Loop-Free Alternate feature, which provides sub-50ms convergence for OSPF and IS-IS networks. It describes the problem of slow convergence in IGP networks and introduces LFA as a solution. The document outlines the operational theory and implementation of LFA, including configuration steps and verification procedures. It also includes examples and details on how LFA provides backup coverage and fast rerouting in the event of link failures.
This user guide provides information about managing and configuring Zend Server Community Edition 4.0.3. It describes the administration interface layout and tabs for monitoring server information and setup. It also provides tasks for working with components, extensions, logs, and other features. Configuration details and API references are included for components like the debugger, optimizer, guard loader, and others. Best practices are also covered for performance, security, development, and deployment.
This document provides guidelines for optimizing accessibility in Ericsson networks. It discusses key performance indicators (KPIs) for measuring accessibility, including call setup success rate and overall service accessibility. It also analyzes factors that can affect accessibility, such as admission control, processor load, and issues after call admission like congestion. Annexes describe user equipment idle mode procedures and call establishment procedures in detail.
The document is an operating manual for the LS-615 Laser System produced by Quantum Composers, Inc. It contains instructions for setup, operation, maintenance and safety procedures. The laser system uses a nitrogen purge and includes components such as a laser head, attenuator assembly, harmonic mixer and more. The system can be operated manually or through an LS-615 Communication Application for laser control and autofocus functions.
This document provides instructions for installing JasperServer Community Edition (CE) release 3.7. It discusses both installing from the installer distribution, which bundles the application server and database, as well as installing from the WAR file distribution, which allows for more flexibility in choosing the application server and database. The document covers prerequisites, installation steps for the bundled installer, starting and stopping JasperServer after installation, and uninstalling JasperServer. It also provides detailed guidance for installing from the WAR file distribution, including configuring build scripts and properties files, deploying to an application server, and setting Java options.
Configuring a highly available Microsoft Lync Server 2013 environment on Dell...Principled Technologies
As this guide has shown, setting up a highly available Microsoft Lync Server 2013 environment on Dell architecture is a straightforward process. In little time, you can deploy two Dell PowerEdge M620 M-series servers, switches, and Dell EqualLogic storage using Microsoft Windows Server 2012, and set up your Lync Server 2013 infrastructure. By setting up a highly available Lync Server 2013 environment on your Dell architecture solution, you can ensure your office communications stay running with little to no downtime to keep your business moving.
Advent Net Manage Engine Service Desk Plus Help Admin Guideguestf80501
The document provides an overview of the key modules and functionality of ManageEngine ServiceDesk Plus. It describes the main modules for managing requests, problems, changes, assets, purchases, and contracts. The request module functions as the help desk where technicians resolve user issues. Problem management seeks to identify the root cause of incidents to prevent recurrences. Change management ensures standardized processes for handling changes. Additional modules allow for asset, purchase, and contract management.
The document provides an overview of the key modules and functionality of ManageEngine ServiceDesk Plus. It describes the main modules for managing requests, problems, changes, assets, purchases, and contracts. The request module functions as the help desk where technicians resolve user issues. Problem management seeks to identify the root cause of incidents to prevent recurrences. Change management ensures standardized processes for handling changes to minimize disruption.
This document provides guidelines for achieving the 100% Pure Java standard. It discusses the purpose of the standard which is to ensure portability and common behavior across Java platforms. The document outlines rules for developing pure Java programs, including avoiding the use of native code, file paths, or platform-specific behavior. It also provides hints and explanations of potential portability problems to help developers create fully portable Java applications.
This document discusses considerations for business managers regarding the total cost of ownership of SOA gateways. It covers factors like cost of implementation, which can be impacted by a gateway's deployability across hardware, software, and virtual form factors, as well as its extensibility through SDKs and standards support. The cost of ongoing operation is also discussed, including manageability, scalability, reliability, and costs associated with updating and upgrading gateways over time. Layer 7 Technologies is highlighted as offering multiple deployment options and a focus on avoiding vendor lock-in.
This document is a book about expert Oracle database architecture for versions 9i and 10g. It discusses programming techniques and solutions. The book covers topics such as developing successful Oracle applications, database architecture overview, database files, memory structures, SQL and PL/SQL, concurrency and locking, performance, backup and recovery, and more. It aims to help readers understand Oracle architecture and how to write efficient and optimized database applications.
This document outlines the machine requirements, prerequisites, and expected outcomes for courses on Linux fundamentals, system administration, networking, shell scripting, and internals.
For the fundamentals course, a Pentium 2 500MHz computer with 32MB RAM is recommended. The system administration course requires completion of fundamentals first. After completing fundamentals and system administration, the equivalent of a junior administrator would be achieved. Additional courses build on this knowledge towards senior roles.
This document is the user manual for SQL Developer version 2.2.0. It provides information on getting started with SQL Developer, including license registration and configuring database drivers. It describes the main features of the SQL Developer desktop interface, including the main menu, tool bar, window areas, and output window. It provides details on the database navigator, connection dialog, SQL editors, bookmarks, diagram editor, database info, settings, and extensions available in SQL Developer.
This technical report describes a virtualized Microsoft Exchange, SQL Server, and SharePoint solution using VMware vSphere and NetApp storage. It evaluates three storage protocols: Fibre Channel (FC), iSCSI, and NFS. For the FC solution, the report details the hardware requirements, including NetApp FAS storage arrays, disk shelves, Ethernet and FC networking components. It then describes the FC storage architecture and layout, including aggregates, volumes, and VM datastores. Performance tests validate the storage efficiency, workload performance, and backup/restore capabilities of the FC-based virtualized Microsoft application environment.
Aaxeon Technologies is changing its name to Antaira to reflect its expansion into new industries and geographic regions. The company will continue operating as normal and customers will see no changes to service, support, or warranties. The president expresses appreciation for customers and looks forward to continued growth together under the new name, Antaira.
Ammonia Plant Selection Sizing and Troubleshooting Karl Kolmetz
Ammonia is produced basically from water, air, and energy. The energy source is
usually hydrocarbons, thus providing hydrogen as well, but may also be coal or
electricity. Steam reforming of light hydrocarbons is the most efficient route, with
about 77% of world ammonia capacity being based on natural gas.
The total energy consumption for the production of ammonia in a modern steam
reforming plant is 40-50% above the thermodynamic minimum. More than half of the
excess consumption is due to compression losses. The practical minimum
consumption is assumed to be 130% of the theoretical minimum.
This guideline covers items in making an ammonia plant. Starting from the raw
material, equipment, manufacturing process and the economics of ammonia plant.
This document provides an introduction to using Actix Analyzer software for analyzing GSM network performance. It covers loading and viewing drive test and other radio network data, performing queries and filters on the data, configuring cell sites and networks, and generating reports. Key features discussed include mapping cells and drive test data, binning and aggregating data, exploring data on charts and tables, and using preconfigured applications and reports for common analysis tasks.
This EMC Engineering TechBook provides insight and understanding of some options available for managing your data center connectivity, including information on some new software management tools developed to bridge the gap in the I/O consolidation environment.
Product description vital qip next generation v7 2_en_feb09(1)Roy Muy Golfo
VitalQIP is a leading IP management software that provides centralized management of IP addresses, DNS, DHCP, and ENUM. It offers automated IP address assignment, accurate network inventory, flexible user interfaces, redundancy of DNS and DHCP servers, and integration with existing DNS and DHCP environments. The document provides an overview of VitalQIP's capabilities, including its architecture with enterprise server, database, and client components. It describes the user interfaces of the core product, such as the new next-generation web GUI, command line interface, and web services API. The document also outlines many features of VitalQIP related to DNS, DHCP, reporting, user permissions, and high availability.
LoadRunner enables you to test your system under controlled and peak load conditions. To generate load, LoadRunner runs thousands of Virtual Users that are distributed over a network.
This document provides an overview and implementation guide for Juniper's Loop-Free Alternate feature, which provides sub-50ms convergence for OSPF and IS-IS networks. It describes the problem of slow convergence in IGP networks and introduces LFA as a solution. The document outlines the operational theory and implementation of LFA, including configuration steps and verification procedures. It also includes examples and details on how LFA provides backup coverage and fast rerouting in the event of link failures.
This user guide provides information about managing and configuring Zend Server Community Edition 4.0.3. It describes the administration interface layout and tabs for monitoring server information and setup. It also provides tasks for working with components, extensions, logs, and other features. Configuration details and API references are included for components like the debugger, optimizer, guard loader, and others. Best practices are also covered for performance, security, development, and deployment.
This document provides guidelines for optimizing accessibility in Ericsson networks. It discusses key performance indicators (KPIs) for measuring accessibility, including call setup success rate and overall service accessibility. It also analyzes factors that can affect accessibility, such as admission control, processor load, and issues after call admission like congestion. Annexes describe user equipment idle mode procedures and call establishment procedures in detail.
The document is an operating manual for the LS-615 Laser System produced by Quantum Composers, Inc. It contains instructions for setup, operation, maintenance and safety procedures. The laser system uses a nitrogen purge and includes components such as a laser head, attenuator assembly, harmonic mixer and more. The system can be operated manually or through an LS-615 Communication Application for laser control and autofocus functions.
This document provides instructions for installing JasperServer Community Edition (CE) release 3.7. It discusses both installing from the installer distribution, which bundles the application server and database, as well as installing from the WAR file distribution, which allows for more flexibility in choosing the application server and database. The document covers prerequisites, installation steps for the bundled installer, starting and stopping JasperServer after installation, and uninstalling JasperServer. It also provides detailed guidance for installing from the WAR file distribution, including configuring build scripts and properties files, deploying to an application server, and setting Java options.
Configuring a highly available Microsoft Lync Server 2013 environment on Dell...Principled Technologies
As this guide has shown, setting up a highly available Microsoft Lync Server 2013 environment on Dell architecture is a straightforward process. In little time, you can deploy two Dell PowerEdge M620 M-series servers, switches, and Dell EqualLogic storage using Microsoft Windows Server 2012, and set up your Lync Server 2013 infrastructure. By setting up a highly available Lync Server 2013 environment on your Dell architecture solution, you can ensure your office communications stay running with little to no downtime to keep your business moving.
This document provides a list of over 100 third party software licenses included in Sun Storage Unified Storage Systems. It includes licenses for components such as Apache web server, BIND DNS server, MySQL database, NTP time synchronization, and OpenSSH secure shell. The document was published by Sun Microsystems in October 2009.
Network Virtualization and Security with VMware NSX - Business Case White Pap...Błażej Matusik
This document discusses how network virtualization and security solutions from VMware, specifically NSX, can help organizations transform their software-defined data centers to achieve greater speed, agility, and security. It outlines key trends driving adoption of software-defined data centers and hybrid cloud, and how NSX addresses challenges around network limitations, security threats, and manual configuration. Benefits of NSX include micro-segmentation, disaster recovery, self-service clouds, automation, and cost savings.
This document discusses how network virtualization and security solutions from VMware, specifically NSX, can help organizations transform their software-defined data centers to achieve greater speed, agility, and security. It outlines key trends driving adoption of software-defined data centers and hybrid cloud, and how NSX addresses challenges around network limitations, security threats, and manual configuration. Benefits of NSX include micro-segmentation, disaster recovery, self-service clouds, automation, and cost savings.
This document provides an overview of cloud computing. It discusses how cloud computing enables greater IT efficiency through on-demand access and scaling of computing resources, which allows companies to reduce costs. It also enables faster application development through tools like virtualization and rapid deployment of machine images. Finally, the document discusses how technologies from Sun like virtualization, modular systems, and open storage can help drive more efficient large-scale cloud deployments and enable new business opportunities through cloud computing.
This document provides help and documentation for SAP Cloud Integration for data services. It includes information on initial setup such as enabling licenses and creating projects. It also describes key concepts like datastores, tasks, processes and available actions. Specific datastore types and their properties are defined. Guidelines are provided for importing metadata, viewing data, and enabling secure network communications for SAP BW datastores.
The document discusses cloud computing and its advantages. It describes how cloud computing delivers higher efficiency and scalability compared to traditional computing models. It allows for faster and more flexible software development. The document outlines how Sun is developing technologies like interoperability, high-density computing, and data services to provide enterprise-grade qualities for cloud platforms and enable new business opportunities that rely on cloud infrastructure.
This document provides an overview and table of contents for the book "Mastering Oracle PL/SQL: Practical Solutions" by Connor McDonald and others. The summary includes:
- The book contains 10 chapters that provide practical solutions for using PL/SQL effectively, including techniques for writing efficient code, using packages, managing cursors, handling data, optimization, triggers, debugging and more.
- The introduction and first chapter discuss why PL/SQL is useful and how to write efficient PL/SQL code. Subsequent chapters cover topics like packages, cursors, data handling, optimization, triggers, debugging and security.
- Packages, debugging, and security each get their own chapter, while data handling and
This document provides an overview and table of contents for the book "Mastering Oracle PL/SQL: Practical Solutions" by Connor McDonald and others. The summary includes:
1) The book contains 10 chapters that provide practical solutions for using PL/SQL effectively, such as writing efficient code, using packages, managing cursors, handling data, optimization techniques, triggers, and debugging.
2) The table of contents lists the chapter titles and provides a brief high-level description of the topics covered in each chapter.
3) Additional sections at the beginning and end of the document include publisher information, copyright details, and lists of contributors such as authors, technical reviewers, and production staff.
This document provides help and documentation for SAP Cloud Integration for data services. It includes information on initial setup such as enabling licenses, creating projects, and moving data. It also describes datastores for connecting to different data sources, how to create and run tasks and processes, and available actions for tasks and processes. The document is a comprehensive guide for getting started with and using SAP Cloud Integration for data services.
This document provides a summary of the HiPath 4000 V6 system software components, including:
- Descriptions of hardware configurations, licensing, applications, and the hard disk layout.
- Details on the HiPath 4000 Assistant, including its applications, user accounts, and notification mechanisms.
- An overview of the HiPath 4000 CSTA software component.
The document contains technical information on the various software aspects that make up the HiPath 4000 V6 communication system.
This document provides instructions for installing and configuring IBM TotalStorage Productivity Center for Replication on AIX systems. It includes information on hardware and software requirements, pre-installation steps, installing required software like DB2, installing TPC for Replication, configuring storage systems and network connectivity, and creating replication sessions between storage systems using TPC for Replication's graphical user interface and command line. The document is intended to help users effectively set up and use TPC for Replication to manage replication services from a single point of control.
This document provides information about installing and using IBM TotalStorage Productivity Center for Replication on AIX systems. It includes details about hardware and software requirements, installation steps, configuration procedures, and usage basics. The document aims to help effectively set up and manage replication services from a single graphical interface.
This document provides guidance on updating SAP ABAP systems on UNIX and Linux using Software Update Manager 2.0 SP11. It describes the basic concepts of upgrades and updates in SAP systems, planning activities like determining the upgrade schedule and scope, preparing for the update including checking software and hardware requirements, and performing the update using the Software Update Manager user interface. The document contains detailed sections on planning, preparation, and using SUM for both upgrades and updates.
This document provides instructions for installing and configuring Insight Tech Console and Report Server. It covers supported operating systems, installation procedures for Windows and Mac computers, updating Tech Console and managed machines, configuring Tech Console preferences, and basic usage instructions. The guide is intended to help IT and technical staff use Tech Console's central management capabilities for tasks like hardware and software inventory, updating systems, monitoring security, and generating utilization reports.
VMware’s Software Defined Data Center (SDDC) vision leverages core data center virtualization technologies to transform data center economics and business agility through automation and non-disruptive deployment that embraces and extends existing compute, network and storage infrastructure investments. Enterprise data centers are already realizing the tremendous benefits of server and storage virtualization solutions to consolidate and repurpose infrastructure resources, reduce operational complexity and dynamically align and scale their application infrastructure in response to business priorities.
This technical white paper discusses VMware's NSX network virtualization platform. It begins by stating that while server and storage virtualization have provided benefits, networking has not modernized in the same way and remains a barrier to realizing the full potential of virtualization. It then describes how NSX aims to virtualize the network in the same way VMware virtualized servers, allowing virtual networks to be programmatically created and managed without reconfiguring physical network devices. This transforms network operations by greatly simplifying provisioning and management tasks.
Vmware nsx network virtualization platform white paperCloudSyntrix
Cloudsyntrix engineering team has been deploying vmware vcloud extend your data center, VMware vSphere Hypervisor 5.1 services securely extend your data center and applications to the cloud.
This document discusses how Oracle Database is the database of choice for deploying SAP solutions. It highlights Oracle's longstanding relationship with SAP in optimizing products for each other. It describes key features of Oracle Database 11g Release 2 that benefit SAP customers like advanced compression, Oracle Exadata, and high availability and security capabilities. The document provides an overview of how these features reduce costs and improve performance, management, and support for mission critical SAP environments.
This document summarizes the services provided by a company called transtec360. They offer IT services including planning, implementation, deployment, IT consulting, cloud solutions, and multi-vendor hardware services. Their goal is to support customers from initial project design through integration and daily operations. They provide customized IT solutions and aim to help customers optimize costs, improve efficiency, and gain flexibility and speed in their IT environments and operations.
This document discusses a high-performance computing solution from ttec that combines GPU computing using NVIDIA Tesla and GRID cards with Nimble storage virtualization. The solution allows centralized storage and collaboration for engineering teams working on complex projects across remote locations. It provides improved security, performance, and total cost of ownership compared to other solutions through the integration of GPU workstations, NVIDIA virtualization technologies, Citrix and VMware virtualization, and high-performance Nimble storage. The solution was implemented for an automotive manufacturer to create a centralized VDI environment for development workloads with stringent CPU, RAM, IOPS and latency requirements.
1. The document discusses various software-defined storage solutions from vendors like IBM, DataCore, and Nimble that can maximize availability, increase performance, and reduce costs for organizations.
2. It provides an overview of different storage platforms like IBM Storwize, IBM Spectrum Virtualize, DataCore VDSA appliances, and Nimble hybrid storage arrays that offer features like virtualization, high availability, flexibility, efficiency, and automation.
3. Recommendations are provided on which solutions are best suited for different use cases and storage requirements.
Transtec has over 35 years of experience providing scientific computing solutions, starting as a reseller of DEC computers and expanding to include SUN, SPARC, IBM, and RS6000 systems. Transtec is now one of the largest European providers of high performance computing (HPC) solutions, with many of its HPC clusters featured in the TOP500 list. HPC has evolved from proprietary mainframes and workstations to scale-out clusters using standard server hardware that are infinitely scalable through adding more nodes as computational needs increase.
The document discusses how high-performance computing systems are essential for fundamental life sciences and medical research. It provides details on applications used for genomic research, molecular dynamics simulations, and drug development. The company discussed offers customized HPC solutions for life sciences workloads, including compute clusters, visualization workstations, and storage solutions. It aims to provide high productivity and performance through workload management, remote visualization, and expert customer support.
This document discusses the importance of computer-aided engineering (CAE) in product development. It describes how CAE is indispensable for structural dynamics, crash simulation, noise-vibration-harshness analysis, computational fluid dynamics, and other applications. The document promotes a company called ttec that provides high-performance computing solutions to support various CAE software applications. Ttec's services include optimized workstations, compute clusters, storage solutions, job management, remote visualization, and customer support to provide a productive development environment for engineering companies.
Transtec has over 30 years of experience in scientific computing. It originally focused on reselling computers but later began offering customized high performance computing cluster solutions using standard servers. These clusters entered the TOP 500 list of the world's fastest systems. Transtec now looks to provide turnkey HPC solutions that are easy to manage and use, hiding complexity from administrators and users. Its solutions combine components like Bright Cluster Manager, Moab, Intel technology, and Panasas storage to deliver usability, reliability and scalability.
The document discusses IBM's technical computing solutions including IBM Application Ready Solutions. It provides an overview of IBM's portfolio of products for technical computing including servers, storage, software and services. It then describes IBM Application Ready Solutions which are preconfigured, optimized and tested solutions for specific applications like ANSYS, Accelrys and CLC bio. These solutions include IBM servers, Platform Computing software for workload and resource management, and IBM GPFS for high performance storage. The solutions are designed to simplify deployment and improve performance and efficiency for technical computing workloads.
VDI-in-a-Box is a simple virtual desktop solution that reduces costs through integrated connection brokering, provisioning, load balancing and profile management. It provides virtual desktops to any user on any device for less than new PCs. The solution simplifies deployment and management of virtual desktops and scales easily by adding more standard servers. It ensures high availability without complex shared storage or load balancing components.
NexentaStor is enterprise-class unified storage software that runs on standard hardware, providing block and file storage capabilities. It offers superior storage management functionality focused on virtualized environments at 70-80% cost savings compared to proprietary storage. NexentaStor provides features such as unlimited file sizes, snapshots, replication, and end-to-end data integrity.
Platform LSF 8 includes new features that simplify administration and improve productivity. It allows administrators to guarantee resources for jobs to ensure service level agreements are met. Live reconfiguration enables changes to clusters without restarting services. Delegated administration empowers line of business owners to manage their own projects. Fairshare and preemptive scheduling see enhancements. These updates help maximize efficiency in high performance computing environments.
This document discusses the evolution of high performance computing (HPC). It notes that historically HPC involved large proprietary mainframes, but today uses clustered commodity servers in a scale-out approach for greater flexibility and scalability. HPC is now seen as a way to solve complex problems in science, engineering and other fields through computer simulation and modeling. Software companies have adapted their programs to take advantage of these clustered HPC systems, allowing complex simulations and calculations to advance fields like engineering, science and bioinformatics.
The document summarizes the Provigo Dacoria storage appliance. It is a turnkey storage virtualization solution based on Datacore software. It provides features such as thin provisioning, caching, high availability through synchronous mirroring, and easy management. The appliance can be installed quickly and manages storage for both physical and virtual systems.
IBM and VMware can help you implement a highly reliable IT infrastructure with hardware and
software designed for virtualization. As the first authorized reseller of VMware products, IBM has
proven experience in developing and delivering VMware-based solutions that help customers
optimize and simplify their IT infrastructure to drive down operating costs. From virtual desktops to
enterprise-class virtualization and cloud solutions, IBM has you covered with innovative technology
and exceptional services and support.
The IBM Storwize V3700 is an efficient, high-performing virtualized storage system for small and medium businesses. It offers advanced software capabilities, up to 120 drives of storage capacity in a compact 2U frame, and is designed for common storage network technologies. The system provides efficient, entry-level virtualized storage specifically for small and midsize businesses and VMware environments through features like thin provisioning, migration, and VAAI integration.
The document discusses Datacore SANsymphony-V, a software-defined storage solution that provides virtualization, pooling and provisioning, continuous availability, performance acceleration, backup/recovery, and remote replication capabilities. It summarizes key features like virtualization of servers, storage and desktops; infrastructure management functions; benefits like cost savings and high utilization; and integration with VMware vSphere.
Microsoft Hyper-V is a hypervisor-based virtualization platform for Windows Server 2008. It requires an x64 server with hardware virtualization support from AMD or Intel. Hyper-V supports 32-bit and 64-bit virtual machines with large memory and multiple virtual processors. New features in Windows Server 2008 R2 Hyper-V include live migration of virtual machines between servers without downtime, cluster shared volumes for storage migration, and support for new processor virtualization features. Hyper-V provides virtualization for applications and server consolidation.
The document discusses Network Attached Storage (NAS) provided by Open-E. It highlights key features of Open-E's Data Storage Software V6 (DSS V6), including its support for NAS, iSCSI, Fibre Channel and InfiniBand protocols. The software provides high performance, security and scalability at a lower cost than alternatives. It also supports various hardware RAID platforms and virtualization. The document then discusses business continuity functions like iSCSI failover, NFS failover, replication and booting from iSCSI LUNs. It provides examples of hardware that can be used with Open-E's software and different available service options.
The document summarizes improvements in the Sandy Bridge server platform compared to previous generations. Key improvements include support for up to 8 cores per CPU, increased memory capacity up to 4 channels and 1600 MHz, integrated PCIe 3.0 providing better bandwidth and connectivity, and improved power efficiency. Sandy Bridge also introduces Intel Advanced Vector Extensions for improved performance on HPC workloads.
This document discusses ensuring high availability and data security with Datacore software. It notes that companies' data is their most important asset and infrastructures have become more dynamic with server virtualization. As a result, storage systems and networks must also adapt quickly. Datacore provides a future-proof and flexible solution to ensure 24/7 availability even during operations, maintenance, extensions or migrations. It allows for central management and virtualization of storage resources across systems and locations for high performance, security and simplicity.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Skybuffer AI: Advanced Conversational and Generative AI Solution on SAP Busin...Tatiana Kojar
Skybuffer AI, built on the robust SAP Business Technology Platform (SAP BTP), is the latest and most advanced version of our AI development, reaffirming our commitment to delivering top-tier AI solutions. Skybuffer AI harnesses all the innovative capabilities of the SAP BTP in the AI domain, from Conversational AI to cutting-edge Generative AI and Retrieval-Augmented Generation (RAG). It also helps SAP customers safeguard their investments into SAP Conversational AI and ensure a seamless, one-click transition to SAP Business AI.
With Skybuffer AI, various AI models can be integrated into a single communication channel such as Microsoft Teams. This integration empowers business users with insights drawn from SAP backend systems, enterprise documents, and the expansive knowledge of Generative AI. And the best part of it is that it is all managed through our intuitive no-code Action Server interface, requiring no extensive coding knowledge and making the advanced AI accessible to more users.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
leewayhertz.com-AI in predictive maintenance Use cases technologies benefits ...alexjohnson7307
Predictive maintenance is a proactive approach that anticipates equipment failures before they happen. At the forefront of this innovative strategy is Artificial Intelligence (AI), which brings unprecedented precision and efficiency. AI in predictive maintenance is transforming industries by reducing downtime, minimizing costs, and enhancing productivity.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
1. Automotive
Simulation Risk Analysis
High Throughput Computing
Price Modelling
Engineering
HIGH CAE Aerospace
PERFORMANCE
COMPUTING 2012/13
TECHNOLOGY
COMPASS CAD Big Data Analytics
Life Sciences
2. TECHNOLOGY COMPASS INTEL CLUSTER READY ............................................................................62
A Quality Standard for HPC Clusters...................................................... 64
TABLE OF CONTENTS AND INTRODUCTION
Intel Cluster Ready builds HPC Momentum ..................................... 69
The transtec Benchmarking Center ....................................................... 73
HIGH PERFORMANCE COMPUTING .................................................... 4 WINDOWS HPC SERVER 2008 R2 ........................................................74
Performance Turns Into Productivity ......................................................6 Elements of the Microsoft HPC Solution ............................................ 76
Flexible deployment with xCAT ...................................................................8 Deployment, system management, and monitoring ................. 78
Job scheduling..................................................................................................... 80
CLUSTER MANAGEMENT MADE EASY ..............................................12 Service-oriented architecture ................................................................... 82
Bright Cluster Manager ................................................................................. 14 Networking and MPI ........................................................................................ 85
Microsoft Office Excel support ................................................................. 88
INTELLIGENT HPC WORKLOAD MANAGEMENT .........................28
Moab HPC Suite – Enterprise Edition.................................................... 30 PARALLEL NFS ...............................................................................................90
New in Moab 7.0 ................................................................................................. 34 The New Standard for HPC Storage ....................................................... 92
Moab HPC Suite – Basic Edition................................................................ 37 Whats´s new in NFS 4.1? ............................................................................... 94
Moab HPC Suite - Grid Option .................................................................... 43 Panasas HPC Storage ...................................................................................... 99
NICE ENGINE FRAME .................................................................................50 NVIDIA GPU COMPUTING ....................................................................110
A technical portal for remote visualization ...................................... 52 The CUDA Architecture ............................................................................... 112
Application highlights.................................................................................... 54 Codename “Fermi” ......................................................................................... 116
Desktop Cloud Virtualization .................................................................... 57 Introducing NVIDIA Parallel Nsight ..................................................... 122
Remote Visualization...................................................................................... 58 QLogic TrueScale InfiniBand and GPUs ............................................ 126
INFINIBAND .................................................................................................130
High-speed interconnects ........................................................................ 132
Top 10 Reasons to Use QLogic TrueScale InfiniBand ................ 136
Intel MPI Library 4.0 Performance ........................................................ 139
InfiniBand Fabric Suite (IFS) – What’s New in Version 6.0 ...... 141
PARSTREAM .................................................................................................144
Big Data Analytics .......................................................................................... 146
GLOSSARY .....................................................................................................156
2
3. MORE THAN 30 YEARS OF EXPERIENCE IN SCIENTIFIC COMPUTING environment is of a highly heterogeneous nature. Even the
1980 marked the beginning of a decade where numerous startups dynamical provisioning of HPC resources as needed does not
were created, some of which later transformed into big players in constitute any problem, thus further leading to maximal utiliza-
the IT market. Technical innovations brought dramatic changes tion of the cluster.
to the nascent computer market. In Tübingen, close to one of Ger-
many’s prime and oldest universities, transtec was founded. transtec HPC solutions use the latest and most innovative
technology. Their superior performance goes hand in hand with
In the early days, transtec focused on reselling DEC computers energy efficiency, as you would expect from any leading edge IT
and peripherals, delivering high-performance workstations to solution. We regard these basic characteristics.
university institutes and research facilities. In 1987, SUN/Sparc
and storage solutions broadened the portfolio, enhanced by This brochure focusses on where transtec HPC solutions excel.
IBM/RS6000 products in 1991. These were the typical worksta- To name a few: Bright Cluster Manager as the technology leader
tions and server systems for high performance computing then, for unified HPC cluster management, leading-edge Moab HPC
used by the majority of researchers worldwide. Suite for job and workload management, Intel Cluster Ready
certification as an independent quality standard for our sys-
In the late 90s, transtec was one of the first companies to offer tems, Panasas HPC storage systems for highest performance
highly customized HPC cluster solutions based on standard and best scalability required of an HPC storage system. Again,
Intel architecture servers, some of which entered the TOP500 with these components, usability and ease of management
list of the world’s fastest computing systems. are central issues that are addressed. Also, being NVIDIA Tesla
Preferred Provider, transtec is able to provide customers with
Thus, given this background and history, it is fair to say that well-designed, extremely powerful solutions for Tesla GPU
transtec looks back upon a more than 30 years’ experience in computing. QLogic’s InfiniBand Fabric Suite makes managing a
scientific computing; our track record shows nearly 500 HPC large InfiniBand fabric easier than ever before – transtec mas-
installations. With this experience, we know exactly what cus- terly combines excellent and well-chosen components that are
tomers’ demands are and how to meet them. High performance already there to a fine-tuned, customer-specific, and thoroughly
and ease of management – this is what customers require to- designed HPC solution.
day. HPC systems are for sure required to peak-perform, as their
name indicates, but that is not enough: they must also be easy Last but not least, your decision for a transtec HPC solution
to handle. Unwieldy design and operational complexity must be means you opt for most intensive customer care and best ser-
avoided or at least hidden from administrators and particularly vice in HPC. Our experts will be glad to bring in their expertise
users of HPC computer systems. and support to assist you at any stage, from HPC design to daily
cluster operations, to HPC Cloud Services.
transtec HPC solutions deliver ease of management, both in the
Linux and Windows worlds, and even where the customer´s Have fun reading the transtec HPC Compass 2012/13!
3
5. High Performance Computing (HPC) has been with us from the very
beginning of the computer era. High-performance computers were
built to solve numerous problems which the “human computers” could
not handle. The term HPC just hadn’t been coined yet. More important,
some of the early principles have changed fundamentally.
HPC systems in the early days were much different from those we see
today. First, we saw enormous mainframes from large computer manu-
facturers, including a proprietary operating system and job management
system. Second, at universities and research institutes, workstations
made inroads and scientists carried out calculations on their dedicated
Unix or VMS workstations. In either case, if you needed more computing
power, you scaled up, i.e. you bought a bigger machine.
Today the term High-Performance Computing has gained a fundamen-
tally new meaning. HPC is now perceived as a way to tackle complex
mathematical, scientific or engineering problems. The integration of
industry standard, “off-the-shelf” server hardware into HPC clusters fa-
cilitates the construction of computer networks of such power that one
single system could never achieve. The new paradigm for parallelization
is scaling out.
5
6. HIGH PERFORMANCE COMPUTING Computer-supported simulations of realistic processes (so-
called Computer Aided Engineering – CAE) has established itself
PERFORMANCE TURNS INTO PRODUCTIVITY
as a third key pillar in the field of science and research along-
side theory and experimentation. It is nowadays inconceivable
that an aircraft manufacturer or a Formula One racing team
would operate without using simulation software. And scien-
tific calculations, such as in the fields of astrophysics, medicine,
pharmaceuticals and bio-informatics, will to a large extent be
dependent on supercomputers in the future. Software manu-
facturers long ago recognized the benefit of high-performance
computers based on powerful standard servers and ported
their programs to them accordingly.
The main advantages of scale-out supercomputers is just
that: they are infinitely scalable, at least in principle. Since
they are based on standard hardware components, such a
supercomputer can be charged with more power whenever
the computational capacity of the system is not sufficient any
more, simply by adding additional nodes of the same kind. A
“transtec HPC solutions are meant to provide cumbersome switch to a different technology can be avoided
customers with unparalleled ease-of-manage- in most cases.
ment and ease-of-use. Apart from that, deciding
for a transtec HPC solution means deciding for The primary rationale in using HPC clusters is to grow, to scale
the most intensive customer care and the best out computing capacity as far as necessary. To reach that goal,
service imaginable” an HPC cluster returns most of the invest when it is continu-
ously fed with computing problems.
Dr. Oliver Tennert Director Technology Management &
HPC Solutions The secondary reason for building scale-out supercomputers is
to maximize the utilization of the system.
6
7. If the individual processes engage in a large amount of com-
munication, the response time of the network (latency) becomes
important. Latency in a Gigabit Ethernet or a 10GE network is typi-
cally around 10 µs. High-speed interconnects such as InfiniBand,
reduce latency by a factor of 10 down to as low as 1 µs. Therefore,
high-speed interconnects can greatly speed up total processing.
The other frequently used variant is called SMP applications.
VARIATIONS ON THE THEME: MPP AND SMP SMP, in this HPC context, stands for Shared Memory Processing.
Parallel computations exist in two major variants today. Ap- It involves the use of shared memory areas, the specific imple-
plications running in parallel on multiple compute nodes are mentation of which is dependent on the choice of the underlying
frequently so-called Massively Parallel Processing (MPP) applica- operating system. Consequently, SMP jobs generally only run on
tions. MPP indicates that the individual processes can each a single node, where they can in turn be multi-threaded and thus
utilize exclusive memory areas. This means that such jobs are be parallelized across the number of CPUs per node. For many HPC
predestined to be computed in parallel, distributed across the applications, both the MPP and SMP variant can be chosen.
nodes in a cluster. The individual processes can thus utilize the
separate units of the respective node – especially the RAM, the Many applications are not inherently suitable for parallel execu-
CPU power and the disk I/O. tion. In such a case, there is no communication between the in-
dividual compute nodes, and therefore no need for a high-speed
Communication between the individual processes is imple- network between them; nevertheless, multiple computing jobs
mented in a standardized way through the MPI software can be run simultaneously and sequentially on each individual
interface (Message Passing Interface), which abstracts the node, depending on the number of CPUs.
underlying network connections between the nodes from
the processes. However, the MPI standard (current version In order to ensure optimum computing performance for these
2.0) merely requires source code compatibility, not binary applications, it must be examined how many CPUs and cores
compatibility, so an off-the-shelf application usually needs deliver the optimum performance.
specific versions of MPI libraries in order to run. Examples of
MPI implementations are OpenMPI, MPICH2, MVAPICH2, Intel We find applications of this sequential type of work typically in
MPI or – for Windows clusters – MS-MPI. the fields of data analysis or Monte-Carlo simulations.
7
8. HIGH PERFORMANCE COMPUTING
FLEXIBLE DEPLOYMENT WITH XCAT
xCAT as a Powerful and Flexible Deployment Tool
xCAT (Extreme Cluster Administration Tool) is an open source
toolkit for the deployment and low-level administration of HPC
cluster environments, small as well as large ones.
xCAT provides simple commands for hardware control, node dis-
covery, the collection of MAC addresses, and the node deploy-
ment with (diskful) or without local (diskless) installation. The
cluster configuration is stored in a relational database. Node
groups for different operating system images can be defined.
Also, user-specific scripts can be executed automatically at
installation time.
xCAT Provides the Following Low-Level Administrative Features
Remote console support
Parallel remote shell and remote copy commands
Plugins for various monitoring tools like Ganglia or Nagios
Hardware control commands for node discovery, collect-
ing MAC addresses, remote power switching and resetting
of nodes
8
9. Automatic configuration of syslog, remote shell, DNS, DHCP, when the code is self-developed, developers often prefer one
and ntp within the cluster MPI implementation over another.
Extensive documentation and man pages
According to the customer’s wishes, we install various compil-
For cluster monitoring, we install and configure the open ers, MPI middleware, as well as job management systems like
source tool Ganglia or the even more powerful open source Parastation, Grid Engine, Torque/Maui, or the very powerful
solution Nagios, according to the customer’s preferences and Moab HPC Suite for the high-level cluster management.
requirements.
Local Installation or Diskless Installation
We offer a diskful or a diskless installation of the cluster nodes.
A diskless installation means the operating system is hosted
partially within the main memory, larger parts may or may
not be included via NFS or other means. This approach allows
for deploying large amounts of nodes very efficiently, and the
cluster is up and running within a very small timescale. Also,
updating the cluster can be done in a very efficient way. For
this, only the boot image has to be updated, and the nodes have
to be rebooted. After this, the nodes run either a new kernel or
even a new operating system. Moreover, with this approach,
partitioning the cluster can also be very efficiently done, either
for testing purposes, or for allocating different cluster parti-
tions for different users or applications.
Development Tools, Middleware, and Applications
According to the application, optimization strategy, or underlying
architecture, different compilers lead to code results of very
different performance. Moreover, different, mainly commercial,
applications, require different MPI implementations. And even
9
10. HPC solution
benchmarking of application
HIGH PERFORMANCE COMPUTING
different systems installation
PERFORMANCE TURNS INTO PRODUCTIVITY
continual
improvement
maintenance, integration
onsite
customer into
support & hardware
training customer’s
managed services assembly
environment
SERVICES AND CUSTOMER CARE FROM A TO Z
application-, burn-in tests software
individual Presales
customer-, of systems & OS
consulting
site-specific installation
sizing of
HPC solution
benchmarking of application
different systems installation
continual
improvement
maintenance, integration
onsite
customer into
support & hardware
training customer’s
managed services assembly
environment
10
11. to important middleware components like cluster management
or developer tools and the customer’s production applications.
Onsite delivery means onsite integration into the customer’s
production environment, be it establishing network connectivity
to the corporate network, or setting up software and configura-
tion parts.
transtec HPC clusters are ready-to-run systems – we deliver, you
HPC @ TRANSTEC: SERVICES AND CUSTOMER CARE FROM A TO Z turn the key, the system delivers high performance. Every HPC
transtec AG has over 30 years of experience in scientific comput- project entails transfer to production: IT operation processes and
ing and is one of the earliest manufacturers of HPC clusters. policies apply to the new HPC system. Effectively, IT personnel is
For nearly a decade, transtec has delivered highly customized trained hands-on, introduced to hardware components and soft-
High Performance clusters based on standard components to ware, with all operational aspects of configuration management.
academic and industry customers across Europe with all the
high quality standards and the customer-centric approach that transtec services do not stop when the implementation projects
transtec is well known for. ends. Beyond transfer to production, transtec takes care. transtec
offers a variety of support and service options, tailored to the
Every transtec HPC solution is more than just a rack full of hard- customer’s needs. When you are in need of a new installation, a
ware – it is a comprehensive solution with everything the HPC major reconfiguration or an update of your solution – transtec is
user, owner, and operator need. able to support your staff and, if you lack the resources for main-
taining the cluster yourself, maintain the HPC solution for you.
In the early stages of any customer’s HPC project, transtec ex- From Professional Services to Managed Services for daily opera-
perts provide extensive and detailed consulting to the customer tions and required service levels, transtec will be your complete
– they benefit from expertise and experience. Consulting is fol- HPC service and solution provider. transtec’s high standards of
lowed by benchmarking of different systems with either specifi- performance, reliability and dependability assure your productiv-
cally crafted customer code or generally accepted benchmarking ity and complete satisfaction.
routines; this aids customers in sizing and devising the optimal
and detailed HPC configuration. transtec’s offerings of HPC Managed Services offer customers the
possibility of having the complete management and administra-
Each and every piece of HPC hardware that leaves our factory tion of the HPC cluster managed by transtec service specialists,
undergoes a burn-in procedure of 24 hours or more if necessary. in an ITIL compliant way. Moreover, transtec’s HPC on Demand
We make sure that any hardware shipped meets our and our services help provide access to HPC resources whenever they
customers’ quality requirements. transtec HPC solutions are turn- need them, for example, because they do not have the possibility
key solutions. By default, a transtec HPC cluster has everything of owning and running an HPC cluster themselves, due to lacking
installed and configured – from hardware and operating system infrastructure, know-how, or admin staff.
11
13. Bright Cluster Manager removes the complexity from the
installation, management and use of HPC clusters, without
compromizing performance or capability. With Bright Cluster
Manager, an administrator can easily install, use and manage
multiple clusters simultaneously, without the need for expert
knowledge of Linux or HPC.
13
14. CLUSTER MANAGEMENT MADE EASY A UNIFIED APPROACH
Other cluster management offerings take a “toolkit” approach
BRIGHT CLUSTER MANAGER
in which a Linux distribution is combined with many third-party
THE CLUSTER INSTALLER TAKES THE ADMINISTRATOR THROUGH THE
tools for provisioning, monitoring, alerting, etc.
INSTALLATION PROCESS AND OFFERS ADVANCED OPTIONS SUCH AS
“EXPRESS” AND “REMOTE”. This approach has critical limitations because those separate
tools were not designed to work together, were not designed
for HPC, and were not designed to scale. Furthermore, each of
the tools has its own interface (mostly command-line based),
and each has its own daemons and databases. Countless hours
of scripting and testing from highly skilled people are required
to get the tools to work for a specific cluster, and much of it
goes undocumented.
Bright Cluster Manager takes a much more fundamental, inte-
grated and unified approach. It was designed and written from
the ground up for straightforward, efficient, comprehensive clus-
ter management. It has a single lightweight daemon, a central
database for all monitoring and configuration data, and a single
BY SELECTING A CLUSTER NODE IN THE TREE ON THE LEFT AND THE TASKS
CLI and GUI for all cluster management functionality.
TAB ON THE RIGHT, THE ADMINISTRATOR CAN EXECUTE A NUMBER OF
POWERFUL TASKS ON THAT NODE WITH JUST A SINGLE MOUSE CLICK.. This approach makes Bright Cluster Manager extremely easy to
use, scalable, secure and reliable, complete, flexible, and easy to
maintain and support.
EASE OF INSTALLATION
Bright Cluster Manager is easy to install. Typically, system admin-
istrators can install and test a fully functional cluster from “bare
metal” in less than an hour. Configuration choices made during
the installation can be modified afterwards. Multiple installation
modes are available, including unattended and remote modes.
Cluster nodes can be automatically identified based on switch
ports rather than MAC addresses, improving speed and reliability
of installation, as well as subsequent maintenance.
14
15. EASE OF USE are performed through one intuitive, visual interface.
Bright Cluster Manager is easy to use. System administrators Multiple clusters can be managed simultaneously. The CMGUI
have two options: the intuitive Cluster Management Graphical runs on Linux, Windows and MacOS (coming soon) and can be
User Interface (CMGUI) and the powerful Cluster Management extended using plugins. The CMSH provides practically the same
Shell (CMSH). The CMGUI is a standalone desktop application functionality as the Bright CMGUI, but via a command-line inter-
that provides a single system view for managing all hardware face. The CMSH can be used both interactively and in batch mode
and software aspects of the cluster through a single point of via scripts. Either way, system administrators now have unprec-
control. Administrative functions are streamlined as all tasks edented flexibility and control over their clusters.
CLUSTER METRICS, SUCH AS GPU AND CPU TEMPERATURES, FAN SPEEDS AND NETWORKS STATISTICS CAN BE VISUALIZED BY SIMPLY DRAGGING AND DROPPING THEM FROM
THE LIST ON THE LEFT INTO A GRAPHING WINDOW ON THE RIGHT. MULTIPLE METRICS CAN BE COMBINED IN ONE GRAPH AND GRAPHS CAN BE ZOOMED INTO. GRAPH LAYOUT
AND COLORS CAN BE TAILORED TO YOUR REQUIREMENTS.
15
16. CLUSTER MANAGEMENT MADE EASY SUPPORT FOR LINUX AND WINDOWS
Bright Cluster Manager is based on Linux and is available
BRIGHT CLUSTER MANAGER
with a choice of pre-integrated, pre-configured and opti-
mized Linux distributions, including SUSE Linux Enterprise
THE STATUS OF CLUSTER NODES, SWITCHES, OTHER HARDWARE, AS WELL AS UP TO
SIX METRICS CAN BE VISUALIZED IN THE RACKVIEW. A ZOOM-OUT OPTION IS AVAIL-
ABLE FOR CLUSTERS WITH MANY RACKS.
THE OVERVIEW TAB PROVIDES INSTANT, HIGH-LEVEL INSIGHT INTO
THE STATUS OF THE CLUSTER.
Server, Red Hat Enterprise Linux, CentOS and Scientific
Linux. Dual-boot installations with Windows HPC Server are
supported as well, allowing nodes to either boot from the
Bright-managed Linux head node, or the Windows-managed
head node.
EXTENSIVE DEVELOPMENT ENVIRONMENT
Bright Cluster Manager provides an extensive HPC development
environment for both serial and parallel applications, including
the following (some optional):
16
17. Compilers, including full suites from GNU, Intel, AMD and THE PARALLEL SHELL ALLOWS FOR SIMULTANEOUS EXECUTION OF COMMANDS OR
SCRIPTS ACROSS NODE GROUPS OR ACROSS THE ENTIRE CLUSTER.
Portland Group
Debuggers and profilers, including the GNU debugger and
profiler, TAU, TotalView, Allinea DDT and Allinea OPT
GPU libraries, including CUDA and OpenCL
MPI libraries, including OpenMPI, MPICH, MPICH2, MPICH-
MX, MPICH2-MX, MVAPICH and MVAPICH2; all cross-compiled
with the compilers installed on Bright Cluster Manager, and
optimized for high-speed interconnects such as InfiniBand
and Myrinet
Mathematical libraries, including ACML, FFTW, GMP,
GotoBLAS, MKL and ScaLAPACK
Other libraries, including Global Arrays, HDF5, IIPP, TBB, Net-
CDF and PETSc
Bright Cluster Manager also provides Environment Modules to Linux kernels can be assigned to individual images. Incremen-
make it easy to maintain multiple versions of compilers, librar- tal changes to images can be deployed to live nodes without
ies and applications for different users on the cluster, without rebooting or re-installation.
creating compatibility conflicts. Each Environment Module file The provisioning system propagates only changes to the
contains the information needed to configure the shell for an images, minimizing time and impact on system performance
application, and automatically sets these variables correctly and availability. Provisioning capability can be assigned to
for the particular application when it is loaded. Bright Cluster any number of nodes on-the-fly, for maximum flexibility and
Manager includes many preconfigured module files for many scalability. Bright Cluster Manager can also provision over
scenarios, such as combinations of compliers, mathematical InfiniBand and to RAM disk.
and MPI libraries.
COMPREHENSIVE MONITORING
POWERFUL IMAGE MANAGEMENT AND PROVISIONING With Bright Cluster Manager, system administrators can collect,
Bright Cluster Manager features sophisticated software image monitor, visualize and analyze a comprehensive set of metrics.
management and provisioning capability. A virtually unlimited Practically all software and hardware metrics available to the
number of images can be created and assigned to as many Linux kernel, and all hardware management interface metrics
different categories of nodes as required. Default or custom (IPMI, iLO, etc.) are sampled.
17
18. CLUSTER MANAGEMENT MADE EASY
BRIGHT CLUSTER MANAGER
HIGH PERFORMANCE MEETS EFFICIENCY
Initially, massively parallel systems constitute a challenge to
both administrators and users. They are complex beasts. Any-
one building HPC clusters will need to tame the beast, master
the complexity and present users and administrators with an
easy-to-use, easy-to-manage system landscape.
Leading HPC solution providers such as transtec achieve this
goal. They hide the complexity of HPC under the hood and
match high performance with efficiency and ease-of-use for
both users and administrators. The “P” in “HPC” gains a double
meaning: “Performance” plus “Productivity”.
Cluster and workload management software like Moab HPC
Suite, Bright Cluster Manager or QLogic IFS provide the means
to master and hide the inherent complexity of HPC systems. For
administrators and users, HPC clusters are presented as single,
large machines, with many different tuning parameters. The
software also provides a unified view of existing clusters when-
ever unified management is added as a requirement by the
customer at any point in time after the first installation. Thus,
daily routine tasks such as job management, user management,
queue partitioning and management, can be performed easily
with either graphical or web-based tools, without any advanced
scripting skills or technical expertise required from the adminis-
trator or user.
18
19. Powerful cluster automation functionality allows
preemptive actions based on monitoring thresholds
Comprehensive cluster monitoring and health checking
framework, including automatic sidelining of unhealthy
nodes to prevent job failure
Scalability from Deskside to TOP500
Off-loadable provisioning for maximum scalability
THE BRIGHT ADVANTAGE Proven on some of the world’s largest clusters
Bright Cluster Manager offers many advantages that lead to
improved productivity, uptime, scalability, performance and Minimum Overhead/Maximum Performance
security, while reducing total cost of ownership. Single lightweight daemon drives all functionality
Daemon heavily optimized to minimize effect on operating
Rapid Productivity Gains system and applications
Easy to learn and use, with an intuitive GUI Single database stores all metric and configuration data
Quick installation: from bare metal to a cluster ready to use,
in less than an hour Top Security
Fast, flexible provisioning: incremental, live, disk-full, disk- Automated security and other updates from key-signed
less, provisioning over InfiniBand, auto node discovery repositories
Comprehensive monitoring: on-the-fly graphs, rackview, Encrypted external and internal communications (optional)
multiple clusters, custom metrics X.509v3 certificate-based public-key authentication
Powerful automation: thresholds, alerts, actions Role-based access control and complete audit trail
Complete GPU support: NVIDIA, AMD ATI, CUDA, OpenCL Firewalls and secure LDAP
On-demand SMP: instant ScaleMP virtual SMP deployment
Powerful cluster management shell and SOAP API for auto-
mating tasks and creating custom capabilities
Seamless integration with leading workload managers: PBS
Pro, Moab, Maui, SLURM, Grid Engine, Torque, LSF
Integrated (parallel) application development environment.
Easy maintenance: automatically update your cluster from
Linux and Bright Computing repositories
Web-based user portal
Bright Computing
Maximum Uptime
Unattended, robust head node failover to spare head node
19
20. CLUSTER MANAGEMENT MADE EASY Examples include CPU and GPU temperatures, fan speeds,
switches, hard disk SMART information, system load, memory
BRIGHT CLUSTER MANAGER
utilization, network statistics, storage metrics, power systems
statistics, and workload management statistics. Custom metrics
can also easily be defined.
Metric sampling is done very efficiently – in one process, or
out-of-band where possible. System administrators have full
flexibility over how and when metrics are sampled, and historic
data can be consolidated over time to save disk space.
THE AUTOMATION CONFIGURATION WIZARD GUIDES THE SYSTEM ADMINISTRATOR
THROUGH THE STEPS OF DEFINING A RULE: SELECTING METRICS, DEFINING THRESH-
OLDS AND SPECIFYING ACTIONS.
CLUSTER MANAGEMENT AUTOMATION
Cluster management automation takes preemptive actions
when predetermined system thresholds are exceeded, sav-
ing time and preventing hardware damage. System thresh-
olds can be configured on any of the available metrics. The
built-in configuration wizard guides the system administra-
20
21. tor through the steps of defining a rule: selecting metrics, EXAMPLE GRAPHS THAT VISUALIZE METRICS ON A GPU CLUSTER.
defining thresholds and specifying actions. For example,
a temperature threshold for GPUs can be established that
results in the system automatically shutting down an over-
heated GPU unit and sending an SMS message to the system
administrator’s mobile phone. Several predefined actions are
available, but any Linux command or script can be config-
ured as an action.
COMPREHENSIVE GPU MANAGEMENT
Bright Cluster Manager radically reduces the time and ef-
fort of managing GPUs, and fully integrates these devices
into the single view of the overall system. Bright includes
powerful GPU management and monitoring capability that
leverages functionality in NVIDIA Tesla GPUs. System admin-
istrators can easily assume maximum control of the GPUs
and gain instant and time-based status insight. In addition
to the standard cluster management capabilities, Bright
Cluster Manager monitors the full range of GPU metrics,
including: MULTI-TASKING VIA PARALLEL SHELL
GPU temperature, fan speed, utilization The parallel shell allows simultaneous execution of multiple
GPU exclusivity, compute, display, persistance mode commands and scripts across the cluster as a whole, or across
GPU memory utilization, ECC statistics easily definable groups of nodes. Output from the executed
Unit fan speed, serial number, temperature, power commands is displayed in a convenient way with variable levels
usage, voltages and currents, LED status, firmware of verbosity. Running commands and scripts can be killed easily
Board serial, driver version, PCI info if necessary. The parallel shell is available through both the
CMGUI and the CMSH.
Beyond metrics, Bright Cluster Manager features built-in
support for GPU computing with CUDA and OpenCL libraries. INTEGRATED WORKLOAD MANAGEMENT
Switching between current and previous versions of CUDA and Bright Cluster Manager is integrated with a wide selection of
OpenCL has also been made easy. free and commercial workload managers. This integration
21
22. CLUSTER MANAGEMENT MADE EASY provides a number of benefits:
The selected workload manager gets automatically installed
BRIGHT CLUSTER MANAGER
and configured
Many workload manager metrics are monitored
The GUI provides a user-friendly interface for configuring,
monitoring and managing the selected workload manager
The CMSH and the SOAP API provide direct and powerful access
to a number of workload manager commands and metrics
WORKLOAD MANAGEMENT QUEUES CAN BE VIEWED AND CON- CREATING AND DISMANTLING A VIRTUAL SMP NODE CAN BE ACHIEVED WITH JUST
FIGURED FROM THE GUI, WITHOUT THE NEED FOR WORKLOAD A FEW CLICKS WITHIN THE GUI OR A SINGLE COMMAND IN THE CLUSTER MANAGE-
MANAGEMENT EXPERTISE. MENT SHELL.
22
23. Reliable workload manager failover is properly configured MAXIMUM UPTIME WITH HEALTH CHECKING
The workload manager is continuously made aware of the Bright Cluster Manager – Advanced Edition includes a powerful
health state of nodes (see section on Health Checking) cluster health checking framework that maximizes system uptime.
It continually checks multiple health indicators for all hardware
The following user-selectable workload managers are tightly and software components and proactively initiates corrective
integrated with Bright Cluster Manager: actions. It can also automatically perform a series of standard
PBS Pro, Moab, Maui, LSF and user-defined tests just before starting a new job, to ensure
SLURM, Grid Engine, Torque a successful execution. Examples of corrective actions include
autonomous bypass of faulty nodes, automatic job requeuing to
Alternatively, Lava, LoadLeveler or other workload managers can avoid queue flushing, and process “jailing” to allocate, track, trace
be installed on top of Bright Cluster Manager. and flush completed user processes. The health checking frame-
work ensures the highest job throughput, the best overall cluster
INTEGRATED SMP SUPPORT efficiency and the lowest administration overhead.
Bright Cluster Manager – Advanced Edition dynamically ag-
gregates multiple cluster nodes into a single virtual SMP node, WEB-BASED USER PORTAL
using ScaleMP’s Versatile SMP™ (vSMP) architecture. Creating The web-based user portal provides read-only access to essential
and dismantling a virtual SMP node can be achieved with just cluster information, including a general overview of the cluster
a few clicks within the CMGUI. Virtual SMP nodes can also be status, node hardware and software properties, workload manager
launched and dismantled automatically using the scripting statistics and user-customizable graphs. The User Portal can easily
capabilities of the CMSH. In Bright Cluster Manager a virtual be customized and expanded using PHP and the SOAP API.
SMP node behaves like any other node, enabling transparent,
on-the-fly provisioning, configuration, monitoring and man- USER AND GROUP MANAGEMENT
agement of virtual SMP nodes as part of the overall system Users can be added to the cluster through the CMGUI or the
management. CMSH. Bright Cluster Manager comes with a pre-configured
LDAP database, but an external LDAP service, or alternative
MAXIMUM UPTIME WITH HEAD NODE FAILOVER authentication system, can be used instead.
Bright Cluster Manager – Advanced Edition allows two head
nodes to be configured in active-active failover mode. Both ROLE-BASED ACCESS CONTROL AND AUDITING
head nodes are on active duty, but if one fails, the other takes Bright Cluster Manager’s role-based access control mechanism
over all tasks, seamlessly. allows administrator privileges to be defined on a per-role basis.
23
24. CLUSTER MANAGEMENT MADE EASY Administrator actions can be audited using an audit file which
stores all their write action.
BRIGHT CLUSTER MANAGER
TOP CLUSTER SECURITY
Bright Cluster Manager offers an unprecedented level of secu-
rity that can easily be tailored to local requirements. Security
features include:
Automated security and other updates from key-signed
Linux and Bright Computing repositories
Encrypted internal and external communications
X.509v3 certificate based public-key authentication to the
cluster management infrastructure
THE WEB-BASED USER PORTAL PROVIDES READ-ONLY ACCESS TO ESSENTIAL CLUSTER
INFORMATION, INCLUDING A GENERAL OVERVIEW OF THE CLUSTER STATUS, NODE
HARDWARE AND SOFTWARE PROPERTIES, WORKLOAD MANAGER STATISTICS AND
USER-CUSTOMIZABLE GRAPHS.
“The building blocks for transtec HPC solu-
tions must be chosen according to our goals
ease-of-management and ease-of-use. With
Bright Cluster Manager, we are happy to have
the technology leader at hand, meeting these
requirements, and our customers value that.”
Armin Jäger HPC Solution Engineer
24
25. Role-based access control and complete audit trail STANDARD AND ADVANCED EDITIONS
Firewalls and secure LDAP Bright Cluster Manager is available in two editions: Standard
Secure shell access and Advanced. The table on this page lists the differences. You
can easily upgrade from the Standard to the Advanced Edition
MULTI-CLUSTER CAPABILITY as your cluster grows in size or complexity.
Bright Cluster Manager is ideal for organizations that need to
manage multiple clusters, either in one or in multiple locations. DOCUMENTATION AND SERVICES
Capabilities include: A comprehensive system administrator manual and user manu-
All cluster management and monitoring functionality availa- al are included in PDF format. Customized training and profes-
ble for all clusters through one GUI sional services are available. Services include various levels of
Selecting any set of configurations in one cluster and support, installation services and consultancy.
export them to any or all other clusters with a few mouse
clicks
Making node images available to other clusters.
BRIGHT CLUSTER MANAGER CAN MANAGE MULTIPLE CLUSTERS SIMULTANEOUSLY. CLUSTER HEALTH CHECKS CAN BE VISUALIZED IN THE RACKVIEW. THIS SCREENSHOT
THIS OVERVIEW SHOWS CLUSTERS IN OSLO, ABU DHABI AND HOUSTON, ALL MAN- SHOWS THAT GPU UNIT 41 FAILS A HEALTH CHECK CALLED “ALLFANSRUNNING”.
AGED THROUGH ONE GUI.
25
26. CLUSTER MANAGEMENT MADE EASY
BRIGHT CLUSTER MANAGER
FEATURE STANDARD ADVANCED
Choice of Linux distributions x x
Intel Cluster Ready x x
Cluster Management GUI x x
Cluster Management Shell x x
Web-Based User Portal x x
SOAP API x x
Node Provisioning x x
Node Identification x x
Cluster Monitoring x x
Cluster Automation x x
User Management x x
Parallel Shell x x
Workload Manager Integration x x
Cluster Security x x
Compilers x x
Debuggers & Profilers x x
MPI Libraries x x
Mathematical Libraries x x
Environment Modules x x
NVIDIA CUDA & OpenCL x x
GPU Management & Monitoring x x
ScaleMP Management & Monitoring - x
Redundant Failover Head Nodes - x
Cluster Health Checking - x
Off-loadable Provisioning - x
Suggested Number of Nodes 4–128 129–10,000+
Multi-Cluster Management - x
Standard Support x x
Premium Support Optional Optional
26
29. While all HPC systems face challenges in workload demand,
resource complexity, and scale, enterprise HPC systems face
more stringent challenges and expectations. Enterprise HPC
systems must meet mission-critical and priority HPC workload
demands for commercial businesses and business-oriented
research and academic organizations. They have complex SLAs
and priorities to balance. Their HPC workloads directly impact
the revenue, product delivery, and organizational objectives
of their organizations.
29
30. INTELLIGENT MOAB HPC SUITE
Moab is the most powerful intelligence engine for policy-based,
HPC WORKLOAD MANAGEMENT predictive scheduling across workloads and resources. Moab
MOAB HPC SUITE – ENTERPRISE EDITION HPC Suite accelerates results delivery and maximize utiliza-
tion while simplifying workload management across complex,
heterogeneous cluster environments. The Moab HPC Suite
products leverage the multi-dimensional policies in Moab to
continually model and monitor workloads, resources, SLAs,
and priorities to optimize workload output. And these policies
utilize the unique Moab management abstraction layer that
integrates data across heterogeneous resources and resource
managers to maximize control as you automate workload man-
agement actions.
Managing the World’s Top Systems, Ready to Manage Yours
Moab manages the world’s largest, most scale-intensive and
complex HPC environments in the world including 40% of the top
10 supercomputing systems, nearly 40% of the top 25 and 36%
of the compute cores in the top 100 systems based on rankings
from www.Top500.org. So you know it is battle-tested and ready
“With Moab HPC Suite, we can meet very de- to efficiently and intelligently manage the complexities of your
manding customers’ requirements as regards environment.
unified management of heterogeneous cluster
environments, grid management, and provide MOAB HPC SUITE – ENTERPRISE EDITION
them with flexible and powerful configuration Moab HPC Suite - Enterprise Edition provides enterprise-ready
and reporting options. Our customers value HPC workload management that self-optimizes the productivity,
that highly.” workload uptime and meeting of SLAs and business priorities
for HPC systems and HPC cloud. It uses the battle-tested and
patented Moab intelligence engine to automate the mission-
Thomas Gebert HPC Solution Architect critical workload priorities of enterprise HPC systems. Enterprise
customers benefit from a single integrated product that brings
30
31. together key enterprise HPC capabilities, implementation, train- achievement of business objectives and outcomes that depend
ing, and 24x7 support services to speed the realization of benefits on the results the enterprise HPC systems deliver. Moab HPC
from their HPC system for their business. Moab HPC Suite – En- Suite Enterprise Edition delivers:
terprise Edition delivers:
Productivity acceleration Productivity acceleration to get more results faster and at a
Uptime automation lower cost
Auto-SLA enforcement Moab HPC Suite – Enterprise Edition gets more results delivered
Grid- and cloud-ready HPC management faster from HPC resources to lower costs while accelerating
overall system, user and administrator productivity. Moab
Designed to Solve Enterprise HPC Challenges provides the unmatched scalability, 90-99 percent utilization,
While all HPC systems face challenges in workload and resource and fast and simple job submission that is required to maximize
complexity, scale and demand, enterprise HPC systems face productivity in enterprise HPC organizations. The Moab intel-
more stringent challenges and expectations. Enterprise HPC ligence engine optimizes workload scheduling and orchestrates
systems must meet mission-critical and priority HPC workload resource provisioning and management to maximize workload
demands for commercial businesses and business-oriented speed and quantity. It also unifies workload management
research and academic organizations. These organizations have across heterogeneous resources, resource managers and even
complex SLA and priorities to balance. And their HPC workloads multiple clusters to reduce management complexity and costs.
directly impact the revenue, product delivery, and organization-
al objectives of their organizations. Uptime automation to ensure workload completes successfully
Enterprise HPC organizations must eliminate job delays and HPC job and resource failures in enterprise HPC systems lead to
failures. They are also seeking to improve resource utilization delayed results and missed organizational opportunities and
and workload management efficiency across multiple heteroge- objectives. Moab HPC Suite – Enterprise Edition intelligently
neous systems. To maximize user productivity, they are required automates workload and resource uptime in HPC systems to en-
to make it easier to access and use HPC resources for users and sure that workload completes reliably and avoids these failures.
even expand to other clusters or HPC cloud to better handle
workload demand and surges. Auto-SLA enforcement to consistently meet service guaran-
tees and business priorities
BENEFITS Moab HPC Suite – Enterprise Edition uses the powerful Moab
Moab HPC Suite - Enterprise Edition offers key benefits to intelligence engine to optimally schedule and dynamically
reduce costs, improve service performance, and accelerate the adjust workload to consistently meet service level agreements
productivity of enterprise HPC systems. These benefits drive the (SLAs), guarantees, and business priorities. This automatically
31
32. INTELLIGENT ensures that the right workloads are completed at the optimal
times, taking into account the complex number of departments,
HPC WORKLOAD MANAGEMENT priorities and SLAs to be balanced.
MOAB HPC SUITE – ENTERPRISE EDITION
Grid- and Cloud-ready HPC management to more efficiently
manage and meet workload demand
The benefits of a traditional HPC environment can be extended
to more efficiently manage and meet workload and resource
demand by sharing workload across multiple clusters through
grid management and the HPC cloud management capabilities
provided in Moab HPC Suite – Enterprise Edition.
CAPABILITIES
Moab HPC Suite – Enterprise Edition brings together key en-
terprise HPC capabilities into a single integrated product that
self-optimizes the productivity, workload uptime, and meeting
of SLA’s and priorities for HPC systems and HPC Cloud.
Productivity acceleration capabilities deliver more results
faster, lower costs, and increase resource, user and administra-
tor productivity
ARCHITECTURE Massive scalability accelerates job response and through-
put, including support for high throughput computing
Workload-optimized allocation policies and provisioning
gets more results out of existing heterogeneous resources to
reduce costs
Workload unification across heterogeneous clusters maxi-
mizes resource availability for workloads and administration
efficiency by managing workload as one cluster
Simplified HPC submission and control for both users and ad-
ministrators with job arrays, templates, self-service submission
32
33. portal and administrator dashboard (i.e. usage limits, usage reports, etc.)
Optimized intelligent scheduling that packs workloads and SLA and priority polices ensure the highest priority workloads
backfills around priority jobs and reservations while balancing are processed first (i.e. quality of service, hierarchical priority
SLAs to efficiently use all available resources weighting, dynamic fairshare policies, etc.)
Advanced scheduling and management of GPGPUs for jobs to Continuous plus future scheduling ensures priorities and gua-
maximize their utilization including auto-detection, policy-based rantees are proactively met as conditions and workload levels
GPGPU scheduling and GPGPU metrics reporting change (i.e. future reservations, priorities, and pre-emption)
Workload-aware auto-power management reduces energy use
and costs by 30-40 percent with intelligent workload consolidati- Grid- and cloud-ready HPC management extends the benefits of
on and auto-power management your traditional HPC environment to more efficiently manage
workload and better meet workload demand
Uptime automation capabilities ensure workload completes suc- Pay-for-use showback and chargeback capabilities track
cessfully and reliably, avoiding failures and missed organizational actual resource usage with flexible chargeback options and
opportunities and objectives reporting by user or department
Intelligent resource placement prevents job failures with gra- Manage and share workload across multiple remote
nular resource modeling that ensures all workload requirements clusters to meet growing workload demand or surges with
are met while avoiding at-risk resources the single self-service portal and intelligence engine with
Auto-response to incidents and events maximizes job and sys- purchase of Moab HPC Suite - Grid Option
tem uptime with configurable actions to pre-failure conditions,
amber alerts, or other metrics and monitors ARCHITECTURE
Workload-aware maintenance scheduling helps maintain a Moab HPC Suite - Enterprise Edition is architected to integrate
stable HPC system without disrupting workload productivity on top of your existing job resource managers and other types
Real-world services expertise ensures fast time to value and of resource managers in your environment. It provides policy-
system uptime with included package of implementation, trai- based scheduling and management of workloads as well as
ning, and 24x7 remote support services resource allocation and provisioning orchestration. The Moab
intelligence engine makes complex scheduling and manage-
Auto-SLA enforcement schedules and adjusts workload to con- ment decisions based on all of the data it integrates from the
sistently meet service guarantees and business priorities so the various resource managers and then orchestrates the job and
right workloads are completed at the optimal times management actions through those resource managers. It
Department budget enforcement schedules resources in does this without requiring any additional agents. This makes
line with resource sharing agreements and budgets it the ideal choice to integrate with existing and new systems
33
34. INTELLIGENT
HPC WORKLOAD MANAGEMENT
NEW IN MOAB 7.0
NEW MOAB HPC SUITE 7.0
The new Moab HPC Suite 7.0 releases deliver continued break-
through advancements in scalability, reliability, and job array
management to accelerate system productivity as well as ex-
tended database support. Here is a look at the new capabilities
and the value they offer customers:
TORQUE Resource Manager Scalability and Reliability Ad-
vancements for Petaflop and Beyond
As part of the Moab HPC Suite 7.0 releases, the TORQUE 4.0
resource manager features scalability and reliability advance-
ments to fully exploit Moab scalability. These advancements
maximize your use of increasing hardware capabilities and
enable you to meet growing HPC user needs. Key advancements
in TORQUE 4.0 for Moab HPC Suite 7.0 include:
The new Job Radix enables you to efficiently run jobs that span
tens of thousands or even hundreds of thousands of nodes.
Each MOM daemon now cascades job communication with
multiple other MOM daemons simultaneously to reduce the
job start-up process time to a small fraction of what it would
normally take across a large number of nodes. The Job Radix
eliminates lost jobs and job start-up bottlenecks caused by
having all nodes MOM daemons communicating with only one
head MOM node. This saves critical minutes on job start-up
process time and allows for higher job throughput.
34
35. New MOM daemon communication hierarchy increases gration with existing user portals, plug-ins of resource manag-
the number of nodes supported and reduces the overhead ers for rich data integration, and script integration. Customers
of cluster status updates by distributing communication now have a standard interface to Moab with REST APIs.
across multiple nodes instead of a single TORQUE head
node. This makes status updates more efficient faster sched- Simplified Self-Service and Admin Dashboard Portal Experience
uling and responsiveness. Moab HPC 7.0 features an enhanced self-service and admin
New multi-threading improves response and reliability, dashboard portal with simplified “click-based” job submission
allowing for instant feedback to user requests as well as the for end users as well as new visual cluster dashboard views of
ability to continue work even if some processes linger. nodes, jobs, and reservations for more efficient management. The
Improved network communications with all UDP-based new Visual Cluster dashboard provides administrators and users
communication replaced with TCP to make data transfers views of their cluster resources that are easily filtered by almost
from node to node more reliable. any factors including id, name, IP address, state, power, pending
actions, reservations, load, memory, processors, etc. Users can
Job Array Auto-Cancellation Policies Improve System Productivity also quickly filter and view their jobs by name, state, user, group,
Moab HPC Suite 7.0 improves system productivity with new job ar- account, wall clock requested, memory requested, start date/
ray auto-cancellation policies that cancel remaining sub-jobs in an time, submit date/time, etc. One-click drill-downs provide addi-
array once the solution is found in the array results. This frees up tional details and options for management actions.
resources, which would otherwise be running irrelevant jobs, to run
other jobs in the queue jobs quicker. The job array auto-cancellation Resource Usage Accounting Flexibility
policies allow you to set auto-cancellations of sub-jobs based on Moab HPC Suite 7.0 includes more flexible resource usage ac-
first, any instance of results success or failure, or specific exit codes. counting options that enable administrators to easily duplicate
custom organizational hierarchies such as organization, groups,
Extended Database Support Now Includes PostgreSQL and projects, business units, cost centers etc. in the Moab Account-
Oracle in Addition to MySQL ing Manager usage budgets and charging structure. This ensures
The extended database support in Moab HPC Suite 7.0 enables resource usage is budgeted , tracked, and reported or charged
customers to use ODBC-compliant PostgreSQL and Oracle back for in the most useful way to admins and their customer
databases in addition to MySQL. This provides customers the groups and users.
flexibility to use the database that best meets their needs or is
the standard for their system.
New Moab Web Services Provide Easier Standard Integration
and Customization
New Moab Web Services provide easier standard integration
and customization for a customer’s environment such as inte-
35
36. INTELLIGENT as well as to manage your HPC system as it grows and expands
in the future.
HPC WORKLOAD MANAGEMENT
MOAB HPC SUITE – BASIC EDITION Moab HPC Suite – Enterprise Edition includes the patented
Moab intelligence engine that enables it to integrate with and
automate management across existing heterogeneous environ-
ments to optimize management and workload efficiency. This
unique intelligence engine includes:
Industry leading multi-dimensional policies that automate
the complex real-time decisions and actions for scheduling
workload and allocating and adapting resources. These mul-
ti-dimensional policies can model and consider the workload
requirements, resource attributes and affinities, SLAs and
priorities to enable more complex and efficient decisions to
be automated.
Real-time and predictive future environment scheduling
that drives more accurate and efficient decisions and service
guarantees as it can proactively adjust scheduling and re-
source allocations as it projects the impact of workload and
resource condition changes.
Open & flexible management abstraction layer lets you
integrate the data and orchestrate workload actions across
the chaos of complex heterogeneous cluster environments
and management middleware to maximize workload control,
automation, and optimization.
COMPONENTS
Moab HPC Suite – Enterprise Edition includes the following inte-
grated products and technologies for a complete HPC workload
management solution:
Moab Workload Manager: Patented multi-dimensional
36
37. intelligence engine that automates the complex decisions based workload management system that accelerates and auto-
and orchestrates policy-based workload placement and mates the scheduling, managing, monitoring, and reporting of
scheduling as well as resource allocation, provisioning and HPC workloads on massive scale, multi-technology installations.
energy management The Moab HPC Suite – Basic Edition patented multi-dimensional
Moab Cluster Manager: Graphical desktop administrator decision engine accelerates both the decisions and orchestrati-
application for managing, configuring, monitoring, and on of workload across the ideal combination of diverse resour-
reporting for Moab managed clusters ces, including specialized resources like GPGPUs. The speed and
Moab Viewpoint: Web-based user self-service job submis- accuracy of the decisions and scheduling automation optimizes
sion and management portal and administrator dashboard workload throughput and resource utilization so more work
portal is accomplished in less time with existing resources to control
Moab Accounting Manager: HPC resource use budgeting costs and increase the value out of HPC investments.
and accounting tool that enforces resource sharing agree-
ments and limits based on departmental budgets and provi- Moab HPC Suite – Basic Edition enables you to address pressing
des showback and chargeback reporting for resource usage HPC challenges including:
Moab Services Manager: Integration interfaces to resource Delays to workload start and end times slowing results
managers and third-party tools Inconsistent delivery on service guarantees and SLA commit-
ments
Moab HPC Suite – Enterprise Edition is also integrated with Under-utilization of resources
TORQUE which is available as a free download on AdaptiveCom- How to efficiently manage workload across heterogeneous and
puting.com. TORQUE is an open-source job/resource manager hybrid systems of GPGPUs, hardware, and middleware
that provides continually updated information regarding the How to simplify job submission & management for users and
state of nodes and workload status. Adaptive Computing is the administrators to maximize productivity
custodian of the TORQUE project and is actively developing
the code base in cooperation with the TORQUE community to Moab HPC Suite – Basic Edition acts as the “brain” of an HPC
provide state of the art resource management. Each Moab HPC system to accelerate and automate complex decision making
Suite product subscription includes support for the Moab HPC processes. The patented decision engine is capable of making
Suite as well as TORQUE, if you choose to use TORQUE as the the complex multi-dimensional policy-based decisions needed to
job/resource manager for your cluster. schedule workload to optimize job speed, job success and resource
utilization. Moab HPC Suite – Basic Edition integrates decision-
MOAB HPC SUITE – BASIC EDITION making data from and automates actions through your system’s
Moab HPC Suite – Basic Edition is a multi-dimensional policy- existing mix of resource managers. This enables all the dimensions
37
38. INTELLIGENT of real-time granular resource attributes and state as well as the
timing of current and future resource commitments to be factored
HPC WORKLOAD MANAGEMENT into more efficient and accurate scheduling and allocation decisi-
MOAB HPC SUITE – BASIC EDITION ons. It also dramatically simplifies the management tasks and pro-
cesses across these complex, heterogeneous environments. Moab
works with many of the major resource management and industry
standard resource monitoring tools covering mixed hardware,
MOAB HPC SUITE - BASIC EDITION network, storage and licenses.
Moab HPC Suite – Basic Edition policies are also able to factor
in organizational priorities and complexities when scheduling
workload and allocating resources. Moab ensures workload is pro-
cessed according to organizational priorities and commitments
and that resources are shared fairly across users, groups and even
multiple organizations. This enables organizations to automati-
cally enforce service guarantees and effectively manage organiza-
tional complexities with simple policy-based settings.
BENEFITS
Moab HPC Suite – Basic Edition drives more ROI and results
from your HPC environment including:
Improved job response times and job throughput with a
workload decision engine that accelerates complex wor-
kload scheduling decisions to enable faster job start times
and high throughput computing
Optimized resource utilization to 90-99 percent with multi-
dimensional and predictive workload scheduling to accomp-
lish more with your existing resources
Automated enforcement of service guarantees, priorities,
and resource sharing agreements across users, groups, and
projects
Increased productivity by simplifying HPC use, access, and
38
39. control for both users and administrators with job arrays, affinity- and node topology-based placement
job templates, optional user portal, and GUI administrator Backfill job scheduling speeds job throughput and maximi-
management and monitoring tool zes utilization by scheduling smaller or less demanding jobs
Streamline job turnaround and reduce administrative as they can fit around priority jobs and reservations to use
burden by unifying and automating workload tasks and re- all available resources
source processes across diverse resources and mixed-system Security policies control which users and groups can access
environments including GPGPUs which resources
Provides a scalable workload management architecture Checkpointing
that can manage peta-scale and beyond, is grid-ready,
compatible with existing infrastructure, and extensible to Real-time and predictive scheduling ensure job priorities and
manage your environment as it grows and evolves guarantees are proactively met as conditions and workload
levels change
CAPABILITIES Advanced reservations guarantee that jobs run when required
Moab HPC Suite – Basic Edition accelerates workload pro- Maintenance reservations reserve resources for planned fu-
cessing with a patented multi-dimensional decision engine ture maintenance to avoid disruption to business workloads
that self-optimizes workload placement, resource utilization Predictive scheduling enables the future workload schedule
and results output while ensuring organizational priorities to be continually forecasted and adjusted along with resour-
are met across the users and groups leveraging the HPC ce allocations to adapt to changes in conditions and new job
environment. and reservation requests
Policy-driven scheduling intelligently places workload on op- Advanced scheduling and management of GPGPUs for jobs to
timal set of diverse resources to maximize job throughput and maximize their utilization
success as well as utilization and the meeting of workload and Automatic detection and management of GPGPUs in envi-
group priorities ronment to eliminate manual configuration and make them
Priority, SLA and resource sharing policies ensure the highest immediately available for scheduling
priority workloads are processed first and resources are Exclusively allocate and schedule GPGPUs on a per-job basis
shared fairly across users and groups such as quality of Policy-based management & scheduling using GPGPU
service, hierarchical priority weighting, and fairshare targets, metrics
limits and weights policies Quick access to statistics on GPGPU utilization and key
Allocation policies optimize resource utilization and prevent metrics for optimal management and issue diagnosis such as
job failures with granular resource modeling and scheduling, error counts, temperature, fan speed, and memory
39
40. INTELLIGENT Easier submission, management, and control of job arrays im-
prove user productivity and job throughput efficiency
HPC WORKLOAD MANAGEMENT Users can easily submit thousands of sub-jobs with a single
MOAB HPC SUITE – BASIC EDITION job submission with an array index differentiating each array
sub-job
Job array usage limit policies enforce number of job maxi-
mums by credentials or class
Simplified reporting and management of job arrays for end
users filters jobs to summarize, track and manage at the
master job level
Scalable job performance to large-scale, extreme-scale, and
high-throughput computing environments
Efficiently manages the submission and scheduling of hund-
reds of thousands of queued job submissions to support
high throughput computing
Fast scheduler response to user commands while scheduling
so users and administrators get the real-time job informati-
on they need
Fast job throughput rate to get results started and delivered
faster and keep utilization of resources up
Open and flexible management abstraction layer easily integrates
with and automates management across existing heterogeneous
resources and middleware to improve management efficiency
Rich data integration and aggregation enables you to set pow-
erful, multi-dimensional policies based on the existing real-time
resource data monitored without adding any new agents
Heterogeneous resource allocation & management for wor-
kloads across mixed hardware, specialty resources such as
40