This document discusses hardware offloading of VXLAN encapsulation and decapsulation in OVS-DPDK. It proposes representing virtual ports (vPorts) as tables to enable hardware offloading of VXLAN processing. Matching and actions on the vPort table would occur in hardware before decapsulation. Fallback processing using software would be used if full hardware offloading is not possible. The goal is to leverage intelligent NIC capabilities to accelerate VXLAN tunnel processing and improve performance for cloud, NFV, and storage workloads.
Ibm spectrum scale fundamentals workshop for americas part 5 ess gnr-usecases...xKinAnx
Ā
This document provides an overview of Spectrum Scale 4.1 system administration. It describes the Elastic Storage Server options and components, Spectrum Scale native RAID (GNR), and tips for best practices. GNR implements sophisticated data placement and error correction algorithms using software RAID to provide high reliability and performance without additional hardware. It features auto-rebalancing, low rebuild overhead through declustering, and end-to-end data checksumming.
import rdma: zero-copy networking with RDMA and Pythongroveronline
Ā
The document discusses using RDMA (Remote Direct Memory Access) for high performance networking in Python. It describes how traditional sockets involve multiple memory copies that reduce performance. RDMA allows data to be directly placed in the receiving application's memory without intermediate copies. The document outlines an implementation of RDMA sockets in Python, called RdmaSocket, that exposes an interface similar to regular sockets but uses RDMA under the hood. Examples are given showing how it can simplify high performance networking applications in Python.
This document discusses disaggregating Ceph storage using NVMe over Fabrics (NVMeoF). It motivates using NVMeoF by showing the performance limitations of directly attaching multiple NVMe drives to individual compute nodes. It then proposes a design to leverage the full resources of a cluster by distributing NVMe drives across dedicated storage nodes and connecting them to compute nodes over a high performance fabric using NVMeoF and RDMA. Some initial Ceph performance measurements using this model show improved IOPS and latency compared to the direct attached approach. Future work could explore using SPDK and Linux kernel improvements to further optimize performance.
The document discusses applying RDMA (Remote Direct Memory Access) to improve performance in distributed deep learning frameworks. It describes implementing RDMA in MXNet, a distributed deep learning framework that uses a parameter server model. The implementation reduces memory copies and network overhead. Initial results showed a 1.5x speedup over the initial RDMA implementation, but the existing implementation using ZeroMQ was still faster. Further optimizations to RDMA are needed to fully realize its performance benefits.
Pushing Packets - How do the ML2 Mechanism Drivers Stack UpJames Denton
Ā
Architecting a private cloud to meet the use cases of its users can be a daunting task. How do you determine which of the many L2/L3 Neutron plugins and drivers to implement? Does network performance outweigh reliability? Are overlay networks just as performant as VLAN networks? The answers to these questions will drive the appropriate technology choice.
In this presentation, we will look at many of the common drivers built around the ML2 framework, including LinuxBridge, OVS, OVS+DPDK, SR-IOV, and more, and will provide performance data to help drive decisions around selecting a technology that's right for the situation. We will discuss our experience with some of these technologies, and the pros and cons of one technology over another in a production environment.
Haodong Tang from Intel gave this talk at the 2018 Open Fabrics Workshop.
"Efficient network messenger is critical for todayās scale-out storage systems. Ceph is one of the most popular distributed storage system providing a scalable and reliable object, block and file storage services. As the explosive growth of Big Data continues, there're strong demands leveraging Ceph build high performance & ultra-low latency storage solution in the cloud and bigdata environment. The traditional TCP/IP cannot satisfy this requirement, but Remote Direct Memory Access (RDMA) can.
"In this session, we'll present the challenges in today's distributed storage system posed by network messenger with the profiling results of Ceph All Flash Array system showing the networking already become the bottleneck and introduce how we achieved 8% performance benefit with Ethernet RDMA protocol iWARP. We'll first present the design of integrating iWARP to Ceph networking module together with performance characterization results with iWARP enabled IO intensive workload. The send part, we will explore the proof-of-concept solution of Ceph on NVMe over iWARP to build high-performance and high-density storage solution. Finally, we will showcase how these solutions can improve OSD scalability, and whatās the next optimization opportunities based on current analysis."
Watch the video: https://wp.me/p3RLHQ-ikV
Learn more: http://intel.com
and
https://insidehpc.com/2018/04/amazon-libfabric-case-study-flexible-hpc-infrastructure/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Ceph data services in a multi- and hybrid cloud worldSage Weil
Ā
IT organizations of the future (and present) are faced with managing infrastructure that spans multiple private data centers and multiple public clouds. Emerging tools and operational patterns like kubernetes and microservices are easing the process of deploying applications across multiple environments, but the achilles heel of such efforts remains that most applications require large quantities of state, either in databases, object stores, or file systems. Unlike stateless microservices, state is hard to move.
Ceph is known for providing scale-out file, block, and object storage within a single data center, but it also includes a robust set of multi-cluster federation capabilities. This talk will cover how Ceph's underlying multi-site capabilities complement and enable true portability across cloud footprints--public and private--and how viewing Ceph from a multi-cloud perspective has fundamentally shifted our data services roadmap, especially for Ceph object storage.
This document discusses hardware offloading of VXLAN encapsulation and decapsulation in OVS-DPDK. It proposes representing virtual ports (vPorts) as tables to enable hardware offloading of VXLAN processing. Matching and actions on the vPort table would occur in hardware before decapsulation. Fallback processing using software would be used if full hardware offloading is not possible. The goal is to leverage intelligent NIC capabilities to accelerate VXLAN tunnel processing and improve performance for cloud, NFV, and storage workloads.
Ibm spectrum scale fundamentals workshop for americas part 5 ess gnr-usecases...xKinAnx
Ā
This document provides an overview of Spectrum Scale 4.1 system administration. It describes the Elastic Storage Server options and components, Spectrum Scale native RAID (GNR), and tips for best practices. GNR implements sophisticated data placement and error correction algorithms using software RAID to provide high reliability and performance without additional hardware. It features auto-rebalancing, low rebuild overhead through declustering, and end-to-end data checksumming.
import rdma: zero-copy networking with RDMA and Pythongroveronline
Ā
The document discusses using RDMA (Remote Direct Memory Access) for high performance networking in Python. It describes how traditional sockets involve multiple memory copies that reduce performance. RDMA allows data to be directly placed in the receiving application's memory without intermediate copies. The document outlines an implementation of RDMA sockets in Python, called RdmaSocket, that exposes an interface similar to regular sockets but uses RDMA under the hood. Examples are given showing how it can simplify high performance networking applications in Python.
This document discusses disaggregating Ceph storage using NVMe over Fabrics (NVMeoF). It motivates using NVMeoF by showing the performance limitations of directly attaching multiple NVMe drives to individual compute nodes. It then proposes a design to leverage the full resources of a cluster by distributing NVMe drives across dedicated storage nodes and connecting them to compute nodes over a high performance fabric using NVMeoF and RDMA. Some initial Ceph performance measurements using this model show improved IOPS and latency compared to the direct attached approach. Future work could explore using SPDK and Linux kernel improvements to further optimize performance.
The document discusses applying RDMA (Remote Direct Memory Access) to improve performance in distributed deep learning frameworks. It describes implementing RDMA in MXNet, a distributed deep learning framework that uses a parameter server model. The implementation reduces memory copies and network overhead. Initial results showed a 1.5x speedup over the initial RDMA implementation, but the existing implementation using ZeroMQ was still faster. Further optimizations to RDMA are needed to fully realize its performance benefits.
Pushing Packets - How do the ML2 Mechanism Drivers Stack UpJames Denton
Ā
Architecting a private cloud to meet the use cases of its users can be a daunting task. How do you determine which of the many L2/L3 Neutron plugins and drivers to implement? Does network performance outweigh reliability? Are overlay networks just as performant as VLAN networks? The answers to these questions will drive the appropriate technology choice.
In this presentation, we will look at many of the common drivers built around the ML2 framework, including LinuxBridge, OVS, OVS+DPDK, SR-IOV, and more, and will provide performance data to help drive decisions around selecting a technology that's right for the situation. We will discuss our experience with some of these technologies, and the pros and cons of one technology over another in a production environment.
Haodong Tang from Intel gave this talk at the 2018 Open Fabrics Workshop.
"Efficient network messenger is critical for todayās scale-out storage systems. Ceph is one of the most popular distributed storage system providing a scalable and reliable object, block and file storage services. As the explosive growth of Big Data continues, there're strong demands leveraging Ceph build high performance & ultra-low latency storage solution in the cloud and bigdata environment. The traditional TCP/IP cannot satisfy this requirement, but Remote Direct Memory Access (RDMA) can.
"In this session, we'll present the challenges in today's distributed storage system posed by network messenger with the profiling results of Ceph All Flash Array system showing the networking already become the bottleneck and introduce how we achieved 8% performance benefit with Ethernet RDMA protocol iWARP. We'll first present the design of integrating iWARP to Ceph networking module together with performance characterization results with iWARP enabled IO intensive workload. The send part, we will explore the proof-of-concept solution of Ceph on NVMe over iWARP to build high-performance and high-density storage solution. Finally, we will showcase how these solutions can improve OSD scalability, and whatās the next optimization opportunities based on current analysis."
Watch the video: https://wp.me/p3RLHQ-ikV
Learn more: http://intel.com
and
https://insidehpc.com/2018/04/amazon-libfabric-case-study-flexible-hpc-infrastructure/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Ceph data services in a multi- and hybrid cloud worldSage Weil
Ā
IT organizations of the future (and present) are faced with managing infrastructure that spans multiple private data centers and multiple public clouds. Emerging tools and operational patterns like kubernetes and microservices are easing the process of deploying applications across multiple environments, but the achilles heel of such efforts remains that most applications require large quantities of state, either in databases, object stores, or file systems. Unlike stateless microservices, state is hard to move.
Ceph is known for providing scale-out file, block, and object storage within a single data center, but it also includes a robust set of multi-cluster federation capabilities. This talk will cover how Ceph's underlying multi-site capabilities complement and enable true portability across cloud footprints--public and private--and how viewing Ceph from a multi-cloud perspective has fundamentally shifted our data services roadmap, especially for Ceph object storage.
Data Engineer's Lunch #83: Strategies for Migration to Apache IcebergAnant Corporation
Ā
In this talk, Dremio Developer Advocate, Alex Merced, discusses strategies for migrating your existing data over to Apache Iceberg. He'll go over the following:
How to Migrate Hive, Delta Lake, JSON, and CSV sources to Apache Iceberg
Pros and Cons of an In-place or Shadow Migration
Migrating between Apache Iceberg catalogs Hive/Glue -- Arctic/Nessie
This presentation provides an overview of the Dell PowerEdge R730xd server performance results with Red Hat Ceph Storage. It covers the advantages of using Red Hat Ceph Storage on Dell servers with their proven hardware components that provide high scalability, enhanced ROI cost benefits, and support of unstructured data.
1. The document discusses migrating from OpenGL to Vulkan, providing analogies comparing the APIs to fixed-function toys, programmable LEGO kits, and raw materials pine wood derby kits.
2. It outlines scenarios that are likely and unlikely to benefit from Vulkan, such as applications with parallelizable CPU-bound graphics work.
3. Key differences between OpenGL and Vulkan are explained, such as Vulkan requiring explicit management of graphics resources, synchronization, and command buffer queuing. The document emphasizes that transitioning to Vulkan truly rethinks the entire graphics rendering approach.
BlueStore, A New Storage Backend for Ceph, One Year InSage Weil
Ā
BlueStore is a new storage backend for Ceph OSDs that consumes block devices directly, bypassing the local XFS file system that is currently used today. It's design is motivated by everything we've learned about OSD workloads and interface requirements over the last decade, and everything that has worked well and not so well when storing objects as files in local files systems like XFS, btrfs, or ext4. BlueStore has been under development for a bit more than a year now, and has reached a state where it is becoming usable in production. This talk will cover the BlueStore design, how it has evolved over the last year, and what challenges remain before it can become the new default storage backend.
Webinar topic: MPLS on Router OS V7 - Part 1
Presenter: Achmad Mardiansyah & M. Taufik Nurhuda
In this webinar series, How MPLS on Router OS V7 works
Please share your feedback or webinar ideas here: http://bit.ly/glcfeedback
Check our schedule for future events: https://www.glcnetworks.com/en/schedule/
Follow our social media for updates: Facebook, Instagram, YouTube Channel, and telegram also discord
Recording available on Youtube
https://youtu.be/SvZrYNA0-rQ
The document provides step-by-step instructions for building and running Intel DPDK sample applications on a test environment with 3 virtual machines connected by 10G NICs. It describes compiling and running the helloworld, L2 forwarding, and L3 forwarding applications, as well as using the pktgen tool for packet generation between VMs to test forwarding performance. Key steps include preparing the Linux kernel for DPDK, compiling applications, configuring ports and MAC addresses, and observing packet drops to identify performance bottlenecks.
Ceph Pacific is a major release of the Ceph distributed storage system scheduled for March 2021. It focuses on five key themes: usability, performance, ecosystem integration, multi-site capabilities, and quality. New features in Pacific include automated upgrades, improved dashboard functionality, snapshot-based CephFS mirroring, per-bucket replication in RGW, and expanded telemetry collection. Looking ahead, the Quincy release will focus on continued improvements in these areas such as resource-aware scheduling in cephadm and multi-site monitoring capabilities.
DIY Netflow Data Analytic with ELK Stack by CL LeeMyNOG
Ā
This document discusses using the ELK stack (Elasticsearch, Logstash, Kibana) to analyze netflow data. It describes IP ServerOne's infrastructure managing over 5000 servers across multiple data centers. Netflow data is collected and sent to Logstash for processing, then stored in Elasticsearch for querying and visualization in Kibana. Examples are given of how the data can be used, such as identifying top talkers, traffic profiling by ASN, and troubleshooting with IP conversation history. The ELK stack is concluded to be a powerful yet not difficult tool for analyzing netflow traffic.
Crimson: Ceph for the Age of NVMe and Persistent MemoryScyllaDB
Ā
Ceph is a mature open source software-defined storage solution that was created over a decade ago.
During that time new faster storage technologies have emerged including NVMe and Persistent memory.
The crimson project aim is to create a better Ceph OSD that is more well suited to those faster devices. The crimson OSD is built on the Seastar C++ framework and can leverage these devices by minimizing latency, cpu overhead, and cross-core communication. This talk will discuss the project design, our current status, and our future plans.
NFV aims to virtualize network functions that were traditionally run on dedicated hardware appliances. This allows the functions to run on commercial off-the-shelf servers and switches in data centers. NFV promises to reduce costs, increase flexibility and speed up innovation cycles compared to proprietary hardware appliances. A key enabler is virtualization technology which allows multiple virtual machines to run in isolation on shared server hardware through the use of hypervisors.
Large scale overlay networks with ovn: problems and solutionsHan Zhou
Ā
Han Zhou presents problems and solutions for scaling Open Virtual Network (OVN) components in large overlay networks. The key challenges addressed are:
1. Scaling the OVN controller by moving from recomputing all flows to incremental processing based on changes.
2. Scaling the southbound OVN database by increasing probe intervals, enabling fast resync on reconnect, and improving performance of the clustered mode.
3. Further work is planned to incrementally install flows, reduce per-host data, and scale out the southbound database with replicas.
How To Use Kafka and Druid to Tame Your Router Data (Rachel Pedreschi, Imply ...confluent
Ā
Do you know who is knocking on your networkās door? Have new regulations left you scratching your head on how to handle what is happening in your network? Network flow data helps answer many questions across a multitude of use cases including network security, performance, capacity planning, routing, operational troubleshooting and more. Todayās modern day streaming data pipelines need to include tools that can scale to meet the demands of these service providers while continuing to provide responsive answers to difficult questions. In addition to stream processing, data needs to be stored in a redundant, operationally focused database to provide fast, reliable answers to critical questions. Together, Kafka and Druid work together to create such a pipeline.
In this talk Eric Graham and Rachel Pedreschi will discuss these pipelines and cover the following topics:
-Network flow use cases and why this data is important.
-Reference architectures from production systems at a major international Bank.
-Why Kafka and Druid and other OSS tools for Network Flows.
-A demo of one such system.
Ceph Object Storage Reference Architecture Performance and Sizing GuideKaran Singh
Ā
Together with my colleagues at Red Hat Storage Team, i am very proud to have worked on this reference architecture for Ceph Object Storage.
If you are building Ceph object storage at scale, this document is for you.
Zeus: Uberās Highly Scalable and Distributed Shuffle as a ServiceDatabricks
Ā
Zeus is an efficient, highly scalable and distributed shuffle as a service which is powering all Data processing (Spark and Hive) at Uber. Uber runs one of the largest Spark and Hive clusters on top of YARN in industry which leads to many issues such as hardware failures (Burn out Disks), reliability and scalability challenges.
This document discusses bonding interfaces in Mikrotik routers. It begins with an overview of bonding and its benefits of higher throughput and failover. It then covers the different options for link monitoring and bonding modes, including active-backup, load balancing, and broadcast modes. It provides an example configuration of bonding two Ethernet interfaces together. Finally, it proposes testing the bonded bandwidth and provides references for further reading.
The document discusses NVIDIA's efforts to harness virtual reality and AI for scientific applications. It summarizes NVIDIA's development of GPUs, CPUs, and DPUs to rearchitect data centers for AI workloads. It also discusses several AI tools and frameworks developed by NVIDIA, including simulations, digital twins, virtual worlds, AI assistants, and tools for industries like climate science, robotics, and autonomous systems.
Snort_inline is an intrusion prevention system that can be configured to operate inline by parsing network traffic through two network cards in bridge mode. This allows Snort_inline to detect threats in real-time and drop malicious traffic. The document discusses how to configure Snort_inline for different network environments like internal LANs, DMZs, and mixed networks by adjusting preprocessing rules and Snort rules. It also describes tools for monitoring Snort alerts and managing intrusion detection rules.
Data Engineer's Lunch #83: Strategies for Migration to Apache IcebergAnant Corporation
Ā
In this talk, Dremio Developer Advocate, Alex Merced, discusses strategies for migrating your existing data over to Apache Iceberg. He'll go over the following:
How to Migrate Hive, Delta Lake, JSON, and CSV sources to Apache Iceberg
Pros and Cons of an In-place or Shadow Migration
Migrating between Apache Iceberg catalogs Hive/Glue -- Arctic/Nessie
This presentation provides an overview of the Dell PowerEdge R730xd server performance results with Red Hat Ceph Storage. It covers the advantages of using Red Hat Ceph Storage on Dell servers with their proven hardware components that provide high scalability, enhanced ROI cost benefits, and support of unstructured data.
1. The document discusses migrating from OpenGL to Vulkan, providing analogies comparing the APIs to fixed-function toys, programmable LEGO kits, and raw materials pine wood derby kits.
2. It outlines scenarios that are likely and unlikely to benefit from Vulkan, such as applications with parallelizable CPU-bound graphics work.
3. Key differences between OpenGL and Vulkan are explained, such as Vulkan requiring explicit management of graphics resources, synchronization, and command buffer queuing. The document emphasizes that transitioning to Vulkan truly rethinks the entire graphics rendering approach.
BlueStore, A New Storage Backend for Ceph, One Year InSage Weil
Ā
BlueStore is a new storage backend for Ceph OSDs that consumes block devices directly, bypassing the local XFS file system that is currently used today. It's design is motivated by everything we've learned about OSD workloads and interface requirements over the last decade, and everything that has worked well and not so well when storing objects as files in local files systems like XFS, btrfs, or ext4. BlueStore has been under development for a bit more than a year now, and has reached a state where it is becoming usable in production. This talk will cover the BlueStore design, how it has evolved over the last year, and what challenges remain before it can become the new default storage backend.
Webinar topic: MPLS on Router OS V7 - Part 1
Presenter: Achmad Mardiansyah & M. Taufik Nurhuda
In this webinar series, How MPLS on Router OS V7 works
Please share your feedback or webinar ideas here: http://bit.ly/glcfeedback
Check our schedule for future events: https://www.glcnetworks.com/en/schedule/
Follow our social media for updates: Facebook, Instagram, YouTube Channel, and telegram also discord
Recording available on Youtube
https://youtu.be/SvZrYNA0-rQ
The document provides step-by-step instructions for building and running Intel DPDK sample applications on a test environment with 3 virtual machines connected by 10G NICs. It describes compiling and running the helloworld, L2 forwarding, and L3 forwarding applications, as well as using the pktgen tool for packet generation between VMs to test forwarding performance. Key steps include preparing the Linux kernel for DPDK, compiling applications, configuring ports and MAC addresses, and observing packet drops to identify performance bottlenecks.
Ceph Pacific is a major release of the Ceph distributed storage system scheduled for March 2021. It focuses on five key themes: usability, performance, ecosystem integration, multi-site capabilities, and quality. New features in Pacific include automated upgrades, improved dashboard functionality, snapshot-based CephFS mirroring, per-bucket replication in RGW, and expanded telemetry collection. Looking ahead, the Quincy release will focus on continued improvements in these areas such as resource-aware scheduling in cephadm and multi-site monitoring capabilities.
DIY Netflow Data Analytic with ELK Stack by CL LeeMyNOG
Ā
This document discusses using the ELK stack (Elasticsearch, Logstash, Kibana) to analyze netflow data. It describes IP ServerOne's infrastructure managing over 5000 servers across multiple data centers. Netflow data is collected and sent to Logstash for processing, then stored in Elasticsearch for querying and visualization in Kibana. Examples are given of how the data can be used, such as identifying top talkers, traffic profiling by ASN, and troubleshooting with IP conversation history. The ELK stack is concluded to be a powerful yet not difficult tool for analyzing netflow traffic.
Crimson: Ceph for the Age of NVMe and Persistent MemoryScyllaDB
Ā
Ceph is a mature open source software-defined storage solution that was created over a decade ago.
During that time new faster storage technologies have emerged including NVMe and Persistent memory.
The crimson project aim is to create a better Ceph OSD that is more well suited to those faster devices. The crimson OSD is built on the Seastar C++ framework and can leverage these devices by minimizing latency, cpu overhead, and cross-core communication. This talk will discuss the project design, our current status, and our future plans.
NFV aims to virtualize network functions that were traditionally run on dedicated hardware appliances. This allows the functions to run on commercial off-the-shelf servers and switches in data centers. NFV promises to reduce costs, increase flexibility and speed up innovation cycles compared to proprietary hardware appliances. A key enabler is virtualization technology which allows multiple virtual machines to run in isolation on shared server hardware through the use of hypervisors.
Large scale overlay networks with ovn: problems and solutionsHan Zhou
Ā
Han Zhou presents problems and solutions for scaling Open Virtual Network (OVN) components in large overlay networks. The key challenges addressed are:
1. Scaling the OVN controller by moving from recomputing all flows to incremental processing based on changes.
2. Scaling the southbound OVN database by increasing probe intervals, enabling fast resync on reconnect, and improving performance of the clustered mode.
3. Further work is planned to incrementally install flows, reduce per-host data, and scale out the southbound database with replicas.
How To Use Kafka and Druid to Tame Your Router Data (Rachel Pedreschi, Imply ...confluent
Ā
Do you know who is knocking on your networkās door? Have new regulations left you scratching your head on how to handle what is happening in your network? Network flow data helps answer many questions across a multitude of use cases including network security, performance, capacity planning, routing, operational troubleshooting and more. Todayās modern day streaming data pipelines need to include tools that can scale to meet the demands of these service providers while continuing to provide responsive answers to difficult questions. In addition to stream processing, data needs to be stored in a redundant, operationally focused database to provide fast, reliable answers to critical questions. Together, Kafka and Druid work together to create such a pipeline.
In this talk Eric Graham and Rachel Pedreschi will discuss these pipelines and cover the following topics:
-Network flow use cases and why this data is important.
-Reference architectures from production systems at a major international Bank.
-Why Kafka and Druid and other OSS tools for Network Flows.
-A demo of one such system.
Ceph Object Storage Reference Architecture Performance and Sizing GuideKaran Singh
Ā
Together with my colleagues at Red Hat Storage Team, i am very proud to have worked on this reference architecture for Ceph Object Storage.
If you are building Ceph object storage at scale, this document is for you.
Zeus: Uberās Highly Scalable and Distributed Shuffle as a ServiceDatabricks
Ā
Zeus is an efficient, highly scalable and distributed shuffle as a service which is powering all Data processing (Spark and Hive) at Uber. Uber runs one of the largest Spark and Hive clusters on top of YARN in industry which leads to many issues such as hardware failures (Burn out Disks), reliability and scalability challenges.
This document discusses bonding interfaces in Mikrotik routers. It begins with an overview of bonding and its benefits of higher throughput and failover. It then covers the different options for link monitoring and bonding modes, including active-backup, load balancing, and broadcast modes. It provides an example configuration of bonding two Ethernet interfaces together. Finally, it proposes testing the bonded bandwidth and provides references for further reading.
The document discusses NVIDIA's efforts to harness virtual reality and AI for scientific applications. It summarizes NVIDIA's development of GPUs, CPUs, and DPUs to rearchitect data centers for AI workloads. It also discusses several AI tools and frameworks developed by NVIDIA, including simulations, digital twins, virtual worlds, AI assistants, and tools for industries like climate science, robotics, and autonomous systems.
Snort_inline is an intrusion prevention system that can be configured to operate inline by parsing network traffic through two network cards in bridge mode. This allows Snort_inline to detect threats in real-time and drop malicious traffic. The document discusses how to configure Snort_inline for different network environments like internal LANs, DMZs, and mixed networks by adjusting preprocessing rules and Snort rules. It also describes tools for monitoring Snort alerts and managing intrusion detection rules.
DESIGN AND PERFORMANCE ANALYSIS OF ZBT SRAM CONTROLLERVLSICS Design
Ā
Memory is an essential part of electronic industry. Since, the processors used in various high performance
PCs, network applications and communication equipment require high speed memories. The type of
memory used depends on system architecture, and its applications. This paper presents an SRAM
architecture known as Zero Bus Turnaround (ZBT). This ZBT SRAM is mainly developed for networking
applications where frequent READ/WRITE transitions are required. The other single data rate SRAMs are
inefficient as they require idle cycles when they frequently switch between reading and writing to the
memory. This controller is simulated on the Spartan 3 device. And the performance analysis is done on the
basis of area, speed and power.
This document discusses optimizing Linux AMIs for performance at Netflix. It begins by providing background on Netflix and explaining why tuning the AMI is important given Netflix runs tens of thousands of instances globally with varying workloads. It then outlines some of the key tools and techniques used to bake performance optimizations into the base AMI, including kernel tuning to improve efficiency and identify ideal instance types. Specific examples of CFS scheduler, page cache, block layer, memory allocation, and network stack tuning are also covered. The document concludes by discussing future tuning plans and an appendix on profiling tools like perf and SystemTap.
SNMP (Simple Network Management Protocol) allows for remote and local management of devices on a network. It uses a client/server model where agents run on devices and collect information, and managers can request data from agents. SNMP is an open standard protocol that is widely used for network monitoring and management.
OSMC 2009 | net-snmp: The forgotten classic by Dr. Michael SchwartzkopffNETWAYS
Ā
Simple Network Management Protocol (SNMP) is probably the oldest protocol for the management of heterogeneous Networks. As twenty years have passed since the first standards in RFC 1157 were published, we can rank SNMP as a classic. However, its advantages have been long forgotten- perhaps in part, thanks to its trouble free application. This presentation will briefly cover the basics of SNMP and deal with its implementation in Linux. The net-smnp agent offers many tricks and configuration options which can be used to solve most management problems easily.
Introduction to IBM Shared Memory Communications Version 2 (SMCv2) and SMC-Dv2zOSCommserver
Ā
This document discusses SMC Version 2 (SMCv2) which introduces enhancements that allow SMC connections across multiple IP subnets by leveraging Routable RoCE (RoCEv2). Key points include:
- RoCEv2 allows SMC traffic to be encapsulated in UDP/IP packets, making it IP routable and no longer restricted to a single IP subnet.
- SMC-D Version 2 (SMC-Dv2) over Internal Shared Memory Version 2 (ISMv2) will be delivered for z/OS 2.4 on IBM z15 systems, allowing SMC connections across LPARs on different IP subnets.
- Future support for SMC-R
The document discusses using tcpdump and ssldump on an F5 device to analyze network traffic. It provides examples of commands to capture full traffic flows, including specifying filters. It also describes how to use tcpdump to troubleshoot issues like traffic not reaching servers. The document discusses using Wireshark with the F5 plugin to decrypt SSL traffic for analysis and provides instructions for configuring Wireshark. It briefly mentions using sFlow for performance monitoring and analytics.
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Amazon Web Services
Ā
Your AMI is one of the core foundations for running applications and services effectively on Amazon EC2. In this session, you'll learn how to optimize your AMI, including how you can measure and diagnose system performance and tune parameters for improved CPU and network performance. We'll cover application-specific examples from Netflix on how optimized AMIs can lead to improved performance.
Snap ML is a machine learning framework for fast training of generalized linear models (GLMs) that can scale to large datasets. It uses multi-level parallelism across nodes and GPUs. Snap ML implementations include snap-ml-local for single nodes, snap-ml-mpi for multi-node HPC environments, and snap-ml-spark for Apache Spark clusters. Experimental results show Snap ML can train a logistic regression model on a 3TB Criteo dataset within 1.5 minutes using 16 GPUs.
Provides an overview of how LWM2M and DNS-SD/DNS-SEC can be used together to provide for secure communications, remote management and provisioning of constrained devices in the Internet of Things using open source software components leshan and Tiaki created in the eclipse IoT community.
Dryad is a general-purpose distributed execution engine that allows developers to easily write efficient parallel applications. It models distributed programs as directed acyclic graphs of computational vertices connected by data channels. The Dryad runtime system schedules vertices across clusters and handles failures, requiring developers to only focus on the computational logic. Dryad demonstrates high performance on both SQL queries and large-scale data mining tasks running on clusters of up to 1800 computers.
Centralized monitoring station for it computing and network infrastructureMOHD ARISH
Ā
This document describes a centralized monitoring system for an IT network using Linux, SNMP, and MRTG. SNMP is used to collect data from network devices and store it in a MySQL database. MRTG generates graphs of this SNMP data over time through an Apache web server. This allows network administrators to monitor performance, faults, and security across various devices from a central location to maintain network availability.
Five cool ways the JVM can run Apache Spark fasterTim Ellison
Ā
The IBM JVM runs Apache Spark fast! This talk explains some of the findings and optimizations from our experience of running Spark workloads.
The talk was originally presented at the SparkEU Summit 2015 in Amsterdam.
Monitoring pfSense 2.4 with SNMP - pfSense Hangout March 2018Netgate
Ā
This document provides an overview of monitoring pfSense 2.4 with SNMP. It describes what SNMP is, the roles of MIBs and network monitoring systems. It then covers the two options for SNMP in pfSense - the built-in bsnmpd daemon and the NET-SNMP package. The document provides instructions for configuring each option and describes features like SNMP security, custom commands, communities, and users.
2009-01-28 DOI NBC Red Hat on System z Performance ConsiderationsShawn Wells
Ā
Presented with the U.S. Department of the Interior, National Business Center. DOI NBC offered a for-fee Linux on System z to the U.S. Government. This presentation steps through performance management considerations, including: FCP/SCSI single path vs multipath LMV; filesystem striping; crypto express2 accelerator (CEX2A) SSL handshakes; cryptographic performance (WebSEAL SSL Access); and CMM1 & CMMA.
This document discusses various ways to enhance networking performance in high-bandwidth applications. It covers application-level optimizations like adjusting Nagle's algorithm and TCP buffer sizes. It also discusses transport-level options like TCP stack selection and modifications, use of specialized networking hardware, and alternative protocols like SCTP. The goal is to minimize data copies, protocol overhead, and optimize resource usage to increase throughput and reduce latency.
Exploiting Multi Core Architectures for Process Speed UpIJERD Editor
Ā
The document discusses exploiting multi-core architectures to speed up processing of satellite data. It describes developing a multithreaded application using a master-slave model to parallelize preprocessing steps like extracting auxiliary data, Reed-Solomon decoding, and decompression. Performance analysis shows the application scales dynamically based on system resources and achieves maximum speedup when assigning 7 threads per core. While theoretical speedup is limited by Amdahl's law, accounting for parallelization overhead provides a more realistic performance measure.
GraphSummit Paris - The art of the possible with Graph TechnologyNeo4j
Ā
Sudhir Hasbe, Chief Product Officer, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges ā from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
What is Augmented Reality Image Trackingpavan998932
Ā
Augmented Reality (AR) Image Tracking is a technology that enables AR applications to recognize and track images in the real world, overlaying digital content onto them. This enhances the user's interaction with their environment by providing additional information and interactive elements directly tied to physical images.
Preparing Non - Technical Founders for Engaging a Tech AgencyISH Technologies
Ā
Preparing non-technical founders before engaging a tech agency is crucial for the success of their projects. It starts with clearly defining their vision and goals, conducting thorough market research, and gaining a basic understanding of relevant technologies. Setting realistic expectations and preparing a detailed project brief are essential steps. Founders should select a tech agency with a proven track record and establish clear communication channels. Additionally, addressing legal and contractual considerations and planning for post-launch support are vital to ensure a smooth and successful collaboration. This preparation empowers non-technical founders to effectively communicate their needs and work seamlessly with their chosen tech agency.Visit our site to get more details about this. Contact us today www.ishtechnologies.com.au
Quarkus Hidden and Forbidden ExtensionsMax Andersen
Ā
Quarkus has a vast extension ecosystem and is known for its subsonic and subatomic feature set. Some of these features are not as well known, and some extensions are less talked about, but that does not make them less interesting - quite the opposite.
Come join this talk to see some tips and tricks for using Quarkus and some of the lesser known features, extensions and development techniques.
Why Mobile App Regression Testing is Critical for Sustained Success_ A Detail...kalichargn70th171
Ā
A dynamic process unfolds in the intricate realm of software development, dedicated to crafting and sustaining products that effortlessly address user needs. Amidst vital stages like market analysis and requirement assessments, the heart of software development lies in the meticulous creation and upkeep of source code. Code alterations are inherent, challenging code quality, particularly under stringent deadlines.
Mobile App Development Company In Noida | Drona InfotechDrona Infotech
Ā
Looking for a reliable mobile app development company in Noida? Look no further than Drona Infotech. We specialize in creating customized apps for your business needs.
Visit Us For : https://www.dronainfotech.com/mobile-application-development/
Top Features to Include in Your Winzo Clone App for Business Growth (4).pptxrickgrimesss22
Ā
Discover the essential features to incorporate in your Winzo clone app to boost business growth, enhance user engagement, and drive revenue. Learn how to create a compelling gaming experience that stands out in the competitive market.
Software Engineering, Software Consulting, Tech Lead, Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Transaction, Spring MVC, OpenShift Cloud Platform, Kafka, REST, SOAP, LLD & HLD.
When deliberating between CodeIgniter vs CakePHP for web development, consider their respective strengths and your project requirements. CodeIgniter, known for its simplicity and speed, offers a lightweight framework ideal for rapid development of small to medium-sized projects. It's praised for its straightforward configuration and extensive documentation, making it beginner-friendly. Conversely, CakePHP provides a more structured approach with built-in features like scaffolding, authentication, and ORM. It suits larger projects requiring robust security and scalability. Ultimately, the choice hinges on your project's scale, complexity, and your team's familiarity with the frameworks.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
Ā
In this second installment of our Essentials of Automations webinar series, weāll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
Weāll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether youāre tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Letās turn complexity into clarity and make your workspaces work wonders!
A Study of Variable-Role-based Feature Enrichment in Neural Models of CodeAftab Hussain
Ā
Understanding variable roles in code has been found to be helpful by students
in learning programming -- could variable roles help deep neural models in
performing coding tasks? We do an exploratory study.
- These are slides of the talk given at InteNSE'23: The 1st International Workshop on Interpretability and Robustness in Neural Software Engineering, co-located with the 45th International Conference on Software Engineering, ICSE 2023, Melbourne Australia
Enterprise Resource Planning System includes various modules that reduce any business's workload. Additionally, it organizes the workflows, which drives towards enhancing productivity. Here are a detailed explanation of the ERP modules. Going through the points will help you understand how the software is changing the work dynamics.
To know more details here: https://blogs.nyggs.com/nyggs/enterprise-resource-planning-erp-system-modules/
Transform Your Communication with Cloud-Based IVR SolutionsTheSMSPoint
Ā
Discover the power of Cloud-Based IVR Solutions to streamline communication processes. Embrace scalability and cost-efficiency while enhancing customer experiences with features like automated call routing and voice recognition. Accessible from anywhere, these solutions integrate seamlessly with existing systems, providing real-time analytics for continuous improvement. Revolutionize your communication strategy today with Cloud-Based IVR Solutions. Learn more at: https://thesmspoint.com/channel/cloud-telephony
WhatsApp offers simple, reliable, and private messaging and calling services for free worldwide. With end-to-end encryption, your personal messages and calls are secure, ensuring only you and the recipient can access them. Enjoy voice and video calls to stay connected with loved ones or colleagues. Express yourself using stickers, GIFs, or by sharing moments on Status. WhatsApp Business enables global customer outreach, facilitating sales growth and relationship building through showcasing products and services. Stay connected effortlessly with group chats for planning outings with friends or staying updated on family conversations.
2. Outline
ļ§ Overview
ļ§ Types of Network traffic in IBM Spectrum Scale environment
ļ§ Components
ļ§ Component traffic flow
ļ§ Other resources
2
3. Overview
ā
IBM Spectrum Scale, like any distributed software, depends highly on the network
infrastructure in order to operate efficiently.
ā
The goal of this presentation is to provide a description of the various network traffic
flows involved in the IBM Spectrum Scale environment.
ā
Network designers can use this information to design a suitable network for the IBM
Spectrum Scale environments
4. Types of network traff in IBM Speftrum Sfale environment
ā
In general, network traffic in IBM Spectrum Scale can be divided into 3 major types:
ā Data/Metadata Traffic:
This type of traffic is mostly around actual data movement between various IBM Spectrum Scale components. It might travel over TCP/IP or use RDMA depending on the network type and
protocol being used.
Examples include: NSD traffic, GNR synchronization, public NAS protocol traffic, AFM replication traffic , external pools traffic etc.
ā Control traffic:
This type of traffic consists of messages (usually small payloads) that are being used for control purposes. It includes both internal and external traffic (internal traffic involves control traffic
between IBM Spectrum Scale components, while external traffic involves control traffic between IBM Spectrum Scale and external entities).
Examples includes: Token traffic, disk leases traffic, encryption related key traffic, quota management traffic, monitoring traffic, CTDB traffic, configuration management traffic, authentication
traffic etc.
ā Adminitration traffic:
This type of traffic is around administrative operations that are being performed on an IBM Spectrum Scale environment
Examples includes: remote shell traffic, Web GUI traffic (external), REST administration traffic etc.
ā
Different components are using the above mentioned traffic type in different ways. Thus, we will try to describe each traffic type on a per
component basis
Note: In the last couple of years, many new features were introduced on top of the traditional GPFS product, which made IBM Spectrum Scale more
feature-rich. Many of those features required introducing new types of network flows.
5. IBM Speftrum Sfale fomponent - āCoreā
ā
The traditional GPFS is the core component of IBM Spectrum Scale.
ā
All the non ācoreā IBM Spectrum Scale components essentially use the ācoreā
components while enriching its capabilities ( access using standard protocols,
management capabilities etc.) . Thus, the flows discussed in this section apply to the other
components as well.
ā
Some traffic flows depend on the features being used (for example: encryption, quotas
etc.).
ā
Some traffic flows are affected by the configuration being used (for example: replication
traffic).
6. IBM Speftrum Sfale fomponents - āCoreā
Data traff
ā
The basic data traffic type in āCoreā environment is that data transfer between the NSD
servers and various clients (which might have other roles, like protocol nodes, TSM nodes,
AFM GW, TCT nodes etc.). In the context of this document, Data includes filesystem
metadata as well.
ā
Data traffic is using the NSD protocol (Network Shared Disk) over either TCP/IP and/or
RDMA (over Ethernet ā a.k.a. RoCE, over IB and over OPA). RDMA protocol is considered
more efficient, especially for large I/O sizes.
ā
The required characteristics of the data channel highly depend on the customer workload
(throughput oriented, latency oriented etc.). While metadata workload is usually latency-
sensitive.
ā
For sequential workloads, data messages size may be as large as the file system block size.
ā
When using IBM Spectrum Scale sync replication, the client is responsible for writing multiple
copies. Thus, it is required to take those into account as well (2X/3X writes for 2/3 replicas).
For FPO configurations, where enableRepWriteStream is enabled, the first NSD server will
write the third replica instead of the client.
ā
Other data paths, which are still considered ācoreā are the ones related to various features
around moving data out of/into āIBM Spectrum Scaleā environments:
ā AFM GW: Read/write (configuration dependent) data to/from external sources (either
remote GPFS cluster or generic NFS server)
ā Spectrum Protect HSM: When using Spectrum Protect as an external pool for ILM,
data will move to/from IBM Spectrum Scale to the IBM Spectrum Protect
infrastructure.
ā TCT: Transparent Cloud Tiering used as an external pool as well, data will be moved
to/from the external cloud/Object infrastructure
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
TSM
HSM
TSM
HSM
TSM
SRV
TCT
GW
TCT
GW
Cloud
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Remote Cluster
/NFS server
7. IBM Speftrum Sfale fomponents - āCoreā
Control traff
ā
Control traffic in IBM Spectrum Scale environments can
be divided into three major types:
ā Cluster āwell beingā - i.e. disk leases (āheartbeatā)
ā Data access or consistency management
(Tokens, quotas, allocation etc.)
ā Configuration or feature related (key
management, Zimon collectors, configuration
management, AFM related RPCs etc)
ā
Most of these types of communication are latency
sensitive (i.e. slow response time will affect
performance) however, some of them,especially the
āwell beingā ones are less latency sensitive, but highly
impact the availability of nodes in the cluster if message
delay exceeds specific values (e.g. heartbeat timeouts).
ā
Note: In most cases, a node can play multiple roles.
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimonZimon
8. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āCluster well beingā - a.k.a disk leases traffic
ā
In each cluster, the cluster manager manages the disk
leases. The cluster manager is automatically elected from
the list of āquorum nodesā.
ā
Each node (regardless of its role) needs to renew its lease
with the cluster manager server (dynamically elected) at
regular intervals (35 sec. By default).
ā
If a node fails to renew its lease, the cluster manager will
try to ping the āsuspected nodeā before expelling it.
ā
Another expel case might take place if a node canāt talk to
another node. In that case, it will ask the CL mgr to expel
the node. The CL MGR will decide which node needs to be
expelled
ā
All nodes in the local cluster (and potentially on remote
clusters) participate in the lease-renewal mechanism,
regardless of their role.
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimonZimon
9. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āData access/consistency managementā
ā
As a distributed filesystem, IBM Spectrum Scale needs some
mechanisms in order to assure data consistency as well as capacity
management. We describe those cases below:
ā Tokens:
Overview:
One of the basic constructs of IBM Spectrum Scale, as a parallel
fs, is the use of tokens in order to maintain consistency.
Essentially, in order to perform an operation on a filesystem object
(file/directory etc.) a relevant token
is required (for reading, writing, caching etc.).
Workflow:
At least one token server is automatically selected from the list of
āmanagerā nodes. The token servers manage the tokens, and
each client (regardless of its role) needs to contact a token server.
The token servers ādivideā the object responsibility between them
based on inode number, thus achieving load balancing and better
resiliency.
Network Requirements:
In general, token traffic is based on small RPC messages, which
are mostly latency sensitive (i.e. not throughput oriented). Slow
response time (due to latency, network congestion etc.) might
result in āhangsā or āhiccupsā from user/application perspective
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimonZimon
10. IBM Speftrum Sfale fomponents - āCoreā
Metadata traff in presenfe of fle sharing
āData access/consistency managementā
ā Metanode:
Overview:
A corner case in token management (in order to achieve better
performance) is the use of a āmetanodeā.
One of the challenges in multi-node access to the same object is
the ability to maintain metadata consistency. While in theory, the
common token system can be used, updating (or querying) object
metadata might result in many token āstealsā. Thus, we use a
slightly different mechanism, in which one node (usually the first
node that opened the object) will act as a āmetadata nodeā for that
object. Any other node that would like to update/query the objectās
metadata (inode fields) will talk to that node ā thus achieving
better scalability for multiple objects.The metanode will be the only
one updating the objectās metadata through the NSD server.
Workflow:
The main difference in workflow when considering the metanode
role is that there will be metadata traffic not only between NSD
clients and server, but also directly between different NSD clients.
Network Requirements:
Metanode traffic is similar to generic token traffic, i.e. latency-
sensitive for small messages.
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimonZimon
11. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āData access/consistency managementā
ā Filesystem allocation and quota:
Overview:
Managing block and inode allocation in highly distributed filesystem is a
big challenge. While there are similarities between standard allocation and
quotas, the main difference is that quotas operates at finer granularity
(there are separate quota for each user/group/fileset etc.) , Thus different
mechanisms are used that affect the communication impact on both.
Workflow:
For filesystem based block/inode allocation, a set of predefined allocation
regions are used. Thus, when a node needs block or inode allocation, it
just needs to be assigned with an available region. The communication
impact is relatively low until the region is used or āStolenā by another node.
The FS MGR manages the region allocation.
Quota is a bit more challenging as it requires much more frequent
messaging. Thus, an eventual consistent usage was implemented. Each
node will get a āshareā of blocks/inodes for the relevant quota entity
(autotuned post 4.1.1). The quota manager, running on the FS MGR, will
manage those shares assignment upon request.
Network Requirements:
Block/inode allocation is also based on small messages and thus are
latency sensitive. While regular allocation might not be affected by
slowness due to the region based approach, quotas allocation might affect
write/create performance.
In both cases, all nodes will contact the respective FS MGR in order to get
allocation regions and/or quota shares.
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimonZimon
12. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āConfiguration/feature relatedā
ā Key management:
Overview:
When IBM Spectrum Scale introduced the native
filesystem encryption, a new dependency (or
communication flow) was introduced, the ability to talk
to a key server (currently either SKLM or Vormetrics)
Workflow:
Since any node in the environment (either local or
remote cluster) might need to access encrypted data,
all nodes needs to have access to a key server.
While the nodes will cache the keys for some time
(configurable) not being able to fetch keys from the key
server might result in data access error
Network Requirements:
Key fetching from key server/s is mostly around small
messages to the designated set of servers. As
mentioned above, all nodes accessing encrypted data
need to have access to those servers
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimonZimon
13. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āConfiguration/feature relatedā
ā Performance collection (zimon):
Overview:
Starting with IBM Spectrum Scale 4.1.1 a performance collection
solution is included with the product. It is based on the zimon
technology. Zimon collects a variety of performance related metrics
to be later presented using either IBM Spectrum Scale GUI,
Grafana (using a bridge) or CLI (using the mmperfmon command)
Workflow:
Like most performance monitoring systems, zimon (starting with
IBM Spectrum Scale 4.2) can potentially use a federated structure
of reporters (every node) and collectors (specifically defined nodes).
Each node reports to its designated collector allowing the various
collectors to be queried for the collected data. The collectors know
about each other ā thus able to satisfy queries for performance data
stored on other collectors.
Network Requirements:
Zimon reporting is also based on relatively small messages, flowing
between monitored nodes to their designated collector. In order to
achieve HA one client might report to more then one collector.
That said, its important to note that zimon traffic is not in the ādata
pathā meaning, lack of ability to report in a timely fashion will result
in some performance data missing, but wouldnāt affect actual data
access (i.e. no user impact, only admin impact).
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimon
14. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āConfiguration/feature relatedā
ā Configuration Management (CCR):
Overview:
In the past, GPFS used two designated configuration server nodes to
manage the authoritative version of the cluster configuration. Starting
with version 4.1 a new Cluster Configuration Repository (CCR) was
introduced. Today, all quorum nodes act as a CCR nodes, meaning
that they manage the distributed configuration state of the cluster.
CCR also allows using multiple configuration files ā thus allowing IBM
Spectrum Scale configuration repository to manage much more info
(including protocols configuration, authentication configuration etc.).
Workflow:
In general, there are two major operations that can be done on
configuration that are being managed by CCR: update and query.
Both can be done from any node ā but will change (and distribute the
changes) to all CCR nodes in the cluster and potentially (depending
on the specific configuration file) to all cluster nodes.
On update, the whole file will be pushed to the CCR nodes ā and while
the file is usually small, it might grow to several MB on large clusters
Network Requirements:
CCR usually use small messages as well, from all cluster nodes.
Special nodes that utilize the CCR more then others (protocol nodes,
GNR nodes, GUI etc) might introduce more traffic then others.
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimon
15. IBM Speftrum Sfale fomponents - āCoreā
Control traff
āConfiguration/feature relatedā
ā AFM related RPCs:
Overview:
While the data traffic part of AFM was discussed earlier,
AFM also uses client to AFM gateway RPCs in order to
let the GW know about filesystem operations that need
to be replicated by the GW.
Workflow:
When a client node (any client that does I/O on an AFM
fileset) perform read/write operation on a filesystem
object ā it send special RPC to the AFM GW so the GW
can either add the specific op to the fileset queue (for
write/creates etc.) or validate/fetch that data on reads.
In order to perform the various operations, the AFM GW
node might require other type of traffic, data or control
in order to actually perform the required operations
Network Requirements:
The AFM related RPCs are small messages, as only
the op itself is being transmitted, not the data itself.
The traffic might flow to/from any node that access the
AFM fileset (on either local or remote clusters).
NSD
server
NSD
server
NSD
server
AFM
GW
AFM
GW
CL
MGR
FS
MGR
Zimon
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
Token
SRV
Meta
Node
AFM
GW
Token
SRV
Token
SRV
Meta
Node
Meta
Node
KLM
SRV
CCR
SRV
CCR
SRV
CCR
SRV
ZimonZimonZimon
16. IBM Speftrum Sfale fomponents - āCoreā
Administration traff
ā
Administration traffic in the core components can be divided into internal
traffic (i.e. traffic that is used to perform operations on the various cluster
nodes) and external traffic (traffic used by external entities in order to
ātalkā to components performing operations on their behalf).
ā
The first type of traffic mostly uses remote shell commands (ssh on
modern systems). Some parts of internal traffic also use special RPCs
that replaced the remote shell commands on older versions of GPFS (in
order to minimize the dependency on remote shell). Those operations
are mostly around distributing config files on new nodes, NSD creation
etc. Some customers are using dedicated āmanagementā network in
order to perform those (admin interface).
Unless otherwise defined (using the adminMode config option) any node
in the cluster might be the source of those operations, in most cases,
some other node will perform the actual operation on the requester
behalf (cluster manager/filesystem manager etc.).
ā
The second type (external administration) is mostly around web based
GUI access and/or the new REST interface to IBM Spectrum Scale
commands.
ā
Both operations are not in the data path and usually being used by
interactive commands (thus, extremely long delays will cause
complaints by administrators), but their performance requirements are
not critical.
NSD
server
NSD
server
NSD
server
GUI
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
REST
mgmt
External
Management
clients
CL
MGR
FS
MGR
CCR
SRV
CCR
SRV
CCR
SRV
18. IBM Speftrum Sfale fomponents - āProtofolsā
ā
The āprotocolsā component of IBM Spectrum scale provides the ability to access data stored on various
filesystems using standard protocols.
ā
Today, the supported protocols are: NFS, CIFS, Swift (Object), HDFS and iSCSI (boot only) .
ā
āprotocol nodesā are nodes that are running the various services. From IBM Spectrum Scale perspective
they are usually āclientā nodes (function wise, not licensing wise) ā meaning that they access the data as
any other IBM Spectrum Scale client node.
ā
Thus, on one hand they utilize all the data/control/management traffic as mentioned in the ācoreā section,
but on the other hand, uses other protocols data/control traffic (protocol dependent).
ā
Some protocols require special control protocol for cluster awareness (CIFS) and most protocols also
require authentication control traffic (usually to external authentication server/s).
19. IBM Speftrum Sfale fomponents - āProtofolsā
Data traff
ā
As mentioned earlier, protocol nodes are using
the core data traffic on their ābackendā in order
to talk to the NSD servers.
ā
On the āfrontendā protocol nodes are using the
relevant protocol data path.
ā
Depending on the access pattern and dataset
the traffic might be throughput oriented or
latency oriented.
ā
Some protocols might assign different roles to
different nodes (e.g. HDFS datanode vs.
namenode).
ā
It is advised not to use the same interfaces for
both IBM Spectrum Scale (backend) traffic and
protocols (frontend) traffic.
NSD
server
NSD
server
NSD
server
CIFS
server
CIFS
server
NFS
server
NFS
server
SWIFT
server
SWIFT
server
HDFS
Data
HDFS
Data
20. IBM Speftrum Sfale fomponents - āProtofolsā
Control traff
ā
As mentioned earlier, protocol nodes are
using the same control traffic on their
ābackendā as any IBM Spectrum Scale client
node.
ā
On the protocols specific part, there are both
common control traffic (usually authentication)
and potentially protocol specific control traffic:
ā Authentication: Most protocols (unless
using local authentication) will contact the
configured external authentication
servers in order to get authentication
related information. For the most part
those are relatively small, latency
sensitive operations
CIFS
server
CIFS
server
NFS
server
NFS
server
SWIFT
server
SWIFT
server
HDFS
Name
Auth
server
Auth
server
21. IBM Speftrum Sfale fomponents - āProtofolsā
Control traff
ā
Protocol Specific control traffic
ā NFS:
IBM Spectrum Scale uses Ganesha NFS
server for providing NFS services.
Back-end:
Ganesha doesnāt use any special internode
communication, it relies on GPFS as a
clustering mechanism.
Front-end:
For the most part, the NFS protocol is includes
mixed data and control traffic. Control traffic
on NFS includes locking (separate on V3 and
integrated on V4), delegation backchannel (V4
ā not yet supported) and rquota protocol.
All of the front end control data is mostly
around small messages somewhat latency
sensitive (application dependent).
NFS
server
NFS
server
NFS
server
locking
backchannel
rquota
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
NFS
client
22. IBM Speftrum Sfale fomponents - āProtofolsā
Control traff
ā
Protocol Specific control traffic
ā CIFS:
In order to provide CIFS services, IBM Spectrum
Scale is using the SAMBA software together
with the CTDB clustering software.
Back-end:
In order to manage the CIFS locking state inside
the cluster, CTDB needs to communicate
between the nodes. The traffic is based on
latency-sensitive small messages .
Front-end:
For the most part, the CIFS protocols mix both
data and control protocol ā so its a bit hard to
differentiate between them. Control traffic
includes locking and callback channel (i.e. when
the server contacts the client to break oplocks
etc.). Both largely depend on the specific
workload and dataset.
CIFS
server
CIFS
server
CIFS
server
locking
backchannel
rquota
NFS
client
NFS
client
NFS
client
NFS
client
CIFS
client
NFS
client
NFS
client
NFS
client
NFS
client
CIFS
client
NFS
client
NFS
client
NFS
client
NFS
client
CIFS
client
NFS
client
NFS
client
NFS
client
NFS
client
CIFS
client
CTDB
23. IBM Speftrum Sfale fomponents - āProtofolsā
Data traff
ā
Protocol Specific control traffic
ā SWIFT/S3:
In order to provide Object based services, IBM
Spectrum Scale is using OpenStack Swift software
with some enhancements in order to provide better
S3 support.
Back-end:
Currently, the only traffic on the backend network with
the object implementation is the same traffic as the
regular ācoreā components.
Front-end:
The object implementation is using several data and
control messages over the public network:
ā Data: With object protocol data, a client requests a read/write
operation to a protocol node over the front end (CES) network. With
traditional Swift mode, that request may be processed by the same
protocol node or passed to a different protocol node for processing.
With unified file and object access mode (Swift on File), the request is
always processed by the protocol node that received the client request.
The front end (CES) network is used for passing requests between
protocol nodes
SWIFT
server
SWIFT
server
SWIFT
server
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
Data
24. IBM Speftrum Sfale fomponents - āProtofolsā
Control traff
ā
Protocol Specific control traffic
ā SWIFT/S3 (Continued)
Control: Control traffic in Swift is mostly around
authentication traffic, and somewhat depends on the
chosen configuration:
ā
If the keystone authentication service is being provided
by the protocol nodes (the common case), then clients
will contact the keystone node in order to validate the
user and get security token.
When using Swift protocol, the server will cache the
token using memcached based on the token TTL, in
order to reduce the authentication traffic, when using S3,
the server will not cache any tokens.
ā
When clients contact a protocol node with their security
token, the node will contact the keystone server in order
to validate the token. In some cases, the keystone server
might be running on different node or on a 3rd party node
(when using external keystone).
ā Note: Currently, all the authentication traffic will take place
using the front-end (CES) IPs and not the back-end network
SWIFT
server
SWIFT
server
SWIFT
server
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
NFS
client
NFS
client
NFS
client
NFS
client
OBJ
client
keystone
server
AUTH
AUTH
25. IBM Speftrum Sfale fomponents - āProtofolsā
Control traff
ā
Protocol Specific control traffic
ā HDFS:
For Hadoop-like workload (or applications that are
using the DFS API in order to access data, IBM
Spectrum Scale implements the transparent Hadoop
connector. The control data is being managed by a
single ānamenodeā at a time, which provides the client
with info which data nodes to get the actual data from.
Back-end:
There is no special back-end data going on for the
hadoop transparent connector.
Front-end:
Today, a single node act as the namenode at any
given time. Clients will first contact this node, using the
standard HDFS RPCs in order to query location of
specific blocks.
As with most control protocols, those are small
messages, relatively latency sensitive (since actual
data is usually large, the latency might not be that
critical).
HDFS
name
HDFS
Data
HDFS
Data
NFS
client
NFS
client
NFS
client
NFS
client
HDFS
client
NFS
client
NFS
client
NFS
client
NFS
client
HDFS
client
NFS
client
NFS
client
NFS
client
NFS
client
HDFS
client
NFS
client
NFS
client
NFS
client
NFS
client
HDFS
client
27. IBM Speftrum Sfale fomponents - āGNRā
ā
GPFS Native Raid (a.k.a IBM Spectrum Scale RAID) is a software based RAID
implementation that is being used with several storage product, for example Elastic
Storage Server (ESS) from IBM.
ā
From ācoreā IBM Spectrum Scale perspective, GNR node are considered NSD servers
where ābelowā the NSD layer we have the GNR layer itself that takes care of the physical
disks management and data protection using declustered RAID technology.
ā
That said, the GNR layer itself present several new network related flows in order to
provide efficient write caching as well as monitoring of the partner node in each GNR
building block.
28. Main Data Flow
IBM Speftrum Sfale fomponents - āGNRā
Data traff
ā
As mentioned earlier, GNR nodes also act as NSD servers, thus
the main data flow that is going on with GNR is reading/writing
to/from clients. That part is discussed in the ācoreā component.
That said, the high performance that GNR nodes usually provide,
require adequate network bandwidth in order to really utilize GNR
nodes capabilities.
ā
Another special type of traffic we have with GNR is NVRAM
replication. Essentially, GNR uses the internal NVRAM on the
nodes in order to implement a ālogTipā, as an extremely fast small
write cache (technically, its an append to circular log) and for
internal GNR metadata operations. Since the NVRAM is internal to
each server, a special protocol (NSPD) is used in order to replicate
those operations into the building block āpartnerā. The replication is
based on small messages which are highly latency sensitive
(higher latency will impact small write performance). Under some
workload, it might make sense to separate this traffic using fast
interconnect (e.g. IB) in order to enhance the performance and
avoid the competition with the main data stream.
GNR
server
GNR
server
GNR
server
GNR
server
GNR
server
GNR
server
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSPD NSPD NSPD
29. Main Data Flow
IBM Speftrum Sfale fomponents - āGNRā
Control traff
ā
Much of GNRās control traffic is the same
as the rest of the ācoreā components.
That said, due to the special need to
manage large number of disks there are
some special control traffic going on in
the background in order to monitor each
building block partner.
ā
The nature of this traffic is around
relatively small messages, but since they
are performed in the background, latency
if not that critical for those.
GNR
server
GNR
server
GNR
server
GNR
server
GNR
server
GNR
server
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSD
client
NSPD NSPD NSPD