WebLogic Stability; Detect and Analyse Stuck ThreadsMaarten Smeets
Stuck threads are a major cause for stability issues of WebLogic Server environments. Often people in operations and development who are confronted with stuck threads, are at a loss what to do. In this presentation we will talk about what stuck threads actually are and how you can detect them. We will elaborate on how you can get to the root cause of a stuck thread and which tools can help you with that. In order to reduce the impact of having stuck threads in an application, we will talk about using workmanagers. In order to prevent stuck threads we will illustrate several patterns which can be implemented in infrastructure and applications. Next time you see a stuck thread, you will know what to do!
CEPH DAY BERLIN - MASTERING CEPH OPERATIONS: UPMAP AND THE MGR BALANCERCeph Community
This talk will introduce the ceph-mgr balancer and the placement group ""upmap"" features added in Luminous.||Experienced Ceph operators will learn practical methods to:| - achieve perfectly uniform OSD distributions| - painlessly migrate data between servers| - easily add capacity to clusters big or small| - transparently modify CRUSH rules or tunables without fear!|
Making Ceph awesome on Kubernetes with Rook - Bassam TabbaraCeph Community
Rook makes running Ceph storage on Kubernetes easy by extending Kubernetes with custom controllers and types. It automates the deployment, configuration, scaling, upgrading and management of Ceph clusters running within Kubernetes. Rook defines the desired state of the storage cluster and uses an operator to reconcile the actual cluster state with the desired state. This allows Ceph to leverage the full power of Kubernetes' services, deployments and APIs for managing stateful applications at scale.
Linux Block Cache Practice on Ceph BlueStore - Junxin ZhangCeph Community
This document discusses using Linux block caching with Ceph BlueStore. It explains that BlueStore can better utilize fast storage devices like SSDs compared to FileStore. It tested using Bcache and DM-writeboost to cache BlueStore data on HDDs using SSDs. Bcache performed better overall. Issues found were slow requests when caching and BlueStore used the same SSD, and inconsistency in SSD data management between BlueStore and the cache. Future work could have BlueStore control all raw disks and prioritize data saving to fast devices.
Using OpenStack in a traditional hosting environment posed scaling challenges that required automating provisioning across multiple data centers. OpenStack was chosen for its support, scalability, and ability to support future cloud offerings. Bluehost implemented optimizations like using MySQL slaves, custom schedulers, and replacing Qpid with ZeroMQ to address scalability issues with messaging, databases, and APIs under heavy load. The enhanced OpenStack deployment now supports over 10,000 physical servers being added daily.
As a service provider, Rackspace is constantly bringing new OpenStack capacity online. In this session, we will detail a myriad of challenges around adding new compute capacity. These include: planning, automation, organizational, quality assurance, monitoring, security, networking, integration, and more.
WebLogic Stability; Detect and Analyse Stuck ThreadsMaarten Smeets
Stuck threads are a major cause for stability issues of WebLogic Server environments. Often people in operations and development who are confronted with stuck threads, are at a loss what to do. In this presentation we will talk about what stuck threads actually are and how you can detect them. We will elaborate on how you can get to the root cause of a stuck thread and which tools can help you with that. In order to reduce the impact of having stuck threads in an application, we will talk about using workmanagers. In order to prevent stuck threads we will illustrate several patterns which can be implemented in infrastructure and applications. Next time you see a stuck thread, you will know what to do!
CEPH DAY BERLIN - MASTERING CEPH OPERATIONS: UPMAP AND THE MGR BALANCERCeph Community
This talk will introduce the ceph-mgr balancer and the placement group ""upmap"" features added in Luminous.||Experienced Ceph operators will learn practical methods to:| - achieve perfectly uniform OSD distributions| - painlessly migrate data between servers| - easily add capacity to clusters big or small| - transparently modify CRUSH rules or tunables without fear!|
Making Ceph awesome on Kubernetes with Rook - Bassam TabbaraCeph Community
Rook makes running Ceph storage on Kubernetes easy by extending Kubernetes with custom controllers and types. It automates the deployment, configuration, scaling, upgrading and management of Ceph clusters running within Kubernetes. Rook defines the desired state of the storage cluster and uses an operator to reconcile the actual cluster state with the desired state. This allows Ceph to leverage the full power of Kubernetes' services, deployments and APIs for managing stateful applications at scale.
Linux Block Cache Practice on Ceph BlueStore - Junxin ZhangCeph Community
This document discusses using Linux block caching with Ceph BlueStore. It explains that BlueStore can better utilize fast storage devices like SSDs compared to FileStore. It tested using Bcache and DM-writeboost to cache BlueStore data on HDDs using SSDs. Bcache performed better overall. Issues found were slow requests when caching and BlueStore used the same SSD, and inconsistency in SSD data management between BlueStore and the cache. Future work could have BlueStore control all raw disks and prioritize data saving to fast devices.
Using OpenStack in a traditional hosting environment posed scaling challenges that required automating provisioning across multiple data centers. OpenStack was chosen for its support, scalability, and ability to support future cloud offerings. Bluehost implemented optimizations like using MySQL slaves, custom schedulers, and replacing Qpid with ZeroMQ to address scalability issues with messaging, databases, and APIs under heavy load. The enhanced OpenStack deployment now supports over 10,000 physical servers being added daily.
As a service provider, Rackspace is constantly bringing new OpenStack capacity online. In this session, we will detail a myriad of challenges around adding new compute capacity. These include: planning, automation, organizational, quality assurance, monitoring, security, networking, integration, and more.
The document discusses an automatic operation bot for Ceph clusters at eBay. It describes monitoring the clusters with tools like Prometheus and node-exporter. When issues arise, over 95% are due to failed disks, which the bot would automatically remove and replace. It would record failures and perform remediation steps like removing OSDs, offline disks, and lighting indicator LEDs. The bot design includes components for monitoring, alerting, a task queue, and executor to automate common operations and reduce manual work.
The document summarizes Qihoo 360's experience deploying Ceph for storage at scale. They use Ceph RBD for virtual machine images and CephFS for a shared file system. For Ceph RBD, they have over 500 nodes across 30+ clusters storing over 1000 object storage devices. They use both full SSD and hybrid SSD/HDD clusters depending on performance needs. Their experience highlights best practices for deployment, performance, stability and operations. For CephFS, they evaluated metadata performance and discussed considerations for a production deployment.
A Performance Comparison of Container-based Virtualization Systems for MapRed...Marcelo Veiga Neves
This document evaluates the performance of container-based virtualization systems compared to native environments for MapReduce clusters. It finds that container systems like LXC and OpenVZ perform similarly to native environments for HDFS and NameNode evaluations based on common Hadoop benchmarks. The evaluation uses a Hadoop cluster of 4 nodes with 2 processors and 16GB RAM per node. Benchmark results show container systems achieve near-native performance for HDFS throughput tests and NameNode latency tests, indicating their viability for virtualizing MapReduce workloads with low performance overhead.
This document summarizes a transition from a traditional DFT and test generation approach to a 1687-based approach for a mixed-signal IC. The traditional flow uses multiple test buses and handcrafted analog tests, requiring months of effort. The 1687 approach aims to streamline testing by accessing analog blocks via scan chains, potentially reducing test time and increasing coverage. While specification coverage would remain the same, defect coverage may increase. Future benefits include reusable test libraries and improved quality for designs using shared IP blocks. A proof of concept simulation showed increased analog defect coverage through improved 1687 access.
Open Liberty is an open source lightweight Java runtime optimized for cloud-native applications. It provides a small footprint, high performance runtime with only the necessary APIs and features loaded. This allows for fast startup times, low memory usage, and continuous delivery. Open Liberty is optimized for containers and Kubernetes and provides zero migration between versions and platforms.
Ceph QoS: How to support QoS in distributed storage system - Taewoong KimCeph Community
This document discusses supporting quality of service (QoS) in distributed storage systems like Ceph. It describes how SK Telecom has contributed to QoS support in Ceph, including an algorithm called dmClock that controls I/O request scheduling according to administrator-configured policies. It also details an outstanding I/O-based throttling mechanism to measure and regulate load. Finally, it discusses challenges like queue depth that can be addressed by increasing the number of scheduling threads, and outlines plans to improve and test Ceph's QoS features.
Global deduplication for Ceph storage clusters can save up to 40% of total storage space by eliminating redundant data blocks. The document discusses two designs for implementing global deduplication in Ceph without adding a centralized metadata server: 1) using a double distribution hash to map data chunks to objects without redirection, and 2) storing deduplication metadata within self-contained objects to avoid complex linking between systems. It also describes the implementation of an extensible tier in Ceph for deduplication using object manifests and ongoing work to contribute these features to the upstream Ceph project. Remaining tasks are listed along with challenges around small chunk sizes and minimizing performance impacts.
Hadoop at Bloomberg:Medium data for the financial industryMatthew Hunt
Overview of big data systems origins, strengths, weaknesses, and uses at Bloomberg for solving "medium data" timeseries issues. Medium data requires modest clusters with consistently low latency and high availability, and Bloomberg has driven changes to core Hadoop components to address.
(WEB401) Optimizing Your Web Server on AWS | AWS re:Invent 2014Amazon Web Services
Tuning your EC2 web server will help you to improve application server throughput and cost-efficiency as well as reduce request latency. In this session we will walk through tactics to identify bottlenecks using tools such as CloudWatch in order to drive the appropriate allocation of EC2 and EBS resources. In addition, we will also be reviewing some performance optimizations and best practices for popular web servers such as Nginx and Apache in order to take advantage of the latest EC2 capabilities.
Bin repacking scheduling in virtualized datacentersFabien Hermenier
Fabien Hermenier, Sophie Demassey, and Xavier Lorca.
In Proceedings of the 17th international conference on Principles and practice of constraint programming (CP'11). Springer-Verlag, Berlin, Heidelberg, pages 27-41.
Peformance Evaluation of Container-based ViMiguel Xavier
This document evaluates the performance of container-based virtualization for high performance computing environments. It finds that container-based systems like LXC, OpenVZ, and Linux-Vserver achieve near-native CPU performance but Xen incurs a 4.3% overhead. For memory and disk, Xen has a 31% throughput overhead while container systems return unused resources. Network performance is worst for Xen due to virtualized drivers. Container systems isolate CPU well but show poor isolation for memory, disk, and network. Based on these results for HPC applications, LXC is the most suitable container-based system.
This document outlines an agenda for a conference on MySQL and Ceph storage solutions. The agenda includes sessions on MySQL performance on Ceph versus AWS, a head-to-head performance lab comparing the two platforms, and architectural considerations for optimizing MySQL on Ceph. Specific topics covered are MySQL and Ceph capabilities like live migration and snapshots, ensuring a consistent developer experience between private Ceph and public cloud, results from sysbench tests showing Ceph can match or exceed AWS performance on price per IOPS, and how Ceph node configuration like CPU cores and flash storage affect MySQL workload performance.
The document discusses reliability guarantees in Apache Kafka. It explains that Kafka provides reliability through replication of data across multiple brokers. As long as the minimum number of in-sync replicas (ISRs) is maintained, messages will not be lost even if individual brokers fail. It also discusses best practices for producers and consumers to ensure data is not lost such as using acks=all for producers, disabling unclean leader election, committing offsets only after processing is complete, and monitoring for errors, lag and reconciliation of message counts.
Combining Cloud Native & PaaS: Building a Fully Managed Application Platform ...DigitalOcean
DigitalOcean App Platform is a managed Platform as a Service (PaaS) that abstracts infrastructure and encodes cloud-native best practices. It provides declarative specifications for deploying stateless web applications from Git repositories with continuous delivery. Apps benefit from built-in services like auto-scaling, ingress routing, CDN, monitoring and more. The platform is built on Kubernetes and cloud-native technologies to provide a fully managed environment for deploying 12-factor apps.
This document outlines an agenda for a session on running MySQL on Ceph storage. The first part will discuss using MySQL on Ceph versus AWS and include a performance head-to-head. The second part will cover Ceph architecture including components like RADOS, pools, and CRUSH algorithm for data placement. The final part will discuss tuning MySQL and Ceph together for optimal performance including adjusting buffer pool size, transaction flushing, and creating specialized pools for IOPS workloads. An accompanying lab will compare MySQL performance on Ceph versus other cloud platforms.
Apache hadoop 3.x state of the union and upgrade guidance - Strata 2019 NYWangda Tan
The document discusses Apache Hadoop 3.x updates and provides guidance for upgrading to Hadoop 3. It covers community updates, features in YARN, Submarine, HDFS, and Ozone. Release plans are outlined for Hadoop, Submarine, and upgrades from Hadoop 2 to 3. Express upgrades are recommended over rolling upgrades for the major version change. The session summarizes that Hadoop 3 is an eagerly awaited release with many successful production uses, and that now is a good time for those not yet upgraded.
Real-Time Inverted Search NYC ASLUG Oct 2014Bryan Bende
Building real-time notification systems is often limited to basic filtering and pattern matching against incoming records. Allowing users to query incoming documents using Solr’s full range of capabilities is much more powerful. In our environment we needed a way to allow for tens of thousands of such query subscriptions, meaning we needed to find a way to distribute the query processing in the cloud. By creating in-memory Lucene indices from our Solr configuration, we were able to parallelize our queries across our cluster. To achieve this distribution, we wrapped the processing in a Storm topology to provide a flexible way to scale and manage our infrastructure. This presentation will describe our experiences creating this distributed, real-time inverted search notification framework.
Cloud Computing: Safe Haven from the Data Deluge? AGBT 2011Toby Bloom
This document discusses moving a genomic data analysis pipeline to the cloud. It provides context on cloud computing and describes an experiment to port an Illumina sequencing analysis pipeline to Amazon Web Services. The document outlines challenges in moving to the cloud like data transfer speeds, efficient resource utilization, and security concerns for controlled-access data. It also compares performance and costs between running analysis locally versus in the cloud. While the cloud poses challenges, the document concludes it is a viable option especially for small centers and large collaborations, though costs can be difficult to predict.
The document discusses an automatic operation bot for Ceph clusters at eBay. It describes monitoring the clusters with tools like Prometheus and node-exporter. When issues arise, over 95% are due to failed disks, which the bot would automatically remove and replace. It would record failures and perform remediation steps like removing OSDs, offline disks, and lighting indicator LEDs. The bot design includes components for monitoring, alerting, a task queue, and executor to automate common operations and reduce manual work.
The document summarizes Qihoo 360's experience deploying Ceph for storage at scale. They use Ceph RBD for virtual machine images and CephFS for a shared file system. For Ceph RBD, they have over 500 nodes across 30+ clusters storing over 1000 object storage devices. They use both full SSD and hybrid SSD/HDD clusters depending on performance needs. Their experience highlights best practices for deployment, performance, stability and operations. For CephFS, they evaluated metadata performance and discussed considerations for a production deployment.
A Performance Comparison of Container-based Virtualization Systems for MapRed...Marcelo Veiga Neves
This document evaluates the performance of container-based virtualization systems compared to native environments for MapReduce clusters. It finds that container systems like LXC and OpenVZ perform similarly to native environments for HDFS and NameNode evaluations based on common Hadoop benchmarks. The evaluation uses a Hadoop cluster of 4 nodes with 2 processors and 16GB RAM per node. Benchmark results show container systems achieve near-native performance for HDFS throughput tests and NameNode latency tests, indicating their viability for virtualizing MapReduce workloads with low performance overhead.
This document summarizes a transition from a traditional DFT and test generation approach to a 1687-based approach for a mixed-signal IC. The traditional flow uses multiple test buses and handcrafted analog tests, requiring months of effort. The 1687 approach aims to streamline testing by accessing analog blocks via scan chains, potentially reducing test time and increasing coverage. While specification coverage would remain the same, defect coverage may increase. Future benefits include reusable test libraries and improved quality for designs using shared IP blocks. A proof of concept simulation showed increased analog defect coverage through improved 1687 access.
Open Liberty is an open source lightweight Java runtime optimized for cloud-native applications. It provides a small footprint, high performance runtime with only the necessary APIs and features loaded. This allows for fast startup times, low memory usage, and continuous delivery. Open Liberty is optimized for containers and Kubernetes and provides zero migration between versions and platforms.
Ceph QoS: How to support QoS in distributed storage system - Taewoong KimCeph Community
This document discusses supporting quality of service (QoS) in distributed storage systems like Ceph. It describes how SK Telecom has contributed to QoS support in Ceph, including an algorithm called dmClock that controls I/O request scheduling according to administrator-configured policies. It also details an outstanding I/O-based throttling mechanism to measure and regulate load. Finally, it discusses challenges like queue depth that can be addressed by increasing the number of scheduling threads, and outlines plans to improve and test Ceph's QoS features.
Global deduplication for Ceph storage clusters can save up to 40% of total storage space by eliminating redundant data blocks. The document discusses two designs for implementing global deduplication in Ceph without adding a centralized metadata server: 1) using a double distribution hash to map data chunks to objects without redirection, and 2) storing deduplication metadata within self-contained objects to avoid complex linking between systems. It also describes the implementation of an extensible tier in Ceph for deduplication using object manifests and ongoing work to contribute these features to the upstream Ceph project. Remaining tasks are listed along with challenges around small chunk sizes and minimizing performance impacts.
Hadoop at Bloomberg:Medium data for the financial industryMatthew Hunt
Overview of big data systems origins, strengths, weaknesses, and uses at Bloomberg for solving "medium data" timeseries issues. Medium data requires modest clusters with consistently low latency and high availability, and Bloomberg has driven changes to core Hadoop components to address.
(WEB401) Optimizing Your Web Server on AWS | AWS re:Invent 2014Amazon Web Services
Tuning your EC2 web server will help you to improve application server throughput and cost-efficiency as well as reduce request latency. In this session we will walk through tactics to identify bottlenecks using tools such as CloudWatch in order to drive the appropriate allocation of EC2 and EBS resources. In addition, we will also be reviewing some performance optimizations and best practices for popular web servers such as Nginx and Apache in order to take advantage of the latest EC2 capabilities.
Bin repacking scheduling in virtualized datacentersFabien Hermenier
Fabien Hermenier, Sophie Demassey, and Xavier Lorca.
In Proceedings of the 17th international conference on Principles and practice of constraint programming (CP'11). Springer-Verlag, Berlin, Heidelberg, pages 27-41.
Peformance Evaluation of Container-based ViMiguel Xavier
This document evaluates the performance of container-based virtualization for high performance computing environments. It finds that container-based systems like LXC, OpenVZ, and Linux-Vserver achieve near-native CPU performance but Xen incurs a 4.3% overhead. For memory and disk, Xen has a 31% throughput overhead while container systems return unused resources. Network performance is worst for Xen due to virtualized drivers. Container systems isolate CPU well but show poor isolation for memory, disk, and network. Based on these results for HPC applications, LXC is the most suitable container-based system.
This document outlines an agenda for a conference on MySQL and Ceph storage solutions. The agenda includes sessions on MySQL performance on Ceph versus AWS, a head-to-head performance lab comparing the two platforms, and architectural considerations for optimizing MySQL on Ceph. Specific topics covered are MySQL and Ceph capabilities like live migration and snapshots, ensuring a consistent developer experience between private Ceph and public cloud, results from sysbench tests showing Ceph can match or exceed AWS performance on price per IOPS, and how Ceph node configuration like CPU cores and flash storage affect MySQL workload performance.
The document discusses reliability guarantees in Apache Kafka. It explains that Kafka provides reliability through replication of data across multiple brokers. As long as the minimum number of in-sync replicas (ISRs) is maintained, messages will not be lost even if individual brokers fail. It also discusses best practices for producers and consumers to ensure data is not lost such as using acks=all for producers, disabling unclean leader election, committing offsets only after processing is complete, and monitoring for errors, lag and reconciliation of message counts.
Combining Cloud Native & PaaS: Building a Fully Managed Application Platform ...DigitalOcean
DigitalOcean App Platform is a managed Platform as a Service (PaaS) that abstracts infrastructure and encodes cloud-native best practices. It provides declarative specifications for deploying stateless web applications from Git repositories with continuous delivery. Apps benefit from built-in services like auto-scaling, ingress routing, CDN, monitoring and more. The platform is built on Kubernetes and cloud-native technologies to provide a fully managed environment for deploying 12-factor apps.
This document outlines an agenda for a session on running MySQL on Ceph storage. The first part will discuss using MySQL on Ceph versus AWS and include a performance head-to-head. The second part will cover Ceph architecture including components like RADOS, pools, and CRUSH algorithm for data placement. The final part will discuss tuning MySQL and Ceph together for optimal performance including adjusting buffer pool size, transaction flushing, and creating specialized pools for IOPS workloads. An accompanying lab will compare MySQL performance on Ceph versus other cloud platforms.
Apache hadoop 3.x state of the union and upgrade guidance - Strata 2019 NYWangda Tan
The document discusses Apache Hadoop 3.x updates and provides guidance for upgrading to Hadoop 3. It covers community updates, features in YARN, Submarine, HDFS, and Ozone. Release plans are outlined for Hadoop, Submarine, and upgrades from Hadoop 2 to 3. Express upgrades are recommended over rolling upgrades for the major version change. The session summarizes that Hadoop 3 is an eagerly awaited release with many successful production uses, and that now is a good time for those not yet upgraded.
Real-Time Inverted Search NYC ASLUG Oct 2014Bryan Bende
Building real-time notification systems is often limited to basic filtering and pattern matching against incoming records. Allowing users to query incoming documents using Solr’s full range of capabilities is much more powerful. In our environment we needed a way to allow for tens of thousands of such query subscriptions, meaning we needed to find a way to distribute the query processing in the cloud. By creating in-memory Lucene indices from our Solr configuration, we were able to parallelize our queries across our cluster. To achieve this distribution, we wrapped the processing in a Storm topology to provide a flexible way to scale and manage our infrastructure. This presentation will describe our experiences creating this distributed, real-time inverted search notification framework.
Cloud Computing: Safe Haven from the Data Deluge? AGBT 2011Toby Bloom
This document discusses moving a genomic data analysis pipeline to the cloud. It provides context on cloud computing and describes an experiment to port an Illumina sequencing analysis pipeline to Amazon Web Services. The document outlines challenges in moving to the cloud like data transfer speeds, efficient resource utilization, and security concerns for controlled-access data. It also compares performance and costs between running analysis locally versus in the cloud. While the cloud poses challenges, the document concludes it is a viable option especially for small centers and large collaborations, though costs can be difficult to predict.
This document summarizes the DevoFlow paper, which proposes techniques to scale flow management for high-performance networks. It finds that per-flow management in OpenFlow introduces high overheads. DevoFlow aims to balance network control, statistics collection, and switch overhead by devolving most flow control to switches while maintaining partial visibility of significant flows. Simulation results show DevoFlow can reduce flow scheduling overheads compared to per-flow control, while still achieving high performance.
Simulation of Heterogeneous Cloud InfrastructuresCloudLightning
During the last years, except from the traditional CPU based hardware servers, hardware accelerators are widely used in various HPC application areas. More specifically, Graphics Processing Units (GPUs), Many Integrated Cores (MICs) and Field-Programmable Gate Arrays (FPGAs) have shown a great potential in HPC and have been widely mobilised in supercomputing and in HPC-Clouds. This presentation focuses on the development of a cloud simulation framework that supports hardware accelerators. The design and implementation of the framework are also discussed.
This presentation was given by Dr. Konstantinos Giannoutakis (CERTH) at the CloudLightning Conference on 11th April 2017.
A Performance Comparison of Container-based Virtualization Systems for MapRed...Miguel Xavier
The document compares the performance of container-based virtualization systems (LXC, OpenVZ, Linux-Vserver) to native performance in MapReduce clusters. It finds that the container systems perform similarly to native environments for HDFS and NameNode evaluations. For HDFS testing, all container systems achieved throughputs within 3Mbps of native performance. For NameNode latency testing via NNBench, all systems responded effectively like native with average latencies between 48-56ms. The evaluation shows container virtualization is a viable option for MapReduce clusters without significant performance penalties.
Real-time Inverted Search in the Cloud Using Lucene and Stormlucenerevolution
Building real-time notification systems is often limited to basic filtering and pattern matching against incoming records. Allowing users to query incoming documents using Solr's full range of capabilities is much more powerful. In our environment we needed a way to allow for tens of thousands of such query subscriptions, meaning we needed to find a way to distribute the query processing in the cloud. By creating in-memory Lucene indices from our Solr configuration, we were able to parallelize our queries across our cluster. To achieve this distribution, we wrapped the processing in a Storm topology to provide a flexible way to scale and manage our infrastructure. This presentation will describe our experiences creating this distributed, real-time inverted search notification framework.
Elasticsearch Sharding Strategy at Tubular LabsTubular Labs
- The document discusses Tubular Labs' sharding strategy for their Elasticsearch clusters which include 3 search clusters, 1 autocomplete cluster, and 1 Elastic Stack cluster.
- They conducted repeatable experiments using Rally to help determine the optimal shard size and number of shards per node. Tests were run against their 2.5 billion document, 4TB production cluster which was CPU intensive.
- The results showed that query performance dropped as the number of shards per node increased. However, loading the cluster more fully in testing yielded better results than their full production cluster, revealing new questions around load distribution and bottlenecks.
This document summarizes the approval of a seminar titled "Implementation of Advance High performance Bus using verilog" presented by Nirav Desai for the degree of Master of Technology. It lists the examiners and is signed by the supervisor, head of department, and includes the date and place.
The next sections include a declaration signed by Nirav Desai about original work and adherence to academic honesty. An acknowledgment section thanks the seminar guide and head of department for their support and guidance.
The abstract provides a high-level overview, stating that the purpose is to propose a scheme to implement an AMBA bus protocol specification using Verilog. It will cover bus basics, AMBA bus
Training Slides: Intermediate 202: Performing Cluster Maintenance with Zero-D...Continuent
Join us for this intermediate training session as we explore how to leverage the power of the Tungsten Clustering to perform database and OS maintenance with zero-downtime. This training is for anyone new to Continuent without prior experience, but will also serve as a wonderful refresher for any current users. Basic MySQL knowledge is assumed.
AGENDA
- Review the cluster architecture
- Describe the rolling maintenance process
- Explore what happens during a master switch
- Discuss cluster states
- Demonstrate rolling maintenance
- Re-cap commands and resources used during the demo
CSense: A Stream-Processing Toolkit for Robust and High-Rate Mobile Sensing A...Farley Lai
The document describes CSense, a stream-processing toolkit for building robust and high-performance mobile sensing applications on Android devices. CSense addresses challenges like concurrency, resource limitations, and high frame rates. It provides a programming model based on stream flow graphs, a compiler for optimization and code generation, and an efficient runtime. Evaluation shows CSense improves throughput by 19x and reduces CPU usage by 45% compared to a naive Java implementation, with low overhead for a variety of mobile sensing applications.
This document provides an overview of an Amazon EKS hands-on workshop. It introduces the workshop agenda which includes deploying example microservices, logging with Elasticsearch Fluentd and Kibana, monitoring with Prometheus and Grafana, and continuous integration/continuous delivery using GitOps with Weave Flux. Key concepts covered are Kubernetes pods, services, deployments, container networking with CNI plugins, observability tools, and CI/CD approaches.
The document proposes an operation zone based load balancer to improve user responsiveness on multicore embedded systems. It aims to reduce the costs of frequent task migration by the existing load balancers. The proposed approach divides the CPU utilization range into three zones - cold, warm and hot. The load balancer operates less frequently in the cold zone and more frequently in the hot zone, with intermediate behavior in the warm zone. Evaluation shows the approach reduces scheduling latency compared to CPU affinity based and non-affinity based systems under stress tests.
High Speed Design Closure Techniques-Balachander KrishnamurthyMassimo Talia
Digital electronics and electronics are not only theory as many Italians are thinking. The investments in electronic design in Italy are very low, since there's the Asian market which specialized their people to the Electronics culture. So The italian electronic engineers are compared to Sciencists, but they're designers of Manifacturing. This webinar describes the main steps and techniques, in order to design a Digital Circuits, to evaluate the timing constraints and the hardware requirements. The webinar is promoted by Xilinx.
This presentation gives an introduction to analysing ChIP-seq data and is part of a bioinformatics workshop. The accompanying websites are available at http://sschmeier.github.io/bioinf-workshop/#!galaxy-chipseq/
This document provides an overview of a lecture on analyzing pathogens using BEAST (Bayesian Evolutionary Analysis Sampling Trees). The lecture covers what makes pathogens special from an evolutionary perspective, an introduction to Bayesian analysis and Markov chain Monte Carlo (MCMC) methods, an overview of the BEAST software package and its components, and a demonstration of running a BEAST analysis. The lecture discusses building phylogenetic trees incorporating sample time data and estimating parameters like substitution rates and population dynamics from molecular sequences using Bayesian methods in BEAST.
Scan design is currently the most popular structured DFT approach. It is implemented by Connecting selected storage elements present in the design into multiple shift registers, called Scan chains.
Scannability Rules -->
The tool perform basic two check
1) It ensures all the defined clocks including set/Reset are at their off-states, the sequential element remain stable and inactive. (S1)
2) It ensures for each defined clocks can capture data when all other defined clocks are off. (S2)
This document discusses key concepts for distributed systems and cloud architecture. It covers strategies for distributing and scaling systems, such as using load balancers and auto-scaling. It also discusses ways to stabilize systems and prevent failures, such as circuit breakers and request throttling. The document then reviews best practices for deploying changes and handling failures. Finally, it provides an overview of considerations for scaling persistence layers, such as data modeling for NoSQL databases and global replication strategies.
SOC-CH3.pptSOC ProcessorsSOC Processors Used in SOC Used in SOCSnehaLatha68
This document discusses processor design, including selecting a processor core, designing the processor pipeline, buffer design, and dealing with branches. It describes in-order pipelines, techniques for branch prediction like branch target buffers and dynamic prediction, and buffer sizing approaches for maximum and mean rates. More robust processors like vector, VLIW, and superscalar designs are also summarized.
Similar to Use of a Levy Distribution for Modeling Best Case Execution Time Variation (20)
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
"Choosing proper type of scaling", Olena SyrotaFwdays
Imagine an IoT processing system that is already quite mature and production-ready and for which client coverage is growing and scaling and performance aspects are life and death questions. The system has Redis, MongoDB, and stream processing based on ksqldb. In this talk, firstly, we will analyze scaling approaches and then select the proper ones for our system.
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor IvaniukFwdays
At this talk we will discuss DDoS protection tools and best practices, discuss network architectures and what AWS has to offer. Also, we will look into one of the largest DDoS attacks on Ukrainian infrastructure that happened in February 2022. We'll see, what techniques helped to keep the web resources available for Ukrainians and how AWS improved DDoS protection for all customers based on Ukraine experience
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/temporal-event-neural-networks-a-more-efficient-alternative-to-the-transformer-a-presentation-from-brainchip/
Chris Jones, Director of Product Management at BrainChip , presents the “Temporal Event Neural Networks: A More Efficient Alternative to the Transformer” tutorial at the May 2024 Embedded Vision Summit.
The expansion of AI services necessitates enhanced computational capabilities on edge devices. Temporal Event Neural Networks (TENNs), developed by BrainChip, represent a novel and highly efficient state-space network. TENNs demonstrate exceptional proficiency in handling multi-dimensional streaming data, facilitating advancements in object detection, action recognition, speech enhancement and language model/sequence generation. Through the utilization of polynomial-based continuous convolutions, TENNs streamline models, expedite training processes and significantly diminish memory requirements, achieving notable reductions of up to 50x in parameters and 5,000x in energy consumption compared to prevailing methodologies like transformers.
Integration with BrainChip’s Akida neuromorphic hardware IP further enhances TENNs’ capabilities, enabling the realization of highly capable, portable and passively cooled edge devices. This presentation delves into the technical innovations underlying TENNs, presents real-world benchmarks, and elucidates how this cutting-edge approach is positioned to revolutionize edge AI across diverse applications.
Nordic Marketo Engage User Group_June 13_ 2024.pptx
Use of a Levy Distribution for Modeling Best Case Execution Time Variation
1. Use of a Levy Distribution for Modeling
Best Case Execution Time Variation
Jonathan Beard, Roger Chamberlain
SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Work also supported by:
1
6. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Streaming Languages
StreamIt, Auto-Pipe, Brook, Cg, S-Net,
Scala-Pipe, Streams-C and
many others
5
7. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Optimization
Slow
Fast Kernel
Super Fast
Medium
6
8. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Optimization
Kernel 1
Kernel 2
Kernel 3
Kernel 2
multi-core A
1 2
3 4
multi-core B
1 2
3 4
More allocation choices,
NUMA node A or B to
allocate stream.
7
9. 1 2
SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Optimization
Kernel 1
Kernel 2
Kernel 3
Kernel 2
multi-core A
1 2
3 4
multi-core B
1 2
3 4
More allocation choices,
NUMA node A or B to
allocate stream.
7
10. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Optimization
Kernel 1
Kernel 2
Kernel 3
Kernel 2
multi-core A
1 2
3 4
multi-core B
1 2
3 4
More allocation choices,
NUMA node A or B to
allocate stream.
1 2
7
11. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Optimization
A B C
“Stream” is modeled as a Queue
A Q1 B Q2 C
8
12. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Optimization
A B C
“Stream” is modeled as a Queue
A Q1 B Q2 C
8
13. We want good models for streaming systems
on shared multi-core systems (i.e., a cluster)
Problem: Accurate measurement is very difficult. Is there
a way to decide on a model without it.
• Commodity multi-core timer availability and latency
• Frequency scaling and core migration
• Measuring modifies the application behavior
SBS
Streaming on Multi-core Systems
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
9
14. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Derived Information
Expected Observed
10
15. SBS
Is there a pattern of minimal variation within the
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Derived Information
Expected Observed
systems we’re running on?
Avg. Service Time = E[ X ] + Error
10
16. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Goal
Find a distribution that characterizes
the minimum expected variation of a
hardware and software system
Use this characterization to
accept or reject models
11
17. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Process
12
• Measurement!
• Workload definition!
• Find a distribution!
• Utilize the distribution to aid model selection
18. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Timer Mechanism
Timer Thread Code
13
Ask for Time
Receive Time
19. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Timer Mechanism
Timer Thread
rdtsc clock_gettime
14
• x86 assembly
• varying methods
to serialize
• relatively fast
• multiple drift
issues
• POSIX standard
• relatively accurate
• portable
• slower than rdtsc
20. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Two Timing Choices
15
21. SBS
NUMA Node Variations
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
16
22. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Minimize Variation
• Restricting timer to single core
!
• Use the x86 rdtsc instruction with processor
recommended serializers for each processor
type
!
• Keeping processes under test on the same
NUMA node as timer
!
• Run timer thread with altered priority to
minimize core context swaps
17
23. SBS
Best Case Execution Time Variation
• no-op instruction implemented in most processors
!
• usually takes exactly 1 cycle
!
• no real functional units are involved, so least
taxing
!
• variation observed in execution time should be
external to process
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
18
24. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Data Collection
• no-op loops calibrated for various nominal
times, tied to a single core and run
thousands of times
!
• Execution time measured end to end for
each run, environment collected
!
• Parameters include:
Number of processes executing on core
Number of context swaps (voluntary,
involuntary)
Many others
19
25. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Levy Distribution
20
Execution Time Error
( obs - mean )
26. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Levy Distribution
21
Normal Distribution
27. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Levy Distribution
22
Gumbel Distribution
28. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Levy Distribution
23
Levy Distribution
29. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Levy Distribution
23
Levy Distribution
30. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Levy Distribution
• Truncation enables mean calculation, but
requires fitting to each dataset to find where
to truncate
!
• The truncation parameters are correlated to
both the number of processes per core and
the expected execution time
!
• Roughly linear relationship gives an
approximate solution to truncation
parameters without refitting
24
32. Hypothesis: Lower Kullback-Leibler (KL) divergence
SBS
Question: Can we use an M/M/1 queueing model to
estimate the mean queue occupancy of this system?
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Test Setup
A Q1 B
!
between expected and realized distribution is
associated with higher model accuracy.
26
33. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Test Setup
A Q1 B
1. Dedicated thread of execution monitors
27
queue occupancy
2. Calculate the estimated mean queue
occupancy using the M/M/1 model
3. Calculate KL Divergence for the arrival
process distribution using the truncated
Levy distribution noise model
34. SBS
Convolution with Exponential
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
28
35. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Conclusions
• The truncated Levy distribution can be used to
approximate BCETV
!
• The distribution of BCETV can be used as a tool
to accept or reject a stochastic queueing model
based on distributional assumptions
!
• KL divergence between the expected and
convolved distribution highly correlates with
queue model accuracy
29
36. SBS
Stream Based
Supercomputing Lab
http://sbs.wustl.edu
Parting Notes
Slides available here:
sbs.wust.edu
!
Timer C++ template code:
http://goo.gl/ItJ3jP
!
Test harness used to collect data:
http://goo.gl/U1VG6N
30