A presentation about COCOMA, a framework for COntrolled COntentious and MAlicious patterns, presented at MERMAT, 2nd International Workshop on Measurement-based Experimental Research, Methodology and Tools, FIA 2013, Dublin, Ireland
Taming Latency: Case Studies in MapReduce Data AnalyticsEMC
This session discusses how to achieve low latency in MapReduce data analysis, with various industrial and academic case studies. These illustrate various improvements on MapReduce for squeezing out latency from whole data processing stack, covering batch-mode MapReduce system, as well as stream processing systems. This session also introduces our BoltMR project efforts on this topic and discloses some interesting benchmark results.
Objective 1: Understand why low-latency matters for many MapReduce-based big data analytics scenarios.
After this session you will be able to:
Objective 2: Learn the root causes of MapReduce latency, the obstacles to lowering the latency and the various (im)mature solutions.
Objective 3: Understand the extent of MapReduce low-latency that is needed for their own applications and which optimization techniques are potentially applicable.
This document proposes a new distributed processing framework called AROM. AROM uses a data flow graph (DFG) model where jobs are represented as directed acyclic graphs. It introduces several improvements over traditional MapReduce programming. The framework is implemented using Scala and uses asynchronous message passing between operators to allow for pipelining and more flexible job definitions. Evaluation shows AROM achieves better performance than MapReduce for certain types of jobs. Future work aims to improve scalability and add dynamic scheduling capabilities.
Slides for a general webinar about BonFIRE, the features offered, the sites making up this multi-site testbed and the tools available for experimenters using the facility.
A video with audio is available on YouTube: http://youtu.be/0ulgvs32wvI
Este documento presenta una breve historia de los virus informáticos más importantes desde 1971 hasta 2008. Explica virus tempranos como Creeper y Core War, así como virus más recientes y dañinos como I Love You, Sasser, Blaster, Nimda, Melissa y Conficker, describiendo sus características y el impacto que tuvieron. El documento también menciona al precursor de los virus informáticos, el juego Core War creado en 1959.
The BonFIRE architecture was presented at the TridentCom Conference. These are the slides for the paper, which describes the key components and principles of the architecture and also some specific features offered to experimenter that are not available elsewhere.
Este documento describe los dialers, programas que realizan llamadas telefónicas sin el consentimiento del usuario, aumentando sus costos. Explica que los dialers causan daños al conectar el PC a servicios de pago sin aviso y usar números de tarifa alta. También proporciona formas de prevenir, eliminar y atacar los dialers, como usar programas anti-marcador, actualizaciones de software y antivirus.
http://inarocket.com
Learn BEM fundamentals as fast as possible. What is BEM (Block, element, modifier), BEM syntax, how it works with a real example, etc.
Taming Latency: Case Studies in MapReduce Data AnalyticsEMC
This session discusses how to achieve low latency in MapReduce data analysis, with various industrial and academic case studies. These illustrate various improvements on MapReduce for squeezing out latency from whole data processing stack, covering batch-mode MapReduce system, as well as stream processing systems. This session also introduces our BoltMR project efforts on this topic and discloses some interesting benchmark results.
Objective 1: Understand why low-latency matters for many MapReduce-based big data analytics scenarios.
After this session you will be able to:
Objective 2: Learn the root causes of MapReduce latency, the obstacles to lowering the latency and the various (im)mature solutions.
Objective 3: Understand the extent of MapReduce low-latency that is needed for their own applications and which optimization techniques are potentially applicable.
This document proposes a new distributed processing framework called AROM. AROM uses a data flow graph (DFG) model where jobs are represented as directed acyclic graphs. It introduces several improvements over traditional MapReduce programming. The framework is implemented using Scala and uses asynchronous message passing between operators to allow for pipelining and more flexible job definitions. Evaluation shows AROM achieves better performance than MapReduce for certain types of jobs. Future work aims to improve scalability and add dynamic scheduling capabilities.
Slides for a general webinar about BonFIRE, the features offered, the sites making up this multi-site testbed and the tools available for experimenters using the facility.
A video with audio is available on YouTube: http://youtu.be/0ulgvs32wvI
Este documento presenta una breve historia de los virus informáticos más importantes desde 1971 hasta 2008. Explica virus tempranos como Creeper y Core War, así como virus más recientes y dañinos como I Love You, Sasser, Blaster, Nimda, Melissa y Conficker, describiendo sus características y el impacto que tuvieron. El documento también menciona al precursor de los virus informáticos, el juego Core War creado en 1959.
The BonFIRE architecture was presented at the TridentCom Conference. These are the slides for the paper, which describes the key components and principles of the architecture and also some specific features offered to experimenter that are not available elsewhere.
Este documento describe los dialers, programas que realizan llamadas telefónicas sin el consentimiento del usuario, aumentando sus costos. Explica que los dialers causan daños al conectar el PC a servicios de pago sin aviso y usar números de tarifa alta. También proporciona formas de prevenir, eliminar y atacar los dialers, como usar programas anti-marcador, actualizaciones de software y antivirus.
http://inarocket.com
Learn BEM fundamentals as fast as possible. What is BEM (Block, element, modifier), BEM syntax, how it works with a real example, etc.
SnapMirror and SnapVault provide data protection capabilities including disaster recovery, backup storage, and migration between storage tiers. SnapMirror replicates volumes between SVMs, even across clusters, while SnapVault provides efficient long-term retention of backups. Together they reduce costs and allow DR sites to be used for other business purposes. The document then discusses SnapMirror concepts like source and destination volumes, replication policies and schedules, and how SnapMirror leverages Snapshot technology. It provides instructions for exploring an existing SnapMirror relationship and creating a new one using the System Manager tool.
Accelerating Spark Genome Sequencing in Cloud—A Data Driven Approach, Case St...Spark Summit
Spark data processing is shifting from on-premises to cloud service to take advantage of its horizontal resource scalability, better data accessibility and easy manageability. However, fully utilizing the computational power, fast storage and networking offered by cloud service can be challenging without deep understanding of workload characterizations and proper software optimization expertise. In this presentation, we will use a Spark based programing framework – Genome Analysis Toolkit version 4 (GATK4, under development), as an example to present a process of configuring and optimizing a proficient Spark cluster on Google Cloud to speed up genome data processing. We will first introduce an in-house developed data profiling framework named PAT, and discuss how to use PAT to quickly establish the best combination of VM configurations and Spark configurations to fully utilize cloud hardware resources and Spark computational parallelism. In addition, we use PAT and other data profiling tools to identify and fix software hotspots in application. We will show a case study in which we identify a thread scalability issue of Java Instanceof operator. The fix in Scala language hugely improves performance of GATK4 and other Spark based workloads.
Anna Vergeles, Nataliia Manakova "Unsupervised Real-Time Stream-Based Novelty...Fwdays
High-load systems produce lots of telemetry information in every time slot. That is quite a challenge to say if the working load has changed significantly right now or everything runs as expected. This presentation covers the novelty detection technique used for cloud systems that combine non-real-time learning with real-time estimation ensemble.
1. The document discusses using cloud computing for performance testing by provisioning virtual machines and load generation servers in the cloud instead of on-premise servers.
2. Commercial testing products and open-source frameworks like JMeter can be used for cloud-based performance testing, with benefits including lower costs, ability to simulate large-scale loads, and geographic distribution.
3. A case study describes a custom Hailstorm framework built on JMeter that was able to simulate 40,000 concurrent users for a client, providing rapid and cost-effective performance metrics at scale.
Srm suite technical presentation nrm - tim piqueurEMC Nederland
The document discusses EMC's Storage Resource Management Suite, which includes tools to optimize storage resources, monitor storage performance and configurations, and assure storage service levels. It provides overviews of the tools' capabilities for visualizing storage relationships, analyzing capacity and performance, validating configurations, monitoring applications and storage, and reporting on service levels. Screenshots demonstrate using the tools to analyze specific applications, storage environments, issues, and optimize resources.
The document discusses tuning Java for large data workloads. It covers symptoms of memory issues like jobs getting stuck or failing. It then discusses various Java and Hadoop configuration settings to optimize memory usage like mapreduce.child.java.opts and mapreduce.map.memory.mb. Finally, it provides an overview of different garbage collectors in Java and factors like generation sizes and concurrent marking that impact performance.
Characteristics of Remote Persistent Memory – Performance, Capacity, or Local...inside-BigData.com
In this deck from the 2019 OpenFabrics Workshop in Austin, Paul Grun from Cray presents: Characteristics of Remote Persistent Memory – Performance, Capacity, or Locality. Which One(s)?
Persistent Memory exhibits several interesting characteristics including persistence, capacity and others. These (sometimes)competing characteristics may require system and server architects to make tradeoffs in system architecture. A sometimes overlooked tradeoff is in the locality of the persistent memory, i.e. locally-attached persistent memory versus remote(or fabric-attached) persistent memory. In this session, we explore some of those tradeoffs and take an early look at the emerging use cases for Remote Persistent Memory and how those may impact network architecture and API design.
Watch the video: https://wp.me/p3RLHQ-jZR
Learn more: https://www.openfabrics.org/2019-workshop-agenda-and-abstracts/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Ever wonder how Java achieves such success in the “write once, run anywhere” (WORA) promise? In this talk, a senior member of the IBM Java team speaks candidly about the many difficulties Java faces behind the scenes around compatibility, and the various perspectives to consider. We describe areas such as bugs, bug fixes, algorithmic implementation assumptions, optimizations, multiple JVM implementations, and language changes. Hear how IBM is making Java better by championing compatibility and by contributing directly to OpenJDK. By the end of the session you will have clear insights on the complexity of the issue and how it’s addressed in the OpenJDK ecosystem.
Originally presented at JavaOne 2012 San Francisco
A Year of Testing in the Cloud: Lessons LearnedTechWell
Jim Trentadue describes how his organization first used the cloud for its non-production needs including development, testing, training, and production support. Jim begins by describing the components of a cloud environment and how it differs from a traditional physical server structure. To prove the cloud concept, he used a risk-based model for determining which servers would be migrated. The result was a win for the organization from a time-to-market and cost savings perspective. Jim shares his do’s and don’ts for moving to the cloud. Do’s include ensure you identify all costs associated with the new cloud infrastructure, implement a risk-based approach to cloud migration, define a governance model, and define Service Level Agreements for your cloud vendor. Jim warns against creating an open-ended environment without a charge-back model to allocate costs and failing to continuously monitor the overall environment.
Oracle ADF Architecture TV - Development - Performance & TuningChris Muir
Slides from Oracle's ADF Architecture TV series covering the Development phase of ADF projects, an in-depth look at performance and tuning of your ADF applications.
Like to know more? Check out:
- Subscribe to the YouTube channel - http://bit.ly/adftvsub
- Development Playlist - http://www.youtube.com/playlist?list=PLJz3HAsCPVaQfFop-QTJUE6LtjkyP_SOp
- Read the episode index on the ADF Architecture Square - http://bit.ly/adfarchsquare
2689 - Exploring IBM PureApplication System and IBM Workload Deployer Best Pr...Hendrik van Run
IBM IMPACT 2013 presentation
This lecture will provide an overview of a combination of design, development, configuration and deployment best practices for IBM PureApplication System and IBM Workload Deployer captured from customer engagement experiences.
The document discusses Effektives Consulting's performance engineering portfolio, which includes user experience and web performance management, cloud-based commerce recommendations, zero-touch deployments, and emerging augmented reality applications. It focuses on web performance management, covering infrastructure capacity planning, a two-stage performance testing approach using both on-premise and cloud-based resources, application profiling, and reporting.
Clipper: A Low-Latency Online Prediction Serving System: Spark Summit East ta...Spark Summit
Machine learning is being deployed in a growing number of applications which demand real-time, accurate, and robust predictions under heavy query load. However, most machine learning frameworks and systems only address model training and not deployment.
In this talk, we present Clipper, a general-purpose low-latency prediction serving system. Interposing between end-user applications and a wide range of machine learning frameworks, Clipper introduces a modular architecture to simplify model deployment across frameworks. Furthermore, by introducing caching, batching, and adaptive model selection techniques, Clipper reduces prediction latency and improves prediction throughput, accuracy, and robustness without modifying the underlying machine learning frameworks. We evaluated Clipper on four common machine learning benchmark datasets and demonstrate its ability to meet the latency, accuracy, and throughput demands of online serving applications. We also compared Clipper to the Tensorflow Serving system and demonstrate comparable prediction throughput and latency on a range of models while enabling new functionality, improved accuracy, and robustness.
Performance comparison on java technologies a practical approachcsandit
Performance responsiveness and scalability is a make-or-break quality for software. Nearly
everyone runs into performance problems at one time or another. This paper discusses about
performance issues faced during one of the project implemented in java technologies. The
challenges faced during the life cycle of the project and the mitigation actions performed. It
compares 3 java technologies and shows how improvements are made through statistical
analysis in response time of the application. The paper concludes with result analysis.
PERFORMANCE COMPARISON ON JAVA TECHNOLOGIES - A PRACTICAL APPROACHcscpconf
Performance responsiveness and scalability is a make-or-break quality for software. Nearly everyone runs into performance problems at one time or another. This paper discusses about
performance issues faced during one of the project implemented in java technologies. The challenges faced during the life cycle of the project and the mitigation actions performed. It compares 3 java technologies and shows how improvements are made through statistical analysis in response time of the application. The paper concludes with result analysis.
This document discusses different WebLogic topology strategies with varying levels of application isolation and performance. It recommends strategies such as running multiple WebLogic instances, multiple managed servers, or virtual machines on a single physical server for development/test environments, and using clusters, session persistence, or hardware partitions for production environments. The goal is to consolidate applications while balancing isolation and resource utilization.
Three Key Concepts for Understanding JSR-352: Batch Programming for the Java ...timfanelli
In this presentation, Tim Fanelli provides an introduction to JSR352 programming, and builds a simple application utilizing the JSR 352 chunk processing model.
The sample program presented may be downloaded here:
https://www.dropbox.com/s/55fsjt4ylny95hc/MySampleBatch.jar
Or, email Tim Fanelli - the contact information is on slide 3!
AI in the Workplace Reskilling, Upskilling, and Future Work.pptxSunil Jagani
Discover how AI is transforming the workplace and learn strategies for reskilling and upskilling employees to stay ahead. This comprehensive guide covers the impact of AI on jobs, essential skills for the future, and successful case studies from industry leaders. Embrace AI-driven changes, foster continuous learning, and build a future-ready workforce.
Read More - https://bit.ly/3VKly70
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
SnapMirror and SnapVault provide data protection capabilities including disaster recovery, backup storage, and migration between storage tiers. SnapMirror replicates volumes between SVMs, even across clusters, while SnapVault provides efficient long-term retention of backups. Together they reduce costs and allow DR sites to be used for other business purposes. The document then discusses SnapMirror concepts like source and destination volumes, replication policies and schedules, and how SnapMirror leverages Snapshot technology. It provides instructions for exploring an existing SnapMirror relationship and creating a new one using the System Manager tool.
Accelerating Spark Genome Sequencing in Cloud—A Data Driven Approach, Case St...Spark Summit
Spark data processing is shifting from on-premises to cloud service to take advantage of its horizontal resource scalability, better data accessibility and easy manageability. However, fully utilizing the computational power, fast storage and networking offered by cloud service can be challenging without deep understanding of workload characterizations and proper software optimization expertise. In this presentation, we will use a Spark based programing framework – Genome Analysis Toolkit version 4 (GATK4, under development), as an example to present a process of configuring and optimizing a proficient Spark cluster on Google Cloud to speed up genome data processing. We will first introduce an in-house developed data profiling framework named PAT, and discuss how to use PAT to quickly establish the best combination of VM configurations and Spark configurations to fully utilize cloud hardware resources and Spark computational parallelism. In addition, we use PAT and other data profiling tools to identify and fix software hotspots in application. We will show a case study in which we identify a thread scalability issue of Java Instanceof operator. The fix in Scala language hugely improves performance of GATK4 and other Spark based workloads.
Anna Vergeles, Nataliia Manakova "Unsupervised Real-Time Stream-Based Novelty...Fwdays
High-load systems produce lots of telemetry information in every time slot. That is quite a challenge to say if the working load has changed significantly right now or everything runs as expected. This presentation covers the novelty detection technique used for cloud systems that combine non-real-time learning with real-time estimation ensemble.
1. The document discusses using cloud computing for performance testing by provisioning virtual machines and load generation servers in the cloud instead of on-premise servers.
2. Commercial testing products and open-source frameworks like JMeter can be used for cloud-based performance testing, with benefits including lower costs, ability to simulate large-scale loads, and geographic distribution.
3. A case study describes a custom Hailstorm framework built on JMeter that was able to simulate 40,000 concurrent users for a client, providing rapid and cost-effective performance metrics at scale.
Srm suite technical presentation nrm - tim piqueurEMC Nederland
The document discusses EMC's Storage Resource Management Suite, which includes tools to optimize storage resources, monitor storage performance and configurations, and assure storage service levels. It provides overviews of the tools' capabilities for visualizing storage relationships, analyzing capacity and performance, validating configurations, monitoring applications and storage, and reporting on service levels. Screenshots demonstrate using the tools to analyze specific applications, storage environments, issues, and optimize resources.
The document discusses tuning Java for large data workloads. It covers symptoms of memory issues like jobs getting stuck or failing. It then discusses various Java and Hadoop configuration settings to optimize memory usage like mapreduce.child.java.opts and mapreduce.map.memory.mb. Finally, it provides an overview of different garbage collectors in Java and factors like generation sizes and concurrent marking that impact performance.
Characteristics of Remote Persistent Memory – Performance, Capacity, or Local...inside-BigData.com
In this deck from the 2019 OpenFabrics Workshop in Austin, Paul Grun from Cray presents: Characteristics of Remote Persistent Memory – Performance, Capacity, or Locality. Which One(s)?
Persistent Memory exhibits several interesting characteristics including persistence, capacity and others. These (sometimes)competing characteristics may require system and server architects to make tradeoffs in system architecture. A sometimes overlooked tradeoff is in the locality of the persistent memory, i.e. locally-attached persistent memory versus remote(or fabric-attached) persistent memory. In this session, we explore some of those tradeoffs and take an early look at the emerging use cases for Remote Persistent Memory and how those may impact network architecture and API design.
Watch the video: https://wp.me/p3RLHQ-jZR
Learn more: https://www.openfabrics.org/2019-workshop-agenda-and-abstracts/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Ever wonder how Java achieves such success in the “write once, run anywhere” (WORA) promise? In this talk, a senior member of the IBM Java team speaks candidly about the many difficulties Java faces behind the scenes around compatibility, and the various perspectives to consider. We describe areas such as bugs, bug fixes, algorithmic implementation assumptions, optimizations, multiple JVM implementations, and language changes. Hear how IBM is making Java better by championing compatibility and by contributing directly to OpenJDK. By the end of the session you will have clear insights on the complexity of the issue and how it’s addressed in the OpenJDK ecosystem.
Originally presented at JavaOne 2012 San Francisco
A Year of Testing in the Cloud: Lessons LearnedTechWell
Jim Trentadue describes how his organization first used the cloud for its non-production needs including development, testing, training, and production support. Jim begins by describing the components of a cloud environment and how it differs from a traditional physical server structure. To prove the cloud concept, he used a risk-based model for determining which servers would be migrated. The result was a win for the organization from a time-to-market and cost savings perspective. Jim shares his do’s and don’ts for moving to the cloud. Do’s include ensure you identify all costs associated with the new cloud infrastructure, implement a risk-based approach to cloud migration, define a governance model, and define Service Level Agreements for your cloud vendor. Jim warns against creating an open-ended environment without a charge-back model to allocate costs and failing to continuously monitor the overall environment.
Oracle ADF Architecture TV - Development - Performance & TuningChris Muir
Slides from Oracle's ADF Architecture TV series covering the Development phase of ADF projects, an in-depth look at performance and tuning of your ADF applications.
Like to know more? Check out:
- Subscribe to the YouTube channel - http://bit.ly/adftvsub
- Development Playlist - http://www.youtube.com/playlist?list=PLJz3HAsCPVaQfFop-QTJUE6LtjkyP_SOp
- Read the episode index on the ADF Architecture Square - http://bit.ly/adfarchsquare
2689 - Exploring IBM PureApplication System and IBM Workload Deployer Best Pr...Hendrik van Run
IBM IMPACT 2013 presentation
This lecture will provide an overview of a combination of design, development, configuration and deployment best practices for IBM PureApplication System and IBM Workload Deployer captured from customer engagement experiences.
The document discusses Effektives Consulting's performance engineering portfolio, which includes user experience and web performance management, cloud-based commerce recommendations, zero-touch deployments, and emerging augmented reality applications. It focuses on web performance management, covering infrastructure capacity planning, a two-stage performance testing approach using both on-premise and cloud-based resources, application profiling, and reporting.
Clipper: A Low-Latency Online Prediction Serving System: Spark Summit East ta...Spark Summit
Machine learning is being deployed in a growing number of applications which demand real-time, accurate, and robust predictions under heavy query load. However, most machine learning frameworks and systems only address model training and not deployment.
In this talk, we present Clipper, a general-purpose low-latency prediction serving system. Interposing between end-user applications and a wide range of machine learning frameworks, Clipper introduces a modular architecture to simplify model deployment across frameworks. Furthermore, by introducing caching, batching, and adaptive model selection techniques, Clipper reduces prediction latency and improves prediction throughput, accuracy, and robustness without modifying the underlying machine learning frameworks. We evaluated Clipper on four common machine learning benchmark datasets and demonstrate its ability to meet the latency, accuracy, and throughput demands of online serving applications. We also compared Clipper to the Tensorflow Serving system and demonstrate comparable prediction throughput and latency on a range of models while enabling new functionality, improved accuracy, and robustness.
Performance comparison on java technologies a practical approachcsandit
Performance responsiveness and scalability is a make-or-break quality for software. Nearly
everyone runs into performance problems at one time or another. This paper discusses about
performance issues faced during one of the project implemented in java technologies. The
challenges faced during the life cycle of the project and the mitigation actions performed. It
compares 3 java technologies and shows how improvements are made through statistical
analysis in response time of the application. The paper concludes with result analysis.
PERFORMANCE COMPARISON ON JAVA TECHNOLOGIES - A PRACTICAL APPROACHcscpconf
Performance responsiveness and scalability is a make-or-break quality for software. Nearly everyone runs into performance problems at one time or another. This paper discusses about
performance issues faced during one of the project implemented in java technologies. The challenges faced during the life cycle of the project and the mitigation actions performed. It compares 3 java technologies and shows how improvements are made through statistical analysis in response time of the application. The paper concludes with result analysis.
This document discusses different WebLogic topology strategies with varying levels of application isolation and performance. It recommends strategies such as running multiple WebLogic instances, multiple managed servers, or virtual machines on a single physical server for development/test environments, and using clusters, session persistence, or hardware partitions for production environments. The goal is to consolidate applications while balancing isolation and resource utilization.
Three Key Concepts for Understanding JSR-352: Batch Programming for the Java ...timfanelli
In this presentation, Tim Fanelli provides an introduction to JSR352 programming, and builds a simple application utilizing the JSR 352 chunk processing model.
The sample program presented may be downloaded here:
https://www.dropbox.com/s/55fsjt4ylny95hc/MySampleBatch.jar
Or, email Tim Fanelli - the contact information is on slide 3!
AI in the Workplace Reskilling, Upskilling, and Future Work.pptxSunil Jagani
Discover how AI is transforming the workplace and learn strategies for reskilling and upskilling employees to stay ahead. This comprehensive guide covers the impact of AI on jobs, essential skills for the future, and successful case studies from industry leaders. Embrace AI-driven changes, foster continuous learning, and build a future-ready workforce.
Read More - https://bit.ly/3VKly70
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
Getting the Most Out of ScyllaDB Monitoring: ShareChat's TipsScyllaDB
ScyllaDB monitoring provides a lot of useful information. But sometimes it’s not easy to find the root of the problem if something is wrong or even estimate the remaining capacity by the load on the cluster. This talk shares our team's practical tips on: 1) How to find the root of the problem by metrics if ScyllaDB is slow 2) How to interpret the load and plan capacity for the future 3) Compaction strategies and how to choose the right one 4) Important metrics which aren’t available in the default monitoring setup.
inQuba Webinar Mastering Customer Journey Management with Dr Graham HillLizaNolte
HERE IS YOUR WEBINAR CONTENT! 'Mastering Customer Journey Management with Dr. Graham Hill'. We hope you find the webinar recording both insightful and enjoyable.
In this webinar, we explored essential aspects of Customer Journey Management and personalization. Here’s a summary of the key insights and topics discussed:
Key Takeaways:
Understanding the Customer Journey: Dr. Hill emphasized the importance of mapping and understanding the complete customer journey to identify touchpoints and opportunities for improvement.
Personalization Strategies: We discussed how to leverage data and insights to create personalized experiences that resonate with customers.
Technology Integration: Insights were shared on how inQuba’s advanced technology can streamline customer interactions and drive operational efficiency.
ScyllaDB is making a major architecture shift. We’re moving from vNode replication to tablets – fragments of tables that are distributed independently, enabling dynamic data distribution and extreme elasticity. In this keynote, ScyllaDB co-founder and CTO Avi Kivity explains the reason for this shift, provides a look at the implementation and roadmap, and shares how this shift benefits ScyllaDB users.
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
"NATO Hackathon Winner: AI-Powered Drug Search", Taras KlobaFwdays
This is a session that details how PostgreSQL's features and Azure AI Services can be effectively used to significantly enhance the search functionality in any application.
In this session, we'll share insights on how we used PostgreSQL to facilitate precise searches across multiple fields in our mobile application. The techniques include using LIKE and ILIKE operators and integrating a trigram-based search to handle potential misspellings, thereby increasing the search accuracy.
We'll also discuss how the azure_ai extension on PostgreSQL databases in Azure and Azure AI Services were utilized to create vectors from user input, a feature beneficial when users wish to find specific items based on text prompts. While our application's case study involves a drug search, the techniques and principles shared in this session can be adapted to improve search functionality in a wide range of applications. Join us to learn how PostgreSQL and Azure AI can be harnessed to enhance your application's search capability.
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...DanBrown980551
This LF Energy webinar took place June 20, 2024. It featured:
-Alex Thornton, LF Energy
-Hallie Cramer, Google
-Daniel Roesler, UtilityAPI
-Henry Richardson, WattTime
In response to the urgency and scale required to effectively address climate change, open source solutions offer significant potential for driving innovation and progress. Currently, there is a growing demand for standardization and interoperability in energy data and modeling. Open source standards and specifications within the energy sector can also alleviate challenges associated with data fragmentation, transparency, and accessibility. At the same time, it is crucial to consider privacy and security concerns throughout the development of open source platforms.
This webinar will delve into the motivations behind establishing LF Energy’s Carbon Data Specification Consortium. It will provide an overview of the draft specifications and the ongoing progress made by the respective working groups.
Three primary specifications will be discussed:
-Discovery and client registration, emphasizing transparent processes and secure and private access
-Customer data, centering around customer tariffs, bills, energy usage, and full consumption disclosure
-Power systems data, focusing on grid data, inclusive of transmission and distribution networks, generation, intergrid power flows, and market settlement data
Northern Engraving | Modern Metal Trim, Nameplates and Appliance PanelsNorthern Engraving
What began over 115 years ago as a supplier of precision gauges to the automotive industry has evolved into being an industry leader in the manufacture of product branding, automotive cockpit trim and decorative appliance trim. Value-added services include in-house Design, Engineering, Program Management, Test Lab and Tool Shops.
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation F...AlexanderRichford
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation Functions to Prevent Interaction with Malicious QR Codes.
Aim of the Study: The goal of this research was to develop a robust hybrid approach for identifying malicious and insecure URLs derived from QR codes, ensuring safe interactions.
This is achieved through:
Machine Learning Model: Predicts the likelihood of a URL being malicious.
Security Validation Functions: Ensures the derived URL has a valid certificate and proper URL format.
This innovative blend of technology aims to enhance cybersecurity measures and protect users from potential threats hidden within QR codes 🖥 🔒
This study was my first introduction to using ML which has shown me the immense potential of ML in creating more secure digital environments!
The Department of Veteran Affairs (VA) invited Taylor Paschal, Knowledge & Information Management Consultant at Enterprise Knowledge, to speak at a Knowledge Management Lunch and Learn hosted on June 12, 2024. All Office of Administration staff were invited to attend and received professional development credit for participating in the voluntary event.
The objectives of the Lunch and Learn presentation were to:
- Review what KM ‘is’ and ‘isn’t’
- Understand the value of KM and the benefits of engaging
- Define and reflect on your “what’s in it for me?”
- Share actionable ways you can participate in Knowledge - - Capture & Transfer
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
This talk will cover ScyllaDB Architecture from the cluster-level view and zoom in on data distribution and internal node architecture. In the process, we will learn the secret sauce used to get ScyllaDB's high availability and superior performance. We will also touch on the upcoming changes to ScyllaDB architecture, moving to strongly consistent metadata and tablets.
1. A Framework for Modeling and
Execution of
Infrastructure Contention
Experiments
Carmelo Ragusa, Philip Robinson and Sergej Svorobej
MERMAT 2013, FIA, 7 May, Dublin
15. Thank you
Contact information:
Carmelo Ragusa
SAP HANA Cloud Infrastructure, Belfast
carmelo.ragusa@sap.com
COCOMA is released as Open Source under Apache v2 license:
https://github.com/cragusa/cocoma
Editor's Notes
Challenges for multi-tenancy include optimisation of resource sharing and guaranteed isolation against physical limitations, co-located faults and malicious attacks. For these reasons testing the performance and resilience of applications with different hardware, platforms, configurations, resource sharing conditions and user loads is important for increasing the assurance that providers and consumers have in cloud applications.
Scalability: given that customer applications can vary from 1 to 1000s of nodes, it must be possible to readily set up and execute useful testing environments for 1 to n number of independent hosts and network elements, avoiding cumbersome, error-prone configuration.Reproducibility: it must be possible to easily repeat testing conditions and tests in order to perform viable regression testing and make reliable claims about software quality.Portability: as hardware and virtualization technologies change, or as applications may be migrated to alternative data centers and platforms, it should be possible to easily reuse and recreate test designs and conditions during these changes.Extensibility: in addition to portability, test designs will need to be modified over time in order to take into account changes in quality constraints, scope, expected demands and software functionality. Systematic, guided procedures for modifying and extending test designs and mechanisms for these changes are necessary, as opposed to starting from scratch each time or making changes without knowledge of all dependencies.Self-containment: it is desirable to have a single top level solution, operational interface and workflows for designing and executing tests as opposed to the tester having to switch between multiple contexts and tools.Controllability: in spite of abstraction and higher level tooling, there is still a need to have control over the behaviour of resources used in the test, minimising the amount of disturbances that might provide unknown variations in test results.
Each requirement is assessed considering four different perspectives that arise in real-world systems with changing business priorities and technologies:Different test types (e.g. functional, load, security) need to be performed.A variety of resource kinds (i.e. network, disk, CPU, memory) need to be manipulated.Heterogeneity of physical hosts, nodes and devices.Given different customers, different scenarios and workload mixes have to be considered in parallel.
The main principles behind the design of the framework are the abstraction from the lower level tools that are used to create loads over the resources, allowing to emulate the wanted contention, as well as the separation of concerns, providing an effective modularisation of the tool which enables easy extensibility and additions of emulators and distributions.
A distribution creates the trend through runs as in a sampling process
Given a share physical machine, we divide the physical resources to VMs (SuT and COCOMA) each time essentially with a different percentage of the total.
CPU, the SuT CPU benchmark degradation over the percentage of CPU used by COCOMAThree clusters can be identifiedin the highest one there are configurations 1, 4 and 5 which according to table II have all the MH setup for the CPU for COCOMAbelow there are configurations 2, 7 and 8, which have the ML setup for the CPU for COCOMAand in the lowest part there are 6 and 3, with COCOMA CPU setup as LAs expected the more physical CPU is controlled by COCOMA the more the SuT is affected when CPU intensive operations are performedRAM, COCOMA used the maximum assigned RAM while increasing the number of threads performing writing operations on the RAMAmount of RAM and overall number of VMs assigned to COCOMA influence the resultsconfigurations 1, 2, 3 and 6, which have all 2 VMs in total, the more amount of RAM is assigned to COCOMA the more the SuT is affectedif we compare configuration 5 and 1, the difference in results is due only to the total number of VMs, being the only differentiator parameter between the two configurationsIOthe amount of files used for the workload does not make any noticeable difference across all configurationsConfiguration 4 and 5 (with 3 VMs assigned to COCOMA vs 1 to the SuT) suffers most
Performance testers/engineers: practitioners investigating for example new colocation algorithms, and generally in need to create a contentious/malicious environment to conduct their tests;Cloud Service Providers: in this case a service provider may offer performance isolation mechanisms and therefore wants to test the effectiveness of those mechanisms, investigate the possibility of offering those mechanisms, or study what characteristics applications need to coexist;Cloud Application Administrators: administrators may need to check when a system is restored, after some maintenance or a crash, that performance isolation mechanisms are working correctly;Application Developers and Testers: application developers may want to investigate the effects of contention over their system, while testers may want to check if provider’s isolation mechanisms work sufficiently;Benchmarks and Standards Groups: in this case it can be used to validate cloud patterns and workloads under investigation and/or characterisation.
On the malicious part, we are looking into covert channels (or side channels) at cache level to infer other processes information and data, as well as at network level to get information about co-located guests, such as IP addresses. The latter could be used by fuzzers to send malicious workloads, or do a DoS.
BonFIRE is an EU project which is designing, building and operating a multi-site cloud-based facility on top of six infrastructure offering heterogeneous Cloud resources, including compute, storage and network. BonFIRE is geared towards experimentation and research into Cloud/IoS, and offers the facilities to easily create, manage and monitor experiments, whilst giving the experimenters more information and control of Cloud resources than what is offered by other public Cloud providers.