Delphix allows databases to run as software rather than hardware, using less space while maintaining full functionality and performance. It turns database servers into a single, virtual authority that can consolidate databases and instantly provision copies for development, testing, and other non-production uses. This cuts capital expenses by 50% and operational expenses by 90% while accelerating innovation by eliminating the time and costs associated with copying and moving databases between environments.
The tech talk was gieven by Ranjeeth Kathiresan, Salesforce Senior Software Engineer & Gurpreet Multani, Salesforce Principal Software Engineer in June 2017.
This presentation describes how to efficiently load data into Hive. I cover partitioning, predicate pushdown, ORC file optimization and different loading schemes
Amazon Aurora services are MySQL and PostgreSQL -compatible relational database engines with the speed, reliability, and availability of high-end commercial databases at one-tenth the cost. This session introduces you to Amazon Aurora, explores the capabilities and features of Aurora, explains common use cases, and helps you get started with Aurora.
Troubleshooting Kerberos in Hadoop: Taming the BeastDataWorks Summit
Kerberos is the ubiquitous authentication mechanism when it comes to secure any Hadoop Services. With recent updates in Hadoop core and various Apache Hadoop components, inherent Kerberos support has matured and has come a long way.
Understanding & configuring Kerberos is still a challenge but even more painful & frustrating is troubleshooting a Kerberos issue. There are lot of things (small & big) that can go wrong (and will go wrong!). This talk covers the Kerberos debugging part in detail and discusses the tools & tricks that can be used to narrow down any Kerberos issue.
Rather than discussing the issues and their resolution, we will focus on how to approach a Kerberos problem and do's / dont's in Kerberos scene. This talk will provide a step by step guide that will equip the audience for troubleshooting future Kerberos problems.
Agenda is to discuss:
- Systematic approach to Kerberos troubleshooting
- Kerberos Tools available in Hadoop arsenal
- Tips & Tricks to narrow down Kerberos issues quickly
- Some nasty Kerberos issues from Support trenches
Some prior knowledge on Kerberos basics will be appreciated but is not a prerequisite.
Speaker:
Vipin Rathor, Sr. Product Specialist (HDP Security), Hortonworks
The tech talk was gieven by Ranjeeth Kathiresan, Salesforce Senior Software Engineer & Gurpreet Multani, Salesforce Principal Software Engineer in June 2017.
This presentation describes how to efficiently load data into Hive. I cover partitioning, predicate pushdown, ORC file optimization and different loading schemes
Amazon Aurora services are MySQL and PostgreSQL -compatible relational database engines with the speed, reliability, and availability of high-end commercial databases at one-tenth the cost. This session introduces you to Amazon Aurora, explores the capabilities and features of Aurora, explains common use cases, and helps you get started with Aurora.
Troubleshooting Kerberos in Hadoop: Taming the BeastDataWorks Summit
Kerberos is the ubiquitous authentication mechanism when it comes to secure any Hadoop Services. With recent updates in Hadoop core and various Apache Hadoop components, inherent Kerberos support has matured and has come a long way.
Understanding & configuring Kerberos is still a challenge but even more painful & frustrating is troubleshooting a Kerberos issue. There are lot of things (small & big) that can go wrong (and will go wrong!). This talk covers the Kerberos debugging part in detail and discusses the tools & tricks that can be used to narrow down any Kerberos issue.
Rather than discussing the issues and their resolution, we will focus on how to approach a Kerberos problem and do's / dont's in Kerberos scene. This talk will provide a step by step guide that will equip the audience for troubleshooting future Kerberos problems.
Agenda is to discuss:
- Systematic approach to Kerberos troubleshooting
- Kerberos Tools available in Hadoop arsenal
- Tips & Tricks to narrow down Kerberos issues quickly
- Some nasty Kerberos issues from Support trenches
Some prior knowledge on Kerberos basics will be appreciated but is not a prerequisite.
Speaker:
Vipin Rathor, Sr. Product Specialist (HDP Security), Hortonworks
Apache Flink is a popular stream computing framework for real-time stream computing. Many stream compute algorithms require trailing data in order to compute the intended result. One example is computing the number of user logins in the last 7 days. This creates a dilemma where the results of the stream program are incomplete until the runtime of the program exceeds 7 days. The alternative is to bootstrap the program using historic data to seed the state before shifting to use real-time data.
This talk will discuss alternatives to bootstrap programs in Flink. Some alternatives rely on technologies exogenous to the stream program, such as enhancements to the pub/sub layer, that are more generally applicable to other stream compute engines. Other alternatives include enhancements to Flink source implementations. Lyft is exploring another alternative using orchestration of multiple Flink programs. The talk will cover why Lyft pursued this alternative and future directions to further enhance bootstrapping support in Flink.
Speaker
Gregory Fee, Principal Engineer, Lyft
Oracle GoldenGate is the leading real-time data integration software provider in the industry - customers include 3 of the top 5 commercial banks, 3 of the top 3 busiest ATM networks, and 4 of the top 5 telecommunications providers.
Oracle GoldenGate moves transactional data in real-time across heterogeneous database, hardware and operating systems with minimal impact. The software platform captures, routes, and delivers data in real time, enabling organizations to maintain continuous uptime for critical applications during planned and unplanned outages.
Additionally, it moves data from transaction processing environments to read-only reporting databases and analytical applications for accurate, timely reporting and improved business intelligence for the enterprise.
Exactly-once Stream Processing with Kafka StreamsGuozhang Wang
I will present the recent additions to Kafka to achieve exactly-once semantics (0.11.0) within its Streams API for stream processing use cases. This is achieved by leveraging the underlying idempotent and transactional client features. The main focus will be the specific semantics that Kafka distributed transactions enable in Streams and the underlying mechanics to let Streams scale efficiently.
Lambda architecture is a popular technique where records are processed by a batch system and streaming system in parallel. The results are then combined during query time to provide a complete answer. Strict latency requirements to process old and recently generated events made this architecture popular. The key downside to this architecture is the development and operational overhead of managing two different systems.
There have been attempts to unify batch and streaming into a single system in the past. Organizations have not been that successful though in those attempts. But, with the advent of Delta Lake, we are seeing lot of engineers adopting a simple continuous data flow model to process data as it arrives. We call this architecture, The Delta Architecture.
Delta Lake, an open-source innovations which brings new capabilities for transactions, version control and indexing your data lakes. We uncover how Delta Lake benefits and why it matters to you. Through this session, we showcase some of its benefits and how they can improve your modern data engineering pipelines. Delta lake provides snapshot isolation which helps concurrent read/write operations and enables efficient insert, update, deletes, and rollback capabilities. It allows background file optimization through compaction and z-order partitioning achieving better performance improvements. In this presentation, we will learn the Delta Lake benefits and how it solves common data lake challenges, and most importantly new Delta Time Travel capability.
At Salesforce, we have deployed many thousands of HBase/HDFS servers, and learned a lot about tuning during this process. This talk will walk you through the many relevant HBase, HDFS, Apache ZooKeeper, Java/GC, and Operating System configuration options and provides guidelines about which options to use in what situation, and how they relate to each other.
HDFS has several strengths: horizontally scale its IO bandwidth and scale its storage to petabytes of storage. Further, it provides very low latency metadata operations and scales to over 60K concurrent clients. Hadoop 3.0 recently added Erasure Coding. One of HDFS’s limitations is scaling a number of files and blocks in the system. We describe a radical change to Hadoop’s storage infrastructure with the upcoming Ozone technology. It allows Hadoop to scale to tens of billions of files and blocks and, in the future, to every larger number of smaller objects. Ozone fundamentally separates the namespace layer and the block layer allowing new namespace layers to be added in the future. Further, the use of RAFT protocol has allowed the storage layer to be self-consistent. We show how this technology helps a Hadoop user and also what it means for evolving HDFS in the future. We will also cover the technical details of Ozone.
Speaker: Sanjay Radia, Chief Architect, Founder, Hortonworks
Automating Your Clone in E-Business Suite R12.2Michael Brown
It is possible to automate the cloning process in Oracle E-Business Suite 12.2. This presentation discusses how to accomplish that and gives some warnings about when it is not possible to run a clone.
For OAUG members, the slides and a recording of the presentation are available on www.oaug.org.
Automate Your Kafka Cluster with Kubernetes Custom Resources confluent
(Sam Obeid, Shopify) Kafka Summit SF 2018
At Shopify we manage multiple Apache Kafka clusters in multiple locations in Google’s cloud platform. We deploy our Kafka clusters as Kubernetes StatefulSets, and we use other K8s workloads to implement different tasks. Automating critical and repetitive operational tasks is one of our top priorities.
In this talk we’ll discuss how we leveraged Kubernetes Custom Resources and Controllers to automate some of the key cluster operational tasks, to detect clusters configuration changes and react to these changes with required actions. We will go through actual examples we implemented at Shopify, how we solved the problem of cluster discovery and how we automated topics creation across different clusters with zero human intervention and safety controls.
Flink Forward San Francisco 2022.
The Table API is one of the most actively developed components of Flink in recent time. Inspired by databases and SQL, it encapsulates concepts many developers are familiar with. It can be used with both bounded and unbounded streams in a unified way. But from afar it can be difficult to keep track of what this API is capable of and how it relates to Flink's other APIs. In this talk, we will explore the current state of Table API. We will show how it can be used as a batch processor, a changelog processor, or a streaming ETL tool with many built-in functions and operators for deduplicating, joining, and aggregating data. By comparing it to the DataStream API we will highlight differences and elaborate on when to use which API. We will demonstrate hybrid pipelines in which both APIs interact with one another and contribute their unique strengths. Finally, we will take a look at some of the most recent additions as a first step to stateful upgrades.
by
David Andreson
"Maximum Availability Architecture (MAA) for Oracle Database, Exadata and the Cloud" was first presented during Oracle Open World (OOW) 2019. This version of the deck has been updated for OOW London 2020 including the latest information regarding patching and upgrading the Oracle Database with Zero Downtime.
Interactive real-time dashboards on data streams using Kafka, Druid, and Supe...DataWorks Summit
When interacting with analytics dashboards, in order to achieve a smooth user experience, two major key requirements are quick response time and data freshness. To meet the requirements of creating fast interactive BI dashboards over streaming data, organizations often struggle with selecting a proper serving layer.
Cluster computing frameworks such as Hadoop or Spark work well for storing large volumes of data, although they are not optimized for making it available for queries in real time. Long query latencies also make these systems suboptimal choices for powering interactive dashboards and BI use cases.
This talk presents an open source real-time data analytics stack using Apache Kafka, Druid, and Superset. The stack combines the low-latency streaming and processing capabilities of Kafka with Druid, which enables immediate exploration and provides low-latency queries over the ingested data streams. Superset provides the visualization and dashboarding that integrates nicely with Druid. In this talk we will discuss why this architecture is well suited to interactive applications over streaming data, present an end-to-end demo of complete stack, discuss its key features, and discuss performance characteristics from real-world use cases. NISHANT BANGARWA, Software engineer, Hortonworks
This session introduces the Kafka Connector for Redis Enterprise. During the presentation we will discover how Kafka in combination with the multi-model database platform from Redis Labs opens up new possibilities for developers. We will conclude this presentation with a live demo of the connector that showcases the multiple database models enabled by the Redis Enterprise Kafka connector.
Serverless Kafka and Spark in a Multi-Cloud Lakehouse ArchitectureKai Wähner
Apache Kafka in conjunction with Apache Spark became the de facto standard for processing and analyzing data. Both frameworks are open, flexible, and scalable.
Unfortunately, the latter makes operations a challenge for many teams. Ideally, teams can use serverless SaaS offerings to focus on business logic. However, hybrid and multi-cloud scenarios require a cloud-native platform that provides automated and elastic tooling to reduce the operations burden.
This session explores different architectures to build serverless Apache Kafka and Apache Spark multi-cloud architectures across regions and continents.
We start from the analytics perspective of a data lake and explore its relation to a fully integrated data streaming layer with Kafka to build a modern data Data Lakehouse.
Real-world use cases show the joint value and explore the benefit of the "delta lake" integration.
Apache Flink is a popular stream computing framework for real-time stream computing. Many stream compute algorithms require trailing data in order to compute the intended result. One example is computing the number of user logins in the last 7 days. This creates a dilemma where the results of the stream program are incomplete until the runtime of the program exceeds 7 days. The alternative is to bootstrap the program using historic data to seed the state before shifting to use real-time data.
This talk will discuss alternatives to bootstrap programs in Flink. Some alternatives rely on technologies exogenous to the stream program, such as enhancements to the pub/sub layer, that are more generally applicable to other stream compute engines. Other alternatives include enhancements to Flink source implementations. Lyft is exploring another alternative using orchestration of multiple Flink programs. The talk will cover why Lyft pursued this alternative and future directions to further enhance bootstrapping support in Flink.
Speaker
Gregory Fee, Principal Engineer, Lyft
Oracle GoldenGate is the leading real-time data integration software provider in the industry - customers include 3 of the top 5 commercial banks, 3 of the top 3 busiest ATM networks, and 4 of the top 5 telecommunications providers.
Oracle GoldenGate moves transactional data in real-time across heterogeneous database, hardware and operating systems with minimal impact. The software platform captures, routes, and delivers data in real time, enabling organizations to maintain continuous uptime for critical applications during planned and unplanned outages.
Additionally, it moves data from transaction processing environments to read-only reporting databases and analytical applications for accurate, timely reporting and improved business intelligence for the enterprise.
Exactly-once Stream Processing with Kafka StreamsGuozhang Wang
I will present the recent additions to Kafka to achieve exactly-once semantics (0.11.0) within its Streams API for stream processing use cases. This is achieved by leveraging the underlying idempotent and transactional client features. The main focus will be the specific semantics that Kafka distributed transactions enable in Streams and the underlying mechanics to let Streams scale efficiently.
Lambda architecture is a popular technique where records are processed by a batch system and streaming system in parallel. The results are then combined during query time to provide a complete answer. Strict latency requirements to process old and recently generated events made this architecture popular. The key downside to this architecture is the development and operational overhead of managing two different systems.
There have been attempts to unify batch and streaming into a single system in the past. Organizations have not been that successful though in those attempts. But, with the advent of Delta Lake, we are seeing lot of engineers adopting a simple continuous data flow model to process data as it arrives. We call this architecture, The Delta Architecture.
Delta Lake, an open-source innovations which brings new capabilities for transactions, version control and indexing your data lakes. We uncover how Delta Lake benefits and why it matters to you. Through this session, we showcase some of its benefits and how they can improve your modern data engineering pipelines. Delta lake provides snapshot isolation which helps concurrent read/write operations and enables efficient insert, update, deletes, and rollback capabilities. It allows background file optimization through compaction and z-order partitioning achieving better performance improvements. In this presentation, we will learn the Delta Lake benefits and how it solves common data lake challenges, and most importantly new Delta Time Travel capability.
At Salesforce, we have deployed many thousands of HBase/HDFS servers, and learned a lot about tuning during this process. This talk will walk you through the many relevant HBase, HDFS, Apache ZooKeeper, Java/GC, and Operating System configuration options and provides guidelines about which options to use in what situation, and how they relate to each other.
HDFS has several strengths: horizontally scale its IO bandwidth and scale its storage to petabytes of storage. Further, it provides very low latency metadata operations and scales to over 60K concurrent clients. Hadoop 3.0 recently added Erasure Coding. One of HDFS’s limitations is scaling a number of files and blocks in the system. We describe a radical change to Hadoop’s storage infrastructure with the upcoming Ozone technology. It allows Hadoop to scale to tens of billions of files and blocks and, in the future, to every larger number of smaller objects. Ozone fundamentally separates the namespace layer and the block layer allowing new namespace layers to be added in the future. Further, the use of RAFT protocol has allowed the storage layer to be self-consistent. We show how this technology helps a Hadoop user and also what it means for evolving HDFS in the future. We will also cover the technical details of Ozone.
Speaker: Sanjay Radia, Chief Architect, Founder, Hortonworks
Automating Your Clone in E-Business Suite R12.2Michael Brown
It is possible to automate the cloning process in Oracle E-Business Suite 12.2. This presentation discusses how to accomplish that and gives some warnings about when it is not possible to run a clone.
For OAUG members, the slides and a recording of the presentation are available on www.oaug.org.
Automate Your Kafka Cluster with Kubernetes Custom Resources confluent
(Sam Obeid, Shopify) Kafka Summit SF 2018
At Shopify we manage multiple Apache Kafka clusters in multiple locations in Google’s cloud platform. We deploy our Kafka clusters as Kubernetes StatefulSets, and we use other K8s workloads to implement different tasks. Automating critical and repetitive operational tasks is one of our top priorities.
In this talk we’ll discuss how we leveraged Kubernetes Custom Resources and Controllers to automate some of the key cluster operational tasks, to detect clusters configuration changes and react to these changes with required actions. We will go through actual examples we implemented at Shopify, how we solved the problem of cluster discovery and how we automated topics creation across different clusters with zero human intervention and safety controls.
Flink Forward San Francisco 2022.
The Table API is one of the most actively developed components of Flink in recent time. Inspired by databases and SQL, it encapsulates concepts many developers are familiar with. It can be used with both bounded and unbounded streams in a unified way. But from afar it can be difficult to keep track of what this API is capable of and how it relates to Flink's other APIs. In this talk, we will explore the current state of Table API. We will show how it can be used as a batch processor, a changelog processor, or a streaming ETL tool with many built-in functions and operators for deduplicating, joining, and aggregating data. By comparing it to the DataStream API we will highlight differences and elaborate on when to use which API. We will demonstrate hybrid pipelines in which both APIs interact with one another and contribute their unique strengths. Finally, we will take a look at some of the most recent additions as a first step to stateful upgrades.
by
David Andreson
"Maximum Availability Architecture (MAA) for Oracle Database, Exadata and the Cloud" was first presented during Oracle Open World (OOW) 2019. This version of the deck has been updated for OOW London 2020 including the latest information regarding patching and upgrading the Oracle Database with Zero Downtime.
Interactive real-time dashboards on data streams using Kafka, Druid, and Supe...DataWorks Summit
When interacting with analytics dashboards, in order to achieve a smooth user experience, two major key requirements are quick response time and data freshness. To meet the requirements of creating fast interactive BI dashboards over streaming data, organizations often struggle with selecting a proper serving layer.
Cluster computing frameworks such as Hadoop or Spark work well for storing large volumes of data, although they are not optimized for making it available for queries in real time. Long query latencies also make these systems suboptimal choices for powering interactive dashboards and BI use cases.
This talk presents an open source real-time data analytics stack using Apache Kafka, Druid, and Superset. The stack combines the low-latency streaming and processing capabilities of Kafka with Druid, which enables immediate exploration and provides low-latency queries over the ingested data streams. Superset provides the visualization and dashboarding that integrates nicely with Druid. In this talk we will discuss why this architecture is well suited to interactive applications over streaming data, present an end-to-end demo of complete stack, discuss its key features, and discuss performance characteristics from real-world use cases. NISHANT BANGARWA, Software engineer, Hortonworks
This session introduces the Kafka Connector for Redis Enterprise. During the presentation we will discover how Kafka in combination with the multi-model database platform from Redis Labs opens up new possibilities for developers. We will conclude this presentation with a live demo of the connector that showcases the multiple database models enabled by the Redis Enterprise Kafka connector.
Serverless Kafka and Spark in a Multi-Cloud Lakehouse ArchitectureKai Wähner
Apache Kafka in conjunction with Apache Spark became the de facto standard for processing and analyzing data. Both frameworks are open, flexible, and scalable.
Unfortunately, the latter makes operations a challenge for many teams. Ideally, teams can use serverless SaaS offerings to focus on business logic. However, hybrid and multi-cloud scenarios require a cloud-native platform that provides automated and elastic tooling to reduce the operations burden.
This session explores different architectures to build serverless Apache Kafka and Apache Spark multi-cloud architectures across regions and continents.
We start from the analytics perspective of a data lake and explore its relation to a fully integrated data streaming layer with Kafka to build a modern data Data Lakehouse.
Real-world use cases show the joint value and explore the benefit of the "delta lake" integration.
21st Century Service Oriented ArchitectureBob Rhubart
Service Oriented Architecture has evolved from concept to reality in the last decade. The right methodology coupled with mature SOA technologies has helped customers demonstrate success in both innovation and ROI. In this session you will learn how Oracle SOA Suite’s orchestration, virtualization, and governance capabilities provide the infrastructure to run mission critical business and system applications. And we’ll take a special look at the convergence of SOA & BPM using Oracle’s Unified technology stack.
(As presented by Samrat Ray at Oracle Technology Network Architect Day in Chicago, October 24, 2011.)
Overview of Microsoft Appliances: Scaling SQL Server to Hundreds of TerabytesJames Serra
Learn how SQL Server can scale to HUNDREDS of terabytes for BI solutions. This session will focus on Fast Track Solutions and Appliances, Reference Architectures, and Parallel Data Warehousing (PDW). Included will be performance numbers and lessons learned on a PDW implementation and how a successful BI solution was built on top of it using SSAS.
Netapp Evento Virtual Business Breakfast 20110616Bruno Banha
Apresentação efectuada pela Netapp no evento Virtual Business Breakfast, realizado no dia 16 de Junho de 2011, no Porto.
O evento realizado em conjunto com a VMware e a NextiraOne Portugal.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
2. Confidentiality
The information presented here is
confidential and proprietary to Delphix
Corp.
Please do not share the information
contained herein with other vendors or
potential competitors.
In case you have received a copy of this
presentation without prior authorization
from Delphix Corp. please do not read,
duplicate, or distribute this presentation
and contact us at: legal@delphix.com.
2 >> Strictly Confidential
3. Database Business Critical Applications: Change
Infrastructure: Happens
Complex,
Redundant, Production
Dynamic Dev: Prod QA: Prod
Support Support
Dev: Test: QA:
Project Project Project
(Upgrade) (Upgrade) (Upgrade)
Operation
UAT,
al Standby,
Pilots,
Reporting, Stage
Training
ODS
DB Ecosystem: Highest Cost
Unstructured Data: High Volume, Low Cost
3 >> Strictly Confidential
4. Database
Virtualization:
Production Dev: Prod QA: Prod
Why make Database Support Support
and move
around all
those copies?
Dev: Test: QA:
Project Project Project An Elegant,
Simple Solution
(Upgrade) (Upgrade) (Upgrade)
A Single, Virtual Authority
Top line: accelerate innovation,
Operation
UAT,efficiency
drive business al Standby,
Pilots,
Bottom line: cut CapEx by 50%,
Training
Reporting, Stage
slash OpEx by 90% ODS
➠ High return, low risk
Consolidate. Instantly Provision. Maximize Performance.
4 >> Strictly Confidential
5. CrossFlow: Turn Hardware Into Software
ERP Database CRM Database Data Warehouse
DB API
PHYSICAL
CrossFlow Engine
Link, Shrink, and TimeFlow
Load • Input: LDB Streams, Logs
No Empty, Temporary, • Map into TimeFlow
or Scratch Blocks (snapshots, compression,
VIRTUAL
log management)
• Output: Provision VDBs
Mask Prior to from Any Time
Provisioning • Optimize Performance,
ERP VDB CRM VDB DW VDB Enforce QoS
• End-to-end Automation
NFS NFS NFS
➠ 50% CapEx Reduction
➠ 90% OpEx Reduction
5 >> Strictly Confidential
6. Market Opportunity
2007 Storage Software Revenue: $10.9 Billion
(Source: IDC) • TAM (2007): $10.93 billion,
10.4% growth/year
$2,867, 26% $2,825, 26%
• SAM (2007): $4.373 billion,
>10% growth/year
• Database only and
related applications
$483, 4%
$521, 5% $1,930, 18%
$887 , 8% $1,418, 13%
EMC Symantec IBM NetApp Hewlett-Packard CA Other
• 2010 target market: 280,000 Oracle customers
• Additional market drivers
• Datacenter OpEx for DBs
• Elimination of downstream solutions (de-dupe)
• Replacement of DB features
>> Strictly Confidential for Deutsche Bank
6 >> Strictly Confidential
7. Virtualization Accelerates Innovation
Software
Pricing: Annual,
Upfront $400,000
Subscription by Lost Opportunity: Time, Costs Stop Innovation
$300,000
Server VDBs: Instant Sandboxes
CPU/Socket $200,000
Innovation
$100,000 Value
Purchase: via
$0
Partners or
Direct -$100,000
-$100,000
Cut CapEx by 50%
-$100,000
Slash OpEx by 90%
-$400,000
-$500,000
-$600,000
* 5 TB Oracle Environment on SAN
7 >> Strictly Confidential for Lightspeed
>> Strictly Confidential
8. Comparing Delphix
Consolidate non-production DBs, automatically refresh and provision
>> Delivering Agility for Business-Critical Applications and Databases: Complete the Chain
8 >> Strictly Confidential for Internal Use Only
10. Team, Board, Advisors
Jedidiah Yueh, President and CEO
Founded Avamar in 1999, sold to EMC in 2006, VP Product Management at EMC
Pioneered de-duplication industry (>$1 billion ’09, Avamar >$1 billion lifetime revenue ‘11)
Harvard, US Presidential Scholar, 10+ patents and patents pending
Alok Srivastava: VP Engineering
Director of Engineering, Clusters and Parallel Storage Technology (RAC), Oracle
Products responsible for ~$2 billion in revenue for Oracle
IIT BS, Wharton MBA, 30+ patents and patents pending
Boris Klots: Chief Architect
Inventor of Cache Fusion, Oracle; 50+ patents and patents pending
PhD Mathematics, Moscow State University
Arnold Silverman: Independent Board Member
Founding board member, Oracle; Co-founder, Business Objects
Other investments: Informatica, Kiva, TimesTen
Chris Schaepe, Lightspeed Managing Director, Board Member
Founder of Lightspeed; Board member, Riverbed
AsheemChandna, Greylock Parnter, Board Member
First US executive at Checkpoint Software; Board Member, Imperva, etc.
Kirk Bowman, Advisor
EVP Field Operations, VMware and EqualLogic
10 >> Strictly Confidential
11. Database Virtualization Summary
Consolidate. Instantly Provision. Maximize Performance.
Database Virtualization
Next step in evolution of datacenter Production
Support
Delphix: simple, powerful, non-intrusive Instant
Business
Sandboxes
Powerful ROI, Low Risk Storage
Reduction
Make an impact
Top line: enable the business, accelerate Accelerate
Application
Compliance,
Legal
innovation Development Database Repository
Virtualization
Bottom line: cut CapEx by 50%, slash
OpEx by 90%
Proof: Try the Product
Offshore
Real-time
Dev
SAP Reporting
Identify deployment opportunity
Landscape
Provisioning
Build business case and ROI analysis
What if you could instantly create full,
secure copies of your databases with no
additional hardware and no impact to
production?
11 >> Strictly Confidential
Delphix is a database virtualization software company that turns expensive database infrastructure into software that can run in a fraction of the space, while preserving full functionality and performance.
Problem: database ecosystem redundant, dynamic, expensive Most datacenters spend a great deal managing unstructured data, but they spend the most pound-for-pound in the database ecosystem. If you zoom into the database ecosystem, the production application and database is only the tip of the iceberg, as little as 10% of the total infrastructure footprint and costs. Most enterprises are unaware of the full extent of the redundancy and complexity required in regular operations. Due to the high business criticality of applications like ERP and CRM, enterprises generally create multiple copies of their database infrastructure before making even small changes or improvements. For most important applications, a full copy of infrastructure exists for development and QA for production support—to help troubleshoot problems if errors occur in the database or application. In addition, typical environments will create two more copies for dev and QA for projects—such as an application or database upgrade, update, or patch, which come out from vendors monthly or quarterly. In enterprises that develop or customize applications, even more copies are created in non-production environment for dev, test, QA, and staging. Even worse, all this redundancy gets magnified by the need to constantly move the data. As applications grow and change, databases have to be refreshed in development, then moved to testing, QA, staging, and back to production. For projects, which come regularly, development has to be frequently refreshed and moved through the cycle; using stale or partial data can prolong timelines or impede proper testing. Other copies, such as staging for data warehouses, have to be refreshed weekly, nightly, or even more frequently. It takes the average IT organization 10 hours to 10 weeks to provision a new copy of a database from one environment to another, navigating internal procurement processes and coordinating among DBAs, server, storage, and backup teams—an enormous operational expense.
Solution and benefits: database virtualization At Delphix, we asked a simple question: why make and move around all these copies? What if you could have a single virtual authority that stayed synchronized with production? You could then serve all the copies needed in enterprises off a single shared footprint. If you could do that, you could really empower IT to better service the business—delivering on critical requests from lines of business to improve applications that drive sales or operational efficiency. You could spend less on infrastructure, while simplifying operational IT complexity. Best of all, if you designed the solution correctly, you could achieve all of this with no changes required to production, so you could get a high return on investment, at very low risk.
Technology: CrossFlow Engine—turn hardware into software The CrossFlow Engine sits at the heart of the Delphix Server. CrossFlow takes streams from physical databases and turns them into software virtual databases that can sit in a fraction of the space, but with full functionality and performance. CrossFlow dynamically maps file changes and database logs into a TimeFlow—a rolling window of changes recorded for an LDB that is always application consistent. With content aware reduction, Delphix compresses and eliminates unnecessary, empty, and transient data blocks that are not used—shrinking even the first copy of an LDB by up to 75%. Underneath TimeFlow, Delphix leverages highly mature snapshot and cloning technologies to present VDBs to target servers as a set of pointers that map to already existing blocks, eliminating the need to create new, physical copies. Snapshots have been in existence for several years but have lacked the application awareness to fully harness their power for the database ecosystem. Changes to Delphix VDBs are written as new, compressed blocks and maintained transparently by the Delphix Server. Once opened, VDBs are not updated by Delphix—they are only changed by database users and applications. Only the TimeFlow for the LDB continues to be updated, and new VDBs can be instantly provisioned from that TimeFlow, or open VDBs can be refreshed in place. [Primary Confusion] Even VDBs can have their own TimeFlow, and VDBs can be easily provisioned from VDBs, enabling users to promote databases from development to testing. In some environments, data security may be a concern, so users can automatically mask sensitive data, such as Social Security Numbers, through integrated support for pre and post scripting—obfuscating private information before VDBs become accessible by users. Importantly, Delphix automates all of the highly complex, fault prone parameterization required to provision copies of databases, from changing Oracle SIDs to editing cache settings. The end result is a technology engine that is designed to consolidate all the dynamic, lifecycle copies of databases created in enterprise environments, while reducing the time and complexity of provisioning to nearly zero.
Served Addressable Market (SAM): Bottoms Up Assumptions: Database or structured market for storage software only Unstructured vs. structured data in general: 90:10 to 70:30. Most unstructured data not in corporate data centers, etc. Most unstructured data (media, etc.) not on enterprise storage (grid, commodity architectures like GooglePlex). Estimated storage software for file system vs. database data at major hardware manufacturers: 60:40. File servers: 70:30; mid-tier arrays: 50:50, high-end arrays: 25:75, but spending skewed toward higher end arrays: total spend estimated at 50:50 Storage management, storage device management, storage infrastructure, file system, and other non addressable markets total: 20% of storage software revenue. Hence, 80% and then 50% of storage software market: $4.373 billion, growing at 10.4% annually 75% of respondents in Forrester survey say critical databases growing 10 to 49% annually. Trend of moving unstructured data into structured databases for better manageability (e.g. MS SharePoint) Market Size for Delphix Server 1.0 TAM Oriented for mid-market and departmental or test/development enterprise sales Same TAM as Overall Market Size SAM $1 billion, growing >10% annually ($1.63 billion with Microsoft) Oracle only for 1.0: 38% DBMS Market Shares Oracle: 38% IBM: 32% Microsoft: 24% Sybase: 2% Others: 5% Market Shares by Size of Database Under 2 TB databases: 60% More than 5 TB: 21% 2 TB to 5 TB: 19% 1 TB to <2 TB: 7% 500 GB to <1 TB: 24% 100 GB to 499 GB: 17% <100 GB: 12% Overall SAM: $4.373 billion, growing at 10.4% annually $4.373 billion time 38% (Oracle only) times 60% (under 2TB): $1 billion With Microsoft: $1.627 billion
Return on investment (ROI): business enablement, OpEx, and CapEx By combining consolidation and data reduction technologies, Delphix cuts capital expenditures by over 50% and can even pay for itself instantly if new database storage needs to be procured in the same quarter. Many IT organizations have been forced to do more with less; the ability to slash complexity and time for provisioning by over 90% allows IT personnel to focus on higher priorities and projects with higher returns. Finally, by reducing the time and cost of provisioning database sandboxes to nearly zero, Delphix enables IT to better service business needs. Today, many business ideas—like an end-of-quarter adjustment to a CRM application to drive sales efficiency—never become reality due to the hurdles posed by physical infrastructure. Software infrastructure like VMware and Delphix, however, facilitate innovation, allowing businesses to capture potentially lost opportunity value. With Delphix, enterprises can spend less and move faster than the competition.
Summary Database virtualization is the next logical step in the evolution of the datacenter. In many environments, it may be the single largest opportunity for operational and capital cost savings available. By designing a solution that targets non-production infrastructure, a Delphix investment can provide high returns with very low risk. With Delphix, enterprises can move faster and spend less than the competition. Delphix can be deployed in less than an hour—less time than it takes to provision a single physical database—so schedule a free trial with sales today. Contact us by email at [email_address] .
Usability and simplicity: 3 clicks to provisioning Most enterprise software products build their user interfaces last, so they reflect the mechanics of the technology. At Delphix, we believe in building interfaces around users, providing intuitive usability and enabling end-user self service. Usability isn’t just about elegant interfaces—it reduces the long term operational cost and complexity of a product, increasing customer value over time. After selecting an LDB in Delphix, it takes only three clicks to provision a VDB with no additional hardware required—a dramatic improvement over the days and weeks often required with physical infrastructure.