This session will focus on basics of what Machine Learning is, different types of Machine Learning and Neural Networks, supervised and unsupervised machine learning, AutoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use many examples from the ADW Oracle Autonomous Database offering, Oracle Machine Learning library to make this a session with lots of code examples in addition to the theory of Machine Learning and you will walk out having a definitive path to being a data scientist
20 tips and tricks with the Autonomous DatabaseSandesh Rao
This covers the top 20 questions most DBA’s , Developers will have on the Autonomous Database from provisioning to backups , troubleshooting , tips and tricks , security and HA . This is a good introduction for on-prem DBA’s who want to learn how this works quickly without spending too much time on it . Questions like what does the free tier cover , how do I do backup or if its automated how do I manage it , how to scale up and down , how to use tools like SQLDeveloper and SQLModeler , endpoints , terraform all in a quick 45 minute session which might take weeks to pickup reading documentation or spanning several presentations
Guiding through a typical Machine Learning PipelineMichael Gerke
Many People are talking about AI and Machine Learning. Here's a quick guideline how to manage ML Projects and what to consider in order to implement machine learning use cases.
R Tool for Visual Studio และการทำงานร่วมกันเป็นทีม โดย เฉลิมวงศ์ วิจิตรปิยะกุ...BAINIDA
R Tool for Visual Studio และการทำงานร่วมกันเป็นทีม โดย เฉลิมวงศ์ วิจิตรปิยะกุล MVP, Microsoft Thailand
THE FIRST NIDA BUSINESS ANALYTICS AND DATA SCIENCES CONTEST/CONFERENCE
microsoft r server for distributed computingBAINIDA
microsoft r server for distributed computing กฤษฏิ์ คำตื้อ,
Technical Evangelist,
Microsoft (Thailand)
ในงาน THE FIRST NIDA BUSINESS ANALYTICS AND DATA SCIENCES CONTEST/CONFERENCE จัดโดย คณะสถิติประยุกต์และ DATA SCIENCES THAILAND
[db tech showcase Tokyo 2018] #dbts2018 #B27 『Discover Machine Learning and A...Insight Technology, Inc.
[db tech showcase Tokyo 2018] #dbts2018 #B27
『Discover Machine Learning and ADWC - The Perfect Combination』
Data Intensity - Director of Innovation Francisco Munoz Alvarez 氏
Informatica PowerCenter Tutorial | Informatica Tutorial for Beginners | EdurekaEdureka!
This Edureka Informatica PowerCenter Tutorial will help you in understanding the various components of Informatica PowerCenter in detail with examples. You will be given a detailed understanding of each client and administrator tool. You will also understand the role of these tools in various phases to solve a use case. Below are the topics covered in this tutorial:
1. Informatica PowerCenter Overview
2. Why Do We Need Data Integration?
3. ETL Process
4. Informatica PowerCenter Administrator Console.
5. Informatica PowerCenter Repository Manager.
6. Informatica PowerCenter Designer
7. Informatica PowerCenter Workflow Manager
8. Informatica PowerCenter Workflow Monitor
Re-imagine Data Monitoring with whylogs and SparkDatabricks
In the era of microservices, decentralized ML architectures and complex data pipelines, data quality has become a bigger challenge than ever. When data is involved in complex business processes and decisions, bad data can, and will, affect the bottom line. As a result, ensuring data quality across the entire ML pipeline is both costly, and cumbersome while data monitoring is often fragmented and performed ad hoc. To address these challenges, we built whylogs, an open source standard for data logging. It is a lightweight data profiling library that enables end-to-end data profiling across the entire software stack. The library implements a language and platform agnostic approach to data quality and data monitoring. It can work with different modes of data operations, including streaming, batch and IoT data.
In this talk, we will provide an overview of the whylogs architecture, including its lightweight statistical data collection approach and various integrations. We will demonstrate how the whylogs integration with Apache Spark achieves large scale data profiling, and we will show how users can apply this integration into existing data and ML pipelines.
20 tips and tricks with the Autonomous DatabaseSandesh Rao
This covers the top 20 questions most DBA’s , Developers will have on the Autonomous Database from provisioning to backups , troubleshooting , tips and tricks , security and HA . This is a good introduction for on-prem DBA’s who want to learn how this works quickly without spending too much time on it . Questions like what does the free tier cover , how do I do backup or if its automated how do I manage it , how to scale up and down , how to use tools like SQLDeveloper and SQLModeler , endpoints , terraform all in a quick 45 minute session which might take weeks to pickup reading documentation or spanning several presentations
Guiding through a typical Machine Learning PipelineMichael Gerke
Many People are talking about AI and Machine Learning. Here's a quick guideline how to manage ML Projects and what to consider in order to implement machine learning use cases.
R Tool for Visual Studio และการทำงานร่วมกันเป็นทีม โดย เฉลิมวงศ์ วิจิตรปิยะกุ...BAINIDA
R Tool for Visual Studio และการทำงานร่วมกันเป็นทีม โดย เฉลิมวงศ์ วิจิตรปิยะกุล MVP, Microsoft Thailand
THE FIRST NIDA BUSINESS ANALYTICS AND DATA SCIENCES CONTEST/CONFERENCE
microsoft r server for distributed computingBAINIDA
microsoft r server for distributed computing กฤษฏิ์ คำตื้อ,
Technical Evangelist,
Microsoft (Thailand)
ในงาน THE FIRST NIDA BUSINESS ANALYTICS AND DATA SCIENCES CONTEST/CONFERENCE จัดโดย คณะสถิติประยุกต์และ DATA SCIENCES THAILAND
[db tech showcase Tokyo 2018] #dbts2018 #B27 『Discover Machine Learning and A...Insight Technology, Inc.
[db tech showcase Tokyo 2018] #dbts2018 #B27
『Discover Machine Learning and ADWC - The Perfect Combination』
Data Intensity - Director of Innovation Francisco Munoz Alvarez 氏
Informatica PowerCenter Tutorial | Informatica Tutorial for Beginners | EdurekaEdureka!
This Edureka Informatica PowerCenter Tutorial will help you in understanding the various components of Informatica PowerCenter in detail with examples. You will be given a detailed understanding of each client and administrator tool. You will also understand the role of these tools in various phases to solve a use case. Below are the topics covered in this tutorial:
1. Informatica PowerCenter Overview
2. Why Do We Need Data Integration?
3. ETL Process
4. Informatica PowerCenter Administrator Console.
5. Informatica PowerCenter Repository Manager.
6. Informatica PowerCenter Designer
7. Informatica PowerCenter Workflow Manager
8. Informatica PowerCenter Workflow Monitor
Re-imagine Data Monitoring with whylogs and SparkDatabricks
In the era of microservices, decentralized ML architectures and complex data pipelines, data quality has become a bigger challenge than ever. When data is involved in complex business processes and decisions, bad data can, and will, affect the bottom line. As a result, ensuring data quality across the entire ML pipeline is both costly, and cumbersome while data monitoring is often fragmented and performed ad hoc. To address these challenges, we built whylogs, an open source standard for data logging. It is a lightweight data profiling library that enables end-to-end data profiling across the entire software stack. The library implements a language and platform agnostic approach to data quality and data monitoring. It can work with different modes of data operations, including streaming, batch and IoT data.
In this talk, we will provide an overview of the whylogs architecture, including its lightweight statistical data collection approach and various integrations. We will demonstrate how the whylogs integration with Apache Spark achieves large scale data profiling, and we will show how users can apply this integration into existing data and ML pipelines.
YouTube Link: https://youtu.be/aGu0fbkHhek
** Data Science Master Program: https://www.edureka.co/masters-program/data-scientist-certification **
This Edureka PPT on "Data Science Full Course" provides an end to end, detailed and comprehensive knowledge on Data Science. This Data Science PPT will start with basics of Statistics and Probability and then moves to Machine Learning and Finally ends the journey with Deep Learning and AI. For Data-sets and Codes discussed in this PPT, drop a comment.
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Creating a Data validation and Testing StrategyRTTS
Creating A Data Validation & Testing Strategy
Are you struggling with formulating a strategy for how to validate the massive amount of data continuously entering your data warehouse or data lake?
We can help you!
Learn how RTTS’ Data Validation Assessment provides:
- an evaluation of your current data validation process
- recommendations on how to improve your process and
- a proposal for successful implementation
This slide deck addresses the following issues:
- How do I find out if I have bad data?
- How do I ensure I am testing the proper data permutations?
- How much of my data needs to be validated and automated?
- Which critical data endpoints need to be tested?
- How do I test data in my cloud environments?
And much more!
For more information, visit:
https://www.rttsweb.com/services/solutions/data-validation-assessment
What is a Data Warehouse and How Do I Test It?RTTS
ETL Testing: A primer for Testers on Data Warehouses, ETL, Business Intelligence and how to test them.
Are you hearing and reading about Big Data, Enterprise Data Warehouses (EDW), the ETL Process and Business Intelligence (BI)? The software markets for EDW and BI are quickly approaching $22 billion, according to Gartner, and Big Data is growing at an exponential pace.
Are you being tasked to test these environments or would you like to learn about them and be prepared for when you are asked to test them?
RTTS, the Software Quality Experts, provided this groundbreaking webinar, based upon our many years of experience in providing software quality solutions for more than 400 companies.
You will learn the answer to the following questions:
• What is Big Data and what does it mean to me?
• What are the business reasons for a building a Data Warehouse and for using Business Intelligence software?
• How do Data Warehouses, Business Intelligence tools and ETL work from a technical perspective?
• Who are the primary players in this software space?
• How do I test these environments?
• What tools should I use?
This slide deck is geared towards:
QA Testers
Data Architects
Business Analysts
ETL Developers
Operations Teams
Project Managers
...and anyone else who is (a) new to the EDW space, (b) wants to be educated in the business and technical sides and (c) wants to understand how to test them.
What Is Data Science? Data Science Course - Data Science Tutorial For Beginne...Edureka!
This Edureka Data Science course slides will take you through the basics of Data Science - why Data Science, what is Data Science, use cases, BI vs Data Science, Data Science tools and Data Science lifecycle process. This is ideal for beginners to get started with learning data science.
You can read the blog here: https://goo.gl/OoDCxz
You can also take a complete structured training, check out the details here: https://goo.gl/AfxwBc
Big Data and BI Tools - BI Reporting for Bay Area Startups User GroupScott Mitchell
This presentation was presented at the July 8th 2014 user group meeting for BI Reporting for Bay Area Start Ups
Content - Creation Infocepts/DWApplications
Presented by: Scott Mitchell - DWApplications
Data Warehousing in Pharma: How to Find Bad Data while Meeting Regulatory Req...RTTS
In the U.S., pharmaceutical firms must meet electronic record-keeping regulations set by the Food and Drug Administration (FDA). The regulation is Title 21 CFR Part 11, commonly known as Part 11.
Part 11 requires regulated firms to implement controls for software and systems involved in processing many forms of data as part of business operations and product development.
Enterprise data warehouses are used by the pharmaceutical and medical device industries for storing data covered by Part 11. QuerySurge, the only test tool designed specifically for automating the testing of data warehouses and the ETL process, is the market leader in testing data warehouses used by Part 11-governed companies.
For more on QuerySurge and Pharma, please visit
http://www.querysurge.com/solutions/pharmaceutical-industry
AUSOUG - Introducing New AI Ops Innovations in Oracle 19c Autonomous Health F...Sandesh Rao
Oracle Autonomous Health Framework (AHF) is Oracle’s Artificial Intelligence Operations platform for autonomous database health management. This session will focus on enhancements to current functionality and new features in 18c and coming in 19c. First successfully introduced in Cluster Health Advisor, and extended to Trace File Analyzer and Hang Manager, Oracle AHF’s applied machine learning technology now enhances additional framework components. You will learn how to utilize these features for determining workload footprint, ongoing monitoring, early detection of anomalies and performance issues, their root causes and corrective actions, prevention of node or database failures, and targeted postmortem analysis enabling quick resolution.
QuerySurge - the automated Data Testing solutionRTTS
QuerySurge is the leading Data Testing solution built specifically to automate the testing of Data Warehouses & Big Data. QuerySurge ensures that the data extracted from data sources remains intact in the target data store by analyzing and pinpointing any differences quickly.
And QuerySurge makes it easy for both novice and experienced team members to validate their organization's data quickly through Query Wizards while still allowing power users the flexibility they need.
All with deep dive reporting and data health dashboards that quickly provides you with a holistic view of your project’s data.
Types of Automated Data Testing
--------------------------------------------
QuerySurge provides data testing solutions for all of your automated data testing needs
- Data Warehouse testing & ETL testing
- Big Data (Hadoop, NoSQL) testing
- Data Interface testing
- Data Migration testing
- Database Upgrade testing
FREE TRIAL
www.QuerySurge.com
Real-time Recommendations for Retail: Architecture, Algorithms, and DesignJuliet Hougland
Users are constantly searching for new content and to stay competitive organizations must act immediately based on up-to-date data. Outdated recommendations decrease the likelihood of presenting the right offer and make it harder to maintain customer loyalty. In order to provide the most relevant recommendations and increase engagement, organizations must track customer interactions and re-score recommendations on the fly.
Data sources have expanded dramatically to include a wealth of historical data and a constant influx of behavior data. The key to moving from predictive models, applied in batch, to models that provide responses in real time, is to focus on the efficiency of model application. The speed that recommendations can be served is influenced by:
Architecture of the recommendation serving platform
Choice of recommendation algorithm
Datastore access patterns
In this presentation, we’ll discuss how developers can use open source components like HBase and Kiji to develop low-latency recommendation models that can be easily deployed by e-commerce companies. We will give practical advice on how to choose models and design data stores that make use of the architecture and quickly serve new recommendations.
Testing Big Data: Automated Testing of Hadoop with QuerySurgeRTTS
Are You Ready? Stepping Up To The Big Data Challenge In 2016 - Learn why Testing is pivotal to the success of your Big Data Strategy.
According to a new report by analyst firm IDG, 70% of enterprises have either deployed or are planning to deploy big data projects and programs this year due to the increase in the amount of data they need to manage.
The growing variety of new data sources is pushing organizations to look for streamlined ways to manage complexities and get the most out of their data-related investments. The companies that do this correctly are realizing the power of big data for business expansion and growth.
Learn why testing your enterprise's data is pivotal for success with big data and Hadoop. Learn how to increase your testing speed, boost your testing coverage (up to 100%), and improve the level of quality within your data - all with one data testing tool.
Data Verification In QA Department FinalWayne Yaddow
Data warehouse and ETL testing should be conducted according to a process and checklist. This presentation provides an overview of recommended methods.
Data Warehouse Testing in the Pharmaceutical IndustryRTTS
In the U.S., pharmaceutical firms and medical device manufacturers must meet electronic record-keeping regulations set by the Food and Drug Administration (FDA). The regulation is Title 21 CFR Part 11, commonly known as Part 11.
Part 11 requires regulated firms to implement controls for software and systems involved in processing many forms of data as part of business operations and product development.
Enterprise data warehouses are used by the pharmaceutical and medical device industries for storing data covered by Part 11 (for example, Safety Data and Clinical Study project data). QuerySurge, the only test tool designed specifically for automating the testing of data warehouses and the ETL process, has been effective in testing data warehouses used by Part 11-governed companies. The purpose of QuerySurge is to assure that your warehouse is not populated with bad data.
In industry surveys, bad data has been found in every database and data warehouse studied and is estimated to cost firms on average $8.2 million annually, according to analyst firm Gartner. Most firms test far less than 10% of their data, leaving at risk the rest of the data they are using for critical audits and compliance reporting. QuerySurge can test up to 100% of your data and help assure your organization that this critical information is accurate.
QuerySurge not only helps in eliminating bad data, but is also designed to support Part 11 compliance.
Learn more at www.QuerySurge.com
"How to document your decisions", Dmytro Ovcharenko Fwdays
We will perform architecture kata around a proposed business case. We will review ADD in detail. How usually architecture vision document looks like. How to match your architecture drivers and proposed architecture decisions in architecture views. We will review what is ATAM and how to perform analysis of your decisions in the right way. And finally, we will create an architecture vision document from scratch.
Introduction to Machine Learning and Data Science using Autonomous Database ...Sandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning , autoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use different open source frameworks as well as some of the tools in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automaticall
Machine Learning in Autonomous Data WarehouseSandesh Rao
Machine Learning in Autonomous Data Warehouse: One can use Oracle Autonomous Data Warehouse for machine learning. There are several ways to do this. This presentation explores these different but related options for performing machine learning. Each of these options enables people with different backgrounds to engage with building machine learning solutions on their data. At the end of the session, you will know which option will work best for you
This is from the Bay area Cloud Computing event https://www.meetup.com/All-Things-Cloud-Computing-Bay-Area/events/271017950/
YouTube Link: https://youtu.be/aGu0fbkHhek
** Data Science Master Program: https://www.edureka.co/masters-program/data-scientist-certification **
This Edureka PPT on "Data Science Full Course" provides an end to end, detailed and comprehensive knowledge on Data Science. This Data Science PPT will start with basics of Statistics and Probability and then moves to Machine Learning and Finally ends the journey with Deep Learning and AI. For Data-sets and Codes discussed in this PPT, drop a comment.
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Creating a Data validation and Testing StrategyRTTS
Creating A Data Validation & Testing Strategy
Are you struggling with formulating a strategy for how to validate the massive amount of data continuously entering your data warehouse or data lake?
We can help you!
Learn how RTTS’ Data Validation Assessment provides:
- an evaluation of your current data validation process
- recommendations on how to improve your process and
- a proposal for successful implementation
This slide deck addresses the following issues:
- How do I find out if I have bad data?
- How do I ensure I am testing the proper data permutations?
- How much of my data needs to be validated and automated?
- Which critical data endpoints need to be tested?
- How do I test data in my cloud environments?
And much more!
For more information, visit:
https://www.rttsweb.com/services/solutions/data-validation-assessment
What is a Data Warehouse and How Do I Test It?RTTS
ETL Testing: A primer for Testers on Data Warehouses, ETL, Business Intelligence and how to test them.
Are you hearing and reading about Big Data, Enterprise Data Warehouses (EDW), the ETL Process and Business Intelligence (BI)? The software markets for EDW and BI are quickly approaching $22 billion, according to Gartner, and Big Data is growing at an exponential pace.
Are you being tasked to test these environments or would you like to learn about them and be prepared for when you are asked to test them?
RTTS, the Software Quality Experts, provided this groundbreaking webinar, based upon our many years of experience in providing software quality solutions for more than 400 companies.
You will learn the answer to the following questions:
• What is Big Data and what does it mean to me?
• What are the business reasons for a building a Data Warehouse and for using Business Intelligence software?
• How do Data Warehouses, Business Intelligence tools and ETL work from a technical perspective?
• Who are the primary players in this software space?
• How do I test these environments?
• What tools should I use?
This slide deck is geared towards:
QA Testers
Data Architects
Business Analysts
ETL Developers
Operations Teams
Project Managers
...and anyone else who is (a) new to the EDW space, (b) wants to be educated in the business and technical sides and (c) wants to understand how to test them.
What Is Data Science? Data Science Course - Data Science Tutorial For Beginne...Edureka!
This Edureka Data Science course slides will take you through the basics of Data Science - why Data Science, what is Data Science, use cases, BI vs Data Science, Data Science tools and Data Science lifecycle process. This is ideal for beginners to get started with learning data science.
You can read the blog here: https://goo.gl/OoDCxz
You can also take a complete structured training, check out the details here: https://goo.gl/AfxwBc
Big Data and BI Tools - BI Reporting for Bay Area Startups User GroupScott Mitchell
This presentation was presented at the July 8th 2014 user group meeting for BI Reporting for Bay Area Start Ups
Content - Creation Infocepts/DWApplications
Presented by: Scott Mitchell - DWApplications
Data Warehousing in Pharma: How to Find Bad Data while Meeting Regulatory Req...RTTS
In the U.S., pharmaceutical firms must meet electronic record-keeping regulations set by the Food and Drug Administration (FDA). The regulation is Title 21 CFR Part 11, commonly known as Part 11.
Part 11 requires regulated firms to implement controls for software and systems involved in processing many forms of data as part of business operations and product development.
Enterprise data warehouses are used by the pharmaceutical and medical device industries for storing data covered by Part 11. QuerySurge, the only test tool designed specifically for automating the testing of data warehouses and the ETL process, is the market leader in testing data warehouses used by Part 11-governed companies.
For more on QuerySurge and Pharma, please visit
http://www.querysurge.com/solutions/pharmaceutical-industry
AUSOUG - Introducing New AI Ops Innovations in Oracle 19c Autonomous Health F...Sandesh Rao
Oracle Autonomous Health Framework (AHF) is Oracle’s Artificial Intelligence Operations platform for autonomous database health management. This session will focus on enhancements to current functionality and new features in 18c and coming in 19c. First successfully introduced in Cluster Health Advisor, and extended to Trace File Analyzer and Hang Manager, Oracle AHF’s applied machine learning technology now enhances additional framework components. You will learn how to utilize these features for determining workload footprint, ongoing monitoring, early detection of anomalies and performance issues, their root causes and corrective actions, prevention of node or database failures, and targeted postmortem analysis enabling quick resolution.
QuerySurge - the automated Data Testing solutionRTTS
QuerySurge is the leading Data Testing solution built specifically to automate the testing of Data Warehouses & Big Data. QuerySurge ensures that the data extracted from data sources remains intact in the target data store by analyzing and pinpointing any differences quickly.
And QuerySurge makes it easy for both novice and experienced team members to validate their organization's data quickly through Query Wizards while still allowing power users the flexibility they need.
All with deep dive reporting and data health dashboards that quickly provides you with a holistic view of your project’s data.
Types of Automated Data Testing
--------------------------------------------
QuerySurge provides data testing solutions for all of your automated data testing needs
- Data Warehouse testing & ETL testing
- Big Data (Hadoop, NoSQL) testing
- Data Interface testing
- Data Migration testing
- Database Upgrade testing
FREE TRIAL
www.QuerySurge.com
Real-time Recommendations for Retail: Architecture, Algorithms, and DesignJuliet Hougland
Users are constantly searching for new content and to stay competitive organizations must act immediately based on up-to-date data. Outdated recommendations decrease the likelihood of presenting the right offer and make it harder to maintain customer loyalty. In order to provide the most relevant recommendations and increase engagement, organizations must track customer interactions and re-score recommendations on the fly.
Data sources have expanded dramatically to include a wealth of historical data and a constant influx of behavior data. The key to moving from predictive models, applied in batch, to models that provide responses in real time, is to focus on the efficiency of model application. The speed that recommendations can be served is influenced by:
Architecture of the recommendation serving platform
Choice of recommendation algorithm
Datastore access patterns
In this presentation, we’ll discuss how developers can use open source components like HBase and Kiji to develop low-latency recommendation models that can be easily deployed by e-commerce companies. We will give practical advice on how to choose models and design data stores that make use of the architecture and quickly serve new recommendations.
Testing Big Data: Automated Testing of Hadoop with QuerySurgeRTTS
Are You Ready? Stepping Up To The Big Data Challenge In 2016 - Learn why Testing is pivotal to the success of your Big Data Strategy.
According to a new report by analyst firm IDG, 70% of enterprises have either deployed or are planning to deploy big data projects and programs this year due to the increase in the amount of data they need to manage.
The growing variety of new data sources is pushing organizations to look for streamlined ways to manage complexities and get the most out of their data-related investments. The companies that do this correctly are realizing the power of big data for business expansion and growth.
Learn why testing your enterprise's data is pivotal for success with big data and Hadoop. Learn how to increase your testing speed, boost your testing coverage (up to 100%), and improve the level of quality within your data - all with one data testing tool.
Data Verification In QA Department FinalWayne Yaddow
Data warehouse and ETL testing should be conducted according to a process and checklist. This presentation provides an overview of recommended methods.
Data Warehouse Testing in the Pharmaceutical IndustryRTTS
In the U.S., pharmaceutical firms and medical device manufacturers must meet electronic record-keeping regulations set by the Food and Drug Administration (FDA). The regulation is Title 21 CFR Part 11, commonly known as Part 11.
Part 11 requires regulated firms to implement controls for software and systems involved in processing many forms of data as part of business operations and product development.
Enterprise data warehouses are used by the pharmaceutical and medical device industries for storing data covered by Part 11 (for example, Safety Data and Clinical Study project data). QuerySurge, the only test tool designed specifically for automating the testing of data warehouses and the ETL process, has been effective in testing data warehouses used by Part 11-governed companies. The purpose of QuerySurge is to assure that your warehouse is not populated with bad data.
In industry surveys, bad data has been found in every database and data warehouse studied and is estimated to cost firms on average $8.2 million annually, according to analyst firm Gartner. Most firms test far less than 10% of their data, leaving at risk the rest of the data they are using for critical audits and compliance reporting. QuerySurge can test up to 100% of your data and help assure your organization that this critical information is accurate.
QuerySurge not only helps in eliminating bad data, but is also designed to support Part 11 compliance.
Learn more at www.QuerySurge.com
"How to document your decisions", Dmytro Ovcharenko Fwdays
We will perform architecture kata around a proposed business case. We will review ADD in detail. How usually architecture vision document looks like. How to match your architecture drivers and proposed architecture decisions in architecture views. We will review what is ATAM and how to perform analysis of your decisions in the right way. And finally, we will create an architecture vision document from scratch.
Introduction to Machine Learning and Data Science using Autonomous Database ...Sandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning , autoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use different open source frameworks as well as some of the tools in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automaticall
Machine Learning in Autonomous Data WarehouseSandesh Rao
Machine Learning in Autonomous Data Warehouse: One can use Oracle Autonomous Data Warehouse for machine learning. There are several ways to do this. This presentation explores these different but related options for performing machine learning. Each of these options enables people with different backgrounds to engage with building machine learning solutions on their data. At the end of the session, you will know which option will work best for you
This is from the Bay area Cloud Computing event https://www.meetup.com/All-Things-Cloud-Computing-Bay-Area/events/271017950/
Introduction to Machine learning - DBA's to data scientists - Oct 2020 - OGBEmeaSandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning with examples, AutoML for training models and this ends with an example of how to predict fraud , to determining shopping patterns to Wine picking and different algorithms as an example and also how to predict workload for your databases. We will also use OML in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automatically
Introduction to Machine Learning - From DBA's to Data Scientists - OGBEMEASandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning with examples, AutoML for training models and this ends with an example of how to predict fraud , to determining shopping patterns to Wine picking and different algorithms as an example and also how to predict workload for your databases. We will also use OML in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automatically
Introduction to AutoML and Data Science using the Oracle Autonomous Database ...Sandesh Rao
We are entering a new era in the database with the introduction of the Oracle Autonomous Database. AI and Machine Learning are center stage to most projects and assist in making complex decisions which was not possible before. Most data science projects don’t get beyond the data scientist and rarely operationalize their predictive models. there are new toolsets and methods available everyday which make this an extremely dynamic space. There are different categories of users who want to use the algorithms , the toolsets but don't know where to start. Whether you are a data scientist who wants to play with data and build your own models or make use of the database features with the built in models or use the specific AI services within a specific vertical such as Insurance or Healthcare . We will take a glimpse at Oracle's Machine Learning Zeppelin-based notebooks for Oracle Autonomous Data Warehouse Cloud to how Oracle uses AIOps and Applied Machine learning for its own operations and the Oracle AI Platform Cloud Service to provided an all rounded view of what Oracle is upto in this space
Architecting the Framework for Compliance & Risk Managementjadams6
Privacy and protection of personal information is a hot topic in data governance. However, the compliance challenge is in creating audit defensibility that ensures practices are compliant and performed in a way that is scalable, transparent, and defensible; thus creating “Audit Resilience.” Data practitioners often struggle with viewing the world from the auditor’s perspective. This presentation focuses on how to create the foundational governance framework supporting a data control model required to produce clean audit findings. These capabilities are critical in a world where due diligence and compliance with best practices are critical in addressing the impacts of security and privacy breaches. The companies in the news recently drive home these points.
ML in GRC: Supporting Human Decision Making for Regulatory Adherence with Mac...BigML, Inc
This is a real-life Machine Learning use case about integrated risk.
Speakers: Thomas Rengersen, Product Owner of the Governance Risk and Compliance Tool for Rabobank, and Thomas Alderse Baas, Co-Founder and Director of The Bowmen Group.
*ML in GRC 2021: Virtual Conference.
#ATAGTR2021 Presentation : "Use of AI and ML in Performance Testing" by Adolf...Agile Testing Alliance
Interactive Session on "Use of AI and ML in Performance Testing" by Adolf Patel Performance Test Architect Cognizant at #ATAGTR2021.
#ATAGTR2021 was the 6th Edition of Global Testing Retreat.
The video recording of the session is now available on the following link: https://www.youtube.com/watch?v=ajyPSmmswpM
To know more about #ATAGTR2021, please visit:https://gtr.agiletestingalliance.org/
Learn statistics and expert opinions on the state of the market regarding data quality in 2023.
Learn about:
- statistics and expert opinions
- the key focus of data quality in 2023
- the Data Maturity Model
- DevOps for data and CI/CD pipelines
- data validation and ETL testing
- test automation
The Automation Firehose: Be Strategic & Tactical With Your Mobile & Web TestingPerfecto by Perforce
The widespread adoption of test automation has created many challenges — for everything from development lifecycle integration to scripting strategy.
One pitfall of automation is that teams often rush to automate everything they can. This is the automation firehose.
However, just because a scenario CAN be automated does not mean it SHOULD be automated. For scenarios that should be automated, teams must adopt implementation plans to ensure tests are reliable and deriving value.
Join this webinar led by Perfecto’s Chief Evangelist, Eran Kinsbruner, along with Thomas Haver, Manager of Automation & Delivery. In this session, the audience will:
-Understand which test scenarios to automate.
-Learn how to maximize the benefits of automation.
-Receive a checklist to determine automation feasibility and ROI.
Form I-9/E-Verify and Compliance & Managed Services for HCM Cloud
Presented by the Department of Homeland Security, ERP Risk Advisors, and Smart ERP Solutions
Part 2: Compliance With Managed Services for Human Capital Management (HCM)
In this webinar, discover how Compliance and Managed Services for Human Capital Management (HCM) Cloud can optimize and streamline your HR processes, saving time and resources and staying compliant. ERP Risk Advisors and SmartERP will showcase the benefits of utilizing managed services, focusing on automation, security, scalability, and compliance. Learn how managed services can enhance your HCM Cloud experience and allow HR professionals to focus on strategic initiatives while ensuring data privacy and regulatory compliance. There will be an open Q&A.
Some of the topics that will be presented:
What do we mean by Compliance through Managed Services?
Flexible Models for Managed Services as an extension of your Internal Team
Ensuring Oracle License Compliance and Reducing risk through Role Management
Demonstrate significant risks with the use of Seeded and Hybrid roles
Automated processes and their impact on HR efficiency
Compliance assurance and risk management through enabling and evaluating audit logs
This webinar will focus on content for HCM professionals that are utilizing HCM Cloud or are currently investigating moving to HCM Cloud.
Note: Both webinars are tailored to HR professionals, compliance officers, and managers. The aim is to offer valuable insights and practical tips for handling DHS Form I-9 compliance and leveraging Managed Services to optimize HCM Cloud operations and compliance.
Best Practices for Rating and Policy Administration System ReplacementEdgewater
Edgewater Technology, AQS and ISO joined forces to share best practices for replacing policy administration and rating systems for P&C insurance companies.
In medicine - an MRI can quickly reveal a hidden ailment and actionable insight to get better. For IT and business leaders whose key concern with the mainframe is the platform costs and lean operations - the CA Mainframe Resource Intelligene reveals multiple sources of hidden mainframe costs and operational inefficiencies along with actionable recommendations.View this slideshare to understand how this new SaaS offering from CA brings together automation, speed, analytics and mainframe expertise of 40+ years. CA Mainframe Resource Intelligence reports answer your CIO’s toughest questions about mainframe optimization and potential for digital transformation.
For more information, please contact your account director or mainframe specialist at:
http://ow.ly/PALG50htHgF
Similar to Introduction to Machine Learning and Data Science using the Autonomous database - Sangam 2019 (20)
Whats new in Autonomous Database in 2022Sandesh Rao
This session covers the new features and happenings in the autonomous database world and will help answer more questions DBAs and Developers will have on the Autonomous Database, from provisioning to backups, troubleshooting, tips and tricks, security and HA. This is a good introduction for on-prem DBAs who want to learn how this works quickly without spending too much time on it. Questions like what does the free tier cover, how do I do backup or if it's automated, how do I manage it, how to scale up and down, how to secure their environment, how to use mtls, how to use tools like SQLDeveloper and SQLModeler, performance tuning all in a quick 45-minute session which might take weeks to pick up reading documentation or spanning several presentations
Oracle Database performance tuning using oratopSandesh Rao
Oratop is a text-based user interface tool for monitoring basic database operations in real-time. This presentation will go into depth on how to use the tool and some example scenarios. It can be used for both RAC and single-instance databases and in combination with top to get a more holistic view of system performance and identify any bottlenecks.
Analysis of Database Issues using AHF and Machine Learning v2 - AOUG2022Sandesh Rao
Oracle Autonomous Health Framework (AHF) is Oracle’s Artificial Intelligence Operations platform for autonomous database health management. This session will focus on enhancements to current functionality and new features in 21c. We will discuss how to use the data which is derived from the Bayesian Net framework of AHF to conduct root cause analysis, telemetry and remediations for issues. You will learn to utilize these features to determine workload footprint, ongoing monitoring, early detection of anomalies and performance issues, their root causes and corrective actions, prevention of node or database failures, and targeted postmortem analysis enabling quick resolution.
Session Highlights:
• Insights into AHF enhancements to current functionality and new features in 21c
• Learn early detection of anomalies and performance issues, their root causes and corrective actions
• Targeted postmortem analysis enabling quick resolution
Analysis of Database Issues using AHF and Machine Learning v2 - SOUGSandesh Rao
Oracle Autonomous Health Framework (AHF) is Oracle’s Artificial Intelligence Operations platform for autonomous database health management. This session will focus on enhancements to current functionality and new features in 21c. We will discuss how to use the data which is derived from the Bayesian Net framework of AHF to conduct root cause analysis, telemetry and remediations for issues. You will learn to utilize these features to determine workload footprint, ongoing monitoring, early detection of anomalies and performance issues, their root causes and corrective actions, prevention of node or database failures, and targeted postmortem analysis enabling quick resolution.
Session Highlights:
• Insights into AHF enhancements to current functionality and new features in 21c
• Learn early detection of anomalies and performance issues, their root causes and corrective actions
• Targeted postmortem analysis enabling quick resolution
15 Troubleshooting Tips and Tricks for database 21c - OGBEMEA KSAOUGSandesh Rao
This session will focus on 15 troubleshooting tips and tricks for DBA’s covering tools from the Oracle Autonomous Health Framework (AHF) like Trace file Analyzer (TFA) to collect , organize and analyze log data , Exachk and orachk to perform mass best practices analysis and automation , Cluster Health Advisor to debug node evictions and calibrate the framework , OSWatcher and its analysis engine , oratop for pinpointing performance issues and many others to make one feel like a rockstar DBA
How to use Exachk effectively to manage Exadata environments OGBEmeaSandesh Rao
Exachk is a tool for helping with best practices with an Exadata Box. This presentation will go through setup , usage , options and how to use it more effectively to be more proactive in fixing issues with an Exadata environment. There are features like baselines , scheduler for ongoing automation , Collection Manager an Apex based interface which is used to determine the common problems , how to setup this dashboard all for free and in under 30 minutes to be a rockstar Exadata DBA
Troubleshooting tips and tricks for Oracle Database Oct 2020Sandesh Rao
This talk presents 15 different tips and tricks using tools to better troubleshoot and debug problems with Database , Oracle RAC and Oracle Clusterware , ASM and how to get the right pieces of data with the least of commands which today most people do manually. This session will cover tools from the Oracle Autonomous Health Framework (AHF) like Trace file Analyzer (TFA) to collect , organize and analyze log data , Exachk and orachk to perform mass best practices analysis and automation , Cluster Health Advisor to debug node evictions and calibrate the framework , OSWatcher and its analysis engine , oratop for pinpointing performance issues and many others to make one feel like a rockstar DBA.
Oracle Autonomous Health Service- For Protecting Your On-Premise Databases- F...Sandesh Rao
Oracle Autonomous Health Service is a new feature in Oracle Database 19c that supports centralized diagnostic monitoring of your on-premise Oracle RAC database clusters. In this session learn how to free up local cluster resources and improve monitoring efficiency with this centralized service-based solution. Practical tips and best practices for setup are covered along with a demonstration of its responsive browser-based dashboard and triage console. Specific Oracle RAC database, Oracle Automatic Storage Management, and cluster issues are rapidly triaged utilizing the built-in support for Trace File Analyzer and ORAchk/EXAchk collections, Cluster Health Advisor problem diagnosis, Hang Manager and Memory Guard events, and QoS management workload performance metrics.
The Machine Learning behind the Autonomous Database ILOUG Feb 2020 Sandesh Rao
Autonomous Database is one of the hottest Oracle products where we have attempted to use Machine Learning for several aspects of the service. We take a view on our current state of ML in the Autonomous Database Cloud and how do we process this data in ADW/ATP with zeppelin notebooks to find anomalies in them to troubleshoot them at a scale of several petabytes a year and conduct AIOps. We will cover some sample notebooks to some use cases we will cover are a Log Anomaly timeline which we reduce significant amounts of logs using semi-supervised machine learning techniques to reduce logs and match them in near real time. Some of the other use cases is to use convolution filters...
Troubleshooting Tips and Tricks for Database 19c ILOUG Feb 2020Sandesh Rao
This session will focus on 19 troubleshooting tips and tricks for DBAs covering tools from the Oracle Autonomous Health Framework (AHF) like Trace file Analyzer (TFA) to collect , organize and analyze log data , Exachk and orachk to perform mass best practices analysis and automation , Cluster Health Advisor to debug node evictions and calibrate the framework , OSWatcher and its analysis engine , oratop for pinpointing performance issues and many others to make one feel like a rockstar DBA
Troubleshooting Tips and Tricks for Database 19c - Sangam 2019Sandesh Rao
DBA's always have a bunch of scripts to do their daily tasks. How to find that stuck session, how to find who is consuming the most resources, how do I take a stack of multiple processes? This session will focus on troubleshooting tips and tricks for DBA's covering tools from the Oracle Autonomous Health Framework (AHF) like Trace file Analyzer (TFA) to collect , organize and analyze log data , Exachk and orachk to perform mass best practices analysis and automation , Cluster Health Advisor to debug node evictions and calibrate the framework , OSWatcher and its analysis engine , Oratop for pinpointing performance issues and many other native Database features like short stacks, system state summaries, quickly spot hangs across RAC clusters among some of them to make your jobs a lot more efficient and make you look good to your bosses !!
20 Tips and Tricks with the Autonomous Database Sandesh Rao
This session will focus on the Autonomous Database which is Oracle’s latest Cloud product and will provide the latest news on what is happening in this space. Some of the topics covered will be - How do I scale the database , How to use the machine learning notebooks , details on the free tier of the database and how to use it among some of the tips and tricks to give you all the skills you need to use the database for the first time if you have not used it before or to better improve your skills if you’re already a power user this will extend your skills and also educate you on new features of the Autonomous Database
The Machine Learning behind the Autonomous Database- EMEA Tour Oct 2019 Sandesh Rao
Autonomous Database is one of the hottest Oracle products where we have attempted to use Machine Learning for several aspects of the service. We take a view on our current state of ML in the Autonomous Database Cloud and how do we process this data in ADW/ATP with zeppelin notebooks to find anomalies in them to troubleshoot them at a scale of several petabytes a year and conduct AIOps. We will cover some sample notebooks to some use cases we will cover are a Log Anomaly timeline which we reduce significant amounts of logs using semi-supervised machine learning techniques to reduce logs and match them in near real time. Some of the other use cases is to use convolution filters to determine maintenance windows within the database workloads , determine best times to do database backups , security anomaly timelines and many others. This presentation will accompany several examples with how to apply these techniques , machine learning knowledge is preferred but not a prerequisite
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Introduction to Machine Learning and Data Science using the Autonomous database - Sangam 2019
1. VP AIOps for the Autonomous Database
Sandesh Rao
Sangam AIOUG
Introduction to Machine Learning and Data
Science using the Oracle Autonomous Database
@sandeshr
https://www.linkedin.com/in/raosandesh/
https://www.slideshare.net/SandeshRao4
2. The following is intended to outline our general product direction. It is intended for information
purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any
material, code, or functionality, and should not be relied upon in making purchasing decisions. The
development, release, timing, and pricing of any features or functionality described for Oracle’s
products may change and remains at the sole discretion of Oracle Corporation.
Statements in this presentation relating to Oracle’s future plans, expectations, beliefs, intentions and
prospects are “forward-looking statements” and are subject to material risks and uncertainties. A
detailed discussion of these factors and other risks that affect our business is contained in Oracle’s
Securities and Exchange Commission (SEC) filings, including our most recent reports on Form 10-K and
Form 10-Q under the heading “Risk Factors.” These filings are available on the SEC’s website or on
Oracle’s website at http://www.oracle.com/investor. All information in this presentation is current as of
September 2019 and Oracle undertakes no duty to update any statement in light of new information or
future events.
Safe harbor statement
3. 1. Overview of ML and the Autonomous Database
2. Regression
3. Classification
4. Clustering
5. Anomaly detection
6. Workload prediction
7. Dynamic maintenance windows
8. Oracle Machine Learning examples
Agenda
5. Tasks Specific to Business and Innovation
• Architecture, planning, data modeling
• Data security and lifecycle management
• Application related tuning
• End-to-End service level management
Maintenance Tasks
• Configuration and tuning of systems, network, storage
• Database provisioning, patching
• Database backups, H/A, disaster recovery
• Database optimization
Traditionally DBAs are Responsible for:
Value Scale
Innovation
Maintenance
6. Tasks Specific to Business and Innovation
• Architecture, planning, data modeling
• Data security and lifecycle management
• Application related tuning
• End-to-End service level management
Maintenance Tasks
• Configuration and tuning of systems, network, storage
• Database provisioning, patching
• Database backups, H/A, disaster recovery
• Database optimization
Freedom from Drudgery for DBA: More Time to Innovate and Improve the Business
Autonomous Database Removes Generic Tasks
Value Scale
Innovation
Maintenance
7. Machine Learning
Solving data-driven
problems
Discovering insights
Making predictions
Data Security
Data classification,
Data life-cycle mgmt
Application Tuning
SQL tuning,
connection mgmt
The Evolution of the DBA/Database Developer Role
Data Engineer
Architecture,
“data wrangler”
8. Data extraction
Data wrangling
Deriving new attributes
(“feature engineering”)
…
…
…
Import predictions & insights
Translate and deploy ML models
Automate
You Are Probably Already Doing Most of This Work!
Database Developer to Data Scientist Journey
1 - https://www.infoworld.com/article/3228245/data-science/the-80-20-data-science-dilemma.html
Typically 80% of the work
Most data scientists spend only 20 percent of their time
on actual data analysis and 80 percent of their time
finding, cleaning, and reorganizing huge amounts of
data, which is an inefficient data strategy1
Eliminated or minimized with Oracle
Data Management platform becomes
combine/hybrid DM + machine learning platform
9. Albert Einstein
“If I had an hour to solve a
problem I'd spend 55
minutes thinking about the
problem and 5 minutes
thinking about solutions.”
10. Lots of Data needs to be crunched
• No time to manually sift through the data
Machine Learning has become accessible
• Anyone can be a Data Scientist
• Software and algorithms are available
• Frameworks allow for massive training with no coding
• CI/CD available for MLOps
Business use cases
- How to find value from the data
Why Machine Learning for us and why now?
11. Analytics Value vs. Maturity
Reports &
Dashboards
Data
Information
Predictions & Insights Appls with ML
Analytical Maturity
ValueofAnalytics
Diagnostic
Analysis &
Reports
Predictive /
Machine
Learning
“ML Enabled”
Applications
What Happened?
Why it Happened?
What WILL happen?
Automated ML Appls
13. ML Project Workflow
Set the business objectives
Gather compare and
clean data
Identify and extract features
(important columns) from imported data
This helps us identify the efficiency of
the algorithm
Take the input data which is also called the training
data and apply the algorithm to it
For the algorithm to function efficiently, it is
important to pick the right value for hyper parameters
(algorithm input parameters to the algorithm)
Once the training data in
the algorithm are
combined we get a model
1
2
3
4
5
14. Types of Machine Learning
Supervised Learning
Predict future outcomes with the help of
training data provided by human experts
Semi-Supervised Learning
Discover patterns within raw data and make
predictions, which are then reviewed by
human experts, who provide feedback which
is used to improve the model accuracy
Unsupervised Learning
Find patterns without any external input other
than the raw data
Reinforcement Learning
Take decisions based on past rewards for this
type of action
29. What is Workload
Automatically
check
workload for
past x mins
Decide if
workload is
abnormally
high
Highlight any
abnormal
workload
issues
Optionally run on
demand
Optionally snooze
checking of a
component
Calculated via machine learning
30. Adaptive Learning
Workload Process
Captures metrics for key
performance dimensions across 5 X
1 minute time windows
CAPTURE1
Using semi-supervised learning via
SME threshold rules, the following
models are retrained :
• Isolation Forest
• One-Class Support Vector Machine
• Local Outlier Factor
Model with highest confidence
becomes the primary, if confidence is
high enough
TRAIN2
Straight after capture, the primary model
is used to predict anomalies.
Where anomalies are identified, metrics
are compared to SME threshold rules to
identify the type of anomaly
PREDICT3
Every
5 Mins
Every
Week
Every
5 Mins
31. Prediction (Every 5 minutes)
5 X 1 min metrics captured for
each dimension & ASH report
captured for later analysis
Metrics evaluated by the primary model to
determine if there are anomalies
If there is no primary model
(i.e. <7 days of data or <=95% model confidence)
then SME rules are used for anomaly detection
Each anomaly is compared against
the SME rules to determine which
dimension it applies to
Any anomalies are raised
along with recently
captured ASH report
32. Resource usage prediction
Configurable threshold
boundary – notify Admin of
forecasts above here
Actual values
(Black)
Forecast values
(Blue line)
Upper & lower
forecast range
(light blue area)
Unusual values
(anomalies)
Future forecast
values
34. Identify Relevant Workload Metrics
• Ex: Average Active Sessions, CPU/Mem/IO Utilization
Time Series Decomposition
• Trend
• Seasonality
• Residual
Workload Seasonality Determination Locating Minimas
Optimum Window Identification and Validation
Model Generation and Training Flow
Maintenance Slot Identification
35. Maintenance window identification
START_TIME CNT
2018-04-11 15:00:00 290
2018-04-11 16:00:00 31120
2018-04-11 17:00:00 21530
2018-04-11 18:00:00 26240
2018-04-11 19:00:00 40520
2018-04-11 20:00:00 54270
2018-04-11 21:00:00 51460
2018-04-11 22:00:00 44310
2018-04-11 23:00:00 25690
START_TIME
2018-04-11 15:00:00 -0.226098
2018-04-11 16:00:00 -0.069821
2018-04-11 17:00:00 -0.350088
2018-04-11 18:00:00 -0.187483
2018-04-11 19:00:00 -0.513240
2018-04-11 20:00:00 0.019737
2018-04-11 21:00:00 0.059213
2018-04-11 22:00:00 -0.011312
2018-04-11 23:00:00 -0.179156
START_TIME
2018-04-11 15:00:00 5.669881
2018-04-11 16:00:00 10.345606
2018-04-11 17:00:00 9.977203
2018-04-11 18:00:00 10.175040
2018-04-11 19:00:00 10.609551
2018-04-11 20:00:00 10.901727
2018-04-11 21:00:00 10.848560
2018-04-11 22:00:00 10.698966
2018-04-11 23:00:00 10.153857
Current Date : 2018-05-12 15:00:00
Current Position in Seasonality : -0.22609829742533585
Best Maintenance Period in next Cycle : 2018-05-12 19:00:00
Worst Maintenance Period in next Cycle : 2018-05-13 08:00:00
Original observation data
1
Convolution filter & average
2
Calculate seasonality
3
Use seasonality to
predict best
maintenance window
4
37. Simple SQL Syntax—Statistical Comparisons (t-tests)
Compare AVE Purchase Amounts Men vs. Women Grouped_By INCOME_LEVEL
Statistical Functions
SELECT SUBSTR(cust_income_level, 1, 22) income_level,
AVG(DECODE(cust_gender, 'M', amount_sold, null)) sold_to_men,
AVG(DECODE(cust_gender, 'F', amount_sold, null)) sold_to_women,
STATS_T_TEST_INDEPU(cust_gender, amount_sold, 'STATISTIC', 'F') t_observed,
STATS_T_TEST_INDEPU(cust_gender, amount_sold) two_sided_p_value
FROM customers c, sales s
WHERE c.cust_id = s.cust_id
GROUP BY ROLLUP(cust_income_level)
ORDER BY income_level, sold_to_men, sold_to_women, t_observed;
STATS_T_TEST_INDEPU (SQL) Example;
P_Values < 05 show statistically
significantly differences in the amounts
purchased by men vs. women
38. Simple SQL Syntax—Attribute Importance - ML Model Build (PL/SQL)
OAA Model Build and Real-time SQL Apply Prediction
BEGIN
DBMS_DATA_MINING.CREATE_MODEL(
model_name => 'BUY_INSURANCE_AI',
mining_function => DBMS_DATA_MINING.ATTRIBUTE_IMPORTANCE,
data_table_name => 'CUST_INSUR_LTV',
case_id_column_name => 'cust_id',
target_column_name => 'BUY_INSURANCE',
settings_table_name => 'Att_Import_Mode_Settings');
END;
/
SELECT attribute_name, rank , attribute_value
FROM BUY_INSURANCE_AI
ORDER BY rank, attribute_name;
Model Results (SQL query)
ATTRIBUTE_NAME RANK ATTRIBUTE_VALUE
BANK_FUNDS 1 0.2161
MONEY_MONTLY_OVERDRAWN 2 0.1489
N_TRANS_ATM 3 0.1463
N_TRANS_TELLER 4 0.1156
T_AMOUNT_AUTOM_PAYMENTS 5 0.1095
A1A2A3A4 A5A6 A7
39. Key Features
Collaborative UI for data scientists
• Packaged with Autonomous Data
Warehouse Cloud (V1)
• Easy access to shared notebooks,
templates, permissions, scheduler, etc.
• SQL ML algorithms API (V1)
• Supports deployment of ML analytics
Machine Learning Notebook for Autonomous Data Warehouse Cloud
Oracle Machine Learning
40. Multiple Languages UIs Supported for End Users & Apps Development
Oracle Machine Leaning
Application DevelopersDBAs
R & Python Data Scientists “Citizen” Data ScientistsNotebook Users & DS Teams
New! New!
42. Define Problem Statement
Poorly Defined Better
Data Mining
Technique
Predict employees that leave
• Based on past employees that voluntarily left:
• Create New Attribute EmplTurnover à O/1
Predict customers that churn
• Based on past customers that have churned:
• Create New Attribute Churn à YES/NO
Target “best” customers
• Recency, Frequency Monetary (RFM) Analysis
• Specific Dollar Amount over Time Window:
• Who has spent $500+ in most recent 18 months
How can I make more $$? • What helps me sell soft drinks & coffee?
Which customers are likely to buy? • How much is each customer likely to spend?
Who are my “best customers”? • What descriptive “rules” describe “best customers”?
How can I combat fraud?
• Which transactions are the most anomalous?
• Then roll-up to physician, claimant, employee…
45. Create New Derived Attributes or “Engineered Features”
Feature Engineering
Source Attribute New Attribute/”Engineered Feature”
Date of Birth AGE
Address DISTANCE_TO_DESTINATION
COMMUTE_TIME
Call detail records (CDRs) #_DROPPED_CALLS
PERCENT_iNTERNATIONAL
Salary PERCENT_VS_PEERS
Purchases TOTALS_PER_CATEGORY (e.g. Food,
Clothing)
46. Create new derived attributes to tease more
information out of the data. For example:
• RECENCY, FREQUENCY, MONETARY
(RFM Analysis)
Create New Derived Attributes or “Engineered Features”
Feature Engineering
47. Data remains in Database
• Model building and scoring occur in-
database
• Leverage investment in Oracle IT
• Eliminate data duplication
- Eliminate separate analytical servers
Deliver enterprise-wide
“predictive” applications
Don’t move the Data
Traditional ML
Hours, Days or Weeks
Data Extraction
Data Prep &
Transformation
Data Mining
Model Building
Data Mining
Model “Scoring”
Data Prep. &
Transformation
Data Import
avings
Model “Scoring”
Embedded Data Prep
Data Preparation
Model Building
Oracle’s in-DB Machine Learning
Secs, Mins or Hours
ORACLE
AUTONOMOUS
DATABASE
48. Increasing sources of relevant data can
boost model accuracy
More Data Variety—Better Predictive Models
Model with 20 variables
Model with “Big Data” and
hundreds -- thousands of
input variables including:
• Demographic data
• Purchase POS transactional
data
• “Unstructured data”, text &
comments
• Spatial location data
• Long term vs. recent
historical behavior
• Web visits
• Sensor data
• etc.
Naïve Guess
or Random
100%
0% Population Size
Responders
Model with 75
variables
Model with 250
variables
100%
Engineered Features – Derived attributes/variable
that reflect domain knowledge—key to best models
49. First, Identify the Key Attributes That Most Influence the Target Attribute
Modeling and Machine Learning
Attribute Importance Model
50. Next, Build Predictive Models to Predict Customers who are Likely to Have Good_Credit
Modeling and Machine Learning
Split Data into Train and Test
Build and Test Classification Model
51. Test the ML model’s accuracy
• Randomly selected “hold out” sample
of data that was used to train the ML
model
• Compute Cumulative Gains, Lift,
Accuracy, etc.
• Review the attributes used in the model
and model coefficients
• Make sure the model makes sense
Next, Build Predictive Models to Predict Customers who are Likely to Have Good_Credit
Model Evaluation (Machine Learning)
Model Evaluation
52. Simple SQL Apply scripts run 100% inside the
Database for immediate ML model
deployment
Apply the Models to Predict “Best Customers”
Deployment
Model Apply/”Scoring”
55. Manage and Analyze All Your Data
Big Data SQL / R
SQL / R / Python
Object
Store
“Engineered Features”
– Derived attributes
that reflect domain
knowledge—key to
best models e.g.:
• Counts
• Totals
• Changes
over time
Boil down the Data Lake
Architecturally,
lots of options
and flexibility
56. In-Database Machine Learning
More Models
Better Models
Faster, More Secure
Less Cost
Ready to Deploy!
No Need To Extract and
Move Data
Data stays in Database
Zero time required.
No production impact.
Data Preparation and
Transformation
Accelerated with
Automatic Data Prep
No separate environment
required. Much faster data prep.
Data stays protected and secured.
Data Mining and
Model Building
SQL, R, Python
Oracle Data Miner UI
OML Notebooks
Oracle Data Miner and AutoML
greatly speed model building.
Less skill required. No coding.
No Need to Transform
Production Data
Embedded Data
Preparation
No need for second
production instance.
Model Scoring
Accelerated Via
Exadata Database Machine
Faster model validation
Easy to repeat model building as often as needed
57. • OAA (Oracle Data Mining + Oracle R Enterprise) and ORAAH combined
• OAA includes support for Partitioned Models, Transactional, Unstructured, Geo-spatial, Graph data. etc,
Oracle’s Machine Learning & Adv. Analytics Algorithms
CLASSIFICATION
• Naïve Bayes
• Logistic Regression (GLM)
• Decision Tree
• Random Forest
• Neural Network
• Support Vector Machine
• Explicit Semantic Analysis
CLUSTERING
• Hierarchical K-Means
• Hierarchical O-Cluster
• Expectation Maximization (EM)
ANOMALY DETECTION
• One-Class SVM
TIME SERIES
• State of the art forecasting using
Exponential Smoothing
• Includes all popular models
e.g. Holt-Winters with trends,
seasons, irregularity, missing data
REGRESSION
• Linear Model
• Generalized Linear Model
• Support Vector Machine (SVM)
• Stepwise Linear regression
• Neural Network
• LASSO *
ATTRIBUTE IMPORTANCE
• Minimum Description Length
• Principal Comp Analysis (PCA)
• Unsupervised Pair-wise KL Div
• CUR decomposition for row & AI
ASSOCIATION RULES
• A priori/ market basket
PREDICTIVE QUERIES
• Predict, cluster, detect, features
SQL ANALYTICS
• SQL Windows, SQL Patterns,
SQL Aggregates
FEATURE EXTRACTION
• Principal Comp Analysis (PCA)
• Non-negative Matrix Factorization
• Singular Value Decomposition (SVD)
• Explicit Semantic Analysis (ESA)
TEXT MINING SUPPORT
• Algorithms support text
• Tokenization and theme extraction
• Explicit Semantic Analysis (ESA) for
document similarity
STATISTICAL FUNCTIONS
• Basic statistics: min, max,
median, stdev, t-test, F-test,
Pearson’s, Chi-Sq, ANOVA, etc.
R PACKAGES
• CRAN R Algorithm Packages
through Embedded R Execution
• Spark MLlib algorithm integration
EXPORTABLE ML MODELS
• REST APIs for deployment
X
1
X
2
A
1
A
2
A
3
A
4
A
5
A
6
A
7
58. ANALYTICAL SQL
• SQL Windows
• SQL Aggregate functions
• LAG/LEAD functions
• SQL for Pattern Matching
• Additional approximate
query
processing: APPROX_COUNT
, APPROX_SUM,
APPROX_RANK
• Regular Expressions
• Linear regression
• ANOVA (Analysis of
variance)
• Test Distribution fit
(e.g. Normal distribution
test, Binomial test, Weibull
test, Uniform
test, Exponential
test, Poisson test, etc.)
• Statistical Aggregates (min,
max, mean, median, stdev,
mode, quantiles, plus x
sigma, minus x sigma, top n
outliers, bottom n outliers)
STATISTICAL FUNCTIONS
• Descriptive statistics
(e.g. median, stdev, mode, sum,
etc.)
• Hypothesis testing
(t-test, F-test, Kolmogorov-
Smirnov test, Mann Whitney
test, Wilcoxon Signed Ranks test
• Correlations analysis
(parametric and nonparametric
e.g.
Pearson’s test for
correlation, Spearman's rho
coefficient, Kendall's tau-b
correlation coefficient)
• Ranking functions
• Cross Tabulations with Chi-square
statistics|
Oracle’s Machine Learning & Adv. Analytics Algorithms
59. Algorithms Operate on Data
ML and AI are just “Algorithms”
Move the Algorithms; Not the Data!;
It Changes Everything!
60. Thank You
Any Questions ?
Sandesh Rao
VP AIOps for the Autonomous Database
@sandeshr
https://www.linkedin.com/in/raosandesh/
https://www.slideshare.net/SandeshRao4