This document discusses data science governance and Kensu's product, Adalog, which aims to address it. It defines data science governance as controlling data activities to meet standards and monitoring production data activity. This involves understanding who does what with which data. Kensu collects metadata on all data tools and processes, connects this information to create a map of all activities, and uses this for impact analysis, dependency analysis, and optimization. Adalog does this to provide accountability and transparency as required by GDPR. It collects data on activities and connects them to automatically generate a process registry and provide transparent reports across the processing chain.
Extended discourse on the importance of data science governance for production ML and how GDPR can become the catalyst but also generate value for organizations!
DataEd Slides: Data Management Best PracticesDATAVERSITY
It is clear that Data Management best practices exist, and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This program describes what must be done at the programmatic level to achieve better data use and a way to implement this as part of your data program. The approach combines DMBoK content and CMMI/DMM processes — permitting organizations the opportunity to benefit from the best of both. It also permits organizations to understand:
• Their current Data Management practices
• Strengths that should be leveraged
• Remediation opportunities
DataEd Slides: Getting Data Quality Right – Success StoriesDATAVERSITY
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is “of sufficient quality.” This program provides a useful framework guiding those approaching Data Quality challenges. Specifically, Data Quality must be approached as an engineering discipline. Data Quality engineering must be approached as a specific ROI-based discipline or it cannot effectively support business strategy. Better understanding of how to “do Data Quality right” allows for speedy identification of business problems, delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Program learning objectives include:
• Vivid demonstrations of how chronic business challenges for organizations are often rooted in broader kinds of Data Quality that suggested treatments can address
• Helping you to understand foundational Data Quality concepts, guiding principles, best practices, and an improved approach to Data Quality at your organization
• The basis of a number of specific case studies illustrating the hallmarks and benefits of Data Quality success
Focus on Your Analysis, Not Your SQL CodeDATAVERSITY
Analysts in the line of business deal with a myriad of time-consuming data preparation and analytic challenges that often require IT or DBA intervention to deliver a requested dataset. Others have taught themselves “enough SQL to be dangerous”, learning the necessary code to extract the data needed to answer their business question. Self-service data analytics empowers these business analysts to take control of the entire analytics process, delivering the necessary results for better business decisions.
Join us to learn how self-service data analytics allows analysts to:
- Utilize a drag-and-drop workflow for data and analytic processes without writing code
- Minimize data movement and ensure data integrity through in-database capabilities
- Easily work across relational and non-relational databases to deliver faster business results
Self-service data analytics delivers a repeatable process that is transparent to not only business analysts, but also SQL coders and decision makers across the organization.
DataEd Slides: Exorcising the Seven Deadly Data SinsDATAVERSITY
The difficulty of implementing a new data strategy often goes under-appreciated, particularly the multi-faceted procedural challenges that need to be met while doing so. Deficiencies in organizational readiness and core competence represent clearly visible problems faced by data managers, but beyond that there are several cultural and structural barriers common to virtually all organizations that must be eliminated in order to facilitate effective management of data. This webinar will discuss these barriers – the titular “Seven Deadly Data Sins” – and in the process will also:
• Elaborate upon the three critical factors that lead to strategy failure
• Demonstrate a two-stage Data Strategy implementation process
• Explore the sources and rationales behind the “Seven Deadly Data Sins,” and recommend solutions
Extended discourse on the importance of data science governance for production ML and how GDPR can become the catalyst but also generate value for organizations!
DataEd Slides: Data Management Best PracticesDATAVERSITY
It is clear that Data Management best practices exist, and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This program describes what must be done at the programmatic level to achieve better data use and a way to implement this as part of your data program. The approach combines DMBoK content and CMMI/DMM processes — permitting organizations the opportunity to benefit from the best of both. It also permits organizations to understand:
• Their current Data Management practices
• Strengths that should be leveraged
• Remediation opportunities
DataEd Slides: Getting Data Quality Right – Success StoriesDATAVERSITY
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is “of sufficient quality.” This program provides a useful framework guiding those approaching Data Quality challenges. Specifically, Data Quality must be approached as an engineering discipline. Data Quality engineering must be approached as a specific ROI-based discipline or it cannot effectively support business strategy. Better understanding of how to “do Data Quality right” allows for speedy identification of business problems, delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Program learning objectives include:
• Vivid demonstrations of how chronic business challenges for organizations are often rooted in broader kinds of Data Quality that suggested treatments can address
• Helping you to understand foundational Data Quality concepts, guiding principles, best practices, and an improved approach to Data Quality at your organization
• The basis of a number of specific case studies illustrating the hallmarks and benefits of Data Quality success
Focus on Your Analysis, Not Your SQL CodeDATAVERSITY
Analysts in the line of business deal with a myriad of time-consuming data preparation and analytic challenges that often require IT or DBA intervention to deliver a requested dataset. Others have taught themselves “enough SQL to be dangerous”, learning the necessary code to extract the data needed to answer their business question. Self-service data analytics empowers these business analysts to take control of the entire analytics process, delivering the necessary results for better business decisions.
Join us to learn how self-service data analytics allows analysts to:
- Utilize a drag-and-drop workflow for data and analytic processes without writing code
- Minimize data movement and ensure data integrity through in-database capabilities
- Easily work across relational and non-relational databases to deliver faster business results
Self-service data analytics delivers a repeatable process that is transparent to not only business analysts, but also SQL coders and decision makers across the organization.
DataEd Slides: Exorcising the Seven Deadly Data SinsDATAVERSITY
The difficulty of implementing a new data strategy often goes under-appreciated, particularly the multi-faceted procedural challenges that need to be met while doing so. Deficiencies in organizational readiness and core competence represent clearly visible problems faced by data managers, but beyond that there are several cultural and structural barriers common to virtually all organizations that must be eliminated in order to facilitate effective management of data. This webinar will discuss these barriers – the titular “Seven Deadly Data Sins” – and in the process will also:
• Elaborate upon the three critical factors that lead to strategy failure
• Demonstrate a two-stage Data Strategy implementation process
• Explore the sources and rationales behind the “Seven Deadly Data Sins,” and recommend solutions
DataEd Slides: Data Strategy – Plans Are Useless but Planning Is InvaluableDATAVERSITY
Too often we hear the question – can you help me with a data strategy? Unfortunately, for most, this is the wrong request because it focuses on its least valuable aspect. The more useful request is – can you help me apply data strategically in support of strategy? Yes, at early maturity phases, the process is more important than the product! Trying to write a good (much less perfect) data strategy on the first attempt is generally not productive – particularly giving the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” By refocusing lesson learning on crawl, walk, run approaches to using data strategically, data is able to keep up with agile, evolving strategies. This approach will contribute more to three primary organizational data goals than other efforts. Learn how improving:
• Your organization’s data
• The way your people use data
• The way your people use data to achieve your organizational strategy
contributes more than predetermined plans. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges pervasively includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs, as organizations identify prioritized areas where better assets, literacy, and support (data strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are covered including:
• A cohesive argument for why Data Strategy is necessary for effective Data Governance
• An overview of prerequisites for effective strategic use of Data Strategy, as well as common pitfalls
• A repeatable process for identifying and removing data constraints
• The importance of balancing business operation and innovation
DataEd Slides: Expressing Data Improvements as Business OutcomesDATAVERSITY
Join us and learn how you can better align your Data Management projects with business objectives to justify funding and gain management approval. Failure to successfully monetize Data Management investments sets up an unfortunate loop of fixing symptoms without addressing the underlying problems. As organizations begin to understand that data practices are the root causes of many business problems, they become more willing to make the required investments. However, we need to also approach them. The No. 1 reason that data programs fail to deliver is that they do not set or measure specific objectives that are meaningful to management. While there are opportunities to assist at the project level, data improvements are better able to be leveraged at the organization level. An improvable, dedicated data program can only be achieved by repeated application of data practices in service of specific business objectives. Data improvements typically do not maintain an ROI calculation. ROIs expressed in terms that board/executive management cares about deeply ensure data program viability. Improving organizational execution of specific data practice improvements must lead directly to specific improvements in organizational KPIs. While organizations may not be currently practiced in this ability, it is quite easy to learn. This presentation uses a number of specific examples calculating the business impact of data improvements. Program learning objectives include:
• Coming to grips with the state of practice
• Understanding the need for a comparable baseline measure
• Seeing application in a number of contexts
Key Elements for a Successful Service Analytics ProgramData Con LA
Data Con LA 2020
DescriptionThis talk will focus on providing the key elements that enable the successful roll out of a self service analytics program at any organization. I'll discuss my tenure at Qualcomm where I led a self service program there for 10 years and grew it to 500 developers, 3000 applications and 15000 end users. I'll also go over other client case studies like the California Department of Public Health and Illumina where we are developing similar self service programs and go over what works and what does not work.
Speaker
Steve Rimar, Analytica Consulting, LLC, CEO & Founder
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality management effectively in support of business strategy. This, in turn, allows for speedy identification of business problems, the delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This, in turn, allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Predictive Analytics - How to get stuff out of your Crystal BallDATAVERSITY
Everyone wants to leverage data. The optimal implementation of analytics is an organization-wide set of capabilities. These are called advantageous organizational analytic capabilities in that a clear ROI is demonstrable from these efforts. Turns out that there are a number of prerequisites to advantageous organizational analytics. These include:
Adopting a crawl, walk, run strategy
Understanding current and potential organizational maturity and corresponding capabilities
Achieving an appropriate technology/human capability balance
Implementing useful IT systems development practices
Installing necessary non-IT leadership
This webinar will explore these and other topics using examples drawn from DOD, healthcare researchers, and donation center operations.
DAS Slides: Building a Data Strategy — Practical Steps for Aligning with Busi...DATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task. The opportunity in getting it right can be significant, however, as data drives many of the key initiatives in today’s marketplace from digital transformation, to marketing, to customer centricity, population health, and more. This webinar will help de-mystify data strategy and data architecture and will provide concrete, practical ways to get started.
RWDG Slides: Metadata Governance for Catalogs, Glossaries, Dictionaries, and ...DATAVERSITY
Metadata Governance is the execution and enforcement of authority over the management of Metadata and other data documentation. Organizations that govern their data documentation find it easier to govern their data as a result. There is direct correlation between the use of Data Catalogs, Business Glossaries and Data Dictionaries and successful governance of data and Metadata.
This month’s RWDG webinar with Bob Seiner will focus on governing the use of the mentioned tools and the Metadata that can be managed inside each one. Bob will talk about governing Metadata in existing Metadata resources versus using new tools to handle this function.
In this webinar, Bob will discuss:
- The relationship between Data Governance and Metadata Governance
- Metadata collected in Data Catalogs, Business Glossaries, and Data Dictionaries
- How to maximize use the data documentation in each resource
- Governing data documentation in Catalogs, Glossaries, and Dictionaries
- Measuring the effectiveness of governed Metadata
DI&A Slides: Data Lake vs. Data WarehouseDATAVERSITY
Modern data analysis is moving beyond the Data Warehouse to the Data Lake where analysts are able to take advantage of emerging technologies to manage complex analytics on large data volumes and diverse data types. Yet, for some business problems, a Data Warehouse may still be the right solution.
If you’re on the fence, join this webinar as we compare and contrast Data Lakes and Data Warehouses, identifying situations where one approach may be better than the other and highlighting how the two can work together.
Get tips, takeaways and best practices about:
- The benefits and problems of a Data Warehouse
- How a Data Lake can solve the problems of a Data Warehouse
- Data Lake Architecture
- How Data Warehouses and Data Lakes can work together
Are you your company’s chief data officer? Given the scarcity of the official role, it’s likely that you’re not — at least in title. But that doesn't mean that you shouldn't operate like one. Do you approach data leadership as a C-level executive or a senior data head? Is your team’s output strategic or just operational? In this interactive keynote, one of the Windy City’s foremost data leaders will lead an interactive discussion on what it takes to lead like a chief, what it looks like, and how to get there and get it done.
Data Governance Best Practices and Lessons LearnedDATAVERSITY
Best practices and lessons learned are powerful tools used to assess an organization’s readiness and initial activities associated with delivering a Data Governance program. There are two criteria to determine if something is best practice for your organization. And the definition of data governance best practice is best way to learn from others and begin with the end in mind.
Bob Seiner will share industry data governance best practices in this month’s installment of the RWDG webinar series. Learn how to use the best practices defined in this webinar to address opportunities to improve your organization’s data governance implementation. Attend this webinar and learn that assessing your organization may not be as difficult as you think.
During this webinar Bob will discuss:
How to define data governance best practices for your organization
Criteria used to determine if a practice is best practice
How to assess your organization against industry best practice
Assessing risks associated with best practice gaps
Addressing opportunities to improve gaps uncovered in the assessment
ADV Slides: What Happened of Note in 1H 2020 in Enterprise Advanced AnalyticsDATAVERSITY
Reassessing the information management marketplace for your enterprise direction on an annual basis is too infrequent. The technology is changing too fast. Data and analytic maturity levels rapidly evolve. What is advanced today may be entry-level in two years. Let’s look at the high points for 1H 2020 in information management developments and how that may change what you are doing now. This can also be a strong data point for preparing 2021 budgets.
In a world of big data, many organizations are struggling to understand how they can exploit this asset. In this presentation we share our framework for creating a Data Fluent organization. Building data fluency requires both individual skills in understanding and communicating data as well as a culture, processes, and tools for using your data.
Data Literacy and Data Virtualization: A Step-by-step Guide to Bolstering You...Denodo
Watch full webinar here: https://bit.ly/2KLc1dE
An organization’s effectiveness can only be as good as the understanding of their data. Hence it is important for both the frontline workers as well as the managers to be data literate, so that they can they understand how the business is functioning, decide if any changes need to be made, and quickly make decisions to realize better outcomes. However, successful data literacy requires stringent processes and an effective tool to operationalize them.
Listen to the our replay on the 10-steps to building a data-literate organization, and how data virtualization can help implement the underpinning processes.
Sense Corp and Denodo have partnered to combine state-of-the art professional services with the industry’s most advanced data virtualization platform to streamline data access in support of the most critical business needs.
Watch the replay to learn:
- The 10-steps to data literacy; what you can do to become a high performer.
- How to use data virtualization as the foundation to implementing data literacy processes.
- Examples of companies that have achieved high levels of data literacy.
Download the Sense Corp 10 Steps to Data Literacy eBook to learn more.
Data-Ed Webinar: Data Quality Success StoriesDATAVERSITY
Organizations must realize what it means to utilize data quality management in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Takeaways:
•Understanding foundational data quality concepts based on the DAMA DMBOK
•Utilizing data quality engineering in support of business strategy
•Case Studies illustrating data quality success
•Data Quality guiding principles & best practices
•Steps for improving data quality at your organization
Data-Ed Online Webinar: Data Governance StrategiesDATAVERSITY
The data governance function exercises authority and control over the management of your mission critical assets and guides how all other data management functions are performed. When selling data governance to organizational management, it is useful to concentrate on the specifics that motivate the initiative. This means developing a specific vocabulary and set of narratives to facilitate understanding of your organizational business concepts. This webinar provides you with an understanding of what data governance functions are required and how they fit with other data management disciplines. Understanding these aspects is a necessary pre-requisite to eliminate the ambiguity that often surrounds initial discussions and implement effective data governance and stewardship programs that manage data in support of organizational strategy.
Takeaways:
Understanding why data governance can be tricky for most organizations
Steps for improving data governance within your organization
Guiding principles & lessons learned
Understanding foundational data governance concepts based on the DAMA DMBOK
Smarter businesses apply AI to learn and continuously evolve the way they work. To extract full value from AI, companies need data strategy that gives them access to all their data – no matter where it lives – in an environment that easily scales and applies the latest discovery technology including advanced analytics, visualization and AI. Learn how IBM Watson and Data provides all the tools companies need to embed AI, machine learning and deep learning in their business, while enabling professionals to gain the most from their data to drive smarter business and lead industry-changing transformations.
DataEd Slides: Getting (Re)Started with Data StewardshipDATAVERSITY
In order to find value in your organization’s data assets, heroic data stewards are tasked with saving the day – every single day! Adhering to the organizational Data Governance (DG) framework, they work to ensure that data is captured right the first time, validated through appropriately automated means, and integrated into business processing. Whether it’s data profiling or in-depth root cause analysis, data stewards ensure the organization’s mission-critical data is reliably coordinated. This program will approach this framework and punctuate important facets of a data steward’s role. Learning objectives include:
• Understanding the motivation for full-time data stewards in your organization
• Comprehending how stewards need to be multifunctional and dexterous, especially at first
• Exploring how stewards successfully target SDLC from cadence, approach, simplicity requirements, foundational prerequisite, and perspectives
DataEd Slides: Data Management + Data Strategy = InteroperabilityDATAVERSITY
Few organizations operate without having to exchange data. (Many do it professionally and well!) The larger the data exchange burden (DEB), the greater the organizational overhead incurred. This death by 1,000 cuts must be factored into each organization’s calculations. Unfortunately, most organizations do not know if their organization’s DEB is great or small. A somewhat greater number of organizations have organized Data Management practices. Focusing Data Management efforts on increasing interoperability by decreasing the DEB friction is a good area to “practice.”
Learning Objectives:
• Gaining a good understanding of both important topics
• Understanding that data only operates at a very intricate, specifically dependent intent and what this means
• Understand state-of-the-practice
• Coordination is key, requiring necessary but insufficient interdependencies and sequencing
• Practice makes perfect
Innovation med big data – chr. hansens erfaringerMicrosoft
Mange steder er Big Data stadig det nye og ukendte, der ikke har topprioritet hos IT, da ”vi ikke har store datamængder”. Men Big Data er meget mere end store datamængder. I Chr. Hansen A/S har Forskning og Udvikling (Innovation) afdelingen arbejdet med værdien af data og som resultat etableret et tværfagligt BioInformatik-program på Big Data teknologier fra Microsoft.
DataEd Slides: Data Strategy – Plans Are Useless but Planning Is InvaluableDATAVERSITY
Too often we hear the question – can you help me with a data strategy? Unfortunately, for most, this is the wrong request because it focuses on its least valuable aspect. The more useful request is – can you help me apply data strategically in support of strategy? Yes, at early maturity phases, the process is more important than the product! Trying to write a good (much less perfect) data strategy on the first attempt is generally not productive – particularly giving the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” By refocusing lesson learning on crawl, walk, run approaches to using data strategically, data is able to keep up with agile, evolving strategies. This approach will contribute more to three primary organizational data goals than other efforts. Learn how improving:
• Your organization’s data
• The way your people use data
• The way your people use data to achieve your organizational strategy
contributes more than predetermined plans. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges pervasively includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs, as organizations identify prioritized areas where better assets, literacy, and support (data strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are covered including:
• A cohesive argument for why Data Strategy is necessary for effective Data Governance
• An overview of prerequisites for effective strategic use of Data Strategy, as well as common pitfalls
• A repeatable process for identifying and removing data constraints
• The importance of balancing business operation and innovation
DataEd Slides: Expressing Data Improvements as Business OutcomesDATAVERSITY
Join us and learn how you can better align your Data Management projects with business objectives to justify funding and gain management approval. Failure to successfully monetize Data Management investments sets up an unfortunate loop of fixing symptoms without addressing the underlying problems. As organizations begin to understand that data practices are the root causes of many business problems, they become more willing to make the required investments. However, we need to also approach them. The No. 1 reason that data programs fail to deliver is that they do not set or measure specific objectives that are meaningful to management. While there are opportunities to assist at the project level, data improvements are better able to be leveraged at the organization level. An improvable, dedicated data program can only be achieved by repeated application of data practices in service of specific business objectives. Data improvements typically do not maintain an ROI calculation. ROIs expressed in terms that board/executive management cares about deeply ensure data program viability. Improving organizational execution of specific data practice improvements must lead directly to specific improvements in organizational KPIs. While organizations may not be currently practiced in this ability, it is quite easy to learn. This presentation uses a number of specific examples calculating the business impact of data improvements. Program learning objectives include:
• Coming to grips with the state of practice
• Understanding the need for a comparable baseline measure
• Seeing application in a number of contexts
Key Elements for a Successful Service Analytics ProgramData Con LA
Data Con LA 2020
DescriptionThis talk will focus on providing the key elements that enable the successful roll out of a self service analytics program at any organization. I'll discuss my tenure at Qualcomm where I led a self service program there for 10 years and grew it to 500 developers, 3000 applications and 15000 end users. I'll also go over other client case studies like the California Department of Public Health and Illumina where we are developing similar self service programs and go over what works and what does not work.
Speaker
Steve Rimar, Analytica Consulting, LLC, CEO & Founder
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality management effectively in support of business strategy. This, in turn, allows for speedy identification of business problems, the delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This, in turn, allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Predictive Analytics - How to get stuff out of your Crystal BallDATAVERSITY
Everyone wants to leverage data. The optimal implementation of analytics is an organization-wide set of capabilities. These are called advantageous organizational analytic capabilities in that a clear ROI is demonstrable from these efforts. Turns out that there are a number of prerequisites to advantageous organizational analytics. These include:
Adopting a crawl, walk, run strategy
Understanding current and potential organizational maturity and corresponding capabilities
Achieving an appropriate technology/human capability balance
Implementing useful IT systems development practices
Installing necessary non-IT leadership
This webinar will explore these and other topics using examples drawn from DOD, healthcare researchers, and donation center operations.
DAS Slides: Building a Data Strategy — Practical Steps for Aligning with Busi...DATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task. The opportunity in getting it right can be significant, however, as data drives many of the key initiatives in today’s marketplace from digital transformation, to marketing, to customer centricity, population health, and more. This webinar will help de-mystify data strategy and data architecture and will provide concrete, practical ways to get started.
RWDG Slides: Metadata Governance for Catalogs, Glossaries, Dictionaries, and ...DATAVERSITY
Metadata Governance is the execution and enforcement of authority over the management of Metadata and other data documentation. Organizations that govern their data documentation find it easier to govern their data as a result. There is direct correlation between the use of Data Catalogs, Business Glossaries and Data Dictionaries and successful governance of data and Metadata.
This month’s RWDG webinar with Bob Seiner will focus on governing the use of the mentioned tools and the Metadata that can be managed inside each one. Bob will talk about governing Metadata in existing Metadata resources versus using new tools to handle this function.
In this webinar, Bob will discuss:
- The relationship between Data Governance and Metadata Governance
- Metadata collected in Data Catalogs, Business Glossaries, and Data Dictionaries
- How to maximize use the data documentation in each resource
- Governing data documentation in Catalogs, Glossaries, and Dictionaries
- Measuring the effectiveness of governed Metadata
DI&A Slides: Data Lake vs. Data WarehouseDATAVERSITY
Modern data analysis is moving beyond the Data Warehouse to the Data Lake where analysts are able to take advantage of emerging technologies to manage complex analytics on large data volumes and diverse data types. Yet, for some business problems, a Data Warehouse may still be the right solution.
If you’re on the fence, join this webinar as we compare and contrast Data Lakes and Data Warehouses, identifying situations where one approach may be better than the other and highlighting how the two can work together.
Get tips, takeaways and best practices about:
- The benefits and problems of a Data Warehouse
- How a Data Lake can solve the problems of a Data Warehouse
- Data Lake Architecture
- How Data Warehouses and Data Lakes can work together
Are you your company’s chief data officer? Given the scarcity of the official role, it’s likely that you’re not — at least in title. But that doesn't mean that you shouldn't operate like one. Do you approach data leadership as a C-level executive or a senior data head? Is your team’s output strategic or just operational? In this interactive keynote, one of the Windy City’s foremost data leaders will lead an interactive discussion on what it takes to lead like a chief, what it looks like, and how to get there and get it done.
Data Governance Best Practices and Lessons LearnedDATAVERSITY
Best practices and lessons learned are powerful tools used to assess an organization’s readiness and initial activities associated with delivering a Data Governance program. There are two criteria to determine if something is best practice for your organization. And the definition of data governance best practice is best way to learn from others and begin with the end in mind.
Bob Seiner will share industry data governance best practices in this month’s installment of the RWDG webinar series. Learn how to use the best practices defined in this webinar to address opportunities to improve your organization’s data governance implementation. Attend this webinar and learn that assessing your organization may not be as difficult as you think.
During this webinar Bob will discuss:
How to define data governance best practices for your organization
Criteria used to determine if a practice is best practice
How to assess your organization against industry best practice
Assessing risks associated with best practice gaps
Addressing opportunities to improve gaps uncovered in the assessment
ADV Slides: What Happened of Note in 1H 2020 in Enterprise Advanced AnalyticsDATAVERSITY
Reassessing the information management marketplace for your enterprise direction on an annual basis is too infrequent. The technology is changing too fast. Data and analytic maturity levels rapidly evolve. What is advanced today may be entry-level in two years. Let’s look at the high points for 1H 2020 in information management developments and how that may change what you are doing now. This can also be a strong data point for preparing 2021 budgets.
In a world of big data, many organizations are struggling to understand how they can exploit this asset. In this presentation we share our framework for creating a Data Fluent organization. Building data fluency requires both individual skills in understanding and communicating data as well as a culture, processes, and tools for using your data.
Data Literacy and Data Virtualization: A Step-by-step Guide to Bolstering You...Denodo
Watch full webinar here: https://bit.ly/2KLc1dE
An organization’s effectiveness can only be as good as the understanding of their data. Hence it is important for both the frontline workers as well as the managers to be data literate, so that they can they understand how the business is functioning, decide if any changes need to be made, and quickly make decisions to realize better outcomes. However, successful data literacy requires stringent processes and an effective tool to operationalize them.
Listen to the our replay on the 10-steps to building a data-literate organization, and how data virtualization can help implement the underpinning processes.
Sense Corp and Denodo have partnered to combine state-of-the art professional services with the industry’s most advanced data virtualization platform to streamline data access in support of the most critical business needs.
Watch the replay to learn:
- The 10-steps to data literacy; what you can do to become a high performer.
- How to use data virtualization as the foundation to implementing data literacy processes.
- Examples of companies that have achieved high levels of data literacy.
Download the Sense Corp 10 Steps to Data Literacy eBook to learn more.
Data-Ed Webinar: Data Quality Success StoriesDATAVERSITY
Organizations must realize what it means to utilize data quality management in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Takeaways:
•Understanding foundational data quality concepts based on the DAMA DMBOK
•Utilizing data quality engineering in support of business strategy
•Case Studies illustrating data quality success
•Data Quality guiding principles & best practices
•Steps for improving data quality at your organization
Data-Ed Online Webinar: Data Governance StrategiesDATAVERSITY
The data governance function exercises authority and control over the management of your mission critical assets and guides how all other data management functions are performed. When selling data governance to organizational management, it is useful to concentrate on the specifics that motivate the initiative. This means developing a specific vocabulary and set of narratives to facilitate understanding of your organizational business concepts. This webinar provides you with an understanding of what data governance functions are required and how they fit with other data management disciplines. Understanding these aspects is a necessary pre-requisite to eliminate the ambiguity that often surrounds initial discussions and implement effective data governance and stewardship programs that manage data in support of organizational strategy.
Takeaways:
Understanding why data governance can be tricky for most organizations
Steps for improving data governance within your organization
Guiding principles & lessons learned
Understanding foundational data governance concepts based on the DAMA DMBOK
Smarter businesses apply AI to learn and continuously evolve the way they work. To extract full value from AI, companies need data strategy that gives them access to all their data – no matter where it lives – in an environment that easily scales and applies the latest discovery technology including advanced analytics, visualization and AI. Learn how IBM Watson and Data provides all the tools companies need to embed AI, machine learning and deep learning in their business, while enabling professionals to gain the most from their data to drive smarter business and lead industry-changing transformations.
DataEd Slides: Getting (Re)Started with Data StewardshipDATAVERSITY
In order to find value in your organization’s data assets, heroic data stewards are tasked with saving the day – every single day! Adhering to the organizational Data Governance (DG) framework, they work to ensure that data is captured right the first time, validated through appropriately automated means, and integrated into business processing. Whether it’s data profiling or in-depth root cause analysis, data stewards ensure the organization’s mission-critical data is reliably coordinated. This program will approach this framework and punctuate important facets of a data steward’s role. Learning objectives include:
• Understanding the motivation for full-time data stewards in your organization
• Comprehending how stewards need to be multifunctional and dexterous, especially at first
• Exploring how stewards successfully target SDLC from cadence, approach, simplicity requirements, foundational prerequisite, and perspectives
DataEd Slides: Data Management + Data Strategy = InteroperabilityDATAVERSITY
Few organizations operate without having to exchange data. (Many do it professionally and well!) The larger the data exchange burden (DEB), the greater the organizational overhead incurred. This death by 1,000 cuts must be factored into each organization’s calculations. Unfortunately, most organizations do not know if their organization’s DEB is great or small. A somewhat greater number of organizations have organized Data Management practices. Focusing Data Management efforts on increasing interoperability by decreasing the DEB friction is a good area to “practice.”
Learning Objectives:
• Gaining a good understanding of both important topics
• Understanding that data only operates at a very intricate, specifically dependent intent and what this means
• Understand state-of-the-practice
• Coordination is key, requiring necessary but insufficient interdependencies and sequencing
• Practice makes perfect
Innovation med big data – chr. hansens erfaringerMicrosoft
Mange steder er Big Data stadig det nye og ukendte, der ikke har topprioritet hos IT, da ”vi ikke har store datamængder”. Men Big Data er meget mere end store datamængder. I Chr. Hansen A/S har Forskning og Udvikling (Innovation) afdelingen arbejdet med værdien af data og som resultat etableret et tværfagligt BioInformatik-program på Big Data teknologier fra Microsoft.
Facilitating Collaborative Life Science Research in Commercial & Enterprise E...Chris Dagdigian
This is a talk I put together for a http://www.neren.org/ seminar called "Bridging the Gap: Research Facilitation". Tried to give a biotech/pharma view for a mostly academic audience.
The pioneers in the big data space have battle scars and have learnt many of the lessons in this report the hard way. But if you are a general manger & just embarking on the big data journey, you should now have what they call the 'second mover advantage’. My hope is that this report helps you better leverage your second mover advantage. The goal here is to shed some light on the people & process issues in building a central big data analytics function
Confirming PagesLess managing. More teaching. Greater AlleneMcclendon878
Confirming Pages
Less managing. More teaching. Greater learning.
INSTRUCTORS GET:
• Interactive Applications – book-specific interactive
assignments that require students to APPLY what
they’ve learned.
• Simple assignment management, allowing you to
spend more time teaching.
• Auto-graded assignments, quizzes, and tests.
• Detailed Visual Reporting where student and
section results can be viewed and analyzed.
• Sophisticated online testing capability.
• A filtering and reporting function
that allows you to easily assign and
report on materials that are correlated
to accreditation standards, learning
outcomes, and Bloom’s taxonomy.
• An easy-to-use lecture capture tool.
Would you like your students to show up for class more prepared? (Let’s face it, class
is much more fun if everyone is engaged and prepared…)
Want ready-made application-level interactive assignments, student progress
reporting, and auto-assignment grading? (Less time grading means more time teaching…)
Want an instant view of student or class performance relative to learning
objectives? (No more wondering if students understand…)
Need to collect data and generate reports required for administration or
accreditation? (Say goodbye to manually tracking student learning outcomes…)
Want to record and post your lectures for students to view online?
INSTRUCTORS...
With McGraw-Hill's Connect® MIS,
haa7685X_fm_i-xxxv.indd ihaa7685X_fm_i-xxxv.indd i 12/20/11 9:29 PM12/20/11 9:29 PM
Confirming Pages
Want an online, searchable version of your textbook?
Wish you could reference your textbook online while you’re doing
your assignments?
Want to get more value from your textbook purchase?
Think learning MIS should be a bit more interesting?
Connect® Plus MIS eBook
If you choose to use Connect™ Plus MIS, you have an affordable and
searchable online version of your book integrated with your other
online tools.
Connect® Plus MIS eBook offers features like:
• Topic search
• Direct links from assignments
• Adjustable text size
• Jump to page number
• Print by section
Check out the STUDENT RESOURCES
section under the Connect® Library tab.
Here you’ll find a wealth of resources designed to help you
achieve your goals in the course. You’ll find things like quizzes,
PowerPoints, and Internet activities to help you study.
Every student has different needs, so explore the STUDENT
RESOURCES to find the materials best suited to you.
haa7685X_fm_i-xxxv.indd iihaa7685X_fm_i-xxxv.indd ii 12/20/11 9:29 PM12/20/11 9:29 PM
Confirming Pages
Management Information Systems
FOR THE INFORMATION AGE
NINTH EDITION
Stephen Haag
DANIELS COLLEGE OF BUSINESS
UNIVERSITY OF DENVER
Maeve Cummings
KELCE COLLEGE OF BUSINESS
PITTSBURG STATE UNIVERSITY
haa7685X_fm_i-xxxv.indd iiihaa7685X_fm_i-xxxv.indd iii 12/26/11 5:37 PM12/26/11 5:37 PM
Confirming Pages
MANAGEMENT INFORMATION SYSTEMS FOR THE INF ...
Data Science - An emerging Stream of Science with its Spreading Reach & ImpactDr. Sunil Kr. Pandey
This is my presentation on the Topic "Data Science - An emerging Stream of Science with its Spreading Reach & Impact". I have compiled and collected different statistics and data from different sources. This may be useful for students and those who might be interested in this field of Study.
Businesses make critical decisions using key data assets, but stakeholders often find it difficult to navigate the complex data landscape to ensure they have the right data and understand it correctly. Companies are dealing with a number of different technologies, multiple data formats, and high data volumes, along with the requirements for data security and governance.
TechWise with Eric Kavanagh, Dr. Robin Bloor and Dr. Kirk Borne
Live Webcast on July 23, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=59d50a520542ee7ed00a0c38e8319b54
Analytical applications are everywhere these days, and for good reason. Organizations large and small are using analytics to better understand any aspect of their business: customers, processes, behaviors, even competitors. There are several critical success factors for using analytics effectively: 1) know which kind of apps make sense for your company; 2) figure out which data sets you can use, both internal and external; 3) determine optimal roles and responsibilities for your team; 4) identify where you need help, either by hiring new employees or using consultants 5) manage your program effectively over time.
Register for this episode of TechWise to learn from two of the most experienced analysts in the business: Dr. Robin Bloor, Chief Analyst of The Bloor Group, and Dr. Kirk Borne, Data Scientist, George Mason University. Each will provide their perspective on how companies can address each of the key success factors in building, refining and using analytics to improve their business. There will then be an extensive Q&A session in which attendees can ask detailed questions of our experts and get answers in real time. Registrants will also receive a consolidated deck of slides, not just from the main presenters, but also from a variety of software vendors who provide targeted solutions.
Visit InsideAnlaysis.com for more information.
Story of Bigdata and its Applications in Financial Institutionsijtsrd
The importance of BigData is indeed nothing new, but being able to manage data efficiently is just now becoming more attainable. Although data management has evolved considerably since the 1800's, advancements made in recent years that have made the process even more efficient. Technique of Data mining, is much used in the banking industry, which helps banks compete in the market and provide the right product to the right customer. While collecting and combining different sources of data into a single significant volumetric Golden Source of TRUTH can be achieved by applying the right combination of tools. In this paper Author introduced BIGDATA technologies in brief along with its applications. Phani Bhooshan | Dr. C. Umashankar "Story of Bigdata and its Applications in Financial Institutions" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-6 , October 2019, URL: https://www.ijtsrd.com/papers/ijtsrd29145.pdf Paper URL: https://www.ijtsrd.com/computer-science/database/29145/story-of-bigdata-and-its-applications-in-financial-institutions/phani-bhooshan
Agile Data Science is a lean methodology that is adopted from Agile Software Development. At the core it centers around people, interactions, and building minimally viable products to ship fast and often to solicit customer feedback. In this presentation, I describe how this work was done in the past with examples. Get started today with our help by visiting http://www.alpinenow.com
Similar to Data science governance : what and how (20)
Non-technical talk for managers and Data Protection Officers about how the reasons behind the automation of creating a global data mapping for GDPR (at least), the challenges and possible methodologies using a new concept of Process Mining based on Data Activities
Scala: the unpredicted lingua franca for data scienceAndy Petrella
Talk given at Strata London with Dean Wampler (Lightbend) about Scala as the future of Data Science. First part is an approach of how scala became important, the remaining part of the talk is in notebooks using the Spark Notebook (http://spark-notebook.io/).
The notebooks are available on GitHub: https://github.com/data-fellas/scala-for-data-science.
Agile data science: Distributed, Interactive, Integrated, Semantic, Micro Ser...Andy Petrella
Distributed Data Science…
* A genomics use case
* Spark Notebook
* Interactive Distributed Data Science
Distributed Data Science… Pipeline
* Pipeline: productizing Data Science
* Demo of Distributed Pipeline (ADAM, Akka, Cassandra, Parquet, Spark)
* Why Micro Services?
* Painful points:
* Data science is Discontiguous
* Context Lost in Translation
* Solution: Data Fellas’ Agile Data Science Toolkit
What is a distributed data science pipeline. how with apache spark and friends.Andy Petrella
What was a data product before the world changed and got so complex.
Why distributed computing/data science is the solution.
What problems does that add?
How to solve most of them using the right technologies like spark notebook, spark, scala, mesos and so on in a accompanied framework
Towards a rebirth of data science (by Data Fellas)Andy Petrella
Nowadays, Data Science is buzzing all over the place.
But what is a, so-called, Data Scientist?
Some will argue that a Data Scientist is a person able to report and present insights in a data set. Others will say that a Data Scientist can handle a high throughput of values and expose them in services. Yet another definition includes the capacity to create meaningful visualizations on the data.
However, we enter an age where velocity is a key. Not only the velocity of your data is high, but the time to market is shortened. Hence, the time separating the moment you receive a set of data and the time you’ll be able to deliver added value is crucial.
In this talk, we’ll review the legacy Data Science methodologies, what it meant in terms of delivered work and results.
Afterwards, we’ll slightly move towards different concepts, techniques and tools that Data Scientists will have to learn and appropriate in order to accomplish their tasks in the age of Big Data.
The dissertation is closed by exposing the Data Fellas view on a solution to the challenges, specially thanks to the Spark Notebook and the Shar3 product we develop.
Distributed machine learning 101 using apache spark from a browser devoxx.b...Andy Petrella
A 3 hours session introducing the concept of Machine Learning and Distributed Computing.
It includes many examples running in notebooks of experience run on data exploring models like LM, RF, K-Means, Deep Learning.
Spark Summit Europe: Share and analyse genomic data at scaleAndy Petrella
Share and analyse genomic data
at scale with Spark, Adam, Tachyon & the Spark Notebook
Sharp intro to Genomics data
What are the Challenges
Distributed Machine Learning to the rescue
Projects: Distributed teams
Research: Long process
Towards Maximum Share for efficiency
Leveraging mesos as the ultimate distributed data science platformAndy Petrella
Keynote at the first @MesosCon #Europe on what was Data Science, what are the new challenge and needs and how we target them in Data Fellas with the Spark Notebook and Shar3
Data Enthusiasts London: Scalable and Interoperable data services. Applied to...Andy Petrella
Data science requires so many skills, people and time before the results can be accessed. Moreover, these results cannot be static anymore. And finally, the Big Data comes to the plate and the whole tool chain needs to change.
In this talk Data Fellas introduces Shar3, a tool kit aiming to bridged the gaps to build a interactive distributed data processing pipeline, or loop!
Then the talk covers genomics nowadays problems including data types, processing, discovery by introducing the GA4GH initiative and its implementation using Shar3.
Spark meetup london share and analyse genomic data at scale with spark, adam...Andy Petrella
Genomics and Health data is nowadays one of the hot topics requiring lots of computations and specially machine learning. This helps science with a very relevant societal impact to get even better outcome. That is why Apache Spark and its ADAM library is a must have.
This talk will be twofold.
First, we'll show how Apache Spark, MLlib and ADAM can be plugged all together to extract information from even huge and wide genomics dataset. Everything will be packed into examples from the Spark Notebook, showing how bio-scientists can work interactively with such a system.
Second, we'll explain how these methodologies and even the datasets themselves can be shared at very large scale between remote entities like hospitals or laboratories using micro services leveraging Apache Spark, ADAM, Play Framework 2, Avro and Tachyon.
Distributed machine learning 101 using apache spark from the browserAndy Petrella
Talk given by Xavier Tordoir and myself at Scala Days Amsterdam 2015.
Contains intro to ML, focusing on what is it and models selection via the Bias Variation constraint.
Then switches a gear to show how genomics can be learned using LDA, KMeans and Random Forest.
Finishes with some insight on what we'll change in the future regarding machine learning and modeling.
In this talk, I fly over the different concepts and advantages of Open Source, Open Data, Crowd Sourcing and Coworking in the context of Startups.
Yet, I put the focus on Data science related entrepreneurship, the domain I live in.
BioBankCloud: Machine Learning on Genomics + GA4GH @ Med at ScaleAndy Petrella
A talk given at the BioBankCloud conference in Feb 2015 about distributed computing in the contexts of genomics and health.
In this one, we exposed what results we obtained exploring the 1000genomes data using ADAM, followed by an introduction to our scalable GA4GH server implementation built using ADAM, Apache Spark and Play Framework 2.
What is Distributed Computing, Why we use Apache SparkAndy Petrella
In this talk we introduce the notion of distributed computing then we tackle the Spark advantages.
The Spark core content is very tiny because the whole explanation has been done live using a Spark Notebook (https://github.com/andypetrella/spark-notebook/blob/geek/conf/notebooks/Geek.snb).
This talk has been given together by @xtordoir and myself at the University of Liège, Belgium.
Lightning fast genomics with Spark, Adam and ScalaAndy Petrella
We are at a time where biotech allow us to get personal genomes for $1000. Tremendous progress since the 70s in DNA sequencing have been done, e.g. more samples in an experiment, more genomic coverages at higher speeds. Genomic analysis standards that have been developed over the years weren't designed with scalability and adaptability in mind. In this talk, we’ll present a game changing technology in this area, ADAM, initiated by the AMPLab at Berkeley. ADAM is framework based on Apache Spark and the Parquet storage. We’ll see how it can speed up a sequence reconstruction to a factor 150.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
2. www.kensu.io 2
- CEO & Founder -
Mathematics & Computer Science MsC.
Creator of Spark Notebook
- CSO & Founder -
Physics PhD.
Genomics & Quantitative Finance
XAVIER TORDOIRANDY PETRELLA
KENSU & ME
Started in 2015 as Data Fellas, focus on Data Science consulting
Team of 10 engineers and scientists
Shift toward Product Company in 2016, renamed to Kensu,
Focus on Data Science Governance
Accelerated by Alchemist Accelerator in San Francisco and The Faktory in Belgium
3. www.kensu.io
TOPICS
1. Some thoughts on “Data Science”
2. Data Science Governance: What
3. Data Science Governance: How
4. GDPR: Accountability principle and transparency
3
5. www.kensu.io
MACHINE LEARNING
Pioneers in 1950s
AI Winter in 1970s due pessimism
Resurgence in 1980s
Machine Learning (and related) is used since the 1990s (esp. SVM and RNN)
Deep learning see widespread commercial use in 2000s
Machine learning receives great publicity (read: buzz) in 2010s
5ref: https://en.wikipedia.org/wiki/Timeline_of_machine_learning
6. www.kensu.io
DATA SCIENCE: +ENGINEERING
Claim: “Data Scientist” coined by DJ Patil in 2008.
Pretty much where Machine Learning was part of Softwares
In a way, when we added “engineering” to the mix
Also, engineering is even more prominent with Big Data Distributed
Computing
6
7. www.kensu.io
DATA SCIENCE: +EXPERIMENTATION
So much data available
So many tools, libraries, frameworks, …
So many things we can try
We have distributed computing now, right? => Let’s try everything
Discover new insights (and potentially new businesses)
7
8. www.kensu.io
DATA SCIENCE: RECAP
Maths: stats, machine learning and so on
Engineering: ETL, Databases, Computing framework, Softwares, Platforms, …
Creativity: “From business intelligence To intelligent business”- Michael Fergusson
Data Science is an umbrella on top of all activities on data
8
10. www.kensu.io
DATA PIPELINE
Data pipeline is connecting activities on data, potentially involving
several technologies.
A pipeline is generally thought as an End-to-End processing line to solve
one problem.
But, part of pipelines are reused to save computation, storage, time, …
Thus interdependency between pipeline segments grows with initiatives
10
11. www.kensu.io
GOAL: TAKE DECISION
Data Pipelines, connected together, aren’t created for the beauty of it.
The ultimate goal is always to take decisions.
Decisions are generally taken or linked to humans with responsibilities.
(even for self driving cars, in case of problem)
Given that pipelines are cut-and-wired, interleaved, …
How not to be anxious at deploying the last piece used by the decision maker
11
12. www.kensu.io
SOURCES OF ANXIETY
What if:
• one of the data used in the process has different patterns suddenly?
• one of the tools, projects or similar is modified upstream?
• the insights are deviating from the reality?
• …
12
13. www.kensu.io
DEBUGGING?
To reduce the anxiety or, actually, reducing the risks, we need ways to debug.
In pure engineering, we have unit, function, integrations tests,… but
How do we do when the problems come from the data themselves?
We can’t generate all cases of data variations, right?
How to debug?
Without the big picture, we may try to optimise a model for weeks for nothing
13
14. www.kensu.io
DATA SCIENCE GOVERNANCE
Data governance: controls that data meets precise standards and
involves monitoring against production data.
Data Science Governance: control that data activity meets precise
standards and involves monitoring against production data activity.
A Data Activity is described by at least technologies, users, systems,
data, processing
14
15. www.kensu.io
GOVERNING DATA SCIENCE
Who does what on which data and where it is done?
What is the impact of a process on the global system?
What are the performance metrics (quality, execution,…) of the processes?
15
16. www.kensu.io
CONTINUOUS INTEGRATION FOR DATA SCIENCE
Data Scientists/Citizens have a view on all the activities applied to
the original sources used in his/her own process.
They also have a control on their own results in production
They have the opportunity to analyse and debug a pipeline
involving all activities:
• independently of the technologies
• involving several people in the enterprise
16
19. www.kensu.io
GET THE DATA
As usual, we have to collect the right data to take right decision.
First run an assessment to create a high level map of all the tools
involved into a company.
For each tool, do whatever it takes to collect information about the
activities it is creating.
Information are metadata, lineage, statistics, accuracy measures, …
19
20. www.kensu.io
CONNECT THE DATA
Data Science Governance needs the global picture.
To do that we need to connect all data that can be collected.
So that, it is possible to create a cartography of all on-going processes.
This map tracks all data and their descendants
20
21. www.kensu.io
USE THE DATA
This is where the fun part starts… the map of data activities is an
amazing source of information
Here are a few things you can think of when using this kind of data:
• impact analysis
• dependency analysis
• optimisation
• recommendation
21
23. www.kensu.io
ACCOUNTABILITY PRINCIPLE
Implement appropriate technical and organisational measures that
ensure and demonstrate that you comply. This may include internal
data protection policies such as staff training, internal audits of
processing activities, and reviews of internal HR policies.
23
24. www.kensu.io
TRANSPARENCY
As well as your obligation to provide comprehensive, clear and
transparent privacy policies, if your organisation has more than 250
employees, you must maintain additional internal records of your
processing activities.
24
25. www.kensu.io
ACCOUNTABILITY: DATA SCIENCE GOVERNANCE
To govern data science, we have to:
• collect activities
• connect activities
With this information we can reliably create automatically the
process registry
25
26. www.kensu.io
TRANSPARENCY: DATA SCIENCE GOVERNANCE
To govern data science seen as a continuous integration solution:
we have to explain and measure activities independently of the
technologies.
With this information we can reliably create transparent reports of
activities across the whole chain of processing
26