<!-- wp:paragraph -->
<p>Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of a high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality Management effectively in support of business strategy, which in turn allows for speedy identification of business problems, delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.</p>
<!-- /wp:paragraph -->
<!-- wp:paragraph -->
<p>Learning Objectives:</p>
<!-- /wp:paragraph -->
<!-- wp:list -->
<ul><li>Understand foundational Data Quality concepts based on the DAMA Guide to Data Management Book of Knowledge (DAMA DMBOK), as well as guiding principles, best practices, and steps for improving Data Quality at your organization</li><li>Recognize how chronic business challenges for organizations are often rooted in poor Data Quality</li><li>Share case studies illustrating the hallmarks and benefits of Data Quality success</li></ul>
<!-- /wp:list -->
Tackling data quality problems requires more than a series of tactical, one off improvement projects. By their nature, many data quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process and technology. Join Donna Burbank and Nigel Turner as they provide practical ways to control data quality issues in your organization.
Tackling Data Quality problems requires more than a series of tactical, one-off improvement projects. By their nature, many Data Quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process, and technology. Join Nigel Turner and Donna Burbank as they provide practical ways to control Data Quality issues in your organization.
Data at the Speed of Business with Data Mastering and GovernanceDATAVERSITY
Do you ever wonder how data-driven organizations fuel analytics, improve customer experience, and accelerate business productivity? They are successful by governing and mastering data effectively so they can get trusted data to those who need it faster. Efficient data discovery, mastering and democratization is critical for swiftly linking accurate data with business consumers. When business teams can quickly and easily locate, interpret, trust, and apply data assets to support sound business judgment, it takes less time to see value.
Join data mastering and data governance experts from Informatica—plus a real-world organization empowering trusted data for analytics—for a lively panel discussion. You’ll hear more about how a single cloud-native approach can help global businesses in any economy create more value—faster, more reliably, and with more confidence—by making data management and governance easier to implement.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
The Data Driven University - Automating Data Governance and Stewardship in Au...Pieter De Leenheer
Data Governance and Stewardship requires automation of business semantics management at its nucleus, in order to achieve data trust between business and IT communities in the organization. University divisions operate highly autonomously and decentralized, and are often geographically distributed. Hence, they benefit more from an collaborative and agile approach to Data Governance and Stewardship approach that adapts to its nature.
In this lecture, we start by reviewing 'C' in ICT and reflect on the dilemma: what is the most important quality of data being shared: truth or trust? We review the wide spectrum of business semantics. We visit the different phases of growing data pain as an organization expands, and we map each phase on this spectrum of semantics.
Next, we introduce our principles and framework for business semantics management to support Data Governance and Stewardship focusing on the structural (what), processual (how) and organizational (who) components. We illustrate with use cases from Stanford University, George Washington University and Public Science and Innovation Administrations.
Data protection and privacy regulations such as the EU’s General Data Protection Regulation (GDPR), the California Consumer Privacy Act (CCPA), and Singapore’s Personal Data Protection Act (PDPA) have been major drivers for data governance initiatives and the emergence of data catalog solutions. Organizations have an ever-increasing appetite to leverage their data for business advantage, either through internal collaboration, data sharing across ecosystems, direct commercialization, or as the basis for AI-driven business decision-making. This requires data governance and especially data asset catalog solutions to step up once again and enable data-driven businesses to leverage their data responsibly, ethically, compliantly, and accountably.
This presentation explores how data catalog has become a key technology enabler in overcoming these challenges.
Improving Data Literacy Around Data ArchitectureDATAVERSITY
Data Literacy is an increasing concern, as organizations look to become more data-driven. As the rise of the citizen data scientist and self-service data analytics becomes increasingly common, the need for business users to understand core Data Management fundamentals is more important than ever. At the same time, technical roles need a strong foundation in Data Architecture principles and best practices. Join this webinar to understand the key components of Data Literacy, and practical ways to implement a Data Literacy program in your organization.
Lessons in Data Modeling: Data Modeling & MDMDATAVERSITY
Master Data Management (MDM) can create a 360 view of core business assets such as Customer, Product, Vendor, and more. Data modeling is a core component of MDM in both creating the technical integration between disparate systems and, perhaps more importantly, aligning business definitions & rules.
Join this webcast to learn how to effectively apply a data model in your MDM implementation.
Tackling data quality problems requires more than a series of tactical, one off improvement projects. By their nature, many data quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process and technology. Join Donna Burbank and Nigel Turner as they provide practical ways to control data quality issues in your organization.
Tackling Data Quality problems requires more than a series of tactical, one-off improvement projects. By their nature, many Data Quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process, and technology. Join Nigel Turner and Donna Burbank as they provide practical ways to control Data Quality issues in your organization.
Data at the Speed of Business with Data Mastering and GovernanceDATAVERSITY
Do you ever wonder how data-driven organizations fuel analytics, improve customer experience, and accelerate business productivity? They are successful by governing and mastering data effectively so they can get trusted data to those who need it faster. Efficient data discovery, mastering and democratization is critical for swiftly linking accurate data with business consumers. When business teams can quickly and easily locate, interpret, trust, and apply data assets to support sound business judgment, it takes less time to see value.
Join data mastering and data governance experts from Informatica—plus a real-world organization empowering trusted data for analytics—for a lively panel discussion. You’ll hear more about how a single cloud-native approach can help global businesses in any economy create more value—faster, more reliably, and with more confidence—by making data management and governance easier to implement.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
The Data Driven University - Automating Data Governance and Stewardship in Au...Pieter De Leenheer
Data Governance and Stewardship requires automation of business semantics management at its nucleus, in order to achieve data trust between business and IT communities in the organization. University divisions operate highly autonomously and decentralized, and are often geographically distributed. Hence, they benefit more from an collaborative and agile approach to Data Governance and Stewardship approach that adapts to its nature.
In this lecture, we start by reviewing 'C' in ICT and reflect on the dilemma: what is the most important quality of data being shared: truth or trust? We review the wide spectrum of business semantics. We visit the different phases of growing data pain as an organization expands, and we map each phase on this spectrum of semantics.
Next, we introduce our principles and framework for business semantics management to support Data Governance and Stewardship focusing on the structural (what), processual (how) and organizational (who) components. We illustrate with use cases from Stanford University, George Washington University and Public Science and Innovation Administrations.
Data protection and privacy regulations such as the EU’s General Data Protection Regulation (GDPR), the California Consumer Privacy Act (CCPA), and Singapore’s Personal Data Protection Act (PDPA) have been major drivers for data governance initiatives and the emergence of data catalog solutions. Organizations have an ever-increasing appetite to leverage their data for business advantage, either through internal collaboration, data sharing across ecosystems, direct commercialization, or as the basis for AI-driven business decision-making. This requires data governance and especially data asset catalog solutions to step up once again and enable data-driven businesses to leverage their data responsibly, ethically, compliantly, and accountably.
This presentation explores how data catalog has become a key technology enabler in overcoming these challenges.
Improving Data Literacy Around Data ArchitectureDATAVERSITY
Data Literacy is an increasing concern, as organizations look to become more data-driven. As the rise of the citizen data scientist and self-service data analytics becomes increasingly common, the need for business users to understand core Data Management fundamentals is more important than ever. At the same time, technical roles need a strong foundation in Data Architecture principles and best practices. Join this webinar to understand the key components of Data Literacy, and practical ways to implement a Data Literacy program in your organization.
Lessons in Data Modeling: Data Modeling & MDMDATAVERSITY
Master Data Management (MDM) can create a 360 view of core business assets such as Customer, Product, Vendor, and more. Data modeling is a core component of MDM in both creating the technical integration between disparate systems and, perhaps more importantly, aligning business definitions & rules.
Join this webcast to learn how to effectively apply a data model in your MDM implementation.
Building a Data Strategy – Practical Steps for Aligning with Business GoalsDATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task – but it’s worth the effort. Getting your Data Strategy right can provide significant value, as data drives many of the key initiatives in today’s marketplace – from digital transformation, to marketing, to customer centricity, to population health, and more. This webinar will help demystify Data Strategy and its relationship to Data Architecture and will provide concrete, practical ways to get started.
Data Catalogs Are the Answer – What is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
Driving Data Intelligence in the Supply Chain Through the Data Catalog at TJXDATAVERSITY
Roles and responsibilities are a critical component of every Data Governance program. Building a set of roles that are practical and that will not interfere with people’s “day jobs” is an important consideration that will influence how well your program is adopted. This tutorial focuses on sharing a proven model guaranteed to represent your organization.
Join Bob Seiner for this lively webinar where he will dissect a complete Operating Model of Roles and Responsibilities that encompasses all levels of the organization. Seiner will detail the roles and describe the most effective way to associate people with the roles. You will walk out of this webinar with a model to apply to your organization.
In this session Bob will share:
- The five levels of Data Governance roles
- A proven Operating Model of Roles and Responsibilities
- How to customize the model to meet your requirements
- Setting appropriate role expectations
- How to operationalize the roles and demonstrate value
Customer-Centric Data Management for Better Customer ExperiencesInformatica
With consumer and business buyer expectations growing exponentially, more businesses are competing on the basis of customer experience. But executing preferred customer experiences requires data about who your customers are today and what will they likely need in the future. Every business can benefit from an AI-powered master data management platform to supply this information to line-of-business owners so they can execute great experiences at scale. This same need is true from an internal business process perspective as well. For example, many businesses require better data management practices to deliver preferred employee experiences. Informatica provides an MDM platform to solve for these examples and more.
This presentation reports on data governance best practices. Based on a definition of fundamental terms and the business rationale for data governance, a set of case studies from leading companies is presented. The content of this presentation is a result of the Competence Center Corporate Data Quality (CC CDQ) at the University of St. Gallen, Switzerland.
To take a “ready, aim, fire” tactic to implement Data Governance, many organizations assess themselves against industry best practices. The process is not difficult or time-consuming and can directly assure that your activities target your specific needs. Best practices are always a strong place to start.
Join Bob Seiner for this popular RWDG topic, where he will provide the information you need to set your program in the best possible direction. Bob will walk you through the steps of conducting an assessment and share with you a set of typical results from taking this action. You may be surprised at how easy it is to organize the assessment and may hear results that stimulate the actions that you need to take.
In this webinar, Bob will share:
- The value of performing a Data Governance best practice assessment
- A practical list of industry Data Governance best practices
- Criteria to determine if a practice is best practice
- Steps to follow to complete an assessment
- Typical recommendations and actions that result from an assessment
Data Integration, Access, Flow, Exchange, Transfer, Load And Extract Architec...Alan McSweeney
These notes describe a generalised data integration architecture framework and set of capabilities.
With many organisations, data integration tends to have evolved over time with many solution-specific tactical approaches implemented. The consequence of this is that there is frequently a mixed, inconsistent data integration topography. Data integrations are often poorly understood, undocumented and difficult to support, maintain and enhance.
Data interoperability and solution interoperability are closely related – you cannot have effective solution interoperability without data interoperability.
Data integration has multiple meanings and multiple ways of being used such as:
- Integration in terms of handling data transfers, exchanges, requests for information using a variety of information movement technologies
- Integration in terms of migrating data from a source to a target system and/or loading data into a target system
- Integration in terms of aggregating data from multiple sources and creating one source, with possibly date and time dimensions added to the integrated data, for reporting and analytics
- Integration in terms of synchronising two data sources or regularly extracting data from one data sources to update a target
- Integration in terms of service orientation and API management to provide access to raw data or the results of processing
There are two aspects to data integration:
1. Operational Integration – allow data to move from one operational system and its data store to another
2. Analytic Integration – move data from operational systems and their data stores into a common structure for analysis
Master Data Management – Aligning Data, Process, and GovernanceDATAVERSITY
Master Data Management (MDM) provides organizations with an accurate and comprehensive view of their business-critical data such as customers, products, vendors, and more. While mastering these key data areas can be a complex task, the value of doing so can be tremendous – from real-time operational integration to data warehousing and analytic reporting. This webinar will provide practical strategies for gaining value from your MDM initiative, while at the same time assuring a solid architectural and governance foundation that will ensure long-term, enterprise-wide success.
2013 Data Governance Professionals Organization (DGPO) Digital River WebinarDeepak Bhaskar, MBA, BSEE
Hosted by the Data Governance Professionals Organzation (DGPO) for webinar attendees. Successful Data Governance at Digital River. 2013 DGIQ Data Governance Best Practice Award: Finalist
Enterprise Data Management Framework OverviewJohn Bao Vuu
A solid data management foundation to support big data analytics and more importantly a data-driven culture is necessary for today’s organizations.
A mature Data Management Program can reduce operational costs and enable rapid business growth and development. Data Management program must evolve to monetize data assets, deliver breakthrough innovation and help drive business strategies in new markets.
DAS Slides: Data Quality Best PracticesDATAVERSITY
Tackling Data Quality problems requires more than a series of tactical, one-off improvement projects. By their nature, many Data Quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process, and technology. Join Nigel Turner and Donna Burbank as they provide practical ways to control Data Quality issues in your organization.
Introduction to DCAM, the Data Management Capability Assessment Model - Editi...Element22
DCAM stands for Data management Capability Assessment Model. DCAM is a model to assess data management capabilities within the financial industry. It was created by the EDM Council in collaboration with over 100 financial institutions. This presentation provides an overview of DCAM and how financial institutions leverage DCAM to improve or establish their data management programs and meet regulatory requirements such as BCBS 239. Also the benefits of DCAM are described as part of this presentation.
Data Governance Roles as the Backbone of Your ProgramDATAVERSITY
The method you follow to form your Data Governance roles and responsibilities will impact the success of your program. There are industry-standard roles that require adjustment to fit the culture of your organization when getting started, gaining acceptance, and demonstrating sustained value. Roles are the backbone of a productive Data Governance program.
Bob Seiner will share his updated operating model of roles and responsibilities in this topical RWDG webinar. The model Bob uses is meant to overlay your present organizational structure rather than requiring you to try and plug your organization into someone else’s model. This webinar will provide everything you need to know about Data Governance roles.
Bob will address the following in this webinar:
• An operating model of Data Governance roles and responsibilities
• How to customize the model to mimic your existing structure
• The meaning behind the oft-used “roles pyramid”
• Detailed responsibilities at each level of the organization
• Using the model to influence Data Governance acceptance
Data Governance and Metadata ManagementDATAVERSITY
Metadata is a tool that improves data understanding, builds end-user confidence, and improves the return on investment in every asset associated with becoming a data-centric organization. Metadata’s use has expanded beyond “data about data” to cover every phase of data analytics, protection, and quality improvement. Data Governance and metadata are connected at the hip in every way possible. As the song goes, “You can’t have one without the other.”
In this RWDG webinar, Bob Seiner will provide a way to renew your energy by focusing on the valuable asset that can make or break your Data Governance program’s success. The truth is metadata is already inherent in your data environment, and it can be leveraged by making it available to all levels of the organization. At issue is finding the most appropriate ways to leverage and share metadata to improve data value and protection.
Throughout this webinar, Bob will share information about:
- Delivering an improved definition of metadata
- Communicating the relationship between successful governance and metadata
- Getting your business community to embrace the need for metadata
- Determining the metadata that will provide the most bang for your bucks
- The importance of Metadata Management to becoming data-centric
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of a high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality Management effectively in support of business strategy, which in turn allows for speedy identification of business problems, delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
DataEd Slides: Approaching Data Management TechnologiesDATAVERSITY
Our architecturally solid stool requires three legs: people, process, and technologies. This webinar looks at the most misunderstood of these three components: technology. While most organizations begin with technologies, it turns out that technologies are the last component that should be considered. This webinar will survey a range of Data Management technologies that can be used to increase the productivity of Data Management efforts.
Building a Data Strategy – Practical Steps for Aligning with Business GoalsDATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task – but it’s worth the effort. Getting your Data Strategy right can provide significant value, as data drives many of the key initiatives in today’s marketplace – from digital transformation, to marketing, to customer centricity, to population health, and more. This webinar will help demystify Data Strategy and its relationship to Data Architecture and will provide concrete, practical ways to get started.
Data Catalogs Are the Answer – What is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
Driving Data Intelligence in the Supply Chain Through the Data Catalog at TJXDATAVERSITY
Roles and responsibilities are a critical component of every Data Governance program. Building a set of roles that are practical and that will not interfere with people’s “day jobs” is an important consideration that will influence how well your program is adopted. This tutorial focuses on sharing a proven model guaranteed to represent your organization.
Join Bob Seiner for this lively webinar where he will dissect a complete Operating Model of Roles and Responsibilities that encompasses all levels of the organization. Seiner will detail the roles and describe the most effective way to associate people with the roles. You will walk out of this webinar with a model to apply to your organization.
In this session Bob will share:
- The five levels of Data Governance roles
- A proven Operating Model of Roles and Responsibilities
- How to customize the model to meet your requirements
- Setting appropriate role expectations
- How to operationalize the roles and demonstrate value
Customer-Centric Data Management for Better Customer ExperiencesInformatica
With consumer and business buyer expectations growing exponentially, more businesses are competing on the basis of customer experience. But executing preferred customer experiences requires data about who your customers are today and what will they likely need in the future. Every business can benefit from an AI-powered master data management platform to supply this information to line-of-business owners so they can execute great experiences at scale. This same need is true from an internal business process perspective as well. For example, many businesses require better data management practices to deliver preferred employee experiences. Informatica provides an MDM platform to solve for these examples and more.
This presentation reports on data governance best practices. Based on a definition of fundamental terms and the business rationale for data governance, a set of case studies from leading companies is presented. The content of this presentation is a result of the Competence Center Corporate Data Quality (CC CDQ) at the University of St. Gallen, Switzerland.
To take a “ready, aim, fire” tactic to implement Data Governance, many organizations assess themselves against industry best practices. The process is not difficult or time-consuming and can directly assure that your activities target your specific needs. Best practices are always a strong place to start.
Join Bob Seiner for this popular RWDG topic, where he will provide the information you need to set your program in the best possible direction. Bob will walk you through the steps of conducting an assessment and share with you a set of typical results from taking this action. You may be surprised at how easy it is to organize the assessment and may hear results that stimulate the actions that you need to take.
In this webinar, Bob will share:
- The value of performing a Data Governance best practice assessment
- A practical list of industry Data Governance best practices
- Criteria to determine if a practice is best practice
- Steps to follow to complete an assessment
- Typical recommendations and actions that result from an assessment
Data Integration, Access, Flow, Exchange, Transfer, Load And Extract Architec...Alan McSweeney
These notes describe a generalised data integration architecture framework and set of capabilities.
With many organisations, data integration tends to have evolved over time with many solution-specific tactical approaches implemented. The consequence of this is that there is frequently a mixed, inconsistent data integration topography. Data integrations are often poorly understood, undocumented and difficult to support, maintain and enhance.
Data interoperability and solution interoperability are closely related – you cannot have effective solution interoperability without data interoperability.
Data integration has multiple meanings and multiple ways of being used such as:
- Integration in terms of handling data transfers, exchanges, requests for information using a variety of information movement technologies
- Integration in terms of migrating data from a source to a target system and/or loading data into a target system
- Integration in terms of aggregating data from multiple sources and creating one source, with possibly date and time dimensions added to the integrated data, for reporting and analytics
- Integration in terms of synchronising two data sources or regularly extracting data from one data sources to update a target
- Integration in terms of service orientation and API management to provide access to raw data or the results of processing
There are two aspects to data integration:
1. Operational Integration – allow data to move from one operational system and its data store to another
2. Analytic Integration – move data from operational systems and their data stores into a common structure for analysis
Master Data Management – Aligning Data, Process, and GovernanceDATAVERSITY
Master Data Management (MDM) provides organizations with an accurate and comprehensive view of their business-critical data such as customers, products, vendors, and more. While mastering these key data areas can be a complex task, the value of doing so can be tremendous – from real-time operational integration to data warehousing and analytic reporting. This webinar will provide practical strategies for gaining value from your MDM initiative, while at the same time assuring a solid architectural and governance foundation that will ensure long-term, enterprise-wide success.
2013 Data Governance Professionals Organization (DGPO) Digital River WebinarDeepak Bhaskar, MBA, BSEE
Hosted by the Data Governance Professionals Organzation (DGPO) for webinar attendees. Successful Data Governance at Digital River. 2013 DGIQ Data Governance Best Practice Award: Finalist
Enterprise Data Management Framework OverviewJohn Bao Vuu
A solid data management foundation to support big data analytics and more importantly a data-driven culture is necessary for today’s organizations.
A mature Data Management Program can reduce operational costs and enable rapid business growth and development. Data Management program must evolve to monetize data assets, deliver breakthrough innovation and help drive business strategies in new markets.
DAS Slides: Data Quality Best PracticesDATAVERSITY
Tackling Data Quality problems requires more than a series of tactical, one-off improvement projects. By their nature, many Data Quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process, and technology. Join Nigel Turner and Donna Burbank as they provide practical ways to control Data Quality issues in your organization.
Introduction to DCAM, the Data Management Capability Assessment Model - Editi...Element22
DCAM stands for Data management Capability Assessment Model. DCAM is a model to assess data management capabilities within the financial industry. It was created by the EDM Council in collaboration with over 100 financial institutions. This presentation provides an overview of DCAM and how financial institutions leverage DCAM to improve or establish their data management programs and meet regulatory requirements such as BCBS 239. Also the benefits of DCAM are described as part of this presentation.
Data Governance Roles as the Backbone of Your ProgramDATAVERSITY
The method you follow to form your Data Governance roles and responsibilities will impact the success of your program. There are industry-standard roles that require adjustment to fit the culture of your organization when getting started, gaining acceptance, and demonstrating sustained value. Roles are the backbone of a productive Data Governance program.
Bob Seiner will share his updated operating model of roles and responsibilities in this topical RWDG webinar. The model Bob uses is meant to overlay your present organizational structure rather than requiring you to try and plug your organization into someone else’s model. This webinar will provide everything you need to know about Data Governance roles.
Bob will address the following in this webinar:
• An operating model of Data Governance roles and responsibilities
• How to customize the model to mimic your existing structure
• The meaning behind the oft-used “roles pyramid”
• Detailed responsibilities at each level of the organization
• Using the model to influence Data Governance acceptance
Data Governance and Metadata ManagementDATAVERSITY
Metadata is a tool that improves data understanding, builds end-user confidence, and improves the return on investment in every asset associated with becoming a data-centric organization. Metadata’s use has expanded beyond “data about data” to cover every phase of data analytics, protection, and quality improvement. Data Governance and metadata are connected at the hip in every way possible. As the song goes, “You can’t have one without the other.”
In this RWDG webinar, Bob Seiner will provide a way to renew your energy by focusing on the valuable asset that can make or break your Data Governance program’s success. The truth is metadata is already inherent in your data environment, and it can be leveraged by making it available to all levels of the organization. At issue is finding the most appropriate ways to leverage and share metadata to improve data value and protection.
Throughout this webinar, Bob will share information about:
- Delivering an improved definition of metadata
- Communicating the relationship between successful governance and metadata
- Getting your business community to embrace the need for metadata
- Determining the metadata that will provide the most bang for your bucks
- The importance of Metadata Management to becoming data-centric
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of a high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality Management effectively in support of business strategy, which in turn allows for speedy identification of business problems, delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
DataEd Slides: Approaching Data Management TechnologiesDATAVERSITY
Our architecturally solid stool requires three legs: people, process, and technologies. This webinar looks at the most misunderstood of these three components: technology. While most organizations begin with technologies, it turns out that technologies are the last component that should be considered. This webinar will survey a range of Data Management technologies that can be used to increase the productivity of Data Management efforts.
DataEd Webinar: Reference & Master Data Management - Unlocking Business ValueDATAVERSITY
Data tends to pile up and can be rendered unusable or obsolete without careful maintenance processes. Reference and Master Data Management (MDM) has been a popular Data Management approach to effectively gain mastery over not just the data but the supporting architecture for processing it. This webinar presents MDM as a strategic approach to improving and formalizing practices around those data items that provide context for many organizational transactions—its master data. Too often, MDM has been implemented technology-first and achieved the same very poor track record (one-third succeeding on-time, within budget, and achieving planned functionality). MDM success depends on a coordinated approach typically involving Data Governance and Data Quality activities.
Learning Objectives:
- Understand foundational reference and MDM concepts based on the Data Management Body of Knowledge (DMBOK)
- Understand why these are an important component of your Data Architecture
- Gain awareness of Reference and MDM Frameworks and building blocks
- Know what MDM guiding principles consist of and best practices
- Know how to utilize reference and MDM in support of business strategy
DataEd Slides: Data Modeling is FundamentalDATAVERSITY
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that any and all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, Data Modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important are the data models driving the engineering and architecture activities of your organization. This webinar illustrates Data Modeling as a key activity upon which so much technology depends.
Data-Ed Slides: Best Practices in Data Stewardship (Technical)DATAVERSITY
In order to find value in your organization's data assets, heroic data stewards are tasked with saving the day- every single day! These heroes adhere to a data governance framework and work to ensure that data is: captured right the first time, validated through automated means, and integrated into business processes. Whether its data profiling or in depth root cause analysis, data stewards can be counted on to ensure the organization's mission critical data is reliable. In this webinar we will approach this framework, and punctuate important facets of a data steward’s role.
Learning Objectives:
- Understand the business need for a data governance framework
- Learn why embedded data quality principles are an important part of system/process design
- Identify opportunities to help drive your organization to a data driven culture
DataEd Slides: Exorcising the Seven Deadly Data SinsDATAVERSITY
The difficulty of implementing Data Strategy concepts often goes underappreciated, especially the multifaceted nature of the challenges that need to be met. Deficiencies in organizational readiness and core competence represent clearly visible problems faced by data managers, but beyond that there are several cultural and structural barriers common to virtually all organizations that must be eliminated in order to facilitate effective management of data.
In this webinar, we will discuss these barriers—the titular “Seven Deadly Data Sins”—and in the process will also:
- Elaborate on the three critical factors that lead to strategy failure
- Demonstrate a two-stage Data Strategy implementation process
- Explore the sources and rationales behind the “Seven Deadly Data Sins,” and recommend solutions and alternative approaches
DataEd Slides: Data Strategy Best PracticesDATAVERSITY
Your Data Strategy should be concise, actionable, and understandable by business and IT! Data is not just another resource. It is your most powerful, yet poorly managed and therefore underutilized organizational asset. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Overcoming lack of talent, barriers in organizational thinking, and seven specific data sins are organizational prerequisites to be satisfied before (a measurable) nine out of 10 organizations can achieve the three primary goals of an organizational Data Strategy, which are to:
- Improve the way your people use data
- Improve the way your people use data to achieve your organizational strategy
- Improve your organization’s data
In this manner, your organizational Data Strategy can be used to best focus your data assets in precise support of your organization's strategic objectives. Once past the prerequisites, organizations must develop a disciplined, repeatable means of improving the data literacy, standards, and supply as business objectives in specific areas that become the foci of subsequent Data Governance efforts. This process (based on the theory of constraints) is where the strategic data work really occurs, as organizations identify prioritized areas where better assets, literacy, and support (Data Strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are covered, including:
- A cohesive argument for why Data Strategy is necessary for effective Data Governance
- An overview of prerequisites for effective Data Strategy, as well as common pitfalls that can detract from its implementation, such as the “Seven Deadly Data Sins”
- A repeatable process for identifying and removing data constraints, and the importance of balancing business operation and innovation while doing so
DataEd Slides: Leveraging Data Management TechnologiesDATAVERSITY
Our architecturally solid stool requires three legs: people, process, and technologies. This webinar looks at the most misunderstood of these three components: technology. While most organizations begin with technologies, it turns out that technologies are the last component that should be considered. This webinar will survey a range of technologies that can be used to increase the productivity of Data Management efforts. The goal is to invest in as little infrastructure as possible while still achieving business/program objectives. This program’s learning objectives include:
• Understanding technology considerations
• Appreciating the overview of data technologies and then specifically
• CASE technologies
• Repositories
• Profiling/discovery tools
• Data Quality engineering tools
• Appreciating the complete Data Quality life cycle
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality management effectively in support of business strategy. This, in turn, allows for speedy identification of business problems, delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges can often trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from reoccurring.
Learning objectives:
-Help you understand foundational Data Quality concepts for improving Data Quality at your organization
-Demonstrate how chronic business challenges for organizations are often rooted in poor Data Quality
-Share case studies illustrating the hallmarks and benefits of Data Quality success
Data-Ed Online Webinar: Business Value from MDMDATAVERSITY
This presentation provides you with an understanding of the goals of reference and master data management (MDM), including establishing and implementing authoritative data sources, establishing and implementing more effective means of delivery data to various business processes, as well as increasing the quality of information used in organizational analytical functions (such as BI). You will understand the parallel importance of incorporating data quality engineering into the planning of reference and MDM.
Takeaways:
What is reference and MDM?
Why are reference and MDM important?
Reference and MDM Frameworks
Guiding principles & best practices
This presentation provides you with an understanding of the goals of reference and master data management (MDM), including establishing and implementing authoritative data sources, establishing and implementing more effective means of delivery data to various business processes, as well as increasing the quality of information used in organizational analytical functions (such as BI). You will understand the parallel importance of incorporating data quality engineering into the planning of reference and MDM.
Check out more of our Data-Ed webinars here: http://www.datablueprint.com/resource-center/webinar-schedule/
Data Governance Strategies - With Great Power Comes Great AccountabilityDATAVERSITY
Much like project team management and home improvement, data governance sounds a lot simpler than it actually is. In a nutshell, data governance is the process by which an organization delegates responsibility and exercises control over mission-critical data assets. In practice, though, data governance directs how all other data management functions are performed, meaning that much of your data management strategy’s capacity to function at all depends on your effectiveness in governing its implementation. Understanding these aspects of governance is necessary to eliminate the ambiguity that often surrounds effective data management and stewardship programs, since the goal of governance is to manage the data that supports organizational strategy.
This webinar will:
-Illustrate what data governance functions are required for effective data management, how they fit with other data management disciplines, and why data governance can be tricky for many organizations
-Help you develop a detailed vocabulary and set of narratives to facilitate understanding of your business objectives and imperatives that demand governance
-Provide direction for selling data governance to organizational management as a specifically motivated initiative
DataEd Slides: Data Management Best PracticesDATAVERSITY
It is clear that Data Management best practices exist and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This approach combines the DM BoK and the CMMI/DMM, permitting organizations with the opportunity to benefit from the best of both. The approach permits organizations to understand current Data Management practices, strengths to leverage, and remediation opportunities. In a nutshell, it describes what must be done at the programmatic level to achieve better data use.
Data-Ed Webinar: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Takeaways:
Understanding how to contribute to organizational challenges beyond traditional data architecting
How to utilize data architectures in support of business strategy
Understanding foundational data architecture concepts based on the DAMA DMBOK
Data architecture guiding principles & best practices
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Find out more: http://www.datablueprint.com/resource-center/webinar-schedule/
DAS Slides: Building a Data Strategy - Practical Steps for Aligning with Busi...DATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task. The opportunity in getting it right can be significant, however, as data drives many of the key initiatives in today’s marketplace: digital transformation, marketing, customer centricity, and more. This webinar will help de-mystify Data Strategy and Data Architecture and will provide concrete, practical ways to get started.
Data-Ed Webinar: Monetizing Data Management - Show Me the MoneyDATAVERSITY
Practicality and profitability may share a page in the dictionary, but incorporating both into a data management plan can prove challenging. Many data professionals struggle to demonstrate tangible returns on data management investments, especially in industries such as healthcare where financial results aren’t necessarily an organization’s primary concern. The key to “monetizing” data management, therefore, is thinking about data in a different way: as an information solution rather than simply an IT one, using data to drive decision-making towards increased profits and potentially alternative returns on investment or value outcomes as well. Taking a broader view of data assets facilitates easier sharing of information across organizational silos, and allows for a wider understanding of the investment’s requirements and benefits.
In this webinar—designed to appeal to both business and IT attendees—your presenter will:
Describe multiple types of value produced through data-centric development and management practices
Expand on and beyond metrics meant for increasing revenues or decreasing costs—i.e. investments that directly impact an organization’s financial position
Detail how alternative statistics and valuations can be used to justify data management and quality initiatives
Data Architecture is foundational to an information-based operational environment. Without proper structure and efficiency in organization, data assets cannot be utilized to their full potential, which in turn harms bottom-line business value. When designed well and used effectively, however, a strong Data Architecture can be referenced to inform, clarify, understand, and resolve aspects of a variety of business problems commonly encountered in organizations.
The goal of this webinar is not to instruct you in being an outright Data Architect, but rather to enable you to envision a number of uses for Data Architectures that will maximize your organization’s competitive advantage. With that being said, we will:
Discuss Data Architecture’s guiding principles and best practices
Demonstrate how to utilize Data Architecture to address a broad variety of organizational challenges and support your overall business strategy
Illustrate how best to understand foundational Data Architecture concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
Data-Ed Online: Approaching Data QualityDATAVERSITY
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality management effectively in support of business strategy. This, in turn, allows for speedy identification of business problems, the delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This, in turn, allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Learning Objectives:
Help you understand foundational Data Quality concepts based on the DAMA Guide to Data Management Book of Knowledge (DAMA DMBoK), as well as guiding principles, best practices, and steps for improving Data Quality at your organization
Demonstrate how chronic business challenges for organizations are often rooted in poor Data Quality
Share case studies illustrating the hallmarks and benefits of Data Quality success
DataEd Slides: The Seven Deadly Data SinsDATAVERSITY
While wrath and envy are best left for human resources to address, overcoming the numerous obstacles that often inhibit successful Data Management must be a full organizational effort. The difficulty of implementing a new Data Strategy often goes underappreciated, particularly the multi-faceted nature of the challenges that need to be met. Deficiencies in organizational readiness and core competence represent clearly visible problems faced by data managers, but beyond that there are several cultural and structural barriers common to virtually all organizations that must be eliminated in order to facilitate effective management of data.
In this webinar, we will discuss these barriers—the titular “Seven Deadly Data Sins”—and in the process will also:
Elaborate upon the three critical factors that lead to strategy failure
Demonstrate a two-stage Data Strategy implementation process
Explore the sources and rationales behind the “Seven Deadly Data Sins” and recommend solutions and alternative approaches
Discuss foundational data concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
Architecture, Products, and Total Cost of Ownership of the Leading Machine Le...DATAVERSITY
Organizations today need a broad set of enterprise data cloud services with key data functionality to modernize applications and utilize machine learning. They need a comprehensive platform designed to address multi-faceted needs by offering multi-function data management and analytics to solve the enterprise’s most pressing data and analytic challenges in a streamlined fashion.
In this research-based session, I’ll discuss what the components are in multiple modern enterprise analytics stacks (i.e., dedicated compute, storage, data integration, streaming, etc.) and focus on total cost of ownership.
A complete machine learning infrastructure cost for the first modern use case at a midsize to large enterprise will be anywhere from $3 million to $22 million. Get this data point as you take the next steps on your journey into the highest spend and return item for most companies in the next several years.
What is data literacy? Which organizations, and which workers in those organizations, need to be data-literate? There are seemingly hundreds of definitions of data literacy, along with almost as many opinions about how to achieve it.
In a broader perspective, companies must consider whether data literacy is an isolated goal or one component of a broader learning strategy to address skill deficits. How does data literacy compare to other types of skills or “literacy” such as business acumen?
This session will position data literacy in the context of other worker skills as a framework for understanding how and where it fits and how to advocate for its importance.
Uncover how your business can save money and find new revenue streams.
Driving profitability is a top priority for companies globally, especially in uncertain economic times. It's imperative that companies reimagine growth strategies and improve process efficiencies to help cut costs and drive revenue – but how?
By leveraging data-driven strategies layered with artificial intelligence, companies can achieve untapped potential and help their businesses save money and drive profitability.
In this webinar, you'll learn:
- How your company can leverage data and AI to reduce spending and costs
- Ways you can monetize data and AI and uncover new growth strategies
- How different companies have implemented these strategies to achieve cost optimization benefits
Data Catalogs Are the Answer – What Is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
In this webinar, Bob will focus on:
-Selecting the appropriate metadata to govern
-The business and technical value of a data catalog
-Building the catalog into people’s routines
-Positioning the data catalog for success
-Questions the data catalog can answer
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, data modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization. This webinar illustrates data modeling as a key activity upon which so much technology and business investment depends.
Specific learning objectives include:
- Understanding what types of challenges require data modeling to be part of the solution
- How automation requires standardization on derivable via data modeling techniques
- Why only a working partnership between data and the business can produce useful outcomes
Analytics play a critical role in supporting strategic business initiatives. Despite the obvious value to analytic professionals of providing the analytics for these initiatives, many executives question the economic return of analytics as well as data lakes, machine learning, master data management, and the like.
Technology professionals need to calculate and present business value in terms business executives can understand. Unfortunately, most IT professionals lack the knowledge required to develop comprehensive cost-benefit analyses and return on investment (ROI) measurements.
This session provides a framework to help technology professionals research, measure, and present the economic value of a proposed or existing analytics initiative, no matter the form that the business benefit arises. The session will provide practical advice about how to calculate ROI and the formulas, and how to collect the necessary information.
How a Semantic Layer Makes Data Mesh Work at ScaleDATAVERSITY
Data Mesh is a trending approach to building a decentralized data architecture by leveraging a domain-oriented, self-service design. However, the pure definition of Data Mesh lacks a center of excellence or central data team and doesn’t address the need for a common approach for sharing data products across teams. The semantic layer is emerging as a key component to supporting a Hub and Spoke style of organizing data teams by introducing data model sharing, collaboration, and distributed ownership controls.
This session will explain how data teams can define common models and definitions with a semantic layer to decentralize analytics product creation using a Hub and Spoke architecture.
Attend this session to learn about:
- The role of a Data Mesh in the modern cloud architecture.
- How a semantic layer can serve as the binding agent to support decentralization.
- How to drive self service with consistency and control.
Enterprise data literacy. A worthy objective? Certainly! A realistic goal? That remains to be seen. As companies consider investing in data literacy education, questions arise about its value and purpose. While the destination – having a data-fluent workforce – is attractive, we wonder how (and if) we can get there.
Kicking off this webinar series, we begin with a panel discussion to explore the landscape of literacy, including expert positions and results from focus groups:
- why it matters,
- what it means,
- what gets in the way,
- who needs it (and how much they need),
- what companies believe it will accomplish.
In this engaging discussion about literacy, we will set the stage for future webinars to answer specific questions and feature successful literacy efforts.
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...DATAVERSITY
Change is hard, especially in response to negative stimuli or what is perceived as negative stimuli. So organizations need to reframe how they think about data privacy, security and governance, treating them as value centers to 1) ensure enterprise data can flow where it needs to, 2) prevent – not just react – to internal and external threats, and 3) comply with data privacy and security regulations.
Working together, these roles can accelerate faster access to approved, relevant and higher quality data – and that means more successful use cases, faster speed to insights, and better business outcomes. However, both new information and tools are required to make the shift from defense to offense, reducing data drama while increasing its value.
Join us for this panel discussion with experts in these fields as they discuss:
- Recent research about where data privacy, security and governance stand
- The most valuable enterprise data use cases
- The common obstacles to data value creation
- New approaches to data privacy, security and governance
- Their advice on how to shift from a reactive to resilient mindset/culture/organization
You’ll be educated, entertained and inspired by this panel and their expertise in using the data trifecta to innovate more often, operate more efficiently, and differentiate more strategically.
Data Governance Trends - A Look Backwards and ForwardsDATAVERSITY
As DATAVERSITY’s RWDG series hurdles into our 12th year, this webinar takes a quick look behind us, evaluates the present, and predicts the future of Data Governance. Based on webinar numbers, hot Data Governance topics have evolved over the years from policies and best practices, roles and tools, data catalogs and frameworks, to supporting data mesh and fabric, artificial intelligence, virtualization, literacy, and metadata governance.
Join Bob Seiner as he reflects on the past and what has and has not worked, while sharing examples of enterprise successes and struggles. In this webinar, Bob will challenge the audience to stay a step ahead by learning from the past and blazing a new trail into the future of Data Governance.
In this webinar, Bob will focus on:
- Data Governance’s past, present, and future
- How trials and tribulations evolve to success
- Leveraging lessons learned to improve productivity
- The great Data Governance tool explosion
- The future of Data Governance
Data Governance Trends and Best Practices To Implement TodayDATAVERSITY
Would you share your bank account information on social media? How about shouting your social security number on the New York City subway? We didn’t think so either – that’s why data governance is consistently top of mind.
In this webinar, we’ll discuss the common Cloud data governance best practices – and how to apply them today. Join us to uncover Google Cloud’s investment in data governance and learn practical and doable methods around key management and confidential computing. Hear real customer experiences and leave with insights that you can share with your team. Let’s get solving.
Topics that you will hear addressed in this webinar:
- Understanding the basics of Cloud Incident Response (IR) and anticipated data governance trends
- Best practices for key management and apply data governance to your day-to-day
- The next wave of Confidential Computing and how to get started, including a demo
It is a fascinating, explosive time for enterprise analytics.
It is from the position of analytics leadership that the enterprise mission will be executed and company leadership will emerge. The data professional is absolutely sitting on the performance of the company in this information economy and has an obligation to demonstrate the possibilities and originate the architecture, data, and projects that will deliver analytics. After all, no matter what business you’re in, you’re in the business of analytics.
The coming years will be full of big changes in enterprise analytics and data architecture. William will kick off the fifth year of the Advanced Analytics series with a discussion of the trends winning organizations should build into their plans, expectations, vision, and awareness now.
Too often I hear the question “Can you help me with our data strategy?” Unfortunately, for most, this is the wrong request because it focuses on the least valuable component: the data strategy itself. A more useful request is: “Can you help me apply data strategically?” Yes, at early maturity phases the process of developing strategic thinking about data is more important than the actual product! Trying to write a good (must less perfect) data strategy on the first attempt is generally not productive –particularly given the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” This program refocuses efforts on learning how to iteratively improve the way data is strategically applied. This will permit data-based strategy components to keep up with agile, evolving organizational strategies. It also contributes to three primary organizational data goals. Learn how to improve the following:
- Your organization’s data
- The way your people use data
- The way your people use data to achieve your organizational strategy
This will help in ways never imagined. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges programmatically includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs as organizations identify prioritized areas where better assets, literacy, and support (data strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are also covered, including:
- A cohesive argument for why data strategy is necessary for effective data governance
- An overview of prerequisites for effective strategic use of data strategy, as well as common pitfalls
- A repeatable process for identifying and removing data constraints
- The importance of balancing business operation and innovation
Who Should Own Data Governance – IT or Business?DATAVERSITY
The question is asked all the time: “What part of the organization should own your Data Governance program?” The typical answers are “the business” and “IT (information technology).” Another answer to that question is “Yes.” The program must be owned and reside somewhere in the organization. You may ask yourself if there is a correct answer to the question.
Join this new RWDG webinar with Bob Seiner where Bob will answer the question that is the title of this webinar. Determining ownership of Data Governance is a vital first step. Figuring out the appropriate part of the organization to manage the program is an important second step. This webinar will help you address these questions and more.
In this session Bob will share:
- What is meant by “the business” when it comes to owning Data Governance
- Why some people say that Data Governance in IT is destined to fail
- Examples of IT positioned Data Governance success
- Considerations for answering the question in your organization
- The final answer to the question of who should own Data Governance
It is clear that Data Management best practices exist and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This program describes what must be done at the programmatic level to achieve better data use and a way to implement this as part of your data program. The approach combines DMBoK content and CMMI/DMM processes – permitting organizations with the opportunity to benefit from the best of both. It also permits organizations to understand:
- Their current Data Management practices
- Strengths that should be leveraged
- Remediation opportunities
MLOps – Applying DevOps to Competitive AdvantageDATAVERSITY
MLOps is a practice for collaboration between Data Science and operations to manage the production machine learning (ML) lifecycles. As an amalgamation of “machine learning” and “operations,” MLOps applies DevOps principles to ML delivery, enabling the delivery of ML-based innovation at scale to result in:
Faster time to market of ML-based solutions
More rapid rate of experimentation, driving innovation
Assurance of quality, trustworthiness, and ethical AI
MLOps is essential for scaling ML. Without it, enterprises risk struggling with costly overhead and stalled progress. Several vendors have emerged with offerings to support MLOps: the major offerings are Microsoft Azure ML and Google Vertex AI. We looked at these offerings from the perspective of enterprise features and time-to-value.
Keeping the Pulse of Your Data – Why You Need Data Observability to Improve D...DATAVERSITY
With the explosive growth of DataOps to drive faster and more confident business decisions, proactively understanding the quality and health of your data is more important than ever. Data observability is an emerging discipline within data quality used to expose anomalies in data by continuously monitoring and testing data using artificial intelligence and machine learning to trigger alerts when issues are discovered.
Join Julie Skeen and Shalaish Koul from Precisely, to learn how data observability can be used as part of a DataOps strategy to improve data quality and reliability and to prevent data issues from wreaking havoc on your analytics and ensure that your organization can confidently rely on the data used for advanced analytics and business intelligence.
Topics you will hear addressed in this webinar:
Data observability – what is it and how it can complement your data quality strategy
Why now is the time to incorporate data observability into your DataOps strategy
How data observability helps prevent data issues from impacting downstream analytics
Examples of how data observability can be used to prevent real-world issues
Empowering the Data Driven Business with Modern Business IntelligenceDATAVERSITY
By consolidating data engineering, data warehouse, and data science capabilities under a single fully-managed platform, BigQuery can accelerate computation, reduce data analysis costs, and streamline data management.
Following in-depth interviews with a security services provider and a telecommunications company, Nucleus Research found that customers moving to Google Cloud BigQuery from on-premises data warehouse solutions accelerate data processing by over 75 percent while reducing data ongoing administrative expenses by over 25 percent.
As BigQuery continues to optimize its platform architecture for compute efficiency and multicloud support, Nucleus expects the vendor to see rapid adoption and further penetrate the data warehouse market.
Enterprise Architecture vs. Data ArchitectureDATAVERSITY
Enterprise Architecture (EA) provides a visual blueprint of the organization, and shows key interrelationships between data, process, applications, and more. By abstracting these assets in a graphical view, it’s possible to see key interrelationships, particularly as they relate to data and its business impact across the organization. Join us for a discussion on how Data Architecture is a key component of an overall Enterprise Architecture for enhanced business value and success.
Data Governance Best Practices, Assessments, and RoadmapsDATAVERSITY
When starting or evaluating the present state of your Data Governance program, it is important to focus on best practices such that you don’t take a ready, fire, aim approach. Best practices need to be practical and doable to be selected for your organization, and the program must be at risk if the best practice is not achieved.
Join Bob Seiner for an important webinar focused on industry best practice around standing up formal Data Governance. Learn how to assess your organization against the practices and deliver an effective roadmap based on the results of conducting the assessment.
In this webinar, Bob will focus on:
- Criteria to select the appropriate best practices for your organization
- How to define the best practices for ultimate impact
- Assessing against selected best practices
- Focusing the recommendations on program success
- Delivering a roadmap for your Data Governance program
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfEnterprise Wired
In this guide, we'll explore the key considerations and features to look for when choosing a Trusted analytics platform that meets your organization's needs and delivers actionable intelligence you can trust.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
1. Data Quality Success Stories
1Copyright 2019 by Data Blueprint Slide #
Time:
• timeliness
• currency
• frequency
• time period
Form:
• clarity
• detail
• order
• presentation
• media
Content:
• accuracy
• relevance
• completeness
• conciseness
• scope
• performance
Starting
point
for new
system
development
data performance metadata
data architecture
data
architecture and
data models
shared data updated data
corrected
data
architecture
refinements
facts &
meanings
Metadata &
Data Storage
Starting point
for existing
systems
Metadata Refinement
• Correct Structural Defects
• Update Implementation
Metadata Creation
• Define Data Architecture
• Define Data Model Structures
Metadata Structuring
• Implement Data Model Views
• Populate Data Model Views
Data Refinement
• Correct Data Value Defects
• Re-store Data Values
Data Manipulation
• Manipulate Data
• Updata Data
Data Utilization
• Inspect Data
• Present Data
Data Creation
• Create Data
• Verify Data Values
Data Assessment
• Assess Data Values
• Assess Metadata
Data & Data Relationships
Hypotheses,
Rules
and
Quantifications
Queries
and
Reports
High Probability Data Quality Problem Cause Formulation
Raw Data
Good data is like good water: best served fresh, and ideally well-filtered. Data Management
strategies can produce tremendous procedural improvements and increased profit margins
across the board, but only if the data being managed is of a high quality. Determining how
Data Quality should be engineered provides a useful framework for utilizing Data Quality
Management effectively in support of business strategy, which in turn allows for speedy
identification of business problems, delineation between structural and practice-oriented
defects in Data Management, and proactive prevention of future issues. Organizations must
realize what it means to utilize Data Quality engineering in support of business strategy. This
webinar will illustrate how organizations with chronic business challenges often can trace the
root of the problem to poor Data Quality. Showing how Data Quality should be engineered
provides a useful framework in which to develop an effective approach. This in turn allows
organizations to more quickly identify business problems as well as data problems caused by
structural issues versus practice-oriented defects and prevent these from re-occurring.
Learning Objectives
• Help you understand foundational data quality concepts based on the DAMA Guide to Data
Management Book of Knowledge (DAMA DMBOK), as well as guiding principles, best
practices, and steps for improving data quality at your organization
• Demonstrate how chronic business challenges for organizations
are often rooted in poor data quality
• Share case studies illustrating the hallmarks and benefits of
data quality success
Date: October 8, 2019
Time: 2:00 PM ET/11:00 AM PT UTC-4
Presenter: Peter Aiken, Ph.D.
2Copyright 2019 by Data Blueprint Slide #
Shannon Kempe
Chief Digital Manager at Dataversity.net
Commonly
Asked
Questions
3Copyright 2019 by Data Blueprint Slide #
1) Will I get copies of the
slides after the event?
2) Is this being recorded?
Get Social With Us!
4Copyright 2019 by Data Blueprint Slide #
Like Us on Facebook
www.facebook.com/
datablueprint
Post questions and
comments
Find industry news, insightful
content
and event updates.
Join the Group
Data Management &
Business Intelligence
Ask questions, gain insights
and collaborate with fellow
data management
professionals
Live Twitter Feed
Join the conversation!
Follow us:
@datablueprint
@paiken
Ask questions and
submit your comments:
#dataed
• DAMA International President 2009-2013 / 2018
• DAMA International Achievement Award 2001
(with Dr. E. F. "Ted" Codd
• DAMA International Community Award 2005
Peter Aiken, Ph.D.
5Copyright 2019 by Data Blueprint Slide #
• I've been doing this a long time
• My work is recognized as useful
• Associate Professor of IS (vcu.edu)
• Founder, Data Blueprint (datablueprint.com)
• DAMA International (dama.org)
• 10 books and dozens of articles
• Experienced w/ 500+ data
management practices worldwide
• Multi-year immersions
– US DoD (DISA/Army/Marines/DLA)
– Nokia
– Deutsche Bank
– Wells Fargo
– Walmart
– …
PETER AIKEN WITH JUANITA BILLINGS
FOREWORD BY JOHN BOTTEGA
MONETIZING
DATA MANAGEMENT
Unlocking the Value in Your Organization’s
Most Important Asset.
Data Quality Success Stories
Copyright 2019 by Data Blueprint Slide # 6Peter Aiken, PhD
2. Who is Joan Smith?
http://www.dataflux.com
7Copyright 2019 by Data Blueprint Slide #
Challenges
• Purchased an A4
on June 15 2007
• Had not done
business with the
dealership prior
• "makes them
seem sleazy when
I get a letter in the
mail before I've
even made the
first payment on
the car advertising
lower payments
than I got"
8Copyright 2019 by Data Blueprint Slide #
Letter from the Bank
… so please continue to open
your mail from either Chase or
Bank One
P.S. Please be on the lookout for any
upcoming communications from
either Chase or Bank One regarding
your Bank One credit card and any
other Bank One product you may
have.
Problems
• I initially discarded the letter!
• I became upset after reading it
• It proclaimed that Chase has data
quality challenges
9Copyright 2019 by Data Blueprint Slide #
How to solve this data quality problem using just tools?
Retail price for the unit was $40
10Copyright 2019 by Data Blueprint Slide #
A congratulations
letter from another
bank
Problems
• Bank did not know
it made an error
• Tools alone could
not have prevented
this error
• Lost confidence in
the ability of the
bank to manage
customer funds
11Copyright 2019 by Data Blueprint Slide # 12Copyright 2019 by Data Blueprint Slide #
DropTable
3. 13Copyright 2019 by Data Blueprint Slide # 14Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
Four ways to make your data sparkle!
1.Prioritize the task
– Cleaning data is costly and time consuming
– Identify mission critical/non-mission critical data
2.Involve the data owners
– Seek input of business units on what constitutes "dirty"
data
3.Keep future data clean
– Incorporate processes and technologies that check
every zip code and area code
4.Align your staff with business
– Align IT staff with business units
(Source: CIO JULY 1 2004)
15Copyright 2019 by Data Blueprint Slide # 16Copyright 2019 by Data Blueprint Slide #
• Information transparency
• Analytics
• Business Intelligence
• Increasing efficiencies
• Decreasing costs
• Driving holistic decision-making
across the organization
High
Quality
Data is
Critical
• SQL Server
– 47,000,000,000,000 bytes
– Largest table 34 billion records
• Informix
– 1,800,000,000 queries/day
– 65,000,000 tables / 517,000 databases
• Teradata
– 117 billion records
– 23 TBs for one table
• DB2
– 29,838,518,078 daily queries
• SQL Server
– 47,000,000,000,000 bytes
– Largest table 34 billion records
• Informix
– 1,800,000,000 queries/day
– 65,000,000 tables / 517,000 databases
• Teradata
– 117 billion records
– 23 TBs for one table
• DB2
– 29,838,518,078 daily queries
Data Footprints
17Copyright 2019 by Data Blueprint Slide #
Repeat 100s, thousands, millions, billions of times ...
18Copyright 2019 by Data Blueprint Slide #
4. Death by 1000 Cuts
19Copyright 2019 by Data Blueprint Slide # 20Copyright 2019 by Data Blueprint Slide #
Garbage In ➜ Garbage Out!
My most profound lesson! (so far)
21Copyright 2019 by Data Blueprint Slide #
Perfect
Model
Garbage
Data
Garbage
Results
Data
Warehouse
Machine
Learning
Business
Intelligence
Block ChainAIMDM
Data
Governance
AnalyticsTechnology
GI➜GO!
22Copyright 2019 by Data Blueprint Slide #
Perfect
Model
Garbage
Data
Garbage
Results
Data
Warehouse
Machine
Learning
Block Chain
AI
MDM
Analytics
Technology
Data
Governance
GI➜GO!
Business
Intelligence
23Copyright 2019 by Data Blueprint Slide #
Perfect
Model
Quality
Data
is
founda-
tional
Garbage
Results
Data
Warehouse
Machine
Learning
Block Chain
AI
MDM
Analytics
Technology
Data
Governance
GI➜GO!
Business
Intelligence
24Copyright 2019 by Data Blueprint Slide #
Perfect
Model
Quality
Data
is
founda-
tional
Garbage
Results
Data
Warehouse
Machine
Learning
Business
Intelligence
Block Chain
AI
MDM
Analytics
Technology
Data
Governance
GI➜GO!
5. 25Copyright 2019 by Data Blueprint Slide #
Perfect
Model
Quality
Data
is
founda-
tional
Garbage
Results
Data
Warehouse
Machine
Learning
Block Chain
AI
MDM
Analytics
Technology
Data
Governance
GI➜GO!
Business
Intelligence
26Copyright 2019 by Data Blueprint Slide #
Perfect
Model
Quality
Data
is
founda-
tional
Good
Results
Data
Warehouse
Machine
Learning
Business
Intelligence
Block Chain
AI
MDM
Analytics
Technology
Data
Governance
Quality In ➜ Quality Out!
Data Knowledge is insufficient and informal
• Data management happens 'pretty well' at
the workgroup level
– Defining characteristic of a workgroup
– Without guidance, what are the chances that all
workgroups are pulling toward the same objectives?
– Consider the time spent attempting informal practices
• Data chaff becomes sand in the machinery
– Preventing smooth interoperation and exchanges
– Difficulties that have been hard to account for
• Organizations and individuals lack
– Skills
– Knowledge (architecture)
– Data Engineering (how)
– Data Strategy (why)
27Copyright 2019 by Data Blueprint Slide #
Standard data
Data supply
Data literacy
Making a Better Data Sandwich
28Copyright 2019 by Data Blueprint Slide #
Data literacy
Standard data
Data supply
Making a Better Data Sandwich
29Copyright 2019 by Data Blueprint Slide #
Standard data
Data supply
Data literacy
Making a Better Data Sandwich
30Copyright 2019 by Data Blueprint Slide #
Standard data
Data supply
Data literacy
This cannot happen without engineering and architecture!
Quality engineering/
architecture work products
do not happen accidentally!
6. Making a Better Data Sandwich
31Copyright 2019 by Data Blueprint Slide #
Standard data
Data supply
Data literacy
This cannot happen without data engineering and architecture!
Quality data engineering/
architecture work products
do not happen accidentally!
Our barn had to pass a foundation inspection
32Copyright 2019 by Data Blueprint Slide #
Engineering Standards
33Copyright 2019 by Data Blueprint Slide #
USS Midway &
Pancakes
34Copyright 2019 by Data Blueprint Slide #
• It is tall
• It has a clutch
• It was built in 1942
• It is cemented to the floor
• It is still in regular use!
Why is this an excellent
engineering example?
35Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
Hidden Data Factories are expensive https://hbr.org/2016/09/bad-data-costs-the-u-s-3-trillion-per-year
• Consider these two questions:
– Were your systems explicitly designed to
be integrated or otherwise work together?
– If not then what is the likelihood that they
will just happen to work well together?
• Data must function at the most granular
interaction or it results in things that:
– Take longer (end-of-day job runs 45 hours)
– Cost more (the wrong assets are transferred)
– Deliver less (features are not delivered)
– Present greater risk (billing delayed 30 days, monthly)
• 20-40% of IT budgets are spent evolving data:
– Data migration (changing the location from one place to another)
– Data conversion (changing it into another form, state, or product)
– Data improvement (inspecting, manipulating it, preparing for subsequent use)
36Copyright 2019 by Data Blueprint Slide #
"The choice of data structure and algorithm
can make the difference between software
running in a few seconds or many days."
http://slideplayer.com/slide/7664141/
7.
DQ
challenges
are context
specific!
37Copyright 2019 by Data Blueprint Slide # 38Copyright 2019 by Data Blueprint Slide #
Much more analysis is
required before we can
implement repeatable
solutions to today's data
quality challenges!
TWITTER
USERS SEND
473400TWEETS
SKYPEUSERS MAKE
176220CALLS
INSTAGRAM
GIPHY
USERS POST
PHOTOSSPOTIFYSTREAMS OVER
750,000
SONGS
TUMBLR
USERS PUBLISH
POSTS
USERS WATCH
VIDEOS
SHIPS
PACKAGES
SNAPCHATTHEWEATHER
CHANNEL
NETFLIX
USERS STREAM
97222HRS
OF VIDEO
VENMO
PROCESSES
$68493
PEER-TO-PEER
TRANSACTIONS
TINDER
AMAZON
USERS MATCH
TIMES
TEXTS SENTNEW COMMENTS
RECEIVES
USERS SHARE
SNAPS
YOUTUBE
LINKEDINGAINS
120+NEW
2083333
,
4333560, ,
1388889
,79740,
BITCOIN
NEW
FORECAST
REQUESTS
1.25
ARE CREATED
RECEIVES
,
49380,AMERICANS
USE
OF INTERNET DATA
3138420, , GB,6940,
18055555,,
1111
UBERUSERS TAKE
RIDES
1389,
1944,
SERVES UP
GIFS
12986111,,
PROFESSIONALS
GOOGLECONDUCTS
SEARCHES
3877140, ,
,,
, ,
REDDIT MINUTE
every
DAY
of the
PRESENTED BY DOMO
2018
,
39Copyright 2019 by Data Blueprint Slide #
https://www.domo.com/learn/data-never-sleeps-6
How much Data,
by the minute!
For the entirety of 2018, every minute
of every day:
• 18 million weather forecast requests
• Netflix streams almost 100,000
hours of video
• LinkedIn adds 120+ individuals
• 1,300 Uber rides
• (almost) a half million tweets
• 7,000 Tinder matches
• 1.25 new cryptocurrencies are
created
• ...
Great inspiration towards valuation ...
• How to Measure Anything: Finding the Value of
Intangibles in Business by Douglas Hubbard (ISBN: 0470539399)
• Measurement is a reduction in uncertainty
• Formalizing stuff forces clarity
• Whatever your measurement problem is,
– it's been done before
• You have more data than you think
• You need less data than you think
• Getting data is more economical than you think
• You probably need different data than you think
• Special shout out to Chapter 7
– Measuring the value of additional information to a decision
40Copyright 2019 by Data Blueprint Slide #
Sheena's in color Activity-Based Costing Kills Someone
41Copyright 2019 by Data Blueprint Slide #
Enrico Fermi (Nobel Prize Physics 1938)
42Copyright 2019 by Data Blueprint Slide #
• Tuners in Chicago ≈ Population/people per household
times % households with tuned pianos
times tunings per year
divided by (tunings per tuner per day
times workdays/year)
• How many piano tuners in the city of Chicago?
– Without using existing lists such as yellow pages, google ...
– Current population of Chicago (3 million at the time)
– Average number of people per household (2 or 3)
– Share of households with regularly tuned pianos (1 in 3)
– Required frequency of tuning (1/year)
– How many pianos can a tuner tune daily? (4 or 5)
– How many days/year are worked (250)
8. Monitization: Time & Leave Tracking
At Least 300 employees are
spending 15 minutes/week
tracking leave/time
43Copyright 2019 by Data Blueprint Slide #
Capture Cost of Labor/Category
44Copyright 2019 by Data Blueprint Slide #
District-L (as an example) Leave Tracking Time Accounting
Employees 73 50
Number of documents 1000 2040
Timesheet/employee 13.7 40.8
Time spent 0.08 0.25
Hourly Cost $6.92 $6.92
Additive Rate $11.23 $11.23
Cost per timekeeper $12.31 $114.56
Total timekeeper cost $898.49 $5,727.89
Monthly cost $21,563.83 $137,469.40
Compute Labor Costs
45Copyright 2019 by Data Blueprint Slide #
Annual Organizational Totals
• $100,000 Salem
• $159,000 Lynchburg
• $100,000 Richmond
• $100,000 Suffolk
• $150,000 Fredericksburg
• $100,000 Staunton
• $100,000 NOVA
• $800,000/month or $9,600,000/annually
• Awareness of the cost of things considered overhead
46Copyright 2019 by Data Blueprint Slide #
47Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
48Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
10. Version 3
55Copyright 2019 by Data Blueprint Slide #
Data
Strategy
Data
Governance
BI/
Warehouse
Reference &
Master Data
Perfecting
operations in 3
data
management
practice areas
56Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
Improve Operations
Innovation
Data quality focus should be sequenced
57Copyright 2019 by Data Blueprint Slide # 58Copyright 2019 by Data Blueprint Slide #
59Copyright 2019 by Data Blueprint Slide #
Ubiquitous Mystery Object
60Copyright 2019 by Data Blueprint Slide #
11. Complex Data Quality Problems
• Agency manages (4,000,000 data items)
– Executive in charge requested
a conversion update
– Was told verbally the conversion was "going well"
– Demanded specifics
• Question: "How many items did you attempt to convert?"
• Answer: "100 items"
• Question: "How many were actually converted?"
• Answer: "5"
• Problems
– Not reporting the "right results"
– These "problems" were discovered too late in the project
– Unsophisticated contractor
61Copyright 2019 by Data Blueprint Slide #
Improving Data Quality during System Migration
• Challenge
– Millions of NSN/SKUs
maintained in a catalog
– Key and other data stored in
clear text/comment fields
– Original suggestion was manual
approach to text extraction
– Left the data structuring problem unsolved
• Solution
– Proprietary, improvable text extraction process
– Converted non-tabular data into tabular data
– Saved a minimum of $5 million
– Literally person centuries of work
62Copyright 2019 by Data Blueprint Slide #
Unmatched
Items
Ignorable
Items
Items
Matched
Week # (% Total) (% Total) (% Total)
1 31.47% 1.34% N/A
2 21.22% 6.97% N/A
3 20.66% 7.49% N/A
4 32.48% 11.99% 55.53%
… … … …
14 9.02% 22.62% 68.36%
15 9.06% 22.62% 68.33%
16 9.53% 22.62% 67.85%
17 9.5% 22.62% 67.88%
18 7.46% 22.62% 69.92%
Determining Diminishing Returns
63Copyright 2019 by Data Blueprint Slide #
Before
After
Time needed to review all NSNs once over the life of the project:
NSNs 2,000,000
Average time to review & cleanse (in minutes) 5
Total Time (in minutes) 10,000,000
Time available per resource over a one year period of time:
Work weeks in a year 48
Work days in a week 5
Work hours in a day 7.5
Work minutes in a day 450
Total Work minutes/year 108,000
Person years required to cleanse each NSN once prior to migration:
Minutes needed 10,000,000
Minutes available person/year 108,000
Total Person-Years 92.6
Resource Cost to cleanse NSN's prior to migration:
Avg Salary for SME year (not including overhead) $60,000.00
Projected Years Required to Cleanse/Total DLA Person Year Saved 93
Total Cost to Cleanse/Total DLA Savings to Cleanse NSN's: $5.5 million
Quantitative Benefits
64Copyright 2019 by Data Blueprint Slide #
Time needed to review all NSNs once over the life of the project:
NSNs 2,000,000
Average time to review & cleanse (in minutes) 5
Total Time (in minutes) 10,000,000
Time available per resource over a one year period of time:
Work weeks in a year 48
Work days in a week 5
Work hours in a day 7.5
Work minutes in a day 450
Total Work minutes/year 108,000
Person years required to cleanse each NSN once prior to migration:
Minutes needed 10,000,000
Minutes available person/year 108,000
Total Person-Years 92.6
Resource Cost to cleanse NSN's prior to migration:
Avg Salary for SME year (not including overhead) $60,000.00
Projected Years Required to Cleanse/Total DLA Person Year Saved 93
Total Cost to Cleanse/Total DLA Savings to Cleanse NSN's: $5.5 million
Quantitative Benefits
65Copyright 2019 by Data Blueprint Slide #
Time needed to review all NSNs once over the life of the project:
NSNs 150,000
Average time to review & cleanse (in minutes) 5
Total Time (in minutes) 750,000
Time available per resource over a one year period of time:
Work weeks in a year 48
Work days in a week 5
Work hours in a day 7.5
Work minutes in a day 450
Total Work minutes/year 108,000
Person years required to cleanse each NSN once prior to migration:
Minutes needed 750,000
Minutes available person/year 108,000
Total Person-Years 7
Resource Cost to cleanse NSN's prior to migration:
Avg Salary for SME year (not including overhead) $60,000.00
Projected Years Required to Cleanse/Total DLA Person Year Saved 7
Total Cost to Cleanse/Total DLA Savings to Cleanse NSN's: $420,000
Time needed to review all NSNs once over the life of the project:
NSNs 2,000,000
Average time to review & cleanse (in minutes) 5
Total Time (in minutes) 10,000,000
Time available per resource over a one year period of time:
Work weeks in a year 48
Work days in a week 5
Work hours in a day 7.5
Work minutes in a day 450
Total Work minutes/year 108,000
Person years required to cleanse each NSN once prior to migration:
Minutes needed 10,000,000
Minutes available person/year 108,000
Total Person-Years 92.6
Resource Cost to cleanse NSN's prior to migration:
Avg Salary for SME year (not including overhead) $60,000.00
Projected Years Required to Cleanse/Total DLA Person Year Saved 93
Total Cost to Cleanse/Total DLA Savings to Cleanse NSN's: $5.5 million
Quantitative Benefits
66Copyright 2019 by Data Blueprint Slide #
12. Year 2000 (or Y2K) Bug
67Copyright 2019 by Data Blueprint Slide #
• Before the internet
– Computing resources were expensive
– It was worth the tradeoff to represent the year
field using two digits
– 1959 was represented to the computer as 59
– Subtracting 59 from 99 yields the correct answer
40 (for dates prior to 2000/01/01!)
– No one expected those programs to still be in use
– Documentation was poorly created/maintained
• If all these fields were not expanded to
four digits before 2000/01/01 then date
calculations will not give correct results
– Subtracting 59 from 00 yields the incorrect
answer -41
– No one knew how long this would take or
cost–only when it must be completed!
On the OFFICIAL Clock of the United States at 1
second BEFORE Midnight showed:
December 31, 1999 11:59:59
One SECOND later the OFFICIAL Clock of the United
States showed:
January 1, 19100 00:00:01
• with a PhD in Chemical Engineering
• have to know whether this product was
Y2K compliant?
Why should a knowledge worker
68Copyright 2019 by Data Blueprint Slide #
International Chemical Company Engine Testing
69Copyright 2019 by Data Blueprint Slide #
• $1billion (+) chemical
company
• Develops/manufactures
additives enhancing the
performance of oils and
fuels ...
• ... to enhance engine/
machine performance
– Helps fuels burn cleaner
– Engines run smoother
– Machines last longer
• Tens of thousands of
tests annually
– Test costs range up to
$250,000!
1.Manual transfer of digital data
2.Manual file movement/duplication
3.Manual data manipulation
4.Disparate synonym reconciliation
5.Tribal knowledge requirements
6.Non-sustainable technology
70Copyright 2019 by Data Blueprint Slide #
Data Integration Solution
71Copyright 2019 by Data Blueprint Slide #
• Integrated the existing systems to
easily search on and find similar
or identical tests
• Results:
– Reduced expenses
– Improved competitive edge
and customer service
– Time savings and improve
operational capabilities
• According to our client’s internal
business case development, they
expect to realize a $25 million
gain each year thanks to this
data integration
Lockheed Martin
• 20 years of project email
– Example from Doug Laney
72Copyright 2019 by Data Blueprint Slide #
13. Logistics Company
• Fortune 450
• Room of 100 associates
• Manually correcting every
item on every customer invoice
• Upon noting this to the
responsible manager - the reply was:
– This is the best quarter
– Of the best year
– I've ever had
– Perhaps I need
to double the
number in
that room?
73Copyright 2019 by Data Blueprint Slide # 74Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
US DoD Reverse Engineering Program Manager
• "Your first project is to keep me from
having to testify to a Congressional
Hearing!" (Belkis Leon-Hong former ASD-C3I)
• Problem:
– 37 systems paid personnel within DoD
– How many were needed?
– How many potential losers?
– What do you mean by employee?
• Process modeling
– Inconclusive results
• Data reverse engineering - definitive
– One legged engineer,
working in waist deep waters,
underneath rotating helicopter blades,
on overtime
75Copyright 2019 by Data Blueprint Slide #
Reverse Engineering New Systems
76Copyright 2019 by Data Blueprint Slide #
Reverse Engineering New Systems for Smooth Implementation. IEEE Software. March/April 1999 16(2):36-43
Platform: UniSys
OS: OS
1998 Age: 21
Data Structure: DMS (Network)
Physical Records: 4,950,000
Logical Records: 250,000
Relationships: 62
Entities: 57
Attributes: 1478
Predicting Engineering Problem Characteristics
New System
Legacy System
#1: Payroll
Legacy System
#2: Personnel
Platform: Amdahl
OS: MVS
1998 Age: 15
Data Structure: VSAM/virtual
database tables
Physical Records: 780,000
Logical Records: 60,000
Relationships: 64
Entities: 4/350
Attributes: 683
Characteristics Logical Physical
Platform: WinTel Records: 250,000 600,000
OS: Win'95 Relationships: 1,034 1,020
1998 Age: new Entities: 1,600 2,706
Data Structure: Client/Sever RDBMS Attributes: 15,000 7,073
77Copyright 2019 by Data Blueprint Slide #
TheBudgetTrap(Parts1&2)
78Copyright 2019 by Data Blueprint Slide #
14. Actual Bid From Systems Integrator
79Copyright 2019 by Data Blueprint Slide #
Extreme Data Engineers ...
2 person months = 40 person days
2,000 attributes mapped onto 15,000
2,000/40 person days = 500/person day
or 500/8 hours = 62.5 attributes/hour
and
15,000/40 person days = 375/person day
or 375/8 hours = 46.875 attributes/hour
Locate, identify, understand, map, transform, document
108 attributes/60 minutes
1.8 attributes/minute!
80Copyright 2019 by Data Blueprint Slide #
What did Rolls Royce Learn
• Old model
– Sell jet engines
• New model
– Sell hours of thrust power
– Power-by-the-hour
– No payment for down time
– Wing to wing
– When was it invented?
from Nascar?
81Copyright 2019 by Data Blueprint Slide #
Fan Blade Sensor
82Copyright 2019 by Data Blueprint Slide #
• 1 Sensor
– Probabilistic (generalist) maintenance
forecasts
• 100 Sensors
– Establish optimal monitoring targets
– Finer tuned and safer maintenance
– Mission Readiness ???
– Storage $$$
– Handling $$$
– Opportunity $$$
– Systemic $$$
– Maintenance $$$
– Total > $1.5 Billion
83Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
Armed Force Example
• Lieutenant attempting to
correct a 4 year
underpayment
of his private's pay
– Significant impact on moral
– Immediate cash issues
– Cost tens of man hours over
months of time to resolve
84Copyright 2019 by Data Blueprint Slide #
Nugee, R. and R. S. Seiner (2010, 6/1/2010). "TDAN.com Interview with Brigadier Richard Nugee – The British Army." 2013, from http://www.tdan.com/view-special- features/13897 and personal communications.
15. Friendly Fire deaths traced to Dead Battery
• Date: Tue, 26 Mar 2002 10:47:52 -0500
From:
Subject: Friendly Fire deaths traced to dead battery
In one of the more horrifying incidents I've read about, U.S. soldiers and
allies were killed in December 2001 because of a stunningly poor design of a
GPS receiver, plus "human error."
http://www.washingtonpost.com/wp-dyn/articles/A8853-2002Mar23.html
A U.S. Special Forces air controller was calling in GPS positioning from
some sort of battery-powered device. He "had used the GPS receiver to
calculate the latitude and longitude of the Taliban position in minutes and
seconds for an airstrike by a Navy F/A-18."
• According to the *Post* story, the bomber crew "required" a "second
calculation in 'degree decimals'" -- why the crew did not have equipment to
perform the minutes-seconds conversion themselves is not explained.
• The air controller had recorded the correct value in the GPS receiver when
the battery died. Upon replacing the battery, he called in the
degree-decimal position the unit was showing -- without realizing that the
unit is set up to reset to its *own* position when the battery is replaced.
The 2,000-pound bomb landed on his position, killing three Special Forces
soldiers and injuring 20 others.
• If the information in this story is accurate, the RISKS involve replacing
memory settings with an apparently-valid default value instead of blinking 0
or some other obviously-wrong display; not having a backup battery to hold
values in memory during battery replacement; not equipping users to
translate one coordinate system to another (reminiscent of the Mars Climate
Orbiter slamming into the planet when ground crews confused English with
metric); and using a device with such flaws in a combat situation
85Copyright 2019 by Data Blueprint Slide #
Formalizing the
Role of U.S. Army
Data Governance
86Copyright 2019 by Data Blueprint Slide #
How one inventory item proliferates data throughout the chain
555 Subassemblies & subcomponents
17,659 Repair parts or Consumables
System 1:
18,214 Total items
75 Attributes/item
1,366,050 Total attributes
System 2
47 Total items
15+ Attributes/item
720 Total attributes
System 3
16,594 Total items
73 Attributes/item
1,211,362 Total attributes
System 4
8,535 Total items
16 Attributes/item
136,560 Total attributes
System 5
15,959 Total items
22 Attributes/item
351,098 Total attributes
Total for the five systems show above:
59,350 Items
179 Unique attributes
3,065,790 values
87Copyright 2019 by Data Blueprint Slide # 88Copyright 2019 by Data Blueprint Slide #
Business Implications
• National Stock Number (NSN)
Discrepancies
– If NSNs in LUAF, GABF, and RTLS are
not present in the MHIF, these records
cannot be updated in SASSY
– Additional overhead is created to correct
data before performing the real
maintenance of records
• Serial Number Duplication
– If multiple items are assigned the same
serial number in RTLS, the traceability of
those items is severely impacted
– Approximately $531 million of SAC 3
items have duplicated serial numbers
• On-Hand Quantity Discrepancies
– If the LUAF O/H QTY and number of items serialized in RTLS conflict, there
can be no clear answer as to how many items a unit actually has on-hand
– Approximately $5 billion of equipment does not tie out between the systems
89Copyright 2019 by Data Blueprint Slide #
Best approaches combines manual and automation
Humans Generally Better Machines Generally Better
• Sense low level stimuli
• Detect stimuli in noisy background
• Recognize constant patterns in varying situations
• Sense unusual and unexpected events
• Remember principles and strategies
• Retrieve pertinent details without a priori
connection
• Draw upon experience and adapt decision to
situation
• Select alternatives if original approach fails
• Reason inductively; generalize from observations
• Act in unanticipated emergencies and novel
situations
• Apply principles to solve varied problems
• Make subjective evaluations
• Develop new solutions
• Concentrate on important tasks when overload
occurs
• Adapt physical response to changes in situation
• Sense stimuli outside human's range
• Count or measure physical quantities
• Store quantities of coded information accurately
• Monitor prespecified events, especially infrequent
• Make rapid and consisted responses to input
signals
• Recall quantities of detailed information
accurately
• Retrieve pertinent detailed without a priori
connection
• Process quantitative data in prespecified ways
• Perform repetitive preprogrammed actions
reliably
• Exert great, highly controlled physical force
• Perform several activities simultaneously
• Maintain operations under heavy operation load
• Maintain performance over extended periods of
time
90Copyright 2019 by Data Blueprint Slide #
16. 91Copyright 2019 by Data Blueprint Slide #
Potential Data Sources
92Copyright 2019 by Data Blueprint Slide #
Data Mapping
12
Mental
illness
Deploy
ments
Work
History
Soldier Legal
Issues
Abuse
Suicide
Analysis
FAPDMSS G1 DMDC CID
Data objects
complete?
All sources
identified?
Best source for
each object?
How reconcile
differences
between
sources?
MDR
93Copyright 2019 by Data Blueprint Slide # 94Copyright 2019 by Data Blueprint Slide #
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
95Copyright 2019 by Data Blueprint Slide #
Senior Army Official
• Room full of Stewards
• A very heavy dose of management support
• Advised the group of his opinion on the matter
• Any questions as to future direction
– "They should make an appointment to speak directly with
me!"
• Empower the team
– The conversation turned from "can this be done?" to "how are we going
to accomplish this?"
– Mistakes along the way would be tolerated
– Implement a workable solution in prototype form
96Copyright 2019 by Data Blueprint Slide #
17. 97Copyright 2019 by Data Blueprint Slide #
Managing Data with Guidance?
• Federal employees
• 44 users from whitehouse.gov
• Thousands of military and
government e-mails
• Canadian citizens
• One-fifth of Quebec
98Copyright 2019 by Data Blueprint Slide #
Ashley
Madison
37,000,000
25,000,000
OPM
70,000,000
Target
99Copyright 2019 by Data Blueprint Slide #
Target Corporation's Database Contents
100Copyright 2019 by Data Blueprint Slide #
• Your age
• Marital status
• Part of town you live in
• How long it takes you to drive
to work
• Estimated salary
• If you have recently moved
• Credit cards carried in your
wallet
• What websites you visit
• Your ethnicity
• Your job history
• The magazines you read
• Work commute
• Sexual preferences
• If you’ve ever declared
bankruptcy or got divorced
• The year you bought (or lost)
your house
• Where you went to school(s)
• What kinds of topics you talk
about online
• Whether you prefer certain
brands of coffee, paper
towels, cereal or applesauce
• Your political leanings,
reading habits, charitable
giving and
• The number of cars you own
101Copyright 2019 by Data Blueprint Slide #
https://oversight.house.gov/report/opm-data-breach-government-jeopardized-national-security-generation/
How the Government Jeopardized Our National
Security for More than a Generation
• Preventable
• Leadership failed
– To heed repeated
recommendations
– To sufficiently respond
to growing threats of
sophisticated cyber
attacks, and
– To prioritize resources
for cybersecurity
• 2014 data breaches
were likely connected
and possibly
coordinated to the 2015
data breach
• OPM misled the public
on the extent of the
damage of the breach
and made false
statements to Congress
Key Findings
102Copyright 2019 by Data Blueprint Slide #
Data Quality Success Stories - Program Overview
1. Data quality must be understood as
an engineering challenge
2. Putting a price on data quality
3. DM BoK components compliment
each other well
4. Savings based stories
5. Innovation based stories
6. Non-monetary stories
7. Takeaways and Q&A
18. • Quality data requires a context specific definition
• Most business problems have data challenges (hidden data
factories) at their root
• All advanced data practices depend on quality data
• AI/ML are suffering from lack of training data
• Few 'easy' fixes exist
• Data quality engineering works well when combined with other DM
BoK 'pie wedges'
• Successful data quality stories demonstrate
– Tangible ongoing savings
– Innovative data uses
– Outcomes more important than money
Take Aways
103Copyright 2019 by Data Blueprint Slide #
+ =
104Copyright 2019 by Data Blueprint Slide #
Questions?
10124 W. Broad Street, Suite C
Glen Allen, Virginia 23060
804.521.4056
Copyright 2019 by Data Blueprint Slide #
105