This document is a chapter from a book about data quality and data governance. The chapter discusses how poor or incorrect data can lead to significant financial losses for businesses. It provides examples where small data errors resulted in large consequences, such as a $327 million satellite project failure due to a mix-up between metric and imperial units. The chapter emphasizes that maintaining high quality data in areas like procurement can help businesses reduce costs by 10% or more by enabling strategic sourcing. It also notes that identifying and resolving duplicate records is important but complex.
The Heart of Data Modeling Webinar: State of the Union Data ModelingDATAVERSITY
Welcome to the first webinar in the new DATAVERSITY Webinar Series: The Heart of Data Modeling with Karen Lopez. Each month Karen will provide education in a specific topic on Data Modeling. Come one, come all the Fourth Thursday of each month to discuss and learn how to truly love your data.
This month Karen will kick off the series and the year with a "State of the Union of Data Modeling" address. What happened in 2014 and will trend through 2015? What's happening with Data Modeling tools? What's trending in schema and schemaless designs? NoSQL databases? How is Big Data continuing to impact the world of Data Models?
DAMA Webinar: Taking Information Governance to the Next LevelDATAVERSITY
Our society is rapidly evolving towards being information driven. Pretty much every industry and line of business is confronted with increasing amounts of data and a push towards better decision taking. The domain of data governance has transformed into a much broader field of expertise that is about much more than being an honest broker focusing on better definitions and clarify ownership. The rapid adoption of analytics, increasing sets of data sources, HOV (harder to obtain value) data and compliance needs has resulted in mushrooming of often unrelated initiatives. The next generation of Information Governance needs to embrace these developments and use the domain experience to allow the convergence of the disparate initiatives and to use the energy and enable information centric organizations.
•Setting the scene: Defining ‘Information Centric Organization’
•Positioning Information Governance in a CDO context: Information innovation as a driver
•Drowning in the data lake or having a breach? Data Governance as a safeguard
•Defining and selling the information strategy
So many of us have learned data modeling and database design approaches from working with one database or data technology. We may have used only one data modeling tool. That means our vocabularies around identifiers and keys tend to be product specfic. Do you know the difference between a unique index and a unique key? What about the difference between RI, FK and AK?
In this webinar we'll look at the generic and proprietary terms for these concepts, as well as where they fit in the data modeling and database design process. We'll also look at implementation options across a few commercial DBMSs and datastores.
Bring your developers and DBAs, too. These concepts span data activities and it's important that your team understand each other and where they, their tools and approaches need to support these features, too.
This presentation provides you with an understanding of reference and master data management (MDM) goals, including establishing and implementing authoritative data sources, establishing and implementing more effective means of delivering data to various business processes, and increasing the quality of information used in organizational analytical functions (such as BI). Attendees will learn how to incorporate data quality engineering into the planning of reference and MDM. Finally, we will discuss why MDM is so critical to the organization’s overall data strategy.
Takeaways:
•What is reference and MDM?
•Why are reference and MDM important?
•How to use Reference and MDM Frameworks
•Guiding principles & best practices for MDM
The Heart of Data Modeling: 7 Ways Your Agile Project is Managing Data WrongDATAVERSITY
Is your organization using agile approaches to systems development project? Have you found that there are conflicting opinions with what should be done, when it should be done and who should do it? Is there even a suggestion that data modeling isn’t needed on an Agile project? Are your data architects stuck in a waterfall world? Are you asking for “no more changes” to the data model? Do your developers thing that “just the right documentation” means no modeling allowed? Does anyone even know where the reference data for the application is located? Or how it is updated?
In this month’s webinar, Karen will show you how data modeling and Agile approaches CAN work together to deliver quality information systems and solutions, with fewer dysfunctions and less tears.
Data-Ed Webinar: Data Quality Success StoriesDATAVERSITY
Organizations must realize what it means to utilize data quality management in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Takeaways:
•Understanding foundational data quality concepts based on the DAMA DMBOK
•Utilizing data quality engineering in support of business strategy
•Case Studies illustrating data quality success
•Data Quality guiding principles & best practices
•Steps for improving data quality at your organization
The Heart of Data Modeling: The Best Data Modeler is a Lazy Data ModelerDATAVERSITY
We're under pressure to do more with fewer resources. And organizations are often short on experienced data modelers. So why should we spend time doing things that can be done by robots. Well, not robots, but automation.
In this month's webinar, Karen demonstrates the types of automation techniques available in leading data modeling tools such as ERwin, ER/Studio and PowerDesigner. She will also leave you with 10 tips on being more lazy. What webinar last promised that, anyway?
The Heart of Data Modeling Webinar: State of the Union Data ModelingDATAVERSITY
Welcome to the first webinar in the new DATAVERSITY Webinar Series: The Heart of Data Modeling with Karen Lopez. Each month Karen will provide education in a specific topic on Data Modeling. Come one, come all the Fourth Thursday of each month to discuss and learn how to truly love your data.
This month Karen will kick off the series and the year with a "State of the Union of Data Modeling" address. What happened in 2014 and will trend through 2015? What's happening with Data Modeling tools? What's trending in schema and schemaless designs? NoSQL databases? How is Big Data continuing to impact the world of Data Models?
DAMA Webinar: Taking Information Governance to the Next LevelDATAVERSITY
Our society is rapidly evolving towards being information driven. Pretty much every industry and line of business is confronted with increasing amounts of data and a push towards better decision taking. The domain of data governance has transformed into a much broader field of expertise that is about much more than being an honest broker focusing on better definitions and clarify ownership. The rapid adoption of analytics, increasing sets of data sources, HOV (harder to obtain value) data and compliance needs has resulted in mushrooming of often unrelated initiatives. The next generation of Information Governance needs to embrace these developments and use the domain experience to allow the convergence of the disparate initiatives and to use the energy and enable information centric organizations.
•Setting the scene: Defining ‘Information Centric Organization’
•Positioning Information Governance in a CDO context: Information innovation as a driver
•Drowning in the data lake or having a breach? Data Governance as a safeguard
•Defining and selling the information strategy
So many of us have learned data modeling and database design approaches from working with one database or data technology. We may have used only one data modeling tool. That means our vocabularies around identifiers and keys tend to be product specfic. Do you know the difference between a unique index and a unique key? What about the difference between RI, FK and AK?
In this webinar we'll look at the generic and proprietary terms for these concepts, as well as where they fit in the data modeling and database design process. We'll also look at implementation options across a few commercial DBMSs and datastores.
Bring your developers and DBAs, too. These concepts span data activities and it's important that your team understand each other and where they, their tools and approaches need to support these features, too.
This presentation provides you with an understanding of reference and master data management (MDM) goals, including establishing and implementing authoritative data sources, establishing and implementing more effective means of delivering data to various business processes, and increasing the quality of information used in organizational analytical functions (such as BI). Attendees will learn how to incorporate data quality engineering into the planning of reference and MDM. Finally, we will discuss why MDM is so critical to the organization’s overall data strategy.
Takeaways:
•What is reference and MDM?
•Why are reference and MDM important?
•How to use Reference and MDM Frameworks
•Guiding principles & best practices for MDM
The Heart of Data Modeling: 7 Ways Your Agile Project is Managing Data WrongDATAVERSITY
Is your organization using agile approaches to systems development project? Have you found that there are conflicting opinions with what should be done, when it should be done and who should do it? Is there even a suggestion that data modeling isn’t needed on an Agile project? Are your data architects stuck in a waterfall world? Are you asking for “no more changes” to the data model? Do your developers thing that “just the right documentation” means no modeling allowed? Does anyone even know where the reference data for the application is located? Or how it is updated?
In this month’s webinar, Karen will show you how data modeling and Agile approaches CAN work together to deliver quality information systems and solutions, with fewer dysfunctions and less tears.
Data-Ed Webinar: Data Quality Success StoriesDATAVERSITY
Organizations must realize what it means to utilize data quality management in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Takeaways:
•Understanding foundational data quality concepts based on the DAMA DMBOK
•Utilizing data quality engineering in support of business strategy
•Case Studies illustrating data quality success
•Data Quality guiding principles & best practices
•Steps for improving data quality at your organization
The Heart of Data Modeling: The Best Data Modeler is a Lazy Data ModelerDATAVERSITY
We're under pressure to do more with fewer resources. And organizations are often short on experienced data modelers. So why should we spend time doing things that can be done by robots. Well, not robots, but automation.
In this month's webinar, Karen demonstrates the types of automation techniques available in leading data modeling tools such as ERwin, ER/Studio and PowerDesigner. She will also leave you with 10 tips on being more lazy. What webinar last promised that, anyway?
Data-Ed Webinar: Data Governance StrategiesDATAVERSITY
Data governance exercises authority and control over the management of your mission critical assets and guides how all other data management functions are performed. When selling data governance to organizational management, it is useful to concentrate on the specifics that motivate the initiative. This means developing a specific vocabulary and set of narratives to facilitate understanding of the business objectives and imperatives that demand governance. This webinar also provides you with an understanding of what data governance functions are required and how they fit with other data management disciplines. Understanding these governance aspects is necessary to eliminate the ambiguity that often surrounds effective data governance and stewardship programs. The goal of governance is to manage the data that supports organizational strategy.
Takeaways:
•Understanding why data governance can be tricky for most organizations
•Steps for improving data governance within your organization
•Guiding principles & lessons learned
•Understanding foundational data governance concepts based on the DAMA DMBOK
Data-Ed Online Webinar: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is data architecture that organizes data assets so they can be used in your business strategy to create real business value. Even though this is important, data architectures are still being used ineffectively. The various uses of data architecture are referenced to inform, clarify, understand, and resolve aspects of a variety of business problems commonly encountered in organizations. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecture to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will maximize your organization’s competitive advantage.
Takeaways:
•How to utilize data architecture to address a broad variety of organizational challenges and how to utilize data architectures in support of business strategy
•Understanding foundational data architecture concepts based on the DAMA DMBOK
•Data architecture guiding principles & best practices
Data warehousing Demo PPTS | Over View | Introduction Kernel Training
Module 1:
Introduction to Data Warehouse & Business Intelligence
Module 2: Data Warehouse Architecture
Module 3: Warehouse: D & F – Dimension & Fact Tables
Module 4: Data Modeling
Module 5: Building Data Warehouse with ER Win
Module 6: Introduction to Open Source ETL Tool – Talend DI Open Studio 5.x
Module 7: Building ETL Project with Talend DI Open Studio 5.x
Module 8: Introduction to Data Visualization BI Tool – Tableau 9.x
Module 9: Building Data Visualization BI Project With Tableau 9.x
Module 10: An Integrated Data Ware Housing & BI Project
Which Case-Studies will be a part of data warehousing and business intelligence online training?
Learn Data warehousing and business intelligence online training by real time expert. Be a part of live sessions. Data warehousing and business intelligence classes by Expert.
Data is the lifeblood of just about every organization and functional area today. As businesses struggle to cope with the data flood, it is even more critical to focus on data as an asset that directly supports business imperatives. Organizations across most industries attempt to address data opportunities (e.g. Big Data) and data challenges (e.g. data quality) to enhance business unit performance. Unfortunately, the results of these efforts frequently fall far below expectations due to haphazard approaches. Overall, poor organizational data management capabilities are the root cause of many of these failures. This webinar covers three lessons (illustrated by examples), which will help you to establish realistic expectations, and help demonstrate the value of this process to both internal and external decision makers.
The pre-conference workshop entitled 'Trust is a Terrible Thing to Waste' from the 2010 International Association of Privacy Professionals conference in Washington, D.C. The session reviewed why trust is important, how to handle crisis communications, and how to build trust before a crisis hits.
Data privacy awareness is on the rise. Users become more and more concerned with how online service providers collect and protect their personal information. And so should you. Discover how to balance the risks and benefits of collecting data in the age of customer centricity.
Chapter 12REVENUE AND INVENTORY-RELATED FINANCIAL STATEMENT .docxcravennichole326
Chapter 12
REVENUE AND INVENTORY-
RELATED FINANCIAL STATEMENT FRAUDS
Short Cases
Case 3
After performing the analysis, the accounts that raise questions are Accounts Receivable and Retained Earnings. The percentage changes are shown in the following table:
2007
2008
Percent
Change
2009
Percent
Change
2010
Percent
Change
Cash
$1,000
$1,200
20.0%
$1,400
16.7%
$1,500
7.1%
A/R
250
375
50.0%
600
60.0%
900
50.0%
Inventory
600
700
16.7%
825
17.9%
975
18.2%
PP&E (net)
1,500
1,700
13.3%
1,800
5.9%
1,950
8.3%
Notes Rec.
500
500
0.0%
500
0.0%
500
0.0%
Tot. Assets
$3,850
$4,475
16.2%
$5,125
14.5%
$5,825
13.7%
A/P
$ 700
$ 900
28.6%
$1,000
11.1%
$1,100
10.0%
Other Liab.
200
300
50.0%
350
16.7%
425
21.4%
Notes Pay.
1,200
1,400
16.7%
1,500
7.1%
1,750
16.7%
Tot. Liab.
$2,100
$2,600
23.8%
$2,850
9.6%
$3,275
14.9%
Stock Out.
$1,000
$1,000
0.0%
$1,000
0.0%
$1,000
0.0%
Ret. Earn.
750
875
16.7%
1,275
45.7%
1,550
21.6%
Total Share. Equity
$1,750
$1,875
7.1%
$2,275
21.3%
$2,550
12.1%
Total Liab./
Share. Equ.
$3,850
$4,475
16.2%
$5,125
14.5%
$5,825
13.7%
The Accounts Receivables balances are most questionable because of the huge jump from year to year. Those balances would be even more suspicious if there had been a drop in business incurred by most companies in the technology sector.
Case 6
The following are symptoms and schemes used in this case:
1. Funneling bank loans through third parties to make it look as though customers had paid when they had not.
2. Deliberately providing “false or incomplete information” to auditors and conspiring to obstruct the firm’s audits.
3. Factoring unpaid receivables to banks to obtain up-front cash. Side letters that were concealed from the auditors gave the banks the right to take the money back if they could not collect from the company’s customers.
4. The bulk of the company’s sales came from contracts signed at the end of quarters, so managers could meet ambitious quarterly sales targets and receive multimillion-dollar bonuses.
One of the first questions that needs to be examined is how the company explained the sudden growth in sales from hundreds to millions. The auditors should have investigated this increase. Also, the fact that most sales were recorded in the last days of the quarter should have been investigated. The auditors should have concluded that management had strong motivation to commit fraud because executives were being paid high bonuses based on sales volume.
As discussed in this chapter, the analytical symptoms and the accounting or documentary symptoms, if understood and analyzed by auditors, could have lead to earlier fraud discovery.
Case 7
1. One way to search for possible red flags of fraud would be to determine if the market value of inventories is higher or lower than reported inventory amounts. Doing this reveals the following:
Reported 2010
Market Value
Finished goods inventory
$1,654,500
$2,400,000
(300 million × $.08)
(Approx. 300 million feet—2010)
...
Data-Ed Webinar: Data Governance StrategiesDATAVERSITY
Data governance exercises authority and control over the management of your mission critical assets and guides how all other data management functions are performed. When selling data governance to organizational management, it is useful to concentrate on the specifics that motivate the initiative. This means developing a specific vocabulary and set of narratives to facilitate understanding of the business objectives and imperatives that demand governance. This webinar also provides you with an understanding of what data governance functions are required and how they fit with other data management disciplines. Understanding these governance aspects is necessary to eliminate the ambiguity that often surrounds effective data governance and stewardship programs. The goal of governance is to manage the data that supports organizational strategy.
Takeaways:
•Understanding why data governance can be tricky for most organizations
•Steps for improving data governance within your organization
•Guiding principles & lessons learned
•Understanding foundational data governance concepts based on the DAMA DMBOK
Data-Ed Online Webinar: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is data architecture that organizes data assets so they can be used in your business strategy to create real business value. Even though this is important, data architectures are still being used ineffectively. The various uses of data architecture are referenced to inform, clarify, understand, and resolve aspects of a variety of business problems commonly encountered in organizations. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecture to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will maximize your organization’s competitive advantage.
Takeaways:
•How to utilize data architecture to address a broad variety of organizational challenges and how to utilize data architectures in support of business strategy
•Understanding foundational data architecture concepts based on the DAMA DMBOK
•Data architecture guiding principles & best practices
Data warehousing Demo PPTS | Over View | Introduction Kernel Training
Module 1:
Introduction to Data Warehouse & Business Intelligence
Module 2: Data Warehouse Architecture
Module 3: Warehouse: D & F – Dimension & Fact Tables
Module 4: Data Modeling
Module 5: Building Data Warehouse with ER Win
Module 6: Introduction to Open Source ETL Tool – Talend DI Open Studio 5.x
Module 7: Building ETL Project with Talend DI Open Studio 5.x
Module 8: Introduction to Data Visualization BI Tool – Tableau 9.x
Module 9: Building Data Visualization BI Project With Tableau 9.x
Module 10: An Integrated Data Ware Housing & BI Project
Which Case-Studies will be a part of data warehousing and business intelligence online training?
Learn Data warehousing and business intelligence online training by real time expert. Be a part of live sessions. Data warehousing and business intelligence classes by Expert.
Data is the lifeblood of just about every organization and functional area today. As businesses struggle to cope with the data flood, it is even more critical to focus on data as an asset that directly supports business imperatives. Organizations across most industries attempt to address data opportunities (e.g. Big Data) and data challenges (e.g. data quality) to enhance business unit performance. Unfortunately, the results of these efforts frequently fall far below expectations due to haphazard approaches. Overall, poor organizational data management capabilities are the root cause of many of these failures. This webinar covers three lessons (illustrated by examples), which will help you to establish realistic expectations, and help demonstrate the value of this process to both internal and external decision makers.
The pre-conference workshop entitled 'Trust is a Terrible Thing to Waste' from the 2010 International Association of Privacy Professionals conference in Washington, D.C. The session reviewed why trust is important, how to handle crisis communications, and how to build trust before a crisis hits.
Data privacy awareness is on the rise. Users become more and more concerned with how online service providers collect and protect their personal information. And so should you. Discover how to balance the risks and benefits of collecting data in the age of customer centricity.
Chapter 12REVENUE AND INVENTORY-RELATED FINANCIAL STATEMENT .docxcravennichole326
Chapter 12
REVENUE AND INVENTORY-
RELATED FINANCIAL STATEMENT FRAUDS
Short Cases
Case 3
After performing the analysis, the accounts that raise questions are Accounts Receivable and Retained Earnings. The percentage changes are shown in the following table:
2007
2008
Percent
Change
2009
Percent
Change
2010
Percent
Change
Cash
$1,000
$1,200
20.0%
$1,400
16.7%
$1,500
7.1%
A/R
250
375
50.0%
600
60.0%
900
50.0%
Inventory
600
700
16.7%
825
17.9%
975
18.2%
PP&E (net)
1,500
1,700
13.3%
1,800
5.9%
1,950
8.3%
Notes Rec.
500
500
0.0%
500
0.0%
500
0.0%
Tot. Assets
$3,850
$4,475
16.2%
$5,125
14.5%
$5,825
13.7%
A/P
$ 700
$ 900
28.6%
$1,000
11.1%
$1,100
10.0%
Other Liab.
200
300
50.0%
350
16.7%
425
21.4%
Notes Pay.
1,200
1,400
16.7%
1,500
7.1%
1,750
16.7%
Tot. Liab.
$2,100
$2,600
23.8%
$2,850
9.6%
$3,275
14.9%
Stock Out.
$1,000
$1,000
0.0%
$1,000
0.0%
$1,000
0.0%
Ret. Earn.
750
875
16.7%
1,275
45.7%
1,550
21.6%
Total Share. Equity
$1,750
$1,875
7.1%
$2,275
21.3%
$2,550
12.1%
Total Liab./
Share. Equ.
$3,850
$4,475
16.2%
$5,125
14.5%
$5,825
13.7%
The Accounts Receivables balances are most questionable because of the huge jump from year to year. Those balances would be even more suspicious if there had been a drop in business incurred by most companies in the technology sector.
Case 6
The following are symptoms and schemes used in this case:
1. Funneling bank loans through third parties to make it look as though customers had paid when they had not.
2. Deliberately providing “false or incomplete information” to auditors and conspiring to obstruct the firm’s audits.
3. Factoring unpaid receivables to banks to obtain up-front cash. Side letters that were concealed from the auditors gave the banks the right to take the money back if they could not collect from the company’s customers.
4. The bulk of the company’s sales came from contracts signed at the end of quarters, so managers could meet ambitious quarterly sales targets and receive multimillion-dollar bonuses.
One of the first questions that needs to be examined is how the company explained the sudden growth in sales from hundreds to millions. The auditors should have investigated this increase. Also, the fact that most sales were recorded in the last days of the quarter should have been investigated. The auditors should have concluded that management had strong motivation to commit fraud because executives were being paid high bonuses based on sales volume.
As discussed in this chapter, the analytical symptoms and the accounting or documentary symptoms, if understood and analyzed by auditors, could have lead to earlier fraud discovery.
Case 7
1. One way to search for possible red flags of fraud would be to determine if the market value of inventories is higher or lower than reported inventory amounts. Doing this reveals the following:
Reported 2010
Market Value
Finished goods inventory
$1,654,500
$2,400,000
(300 million × $.08)
(Approx. 300 million feet—2010)
...
In This Issue:
1. Your #1 MUST-DO Resolution For 2017
2. Free Report: What Every Small Business Owner Must Know About Protecting And Preserving their Company’s Critical Data And Computer Systems
3. 3 Ways Smart People Blow The Close
4. STAYING ON TOP
Presentation on the uses & misues of data, embracing illustrations & examples, as presented to the Numis Securities Media Conference in London April 2011
Architecture, Products, and Total Cost of Ownership of the Leading Machine Le...DATAVERSITY
Organizations today need a broad set of enterprise data cloud services with key data functionality to modernize applications and utilize machine learning. They need a comprehensive platform designed to address multi-faceted needs by offering multi-function data management and analytics to solve the enterprise’s most pressing data and analytic challenges in a streamlined fashion.
In this research-based session, I’ll discuss what the components are in multiple modern enterprise analytics stacks (i.e., dedicated compute, storage, data integration, streaming, etc.) and focus on total cost of ownership.
A complete machine learning infrastructure cost for the first modern use case at a midsize to large enterprise will be anywhere from $3 million to $22 million. Get this data point as you take the next steps on your journey into the highest spend and return item for most companies in the next several years.
Data at the Speed of Business with Data Mastering and GovernanceDATAVERSITY
Do you ever wonder how data-driven organizations fuel analytics, improve customer experience, and accelerate business productivity? They are successful by governing and mastering data effectively so they can get trusted data to those who need it faster. Efficient data discovery, mastering and democratization is critical for swiftly linking accurate data with business consumers. When business teams can quickly and easily locate, interpret, trust, and apply data assets to support sound business judgment, it takes less time to see value.
Join data mastering and data governance experts from Informatica—plus a real-world organization empowering trusted data for analytics—for a lively panel discussion. You’ll hear more about how a single cloud-native approach can help global businesses in any economy create more value—faster, more reliably, and with more confidence—by making data management and governance easier to implement.
What is data literacy? Which organizations, and which workers in those organizations, need to be data-literate? There are seemingly hundreds of definitions of data literacy, along with almost as many opinions about how to achieve it.
In a broader perspective, companies must consider whether data literacy is an isolated goal or one component of a broader learning strategy to address skill deficits. How does data literacy compare to other types of skills or “literacy” such as business acumen?
This session will position data literacy in the context of other worker skills as a framework for understanding how and where it fits and how to advocate for its importance.
Building a Data Strategy – Practical Steps for Aligning with Business GoalsDATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task – but it’s worth the effort. Getting your Data Strategy right can provide significant value, as data drives many of the key initiatives in today’s marketplace – from digital transformation, to marketing, to customer centricity, to population health, and more. This webinar will help demystify Data Strategy and its relationship to Data Architecture and will provide concrete, practical ways to get started.
Uncover how your business can save money and find new revenue streams.
Driving profitability is a top priority for companies globally, especially in uncertain economic times. It's imperative that companies reimagine growth strategies and improve process efficiencies to help cut costs and drive revenue – but how?
By leveraging data-driven strategies layered with artificial intelligence, companies can achieve untapped potential and help their businesses save money and drive profitability.
In this webinar, you'll learn:
- How your company can leverage data and AI to reduce spending and costs
- Ways you can monetize data and AI and uncover new growth strategies
- How different companies have implemented these strategies to achieve cost optimization benefits
Data Catalogs Are the Answer – What is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
Data Catalogs Are the Answer – What Is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
In this webinar, Bob will focus on:
-Selecting the appropriate metadata to govern
-The business and technical value of a data catalog
-Building the catalog into people’s routines
-Positioning the data catalog for success
-Questions the data catalog can answer
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, data modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization. This webinar illustrates data modeling as a key activity upon which so much technology and business investment depends.
Specific learning objectives include:
- Understanding what types of challenges require data modeling to be part of the solution
- How automation requires standardization on derivable via data modeling techniques
- Why only a working partnership between data and the business can produce useful outcomes
Analytics play a critical role in supporting strategic business initiatives. Despite the obvious value to analytic professionals of providing the analytics for these initiatives, many executives question the economic return of analytics as well as data lakes, machine learning, master data management, and the like.
Technology professionals need to calculate and present business value in terms business executives can understand. Unfortunately, most IT professionals lack the knowledge required to develop comprehensive cost-benefit analyses and return on investment (ROI) measurements.
This session provides a framework to help technology professionals research, measure, and present the economic value of a proposed or existing analytics initiative, no matter the form that the business benefit arises. The session will provide practical advice about how to calculate ROI and the formulas, and how to collect the necessary information.
How a Semantic Layer Makes Data Mesh Work at ScaleDATAVERSITY
Data Mesh is a trending approach to building a decentralized data architecture by leveraging a domain-oriented, self-service design. However, the pure definition of Data Mesh lacks a center of excellence or central data team and doesn’t address the need for a common approach for sharing data products across teams. The semantic layer is emerging as a key component to supporting a Hub and Spoke style of organizing data teams by introducing data model sharing, collaboration, and distributed ownership controls.
This session will explain how data teams can define common models and definitions with a semantic layer to decentralize analytics product creation using a Hub and Spoke architecture.
Attend this session to learn about:
- The role of a Data Mesh in the modern cloud architecture.
- How a semantic layer can serve as the binding agent to support decentralization.
- How to drive self service with consistency and control.
Enterprise data literacy. A worthy objective? Certainly! A realistic goal? That remains to be seen. As companies consider investing in data literacy education, questions arise about its value and purpose. While the destination – having a data-fluent workforce – is attractive, we wonder how (and if) we can get there.
Kicking off this webinar series, we begin with a panel discussion to explore the landscape of literacy, including expert positions and results from focus groups:
- why it matters,
- what it means,
- what gets in the way,
- who needs it (and how much they need),
- what companies believe it will accomplish.
In this engaging discussion about literacy, we will set the stage for future webinars to answer specific questions and feature successful literacy efforts.
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...DATAVERSITY
Change is hard, especially in response to negative stimuli or what is perceived as negative stimuli. So organizations need to reframe how they think about data privacy, security and governance, treating them as value centers to 1) ensure enterprise data can flow where it needs to, 2) prevent – not just react – to internal and external threats, and 3) comply with data privacy and security regulations.
Working together, these roles can accelerate faster access to approved, relevant and higher quality data – and that means more successful use cases, faster speed to insights, and better business outcomes. However, both new information and tools are required to make the shift from defense to offense, reducing data drama while increasing its value.
Join us for this panel discussion with experts in these fields as they discuss:
- Recent research about where data privacy, security and governance stand
- The most valuable enterprise data use cases
- The common obstacles to data value creation
- New approaches to data privacy, security and governance
- Their advice on how to shift from a reactive to resilient mindset/culture/organization
You’ll be educated, entertained and inspired by this panel and their expertise in using the data trifecta to innovate more often, operate more efficiently, and differentiate more strategically.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Data Governance Trends - A Look Backwards and ForwardsDATAVERSITY
As DATAVERSITY’s RWDG series hurdles into our 12th year, this webinar takes a quick look behind us, evaluates the present, and predicts the future of Data Governance. Based on webinar numbers, hot Data Governance topics have evolved over the years from policies and best practices, roles and tools, data catalogs and frameworks, to supporting data mesh and fabric, artificial intelligence, virtualization, literacy, and metadata governance.
Join Bob Seiner as he reflects on the past and what has and has not worked, while sharing examples of enterprise successes and struggles. In this webinar, Bob will challenge the audience to stay a step ahead by learning from the past and blazing a new trail into the future of Data Governance.
In this webinar, Bob will focus on:
- Data Governance’s past, present, and future
- How trials and tribulations evolve to success
- Leveraging lessons learned to improve productivity
- The great Data Governance tool explosion
- The future of Data Governance
Data Governance Trends and Best Practices To Implement TodayDATAVERSITY
Would you share your bank account information on social media? How about shouting your social security number on the New York City subway? We didn’t think so either – that’s why data governance is consistently top of mind.
In this webinar, we’ll discuss the common Cloud data governance best practices – and how to apply them today. Join us to uncover Google Cloud’s investment in data governance and learn practical and doable methods around key management and confidential computing. Hear real customer experiences and leave with insights that you can share with your team. Let’s get solving.
Topics that you will hear addressed in this webinar:
- Understanding the basics of Cloud Incident Response (IR) and anticipated data governance trends
- Best practices for key management and apply data governance to your day-to-day
- The next wave of Confidential Computing and how to get started, including a demo
It is a fascinating, explosive time for enterprise analytics.
It is from the position of analytics leadership that the enterprise mission will be executed and company leadership will emerge. The data professional is absolutely sitting on the performance of the company in this information economy and has an obligation to demonstrate the possibilities and originate the architecture, data, and projects that will deliver analytics. After all, no matter what business you’re in, you’re in the business of analytics.
The coming years will be full of big changes in enterprise analytics and data architecture. William will kick off the fifth year of the Advanced Analytics series with a discussion of the trends winning organizations should build into their plans, expectations, vision, and awareness now.
Too often I hear the question “Can you help me with our data strategy?” Unfortunately, for most, this is the wrong request because it focuses on the least valuable component: the data strategy itself. A more useful request is: “Can you help me apply data strategically?” Yes, at early maturity phases the process of developing strategic thinking about data is more important than the actual product! Trying to write a good (must less perfect) data strategy on the first attempt is generally not productive –particularly given the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” This program refocuses efforts on learning how to iteratively improve the way data is strategically applied. This will permit data-based strategy components to keep up with agile, evolving organizational strategies. It also contributes to three primary organizational data goals. Learn how to improve the following:
- Your organization’s data
- The way your people use data
- The way your people use data to achieve your organizational strategy
This will help in ways never imagined. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges programmatically includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs as organizations identify prioritized areas where better assets, literacy, and support (data strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are also covered, including:
- A cohesive argument for why data strategy is necessary for effective data governance
- An overview of prerequisites for effective strategic use of data strategy, as well as common pitfalls
- A repeatable process for identifying and removing data constraints
- The importance of balancing business operation and innovation
Who Should Own Data Governance – IT or Business?DATAVERSITY
The question is asked all the time: “What part of the organization should own your Data Governance program?” The typical answers are “the business” and “IT (information technology).” Another answer to that question is “Yes.” The program must be owned and reside somewhere in the organization. You may ask yourself if there is a correct answer to the question.
Join this new RWDG webinar with Bob Seiner where Bob will answer the question that is the title of this webinar. Determining ownership of Data Governance is a vital first step. Figuring out the appropriate part of the organization to manage the program is an important second step. This webinar will help you address these questions and more.
In this session Bob will share:
- What is meant by “the business” when it comes to owning Data Governance
- Why some people say that Data Governance in IT is destined to fail
- Examples of IT positioned Data Governance success
- Considerations for answering the question in your organization
- The final answer to the question of who should own Data Governance
It is clear that Data Management best practices exist and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This program describes what must be done at the programmatic level to achieve better data use and a way to implement this as part of your data program. The approach combines DMBoK content and CMMI/DMM processes – permitting organizations with the opportunity to benefit from the best of both. It also permits organizations to understand:
- Their current Data Management practices
- Strengths that should be leveraged
- Remediation opportunities
MLOps – Applying DevOps to Competitive AdvantageDATAVERSITY
MLOps is a practice for collaboration between Data Science and operations to manage the production machine learning (ML) lifecycles. As an amalgamation of “machine learning” and “operations,” MLOps applies DevOps principles to ML delivery, enabling the delivery of ML-based innovation at scale to result in:
Faster time to market of ML-based solutions
More rapid rate of experimentation, driving innovation
Assurance of quality, trustworthiness, and ethical AI
MLOps is essential for scaling ML. Without it, enterprises risk struggling with costly overhead and stalled progress. Several vendors have emerged with offerings to support MLOps: the major offerings are Microsoft Azure ML and Google Vertex AI. We looked at these offerings from the perspective of enterprise features and time-to-value.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
2. Managing Blind
A Data Quality
and
Data Governance
Vade Mecum
By Peter R. Benson
Project Leader for ISO 8000, the International
Standard for Data Quality
Edited by Melissa M. Hildebrand
rev 2012.06.28
Copyright 2012 by Peter R. Benson
ECCMA Edition
ECCMA Edition License Notes:
This eBook is licensed for your personal enjoyment only. This
eBook may not be re-sold or given away to other people. If you
would like to share this eBook with another person, please
purchase an additional copy for each recipient. If you’re
reading this eBook and did not purchase it, or it was not
purchased for your use only, then please visit eccma.org and
purchase your own copy. It is also available at
Smashwords.com. Thank you for respecting the hard work of
this author.
***~~~***
2
3. Table of Contents
Preface
Basic principles
Chapter 1: Show me the money
Chapter 2: The law of unintended consequences
Chapter 3: Defining data and information
Chapter 4: The characteristics of data and information
Chapter 5: A simplified taxonomy of data
Chapter 6: Defining data quality
Chapter 7: Stating requirements for data
Chapter 8: Building a corporate business language
Chapter 9: Classifications
Chapter 10: Master data record duplication
Chapter 11: Data governance
Chapter 12: Where do we go from here?
Appendix 1: Managing a data cleansing process for assets,
materials or services
Further readings
***~~~***
3
4. Chapter 1: Show me the money
Business is about profit and profit is generated in the short
term by reducing cost and increasing revenue but in the longer
term by managing risk.
Risk management is fundamental to the finance and insurance
industries where the ability to “predict” is at the core of the
business. The difference between an actuary and a gambler is
data. The actuary promotes their ability to record and analyze
data and the gambler must hide any such ability or risk being
asked to leave the casino.
It is not surprising that data plays a key role in risk
management. Taking a “calculated risk” implies there is some
data upon which you can actually perform the calculation.
Other than in the finance and insurance industries, risk
management is a hard sell to all but the most sophisticated
managers. Cost reduction is a management favorite and an
easier sell, but if you can associate data quality and
governance with revenue growth you’ve scored a home run.
Most recorded examples of failures due to missing or incorrect
data fall into the catastrophic loss category. This is only
because of the enormity of the loss compared to the ease with
which the error was made, or the tiny amount of data involved.
There are whole websites devoted to listing the financial
consequences of data errors. Some of my favorites include;
Timo Elliott’s YouTube account of a simple error in the property
17
5. tax records that resulted in school budget cutbacks, as well as,
the Mars Climate Orbiter. The Mars Climate Orbiter was a $327
million project that came to an untimely end because of what
has become known as the “metric mix-up.” The software on the
Mars Climate Orbiter used the metric system, while the ground
crew was entering data using the imperial system. There is also
the story of Napoleon’s army who was able to force the
surrender of the Austrian army at Ulm when the Russians failed
to turn up as scheduled purportedly because they were using
the Julian calendar and not the Gregorian calendar used by the
Austrians; now that is what I call being stood up!
We all have personal stories in having to deal with the
consequences of data errors but my absolute personal favorite,
at least in hindsight, involves the IRS. It all began one morning
when I was handed a crisp envelope from the IRS. Inside the
envelope was a letter explaining that I was going to be audited.
This sort of letter sends chills up your spine. When I recovered
and mustered the courage to call the number on the letter, I
was surprised to be speaking to an eminently reasonable
inspector. She asked me to confirm that I was claiming a
deduction for alimony paid to my ex-wife. Not exactly the sort
of thing you wanted to be reminded of, but I was happy to
confirm that this was indeed the case. “According to our
records you have been claiming this deduction for over ten
years,” again not something I cared to be reminded of, but the
answer was an easy “yes”. There was a worrying silence,
followed by, “I am afraid this is not possible.” The chills quickly
18
6. rolled up my spine again. “The social security number you have
entered on your tax return belongs to a fourteen year old
female living in Utah.” To my utter surprise and after a long
exhale, I was glad to be able to correct the error which turned
out to be no more that a reversal of two digits in the social
security number. You have to be impressed by the ability of the
IRS to connect the dots. I know I was, and I should have quit
while I was ahead. There had been recent news reports about
child brides in Utah, so my reply was “Well at least she was
from Utah.” It did not impress the IRS agent who reminded me
that the IRS office I was speaking to was in Utah; apparently
humor is not a requirement for an IRS agent.
What jumps out from these examples is the multiplier effect. A
simple data error can easily, and all too often does, mushroom
into larger, far reaching and lasting economic fallout. Data
errors are rarely benign; more often than not they are
catastrophic.
As a general rule, most managers are natural risk takers, and
unless you are in the insurance industry, it is an uphill struggle
to associate data quality and governance with meaningful value
in the form of risk management or loss mitigation with one
notable exception. By focusing on resolving frequent small
losses, rather than larger catastrophic losses, it is usually
possible to correlate data quality and governance with reducing
loss. Examples include, reducing production down time and
delivery delays. These are most often considered to be revenue
19
7. generation and not cost reduction. The correlation between
data quality and delivered production capacity or on time
delivery is generally accepted, and the calculation of the
additional revenue generated is straightforward.
The role quality data plays in reducing cost is also generally
accepted, although the specifics are poorly understood. There
is clear evidence that simple vendor rationalization or group
purchasing will drive down price. However this can be easily
overdone to the point of exchanging short term price
advantage for long term reliance on larger suppliers able to
reclaim the price advantage over the longer term. The ultimate
goal is to commoditize goods and services to the point where
there are many competing suppliers. This requires excellent
vendor, material and service master data. The rewards can be
huge, not only in highly competitive pricing but also in a
flexible and resilient supply chain.
As a general rule most companies can save 10% of their total
expenditure on materials and services simply by good
procurement practices which include maintaining up to date
material and service masters supported by negotiated
contracts. The challenge is to maintain the discipline in the face
of urgent and unpredictable requirements for goods or services.
Most companies make it difficult and time consuming to add a
new item to their material or service masters and the result is
“free text” or “maverick spend." These are off contract
purchases where the item purchased is not in the material or
20
8. service master, instead a “free text” description is entered in
the purchase order. Free text descriptions are rarely
accompanied by rigorous classification and as a result
management reports start to lose accuracy as an ever
increasing percentage of spend appears under the
“miscellaneous” or “unclassified” headings, hardly a
management confidence builder. It is interesting that most ERP
systems require absolute unambiguous identification of the
party to be paid, on the pretext that it is required by law, which
it is, but they do not require the unambiguous identification of
the items purchased. As many have found out at their
considerable expense, the law also requires the identification
and unambiguous description of the goods or service
purchased. As federal and state governments go on the hunt
for more tax revenue, we can expect to see greater scrutiny of
purchase order line item descriptions to determine what is and
what is not accepted as an "ordinary and necessary” business
expense.
The most common scenario is a big effort to rationalize
procurement, which is then accompanied by a substantial drop
in free text spend. A big part of this effort is the identification
of duplicates. Vendor master duplicates are actually rare in
terms of the identification of the legal entity that needs to be
paid, but less rare is a lack of understanding of the relationship
between suppliers and how this impacts pricing. Customer
record duplication is actually surprisingly common, and worst of
all is material master duplication. Material master record
21
9. duplication all by itself can easily be responsible for up to a
30% price differential. Chapter 10 deals specifically with the
issue of the identification and resolution of duplicate records
but suffice to say it is not as straight forward of an issue as
many believe. Duplication is a matter of perspective and
timing.
Without good data governance that keeps the master data up
to date, data quality degrades and free text purchasing rises
again. Free text spend is actually a great indicator of the
success of a data quality and data governance program; the
lower the free text spend the more successful the program. It
is not hard to justify a data quality and data governance
program based on the initial measurable savings, but it is
harder to maintain a program as a cost avoidance initiative.
The ultimate goal is to associate a data quality and governance
program with revenue growth, preferably profitable revenue
growth. This can appear challenging but in reality it is not.
In 2010, The Economist Intelligence Unit’s editorial team
conducted a survey of 602 senior executives. Of which, 96% of
the executives surveyed considered data either “extremely
(69%) or somewhat (27%) valuable in creating and
maintaining a competitive advantage.”
Debra D'Agostino, Managing Editor of Business Research at the
Economist Intelligence Unit and editor of the report also states
"It's not enough to merely collect the data; companies need to
create strategies to ensure they can use information to get
22
10. ahead of their competitors."
How do you use data, let alone data quality and governance as
a competitive advantage? The most common answer is to look
inwards and consider data as a source of knowledge to be
mined for business intelligence. This has been done with
phenomenal success. From targeting customers with highly
contextual and relevant offers, to cutting edge logistics, to
product customization and everything in between.
Wal-Mart can rightly be said to be an information company that
uses retail to generate revenue and not a retail outlet that uses
information to maximize revenue. Data itself has value and
many companies have successfully turned their data into a
revenue source.
Roger Ehrenberg states it well when he says, “In today's world,
every business generates potentially valuable data. The
question is, are there ways of turning passive data into an
active asset to increase the value of the business by making its
products better, delivering a better customer experience, or
creating a data stream that can be licensed to someone for
whom it is most valuable?”
I have found that you can often convincingly calculate the
value of data by identifying the data that is essential to a
specific business process. Without the data, the process may
not fail but it would slow down, revenue would be lost and
costs would increase. Data is rarely the only contributing factor
to the efficiency of a specific process however, by looking at
23
11. how data contributes to the efficiency of the process you can
measure the value of the data.
Of course there is nothing like a crisis to focus attention and
liberate financial resources quickly. In order to sell a data
quality or data governance program it helps if you can find a
burning bridge, and if you cannot find one that is actually on
fire, it is not unknown to find one you can set on fire or at the
very least to point to the enormous and imminent risk of fire. It
really does work, ask any politician.
Any good data quality or data governance specialist will tell you
“Show me the data and I will show you the money.”
***~~~***
24