Data analytics, data management, and master data
management are part of an overall imperative
for public-sector organizations. They are central to
organizational competitiveness and relevancy. The City of
Cincinnati, Ohio, has developed a robust master data management
process, and any government can use the city’s
achievements as a best practices model for their own master
data management strategy. This article looks at several
administrative regulations, touching on reasons why master
data management is essential, the benefits it can confer, how
Cincinnati got started, the city’s framework, and the lessons
the city learned along the way
In business, master data management is a method used to define and manage the critical data of an organization to provide, with data integration, a single point of reference.
Analytics Organization Modeling for Maturity Assessment and Strategy DevelopmentVijay Raj
The paper discusses Business Intelligence Organization Modeling as a concept along with practical implementation aspects with reference to Analytics and Business Intelligence Strategy in large enterprises. BI organization modeling revolves around the ability to model the patterns of BI prevalent within a corporate structure to assess organizational capability and maturity, and there by contributing towards BI strategy development and implementation. The paper also details Analytics & BI organization modeling in a predominantly SAP based enterprise ecosystem and is demonstrated with BI systems based on the SAP NetWeaver Business Warehouse (BW) using data discovery and machine learning techniques. The data discovery process for Analytics & BI organization modeling is carried out using SAP Lumira Data Visualization tool connected to an SAP NetWeaver BW based Global Enterprise Data Warehousing and Reporting System.
The Importance of Master Data ManagementDATAVERSITY
Despite its immaterial nature, data has a tendency to pile up as time goes on, and can quickly be rendered unusable or obsolete without careful maintenance and streamlining of processes for its management. This presentation will provide you with an understanding of reference and Master Data Management (MDM), one such method for keeping mass amounts of business data organized and functional towards achieving business goals.
MDM’s guiding principles include the establishment and implementation of authoritative data sources and effective means of delivering data to various business processes, as well as increases to the quality of information used in organizational analytical functions (such as BI). To that end, attendees of this webinar will learn how to:
Structure their Data Management processes around these principles
Incorporate Data Quality engineering into the planning of reference and MDM
Understand why MDM is so critical to their organization’s overall data strategy
Discuss foundational MDM concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
In business, master data management is a method used to define and manage the critical data of an organization to provide, with data integration, a single point of reference.
Analytics Organization Modeling for Maturity Assessment and Strategy DevelopmentVijay Raj
The paper discusses Business Intelligence Organization Modeling as a concept along with practical implementation aspects with reference to Analytics and Business Intelligence Strategy in large enterprises. BI organization modeling revolves around the ability to model the patterns of BI prevalent within a corporate structure to assess organizational capability and maturity, and there by contributing towards BI strategy development and implementation. The paper also details Analytics & BI organization modeling in a predominantly SAP based enterprise ecosystem and is demonstrated with BI systems based on the SAP NetWeaver Business Warehouse (BW) using data discovery and machine learning techniques. The data discovery process for Analytics & BI organization modeling is carried out using SAP Lumira Data Visualization tool connected to an SAP NetWeaver BW based Global Enterprise Data Warehousing and Reporting System.
The Importance of Master Data ManagementDATAVERSITY
Despite its immaterial nature, data has a tendency to pile up as time goes on, and can quickly be rendered unusable or obsolete without careful maintenance and streamlining of processes for its management. This presentation will provide you with an understanding of reference and Master Data Management (MDM), one such method for keeping mass amounts of business data organized and functional towards achieving business goals.
MDM’s guiding principles include the establishment and implementation of authoritative data sources and effective means of delivering data to various business processes, as well as increases to the quality of information used in organizational analytical functions (such as BI). To that end, attendees of this webinar will learn how to:
Structure their Data Management processes around these principles
Incorporate Data Quality engineering into the planning of reference and MDM
Understand why MDM is so critical to their organization’s overall data strategy
Discuss foundational MDM concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
Reference matter data management:
Two categories of structured data :
Master data: is data associated with core business entities such as customer, product, asset, etc.
Transaction data: is the recording of business transactions such as orders in manufacturing, loan and credit card payments in banking, and product sales in retail.
Reference data: is any kind of data that is used solely to categorize other data found in a database, or solely for relating data in a database to information beyond the boundaries of the enterprise .
Suresh Menon, Vice President, Product Management - Information Quality Solutions at Informatica, shares how to master your data and your business from the 2015 Informatica Government Summit.
Master data management (mdm) & plm in context of enterprise product managementTata Consultancy Services
The presentation discusses the classical features and advantages of Master Data Management (MDM) system along with appropriate situations to use it. How do companies apply MDM who design, manufacture and sell their products in several geographies facing challenges in making appropriate decisions on their investment in PLM & MDM space?
Another important aspect covers the comparison/relation between a MDM system (or Product Master System) and Enterprise PLM system. How can you maximize your ROI on both PLM and MDM investments? With examples from different industries the key takeaways include whether your organization requires an MDM solution or not.
MMIS/HealthCare Payer Applications depend upon traditional data base models and structured data analytics to fulfill their needs. These approaches, while adequate in the past, will not suffice to address future requirements. They lack the processing capability to load and query multi-terabyte datasets in a timely fashion and the flexibility to effectively manage unstructured and semi-structured data. Adapting “Big Data” platform to MMIS application will resolve above issues.
Overcoming the Challenges of your Master Data Management JourneyJean-Michel Franco
This Presentaion runs you through all the key steps of an MDM initiative. It considers and showcase the key milestones and building blocks that you will have to roll-out to make your MDM
journey
-> Please contact Talend for a dedicated interactive sessions with a storyboard by customer domain
Developing & Deploying Effective Data Governance FrameworkKannan Subbiah
This is the slide deck presented at the Customer Privacy and Data Protection India Summit 2019 held in Mumbai, India. The specific topics touched upon are the guiding principles, Aligning with Data Architecture, Data Quality & Compliance.
Evidence suggests that the track record of MDM (Master Data Management) initiatives is not very good. Traditional MDM is often a costly, time-consuming, IT-driven activity that is disconnected from business goals and stakeholders. Even MDM projects that initially meet their goals often suffer during sustainment, or are limited to specific divisions and fail to provide value for the rest of the organization.
This webinar will:
- Review the technical and business challenges associated with the traditional MDM lifecycle
- Explain why the technologies and conventional wisdom associated with MDM do not seem to be working
- Discuss use cases of organizations who have achieved success by adopting a new, iterative approach called "streamlined MDM"
Presentation to introduce information governance. This should be used in conjunction with the paper I published on my website. A full information governance methodology, with research included from the foremost authorities on data governance.
Reference matter data management:
Two categories of structured data :
Master data: is data associated with core business entities such as customer, product, asset, etc.
Transaction data: is the recording of business transactions such as orders in manufacturing, loan and credit card payments in banking, and product sales in retail.
Reference data: is any kind of data that is used solely to categorize other data found in a database, or solely for relating data in a database to information beyond the boundaries of the enterprise .
Suresh Menon, Vice President, Product Management - Information Quality Solutions at Informatica, shares how to master your data and your business from the 2015 Informatica Government Summit.
Master data management (mdm) & plm in context of enterprise product managementTata Consultancy Services
The presentation discusses the classical features and advantages of Master Data Management (MDM) system along with appropriate situations to use it. How do companies apply MDM who design, manufacture and sell their products in several geographies facing challenges in making appropriate decisions on their investment in PLM & MDM space?
Another important aspect covers the comparison/relation between a MDM system (or Product Master System) and Enterprise PLM system. How can you maximize your ROI on both PLM and MDM investments? With examples from different industries the key takeaways include whether your organization requires an MDM solution or not.
MMIS/HealthCare Payer Applications depend upon traditional data base models and structured data analytics to fulfill their needs. These approaches, while adequate in the past, will not suffice to address future requirements. They lack the processing capability to load and query multi-terabyte datasets in a timely fashion and the flexibility to effectively manage unstructured and semi-structured data. Adapting “Big Data” platform to MMIS application will resolve above issues.
Overcoming the Challenges of your Master Data Management JourneyJean-Michel Franco
This Presentaion runs you through all the key steps of an MDM initiative. It considers and showcase the key milestones and building blocks that you will have to roll-out to make your MDM
journey
-> Please contact Talend for a dedicated interactive sessions with a storyboard by customer domain
Developing & Deploying Effective Data Governance FrameworkKannan Subbiah
This is the slide deck presented at the Customer Privacy and Data Protection India Summit 2019 held in Mumbai, India. The specific topics touched upon are the guiding principles, Aligning with Data Architecture, Data Quality & Compliance.
Evidence suggests that the track record of MDM (Master Data Management) initiatives is not very good. Traditional MDM is often a costly, time-consuming, IT-driven activity that is disconnected from business goals and stakeholders. Even MDM projects that initially meet their goals often suffer during sustainment, or are limited to specific divisions and fail to provide value for the rest of the organization.
This webinar will:
- Review the technical and business challenges associated with the traditional MDM lifecycle
- Explain why the technologies and conventional wisdom associated with MDM do not seem to be working
- Discuss use cases of organizations who have achieved success by adopting a new, iterative approach called "streamlined MDM"
Presentation to introduce information governance. This should be used in conjunction with the paper I published on my website. A full information governance methodology, with research included from the foremost authorities on data governance.
From Chaos to Clarity: Crafting a Data Strategy Roadmap for Organizational Tr...TekLink International LLC
Discover the power of a data strategy roadmap and BI roadmap strategy in optimizing data utilization, informed decision-making, and achieving business objectives. Gain a competitive edge, enhance operational efficiency, and drive innovation.
Enterprise Information Management Strategy - a proven approachSam Thomsett
Access a proven approach to Enterprise Information Management Strategy - providing a framework for Digital Transformation - by a leader in Information Management Consulting - Entity Group
We offer a guide to change management that enables data quality throughout the organization and a sample operational data quality scorecard. This helps making operational data quality a way of life in your enterprise, from data origination of data sources to transformation
The right approach to data governance plays a crucial role in the success of AI and analytics initiatives within an organization. This is especially true for small to medium-sized companies that must harness the power of data to drive growth, innovation and competitiveness.
This guide aims to provide SMB organizations with a practical roadmap to successfully implement a data governance strategy that ensures data quality, security and compliance. Use it to unlock the full potential of your data assets.
In your cloud transition, don’t overlook the finance and accounting implications, which influence efforts from risk management and security to regulatory compliance. Reap the full benefits of an enterprisewide cloud deployment by following four strategies that will help you consider the holistic impact of the cloud.
Learn more - http://gt-us.co/1wJulWG
Governance and Architecture in Data IntegrationAnalytiX DS
AnalytiX™ Mapping Manager™ provides this discipline and rigor through its dedicated data mapping methodology as well as its metadata management processes and powerful patented mapping technology. AnalytiX™ Mapping Manager™ was designed and developed to not only fill the gap of having the ability to manage and version mapping specifications, but to also streamline and improve current process and drive standards around the entire process and across the enterprise for all integration and governance processes.
Optimizing Local Government Management through Performance and Data AnalyticsHarry Black
Time is of the essence for local governments in this age of
the Internet of things and big data. A fundamental question
is “Can government be disrupted?” The answer is yes, and
it will happen, brought about by driverless cars, drones, big
data-driven algorithms, and robots. The way that we currently
conduct business in government will change, requiring fewer
people, fewer facilities, and less equipment.
Fy17 one page strategic plan final 5 5-16Harry Black
The end product of an exhaustive internal strategic planning process that centered around Five priority goals: Safer Streets, A Growing Economy, Thriving and Healthy Neighborhoods, Innovative Government and Fiscal Sustainability and Strategic Reinvestment.
City of cincinnati fy17 one page strategic plan final 5 5-16Harry Black
The end product of an exhaustive internal strategic planning process that centered around Five priority goals: Safer Streets, A Growing Economy, Thriving and Healthy Neighborhoods, Innovative Government and Fiscal Sustainability and Strategic Reinvestment.
City of Cincinnati One Page Strategic PlanHarry Black
The end product of an exhaustive internal strategic planning process that centered around Five priority goals: Safer Streets, A Growing Economy, Thriving and Healthy Neighborhoods, Innovative Government and Fiscal Sustainability and Strategic Reinvestment.
Cincinnati's Permits and Inspections Streamlining InitiativeHarry Black
This presentation provides an overview of how Cincinnati has transformed its development processes through organizational reconfiguration, business process re-engineering, training, and the strategic utilization of technology.
First annual economic inclusion update 031716 final (2)Harry Black
Progress report on the operationalization of the City's Department of Economic Inclusion and the recommendations of the Economic Inclusion Advisory Council.
Cincinnati customer service impact updateHarry Black
Overview of the impact of the City's Office of Performance and Data Analytics on the City's quest to enhance its overall customer service effectiveness.
City of Cincinnati Capital Acceleration Plan Presentation 012716 final [compa...Harry Black
Cincinnati's innovative approach to meeting its transportation needs enabling it to increase annual road repavement spend by 40% and road preventative maintenance by 100% using an innovative just in time public financial management model.
This article focuses on the
strategic steps taken by the
City of Baltimore to manage the
cost of benefit programs, to
bring the level of benefit offered
more in-line with the other comparable
public entities, and to
take advantage of the opportunities
brought by the evolving
healthcare law in order to satisfy
the City's short- and longterm
goals while maintaining
financial health.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
1. By Harry Black
A Framework for the Public Sector
Master Data Management
2. August 2018 | Government Finance Review 37
D
ata analytics, data management, and master data
management are part of an overall imperative
for public-sector organizations. They are central to
organizational competitiveness and relevancy. The City of
Cincinnati, Ohio, has developed a robust master data man-
agement process, and any government can use the city’s
achievements as a best practices model for their own mas-
ter data management strategy. This article looks at several
administrative regulations, touching on reasons why master
data management is essential,the benefits it can confer,how
Cincinnati got started, the city’s framework, and the lessons
the city learned along the way.
AN ESSENTIAL PROCESS
Master data management integrates technologies and pro-
cesses in a disciplined fashion across the organization, allow-
ing the flow of data from numerous stand-alone systems into
one unified process via an enterprise-wide technology tool.
This allows an organization to make
the transition from a series of non-
unified silos to a master data manage-
ment architecture that integrates the
output of multiple systems to a master
data management structure that inte-
grates the outputs from all systems,
creating an organization-wide man-
agement capability.
A “smart” city is data-driven.
Increasing the scope, volume, quality,
and utility of data can lead to improved
performance, better customer ser-
vice, and greater efficiency through
creative problem solving, while also
promoting government transparency
and accountability.
Governments seek to deliver efficient and effective cus-
tomer services at the lowest possible cost. One way to identify
opportunities for improved delivery of services is by using a
strong analytics infrastructure, which provides a framework
for storing and using data. Because multiple departments
work together to deliver services, a key component of this
analytics infrastructure is a clear data governance policy,
which outlines the expectations for data access, availability,
and management to ensure cross-functional decision making,
accountability, data integrity, and data availability.
Optimizing the availability of quality data can be challeng-
ing, but it’s worth the effort. It forces the organization to ask
itself a set of difficult questions, such as:
n What business process does this data represent?
n How are the data structured?
n What information populates these data categories?
n How is this data input?
n Is this data incomplete?
The specific objectives of analytics infrastructure frame-
work are to establish expectations and apply appropriate
controls over:
n Data inventory and ownership.
n Data collection.
n Data use and disclosure.
n Data availability, retention, and disposal.
Governments are stewards of the
data it collects from residents, custom-
ers, and visitors, so it has a responsibil-
ity to protect those data. At the same
time, the collected data is a valu-
able asset for managing the enterprise,
and it can be critical for identifying
opportunities to improve the quality,
effectiveness, and efficiency of ser-
vice delivery to residents and custom-
ers. An organization’s capability to
achieve these goals relies on a strong
analytics infrastructure and direct
access to enterprise data in order to
perform the following core functions
related to this mission:
n Automated data dashboards that are used to consistently
monitor and evaluate performance.
n Monitoring of operations in real-time.
n Self-service data discovery, which allows departments to
fluidly use data for gaining insight about operations with-
out relying on power users and database administrators
for analysis.
n Predictive analytics, which help organizations to develop
predictive models that enable proactive and preventative
service delivery to enhance operational effectiveness
and efficiency.
Master data management
integrates technologies and
processes in a disciplined
fashion across the organization,
allowing the flow of data from
numerous stand-alone systems
into one unified process
via an enterprise-wide
technology tool.
3. 38 Government Finance Review | August 2018
n Open data to provide transparency
and foster innovation.
To use mater data management
effectively, governments must under-
stand its many nuances and be able
conceptualize they way data should
be arranged to respond to questions
and clarify the underlying business
processes. Benefits fall into four basic
categories: cost savings, efficiency,
effectiveness, and revenue maximiza-
tion. The number of datasets created
can be unlimited; see Exhibit 1 for an
example of some of the ones used in
Cincinnati. (More information is avail-
able at data.cincinnati-oh.gov/.)
GETTING STARTED
There is no one tried-and-true
approach to establishing a master
data management capability. What’s
important is to consider various do’s
and don’ts derived from best prac-
tices and the body of knowledge that
has evolved over the past decade. It
isn’t necessary to reinvent the wheel.
Governments across the country,
including Cincinnati, developed best
practices. Like Cincinnati, other gov-
ernments have developed master data
management policies, regulations,
and procedures; these public docu-
ments can serve as a reference point or template.
In government, the information technology (IT) function
is usually decentralized, which is a problem if your goal is to
establish a master data management structure and or open
data capability. In these cases, the best operating model allows
the central office to lead and maintain policy development,
establish standards, and monitor and enforce compliance,
while allowing some degree of self-determination on the part
of the user community (internal and external). To get started:
n Rationalize, standardize, and optimize the organization’s
IT delivery model.
n Develop an open data requirements and implementation
plan.
n Establish an analytics infrastructure policy.
Rationalizing the I.T. Function. The objectives of IT
service delivery standardization and optimization include the
following:
n Improve IT support and service delivery across the organi-
zation.
n Ensure that the total enterprise has equal and adequate
access to technology systems and support resources.
n Reduce duplication of systems, contracts, and mainte-
nance agreements.
n Optimize staff expertise and align it with enterprise tech-
nology goals and objectives.
n Implement a standardized, enterprise-wide service desk,
procurement process, and inventory system.
Exhibit 1: Representative Types of Datasets
GPS data to track location and status of vehicle assets.
Computer-aided dispatch data to record public safety
responses and in-field activity by first responders.
Records management system case data for police (e.g., geo-
coding, cleaning, and joining with computer-aided dispatch
incident data).
Emergency management services on-scene response data.
Proactive green space cleaning activity.
Health center clinic data (e.g., geo-coding and cleaning up
address data so the health department can identify trends
on a neighborhood level).
Performance and workflow data.
Customer service request survey data from surveys received
from citizens once departments have closed out service
request tickets.
Development data geo-coded for the Community and
Economic Development Department to use in tracking
development trends and activity.
Police traffic stops and citations.
Police use of force.
Assaults on police officers.
Police officer-involved shootings.
Violent crime.
Shootings and police calls for service.
To use master data
management effectively,
governments must understand
its many nuances and be able
conceptualize they way data
should be arranged to respond
to questions and clarify the
underlying business processes.
4. August 2018 | Government Finance Review 39
n Strengthen cybersecurity efforts
through policy compliance and
proactive security initiatives.
n Ensure management accountability.
n Clearly define enterprise-wide IT
and data management leadership
roles and responsibilities.
Delivering on these objectives
requires a strong chief information
officer and data management executives, along with engaged
and committed enterprise executive leadership at the highest
level. This phase of the process requires engaging and break-
ing down silos across the organization.
Open Data Requirements and Implementation Plan.
Citizen access to their government’s information is funda-
mental to transparency and accountability. Therefore, public
enterprises have an obligation to make datasets about the
operations and finances of government publicly available
for review, interpretation, analysis, research, and criticism.
Datasets should be complete, accessible, reflective of primary
data, available to the public in a timely manner, supplied in
a format that can be computer-processed, license-free, and
non-proprietary.
Establishing an Analytics Infrastructure Policy. Set up
an open data executive committee made up of senior-level
staff representing the appropriate cross-section of decision
makers and influencers throughout the government — indi-
viduals who are directly responsible for IT, performance man-
agement, and data management. The executive committee
should meet regularly, to vet data that are specific to security
or privacy concerns and review funding. The committee will
also likely be responsible for conducting an annual review
of all data management/governance policies and making
recommendations for revision. It should also establish a
smaller working group to serve as a forum supporting the
implementation plan. The working group will work to support
the organization’s open government goals, including collabo-
rations with researchers, the private sector, and the public.
The working group will also identify problems or issues that
arise during the course of their work on the implementation
plan and develop recommendations for resolution. Each
sub-organization within the enterprise needs an open data
coordinator, who is responsible for attending meetings of
the open data working group, identifying any potential pri-
vacy or security concerns with a dataset and informing their
departmental leadership. The open
data coordinators, in cooperation with
their department directors, should be
expected to identify and recommend
datasets for public distribution to the
chief data officer.
The chief data officer is responsible
for realizing the implementation plan,
as established in a related enterprise
policy, by providing technical assistance and guidance to the
open data working group, and by identifying needed infra-
structure and resources. The chief data officer establishes
standards and schedules for the publication of datasets, with
the goal of presenting them in an efficient and cost-effective
manner, and reviews existing policies to identify impediments
to the open data initiative. He or she works with relevant staff,
including but not limited to the government’s information
security officer and law department, to guide and/or propose
revisions to existing policies in order to facilitate the orga-
nization’s open data initiative, promoting greater openness
without damaging legal and financial interests. The chief data
officer facilitates the sharing of information among depart-
ments and is responsible for making recommendations to the
executive committee for revisions, with the goal of expanding
the open data initiative and increasing the number of open
datasets. These recommendations should be submitted at
least once a year. The chief data officer also acts as a liaison
between the open data working group and the executive
It isn’t necessary to reinvent
the wheel. Governments
across the country have
developed best practices.
5. 40 Government Finance Review | August 2018
committee, providing department-by-
department summary updates once
a month to the executive committee
and other enterprise leadership.
MORE ON DATA
GOVERNANCE
Data governance policy should out-
line the government’s expectations for
data access, availability, and man-
agement, to ensure cross-functional
decision making, accountability, data
integrity, and data availability. The
scope of data governance policies and procedures should
cover:
n Analytics infrastructure expectations.
n Analytics infrastructure policy.
n Analytics infrastructure procedures.
n Open availability of data.
n Data collection.
n Automatic extraction of data.
n Data use and disclosure.
n Data retention and disposal.
n Responses to public records
requests.
n Third-party IT solutions and IT pro-
curement (evaluating accessibility,
who owns the data, services and
assistance provided, and security).
n Analytics infrastructure roles and
responsibilities.
n Analytics infrastructure procedure
(comprising four main phases:
identify and authorize; develop;
approve; and implement).
Operating Framework. Exhibit 2 shows the basic struc-
ture of a data management model. This is Cincinnati’s model,
and it probably resembles that of other governments with a
mature master data management function. Hardware and
software requirements vary, depending on the organization
and its appetite for investment. Typical cost items associ-
ated with creating the infrastructure include staff resources,
software, open data portal hosting, data visualization soft-
ware, and the direct costs associated with establishing
a data warehouse.
LESSONS LEARNED
All the components of the master data management capa-
bility need to be aligned with the mission, vision, and priority
goals of the enterprise. Then, data management can help
drive the organization’s performance management efforts
and innovation initiatives to optimize transparency and com-
munity engagement while generating operating and service
delivery efficiencies. Alignment can be achieved through a
strategic planning exercise that yields no more than five prior-
ity goals. Strategy should be laser focused.
Every enterprise-wide initiative needs an executive spon-
sor, someone who is from the highest level of the organization
and who believes in the power of data. The executive sponsor
will need to fight for sufficient funding and resources, help
break through silos, and keep the effort going until it’s com-
plete — and then ensure sustainability.
This effort will require the government to do things dif-
ferently, including working more seamlessly across all
departments, so change management is critical. Effective
change management strategy and tactics help in maintaining
Public enterprises have an
obligation to make datasets
about the operations and
finances of government
publicly available for review,
interpretation, analysis,
research, and criticism.
6. August 2018 | Government Finance Review 41
calm, trust, and communications. The strategy and plan
should be flexible and responsive to
situational dynamics. Start planning
for change before starting the master
data management initiative to ensure
readiness and preparedness. The
executive sponsor will be helpful here.
Don’t underestimate the relevance
of data governance, a process that
reflects discipline and alignment
with the government’s priority goals.
Data governance makes it possible to consistently produce
timely and reliable data. Master data
management is a dynamic process
that requires constant tracking, moni-
toring, and measuring to keep up with
the data requirements of its employees
and its external stakeholders. Data
governance allows the government
to identify, capture, manage, and dis-
seminate data, and to establish data
management business rules.
Exhibit 2: Cincinnati’s Data Management Model
Automation
The primary value add of technology.
It allows us to process data the
same way every time, faster, without
human oversight, which reduces
errors and saves time. It also creates
opportunities for more advanced
problem solving.
Data Collection
Benefits of API
Redaction
Makes sure there are no
sensitive data leaks. Only
requires maintenence
and tweaking, not
constant checking and
oversight. The lens is
a safer method for
redaction because it
automates it.
Data Quality
The internal processes
that precede release of
an open data set are
critical to data quality.
These processess
include: geocoding;
address validation;
verifying fields and data
characteristics; verifying
data entry/output.
Open
Data
Platform
Public facing
user interface
for ease of
access
Curated
Data Set
Lens
Redaction
anonymization
data quality
Internal
Analysis
Source
Data
Internal
Data
Repository
API
Public/
Private
Partnership
Machine
Readable
Up-to-Date
Information
Industry
Reliability
All the components of the
master data management
capability need to be aligned
with the mission, vision, and
priority goals of the enterprise.
7. 42 Government Finance Review | August 2018
Establish a means of measuring the
program’s success and impact. A list
of some things that can be tracked
includes:
n How master data management
allows for the optimization of col-
laboration across the organization.
n Cost avoidance and savings.
n Customer satisfaction ratings.
n Insights that have led to innovation.
n Service delivery breakthroughs.
n Community engagement break-
throughs.
n Staffing and resources management impacts.
n Impact as a crime prevention and crime-fighting tool.
n Impact as a tool for creating healthy communities.
CONCLUSIONS
Master data management is the foundation that makes data
analytics and predictive analytics possible. If implemented cor-
rectly, it can be a powerful tool that helps governments work
smarter, faster, and with greater transparency. Establishing
a master data management capability does not have to take
forever — with commitment, it can be done within a window
of six to ten months. As a tool, master
data management makes it possible
for an organization to hone in on what
matters most instead of wasting time on
things that aren’t important. Master data
management can help organizations
innovate and solve service delivery and
operational problems that might have
been intractable before; it also opti-
mizes an organization’s collaborative
capabilities and allows for greater con-
nectivity between governments and the
people they serve. y
HARRY BLACK is the founder and general manager of Maximus
Management Group, LLC. Before that, he was city manager of the
City of Cincinnati, Ohio; finance director of the City of Baltimore,
Maryland; and co-manager of Global Commerce Solutions, Inc., a
government services firm he co-founded. Black also held the position
of vice president and program manager of McKissack McKissack
in Washington D.C., and deputy chief administrative officer for the
City of Richmond, Virginia. Black’s public service career also includes
stints with the New York City Transit Authority, the Port Authority
of New York and New Jersey, the New York City Mayor’s Office of
Contracts, and the District of Columbia Government.
Master data management is a
dynamic process that requires
constant tracking, monitoring,
and measuring to keep up with
the data requirements of its
employees and its external
stakeholders.