This document discusses managing dirty data in organizations using ERP systems. It begins with defining dirty data and how it can negatively impact organizations. It then discusses the costs of using dirty data and how ERP systems can help integrate disparate data sources and clean up dirty data. The document also summarizes lessons learned from a case study of a company that implemented an ERP system, including the importance of understanding how ERP systems change user roles and communicating those changes.
Boosting Cybersecurity with Data Governance (peer reviewed)Guy Pearce
Data Governance has a significant role to play in information security, with special data classes beyond the regular four cyber classes (public, confidential, classified and restricted) being useful in helping the organization identify whether sensitive data was exposed in a breach.
This document provides an overview of key concepts from Chapter 1 of the textbook. It discusses how IT can optimize organizational performance through agility and mobility. Effective use of IT requires understanding the relationships between people, business processes, and organizational culture. The chapter also covers strategic planning, competitive models, business performance measurement, ethics, and career opportunities in the IT field.
This document provides an overview of master data management (MDM) and how organizations can treat data as a key asset. It discusses how as companies grow, their data becomes fragmented across different systems. This leads to inconsistencies and a lack of a single view of critical data like customers. The document defines MDM and explains how it can integrate master data to provide a consistent view. It argues that as data drives business operations, organizations should manage data as strategically as they do physical assets to maximize its value.
This document discusses data, text, and document management. It covers file management systems, database management systems, data warehouses, and enterprise content management. Effective data management is important for organizations to access and use data efficiently. It reduces data redundancy, inconsistency, and isolation issues. Database management systems help solve problems of file-based systems by providing centralized access to data. Data warehouses further organize vast amounts of data for analysis and decision making.
The document discusses various types of decision support systems and knowledge management systems. It describes management information systems, decision support systems, executive support systems, group decision support systems, and knowledge management systems. It also discusses how intelligent techniques like expert systems, case-based reasoning, fuzzy logic, neural networks, genetic algorithms, and intelligent agents can enhance decision making.
Developing A Universal Approach to Cleansing Customer and Product DataFindWhitePapers
Take a look at this review of current industry problems concerning data quality, and learn more about how companies are addressing quality problems with customer, product, and other types of corporate data. Read about products and use cases from SAP to see how vendors are supporting data cleansing.
Management information systems (MIS) collect, process, store, and distribute data to support decision-making and control in organizations. Decision support systems (DSS) are a type of information system that supports business or organizational decision-making activities. DSS provide analysis of information to help decision makers choose among alternative solutions. There are different types of DSS, including communication-driven, data-driven, document-driven, knowledge-driven, model-driven, spreadsheet-based, and web-based systems. DSS use analytical tools like what-if analysis, sensitivity analysis, goal-seeking analysis, and optimization analysis to help decision makers evaluate alternatives.
Boosting Cybersecurity with Data Governance (peer reviewed)Guy Pearce
Data Governance has a significant role to play in information security, with special data classes beyond the regular four cyber classes (public, confidential, classified and restricted) being useful in helping the organization identify whether sensitive data was exposed in a breach.
This document provides an overview of key concepts from Chapter 1 of the textbook. It discusses how IT can optimize organizational performance through agility and mobility. Effective use of IT requires understanding the relationships between people, business processes, and organizational culture. The chapter also covers strategic planning, competitive models, business performance measurement, ethics, and career opportunities in the IT field.
This document provides an overview of master data management (MDM) and how organizations can treat data as a key asset. It discusses how as companies grow, their data becomes fragmented across different systems. This leads to inconsistencies and a lack of a single view of critical data like customers. The document defines MDM and explains how it can integrate master data to provide a consistent view. It argues that as data drives business operations, organizations should manage data as strategically as they do physical assets to maximize its value.
This document discusses data, text, and document management. It covers file management systems, database management systems, data warehouses, and enterprise content management. Effective data management is important for organizations to access and use data efficiently. It reduces data redundancy, inconsistency, and isolation issues. Database management systems help solve problems of file-based systems by providing centralized access to data. Data warehouses further organize vast amounts of data for analysis and decision making.
The document discusses various types of decision support systems and knowledge management systems. It describes management information systems, decision support systems, executive support systems, group decision support systems, and knowledge management systems. It also discusses how intelligent techniques like expert systems, case-based reasoning, fuzzy logic, neural networks, genetic algorithms, and intelligent agents can enhance decision making.
Developing A Universal Approach to Cleansing Customer and Product DataFindWhitePapers
Take a look at this review of current industry problems concerning data quality, and learn more about how companies are addressing quality problems with customer, product, and other types of corporate data. Read about products and use cases from SAP to see how vendors are supporting data cleansing.
Management information systems (MIS) collect, process, store, and distribute data to support decision-making and control in organizations. Decision support systems (DSS) are a type of information system that supports business or organizational decision-making activities. DSS provide analysis of information to help decision makers choose among alternative solutions. There are different types of DSS, including communication-driven, data-driven, document-driven, knowledge-driven, model-driven, spreadsheet-based, and web-based systems. DSS use analytical tools like what-if analysis, sensitivity analysis, goal-seeking analysis, and optimization analysis to help decision makers evaluate alternatives.
This chapter discusses key concepts in data management including data sources, collection and quality; data warehousing; database management systems; and the data life cycle. It describes how data moves from raw to processed to analytical stages and how tools like data profiling, quality management, and integration improve data infrastructure. Finally, it addresses managerial issues around data storage, delivery, security, and ethics.
The document discusses network management and mobility. It covers business networks and their functions of mobility, collaboration, relationships and search. Wireless broadband networks like 3G and 4G are described along with technologies like WiMAX and LTE. Network management portals are discussed as well as collaboration tools. Issues of network infrastructure, performance, intranets, extranets and VPNs are also summarized.
Decision support systems (DSS) are a class of computerized systems that help organizational decision-making. A DSS compiles useful information from data, documents, and business models to help decision-makers identify and solve problems. It has three key functions: capturing past information, data processing, and data retrieval. A DSS has three core components - a database management system, model-based management system, and dialog generation/management system. There are different types of DSS that aid decision-making through various methods like data, models, knowledge, or documents.
The document discusses management information systems and their role in organizations. It defines key concepts like data, information, and information systems. It explains that MIS aims to support decision-making at different management levels by providing the right type of information. The functions of MIS include data capturing, processing, storage and dissemination. MIS must be tailored to an organization's structure, with different reporting structures for line, functional and divisional setups. Finally, the document outlines the positive and negative impacts of information systems on organizations.
An information system is a combination of hardware, software, infrastructure, and trained personnel organized to facilitate planning, control, coordination, decision making in an organization. There are several types of information systems including executive support systems, management information systems, decision support systems, knowledge management systems, transaction processing systems, and office automation systems. The five basic components of an information system are hardware, software, data, procedures, and people.
El campo de DSS / BI esta evolucionando desde sus origenes como una herramienta primariamente de soporte personal y está rapidamente llegando a ser una comodidad compartida a traves de de las organizaciones
An Organizational Memory and Knowledge System (OMKS): Building Modern Decisio...Waqas Tariq
This document presents an approach for building modern decision support systems that integrates data warehouses and organizational memory systems. The approach uses scenarios to capture tacit knowledge from employees and ontology for organizing diverse data and knowledge sources. It is based on theoretical foundations including decision making processes, data warehousing, knowledge management, organizational memory systems, and the knowledge spiral concept. The integration of these systems is expected to enhance organizational decision making and learning.
This document provides information on decision support systems (DSS). It discusses definitions of DSS and how they support decision making. DSS can take many forms, from model-driven to data-driven systems. The document outlines frameworks for developing DSS and describes different types of DSS including passive, active, and cooperative systems. It also discusses applications of DSS in areas like business and agriculture.
Information Systems and Knowledge ManagementMeenakshi Paul
Information Systems and Knowledge
Management, Information, data and Intelligence, The Characteristics of Valuable Information, Relevance, Quality, Timeliness, Completeness, Knowledge Management, Global Information Systems, Decision Support Systems, Databases and Data Warehousing, Input Management, Computerized Data Archives Networks and Electronic Data Interchange, The Internet and Research
The document discusses data security and data management. It defines data security as processes and practices to protect critical IT systems and information. Effective data security uses controls, applications, and techniques to identify important data and apply appropriate security controls. Data security is important for organizations to protect user and customer data from unauthorized access. Common data security methods include access controls, authentication, backups, encryption, and data erasure. Data management techniques aim to ensure data quality, integrate data across systems, and govern data use and access. The document also discusses specific techniques for data cleansing, integration, and other aspects of data management.
This document discusses different types of information systems used in organizations, including transaction processing systems, management information systems, decision support systems, and executive support systems. It also discusses how information systems interact with and support different levels of an organization, including operational, management, and strategic levels. Finally, it examines how information technology impacts and changes organizational structure, processes, boundaries and global operations.
This document discusses organization of information systems. It covers centralized, decentralized, and distributed processing models. It describes the roles and responsibilities of information systems professionals in gathering, analyzing, and reporting data to support business processes, decision making, and competitive advantage. The document also covers security and ethical issues in information systems, including information rights, intellectual property rights, and security risks. It discusses threats like data loss, theft, and damage from viruses. Finally, it outlines some controls for security threats like careful hiring, access restrictions, monitoring, audits, and encryption.
The document discusses different types of information systems used in organizations:
1. Expert systems use knowledge from experts to provide answers to questions.
2. Decision support systems provide interactive support to help with decision making by analyzing multiple data and models.
3. Executive information systems provide structured information to senior executives.
It also discusses key concepts in developing decision support systems like the iterative design process and components such as the user interface, data management, and model management.
This document discusses management information systems (MIS) and their role in organizations. It begins by defining MIS as a system that provides managers with information to help with decision making, planning, and control. It then discusses different types of information systems at various levels, including operational, knowledge, management, and strategic levels. Transaction processing systems, management information systems, decision support systems, and executive information systems are described. The document also discusses digital firms and how they leverage various applications and technologies to digitally enable core business functions.
Integrating user needs into ECM projects is key to success. Whether it is a initial implementation or a reboot or just expanding use, user needs and UX testing should be integrated into every project
Information Governance, Managing Data To Lower Risk and Costs, and E-Discover...David Kearney
Information governance, records and information management, and data disposition policies are ways to help lower costs and mitigate risks for organizations. Policies and procedures to actively manage data are not just an IT "problem," they're a collaborative business initiative that is a must in today's "big data" environment. With electronic discovery rules, government regulations and the Sarbanes-Oxley Act, all organizations must proactively take steps to manage their data with well-governed processes and controls, or be willing to face the risks and costs that come along with keeping everything. Organizations must know what information they have, where it is located, the duration data must be retained and what information would be needed when responding to an event.
There have been numerous instances of severe legal penalties for organizations that did not have an electronic data strategy, tools, processes and controls to locate and understand their own data. In addition, the risks of unmanaged data include skyrocketing infrastructure and personnel costs and an increase in attorney time to manage massive amounts of data when a litigation event occurs.
Information governance is needed much like any business continuity and disaster recovery plans, but with an understanding of data: where data are located, how data are managed, event response, and regular testing of processes and procedures for preparedness.
Architecture Framework for Resolution of System Complexity in an EnterpriseIOSR Journals
This document presents an architecture framework for resolving system complexity in an enterprise. It discusses how enterprise architecture can be used to address issues like requirement complexity, organizational complexity, process complexity, and design complexity. The framework breaks down the enterprise information system into subsystems like back-end systems, front-end systems, management tools, and communication systems. It also separates concerns into different architecture layers - an external enterprise model, conceptual enterprise model, front-end systems, back-end systems, and management tools. The framework is intended to provide a structured approach to managing complexity by organizing enterprise data and functions across the different systems and models.
Decision support systems and business intelligenceShwetabh Jaiswal
This document discusses decision support systems and business intelligence. It describes how the modern business environment requires computerized systems to help with complex decision making. Business intelligence transforms raw data into useful information through methodologies, processes and technologies. Decision support systems couple individual expertise with computer capabilities to improve decision quality for semi-structured problems. Both systems use similar architectures of data warehouses, analytics, and user interfaces to enable analysis and informed decisions.
Towards A Model Of Organisational Prerequistes For Enterprise Wide Sys IntegDonovan Mulder
The document discusses a model of organizational prerequisites for enterprise-wide systems integration projects. It proposes that examining critical factors throughout an integration project's lifecycle can help identify an organization's readiness. Specifically, the model involves assessing an organization's "existing," "planned," and "implemented" states across five factors: systems, data, skills, organizational, and project management. The goal is to help organizations better understand the impacts and complexities of large-scale integration efforts like ERP and data warehousing implementations.
La base monetaria es la suma del efectivo en manos del público y las reservas bancarias. El Banco de la República es el encargado de administrar la base monetaria en Colombia con el objetivo de mantener la estabilidad de precios. La tabla y gráfica muestran que la base monetaria ha crecido constantemente entre 2006 y 2009, alcanzando casi $400.000 millones en 2008.
This chapter discusses key concepts in data management including data sources, collection and quality; data warehousing; database management systems; and the data life cycle. It describes how data moves from raw to processed to analytical stages and how tools like data profiling, quality management, and integration improve data infrastructure. Finally, it addresses managerial issues around data storage, delivery, security, and ethics.
The document discusses network management and mobility. It covers business networks and their functions of mobility, collaboration, relationships and search. Wireless broadband networks like 3G and 4G are described along with technologies like WiMAX and LTE. Network management portals are discussed as well as collaboration tools. Issues of network infrastructure, performance, intranets, extranets and VPNs are also summarized.
Decision support systems (DSS) are a class of computerized systems that help organizational decision-making. A DSS compiles useful information from data, documents, and business models to help decision-makers identify and solve problems. It has three key functions: capturing past information, data processing, and data retrieval. A DSS has three core components - a database management system, model-based management system, and dialog generation/management system. There are different types of DSS that aid decision-making through various methods like data, models, knowledge, or documents.
The document discusses management information systems and their role in organizations. It defines key concepts like data, information, and information systems. It explains that MIS aims to support decision-making at different management levels by providing the right type of information. The functions of MIS include data capturing, processing, storage and dissemination. MIS must be tailored to an organization's structure, with different reporting structures for line, functional and divisional setups. Finally, the document outlines the positive and negative impacts of information systems on organizations.
An information system is a combination of hardware, software, infrastructure, and trained personnel organized to facilitate planning, control, coordination, decision making in an organization. There are several types of information systems including executive support systems, management information systems, decision support systems, knowledge management systems, transaction processing systems, and office automation systems. The five basic components of an information system are hardware, software, data, procedures, and people.
El campo de DSS / BI esta evolucionando desde sus origenes como una herramienta primariamente de soporte personal y está rapidamente llegando a ser una comodidad compartida a traves de de las organizaciones
An Organizational Memory and Knowledge System (OMKS): Building Modern Decisio...Waqas Tariq
This document presents an approach for building modern decision support systems that integrates data warehouses and organizational memory systems. The approach uses scenarios to capture tacit knowledge from employees and ontology for organizing diverse data and knowledge sources. It is based on theoretical foundations including decision making processes, data warehousing, knowledge management, organizational memory systems, and the knowledge spiral concept. The integration of these systems is expected to enhance organizational decision making and learning.
This document provides information on decision support systems (DSS). It discusses definitions of DSS and how they support decision making. DSS can take many forms, from model-driven to data-driven systems. The document outlines frameworks for developing DSS and describes different types of DSS including passive, active, and cooperative systems. It also discusses applications of DSS in areas like business and agriculture.
Information Systems and Knowledge ManagementMeenakshi Paul
Information Systems and Knowledge
Management, Information, data and Intelligence, The Characteristics of Valuable Information, Relevance, Quality, Timeliness, Completeness, Knowledge Management, Global Information Systems, Decision Support Systems, Databases and Data Warehousing, Input Management, Computerized Data Archives Networks and Electronic Data Interchange, The Internet and Research
The document discusses data security and data management. It defines data security as processes and practices to protect critical IT systems and information. Effective data security uses controls, applications, and techniques to identify important data and apply appropriate security controls. Data security is important for organizations to protect user and customer data from unauthorized access. Common data security methods include access controls, authentication, backups, encryption, and data erasure. Data management techniques aim to ensure data quality, integrate data across systems, and govern data use and access. The document also discusses specific techniques for data cleansing, integration, and other aspects of data management.
This document discusses different types of information systems used in organizations, including transaction processing systems, management information systems, decision support systems, and executive support systems. It also discusses how information systems interact with and support different levels of an organization, including operational, management, and strategic levels. Finally, it examines how information technology impacts and changes organizational structure, processes, boundaries and global operations.
This document discusses organization of information systems. It covers centralized, decentralized, and distributed processing models. It describes the roles and responsibilities of information systems professionals in gathering, analyzing, and reporting data to support business processes, decision making, and competitive advantage. The document also covers security and ethical issues in information systems, including information rights, intellectual property rights, and security risks. It discusses threats like data loss, theft, and damage from viruses. Finally, it outlines some controls for security threats like careful hiring, access restrictions, monitoring, audits, and encryption.
The document discusses different types of information systems used in organizations:
1. Expert systems use knowledge from experts to provide answers to questions.
2. Decision support systems provide interactive support to help with decision making by analyzing multiple data and models.
3. Executive information systems provide structured information to senior executives.
It also discusses key concepts in developing decision support systems like the iterative design process and components such as the user interface, data management, and model management.
This document discusses management information systems (MIS) and their role in organizations. It begins by defining MIS as a system that provides managers with information to help with decision making, planning, and control. It then discusses different types of information systems at various levels, including operational, knowledge, management, and strategic levels. Transaction processing systems, management information systems, decision support systems, and executive information systems are described. The document also discusses digital firms and how they leverage various applications and technologies to digitally enable core business functions.
Integrating user needs into ECM projects is key to success. Whether it is a initial implementation or a reboot or just expanding use, user needs and UX testing should be integrated into every project
Information Governance, Managing Data To Lower Risk and Costs, and E-Discover...David Kearney
Information governance, records and information management, and data disposition policies are ways to help lower costs and mitigate risks for organizations. Policies and procedures to actively manage data are not just an IT "problem," they're a collaborative business initiative that is a must in today's "big data" environment. With electronic discovery rules, government regulations and the Sarbanes-Oxley Act, all organizations must proactively take steps to manage their data with well-governed processes and controls, or be willing to face the risks and costs that come along with keeping everything. Organizations must know what information they have, where it is located, the duration data must be retained and what information would be needed when responding to an event.
There have been numerous instances of severe legal penalties for organizations that did not have an electronic data strategy, tools, processes and controls to locate and understand their own data. In addition, the risks of unmanaged data include skyrocketing infrastructure and personnel costs and an increase in attorney time to manage massive amounts of data when a litigation event occurs.
Information governance is needed much like any business continuity and disaster recovery plans, but with an understanding of data: where data are located, how data are managed, event response, and regular testing of processes and procedures for preparedness.
Architecture Framework for Resolution of System Complexity in an EnterpriseIOSR Journals
This document presents an architecture framework for resolving system complexity in an enterprise. It discusses how enterprise architecture can be used to address issues like requirement complexity, organizational complexity, process complexity, and design complexity. The framework breaks down the enterprise information system into subsystems like back-end systems, front-end systems, management tools, and communication systems. It also separates concerns into different architecture layers - an external enterprise model, conceptual enterprise model, front-end systems, back-end systems, and management tools. The framework is intended to provide a structured approach to managing complexity by organizing enterprise data and functions across the different systems and models.
Decision support systems and business intelligenceShwetabh Jaiswal
This document discusses decision support systems and business intelligence. It describes how the modern business environment requires computerized systems to help with complex decision making. Business intelligence transforms raw data into useful information through methodologies, processes and technologies. Decision support systems couple individual expertise with computer capabilities to improve decision quality for semi-structured problems. Both systems use similar architectures of data warehouses, analytics, and user interfaces to enable analysis and informed decisions.
Towards A Model Of Organisational Prerequistes For Enterprise Wide Sys IntegDonovan Mulder
The document discusses a model of organizational prerequisites for enterprise-wide systems integration projects. It proposes that examining critical factors throughout an integration project's lifecycle can help identify an organization's readiness. Specifically, the model involves assessing an organization's "existing," "planned," and "implemented" states across five factors: systems, data, skills, organizational, and project management. The goal is to help organizations better understand the impacts and complexities of large-scale integration efforts like ERP and data warehousing implementations.
La base monetaria es la suma del efectivo en manos del público y las reservas bancarias. El Banco de la República es el encargado de administrar la base monetaria en Colombia con el objetivo de mantener la estabilidad de precios. La tabla y gráfica muestran que la base monetaria ha crecido constantemente entre 2006 y 2009, alcanzando casi $400.000 millones en 2008.
The Royalty Of Loyalty Crm, Quality And RetentionDonovan Mulder
The document discusses the relationship between customer relationship management (CRM), product/service quality, and customer loyalty. It proposes a new 5Qs model to measure quality and loyalty, with the 5Qs being quality of object, processes, infrastructure, interaction, and atmosphere. The model suggests that improving each quality dimension through CRM strategies can increase customer satisfaction and loyalty over time. Effective CRM requires understanding customers, having a well-structured customer database, and linking quality improvements to strategy changes to document their impact on satisfaction and competitive advantage.
10. What Managers Should Know About Erp ErpiiDonovan Mulder
The document discusses enterprise resource planning (ERP) systems and what managers should know about implementing ERP/ERP II systems. Key points include:
- ERP systems integrate different business units like finance, manufacturing, and HR to improve information flow and competitiveness.
- Implementing ERP requires significant investment and organizational change, and is one of the most important decisions managers will make.
- Managers should carefully analyze their needs, develop an implementation plan, and ensure top management support to avoid high failure rates for ERP projects.
The document discusses the V-Model approach to software testing and how it relates different types of testing to stages of development. It provides an overview of the V-Model stages including requirements, objectives, external specification, system design, and module interface specification. It also discusses how test objectives can be defined for each testing level to avoid redundancy. While the V-Model provides a constrained model, the types of testing it identifies are still relevant for more iterative processes.
14. Business Process Approach Towards An Inter Organizational Enterprise SystemDonovan Mulder
The document discusses business process approaches towards inter-organizational enterprise systems. It reviews existing ERP literature and process models, and proposes an extended framework to examine ERP implementation from an inter-organizational perspective. The study uses a focus group methodology to gain insights on critical issues across the ERP planning, project, and post-implementation phases in extended enterprise environments.
16. Erp Ii A Conceptual Framework For Next Generation Enterprise SystemsDonovan Mulder
The document discusses the evolution of enterprise systems and proposes a conceptual framework called ERP II for next-generation enterprise systems. It outlines how enterprise resource planning (ERP) systems have developed over decades from early inventory control systems to integrated ERP packages. The framework is intended to compile concepts from existing enterprise systems into a comprehensive model to analyze and design complex enterprise architecture.
The Hidden Financial Costs Of Erp SoftwareDonovan Mulder
This document discusses how enterprise resource planning (ERP) systems can negatively impact a company's flexibility and decision-making ability. It argues that ERP software often requires companies to conform their core business processes to fit the software, rather than allowing flexibility. This rigidity increases the costs of making changes and implementing innovations. As a result, many cost-saving or efficiency-improving projects may be rejected due to the high costs of altering the ERP software. The document also notes that ERP systems were often sold with promises of future cost savings but led to unanticipated high implementation costs and disruptions for companies.
8. Erp Managing The Implementation ProcessDonovan Mulder
This document discusses a study that investigated factors contributing to successful enterprise resource planning (ERP) implementations, defined as being completed on time and under budget. The researchers conducted case studies of ERP implementations and identified common factors among companies that managed their projects successfully. These factors were grouped into categories of planning effort, implementation decisions, and implementation management. The researchers then conducted a survey of US manufacturing companies to validate the factors identified in the case studies and better understand their individual and collective impacts on implementation timelines and costs. Logistic regression models were used to classify companies based on whether they completed implementations on time and under budget. The results provide insights into factors that managers should consider when implementing major systems like ERP.
Overlooked aspects of data governance: workflow framework for enterprise data...Anastasija Nikiforova
This presentation is a supplementary material for the article "Overlooked aspects of data governance: workflow framework for enterprise data deduplication" (Azeroual, Nikiforova, Shei) presented at The International Conference on Intelligent Computing, Communication, Networking and Services (ICCNS2023).
Abstract of the paper: Data quality in companies is decisive and critical to the benefits their products and services can provide. However, in heterogeneous IT infrastructures where, e.g., different applications for Enterprise Resource Planning (ERP), Customer Relationship Management (CRM), product management, manufacturing, and marketing are used, duplicates, e.g., multiple entries for the same customer or product in a database or information system, occur. There can be several reasons for this, but the result of non-unique or duplicate records is a degraded data quality. This ultimately leads to poorer, inefficient, and inaccurate data-driven decisions. For this reason, in this paper, we develop a conceptual data governance framework for effective and efficient management of duplicate data, and improvement of data accuracy and consistency in large data ecosystems. We present methods and recommendations for companies to deal with duplicate data in a meaningful way.
Encrypted Data Management With Deduplication In Cloud...Angie Jorgensen
The document discusses some disadvantages of Minitrex's current data management system and proposes solutions based on customer relationship management (CRM) theories. It finds that Minitrex's data is siloed across different departments, leading to issues like duplicate customer records and a lack of a holistic view of customers. It suggests integrating CRM across departments to get a unified view of customers. It also recommends utilizing CRM software to consolidate data to improve data quality, gain insights, and better manage customer relationships. Leadership support and an integrated, holistic approach are identified as important for effective use of CRM.
Top 10 guidelines for deploying modern data architecture for the data driven ...LindaWatson19
Enterprises are facing a new revolution, powered by the rapid adoption of data analytics with modern technologies like machine learning and artificial intelligence (A).
Organizations are facing rapidly growing data due to increased data collection and virtualization. This growth strains backup and storage capabilities. IBM Real-time Compression Appliances can reduce primary and backup data by up to 80% in real-time without performance impacts, addressing storage efficiency issues and reducing costs. At Ben-Gurion University, real-time compression provided 65-83% reductions, extending infrastructure lifespan.
A Data Quality Model for Asset Management in Engineering OrganisationsCyrus Sorab
Abstract: Data Quality (DQ) is a critical issue for effective asset management. DQ problems can result in severe negative consequences for an organisation. This paper aims to explore DQ issues associated with the implementation of Enterprise Asset Management (EAM) systems.
Completed Paper - Authors
Andy Koronios
University of South Australia
andy.koronios@unisa.edu.au
Shien Lin
University of South Australia
shien.lin@unisa.edu.au
Jing Gao
University of South Australia
jing.gao@unisa.edu.au
Paper available for download here - https://pdfs.semanticscholar.org/cb3c/9209e2909033e21019157d279d5363b0db76.pdf
data collection, data integration, data management, data modeling.pptxSourabhkumar729579
it contains presentation of data collection, data integration, data management, data modeling.
it is made by sourabh kumar student of MCA from central university of haryana
A System Approach For Defining Data Center Value Proposition.pdfVernette Whiteside
This document discusses defining the value proposition of a data center using a systems approach. It introduces a method to measure a data center's value using a set of metrics that capture the behavior and outcomes of a data center as a system. These metrics would provide measures for variables like performance, investments, operations, and services. Analyzing these metrics would provide a system model to define stakeholder value. Current methods for evaluating IT investments often focus only on financial metrics and lack consideration of external factors, behaviors, and qualitative impacts. A balanced, mixed approach is needed to fully capture a data center's true value proposition.
Estuate - Control Application Data GrowthEstuate, Inc.
Data growth is a significant challenge for most Fortune 1000 companies. The drivers of data growth are organic business growth, mergers and acquisitions, data retention requirements and something the author of this white paper refers to as the “Data Multiplier Effect.”
The document discusses the challenges clients face with bad customer data, including inconsistent data between systems, lack of data standards and ownership, difficulty retrieving archived data, and high costs of data issues. It provides examples of data quality problems that have cost companies millions or billions of dollars. The document advocates implementing data management and architecture practices to address these challenges and ensure accurate, consistent and secure customer data.
The document discusses how investing in data quality can provide a significant return on investment for companies. It outlines five tenets that leading companies embrace to realize this ROI from quality data: 1) view data quality as a business issue, not just an IT issue, 2) establish an explicit data governance strategy, especially at the point of data entry, 3) use a third-party data provider to consolidate and cleanse data, 4) address the challenge of maintaining accurate data given the rapid rate of data changes, and 5) strive for a 360-degree view of customers and suppliers across the organization.
http://www.embarcadero.com
Data yields information when its definition is understood or readily available and it is presented in a meaningful context. Yet even the information that may be gleaned from data is incomplete because data is created to drive applications, not to inform users. Metadata is the data that holds application
data definitions as well as their operational and business context, and so plays a critical role in data and application design and development, as well as in providing an intelligent operational environment that's driven by business meaning.
Strategic Information Management Through Data ClassificationBooz Allen Hamilton
This white paper presents a comprehensive approach to information management programs. It outlines how data growth directly affects the risk posture of critical corporate information assets. In addition, it defines common problems caused by gaps in information management programs as well as consequences associated with immature methodologies.
Building an Effective Data Management StrategyHarley Capewell
In June 2013, Experian hosted a Data
Management Summit in London, with over
100 delegates from the public, private and
third sectors. Speakers from Experian
and across the data industry explored the
challenges of developing and implementing
data quality strategies - and how to
overcome them. Read on for more information.
Gartner Positions Data Flux In The Leaders Quadrant Of The Magic Quadrant For...Cezar Cursaru
The document summarizes a Gartner report on the Magic Quadrant for Data Quality Tools. It discusses trends in the market such as growth, mergers and acquisitions, and the evolution of tools to offer more full-function suites. It evaluates vendors based on capabilities like profiling, parsing, standardization, and matching. The market is expected to grow significantly due to the importance of data quality initiatives.
NFRASTRUCTURE MODERNIZATION REVIEW
Analyze the issues
Hardware
Over-running volume of data is a problem that should be addressed by data management and storage management. Data is being constantly collected but poorly analyzed which leads to excessive amounts of data occupying storage and delay in operations which inevitably affect production, sales and profits. If this remains unresolved, current data may have to be moved to external storage and recovered if needed. There is also the risk of data not being encoded into computers and thus will remain in manual state. This can be a case of redundant or extraneous data that is not yet cleaned and normalized by operations managers with the guidance of IT. This situation is known as data overload where companies actually use only a fraction of the data they capture and store. Many companies simply hoard data to make sure that they are readily available when they are needed. This negatively impacts the Corporation when assessing data relevance, accuracies and timeliness (Marr, 2016).
Software
The Largo Corporation (LC) seems to running on an enterprise resource planning system that is probably as long as 20 years old. Initially, LC has had success with the old system because they were able to establish themselves in various industries such as healthcare, media, government, etc. But due to various concerns, the Corporation is currently running on an outdated system because it is unable to provide services that keeps the Corporation a float. The LC is losing revenue and customers. Complete data without analysis is invaluable because, no information and insights can be produced that will support decisions. Customer data should lead to the best marketing and sales campaigns. The Corporation needs to recognize its weaknesses and implement changes to their software by incorporating funding for a new system that is reliable, secure, and has the ability to run on integrated systems; all of which will streamline data organization and analysis for the enterprise. (Rouse, n.d).
Network/Telecommunications
The network that was built in the 1980’s has become slow and unreliable affecting business operations. The problems caused by the old network are; lack of integration and communication between departments affecting the work flow, supply vs. demand, and inability to analyze data to carry out these operations. The Corporation should have taken into consideration the growth of the company by expanding and upgrading their networks along with their services. They should also take into consideration the number of departments, the number of users and their skill level, storage and bandwidth, and budget (Rasmussen, 2011). The current network does not allow employees to connect on their mobile devices which restricts flexibility and places limitations on productivity and portability.
Management
The responses of both IT and the business group are both juxtaposed against e ...
Global Data Management: Governance, Security and Usefulness in a Hybrid WorldNeil Raden
With Global Data Management methodology and tools, all of your data can be accessed and used no matter where it is or where it is from: on-premises, private cloud, public cloud(s), hybrid cloud, open source, third-party data and any combination of the these, with security, privacy and governance applied as if they were a single entity. Ingenious software products and the economics of computing make it economical to do this. Not free, but feasible.
Master Data Management: Extracting Value from Your Most Important Intangible ...FindWhitePapers
This SAP Insight explores the importance of master data and the barriers to achieving sound master data, describes the ideal master data management solution, and explains the value and benefits of effective management of master data.
DATA VIRTUALIZATION FOR DECISION MAKING IN BIG DATAijseajournal
This document provides an overview of data virtualization techniques used for data analytics and business intelligence. It discusses how data virtualization creates a single virtual view of data across different sources to support decision making. It contrasts data virtualization with traditional ETL approaches, noting that data virtualization does not physically move data but instead queries sources in real-time. The document also outlines how data virtualization can reduce costs and improve scalability compared to ETL for integrating large, heterogeneous data in real-time.
This document discusses big data and the importance of data quality for big data initiatives. It defines big data as large, diverse digital data sets that require new techniques to enable capture, storage, analysis and visualization. The key challenges of big data include integrating diverse structured and unstructured data sources and ensuring high quality data. The document emphasizes that poor data quality can undermine big data analytics efforts and lead to wrong insights. It promotes establishing a data quality framework including profiling, standardization, matching and enrichment to enable valid big data analytics.
Similar to Managing Dirty Data In Organization Using Erp (20)
The Critical Success Factors For Erp Implementation An Organisational Fit Per...Donovan Mulder
This document summarizes a research study on critical success factors for enterprise resource planning (ERP) implementation from an organizational fit perspective. The study explores why ERP projects have unusually high failure rates despite their strategic importance. It defines organizational fit of ERP as the congruence between the original ERP system and its organizational context. The study examines how organizational fit of ERP and certain implementation contingencies like ERP adaptation, process adaptation, and organizational resistance impact ERP implementation success. The results from a field survey of 34 organizations show that ERP implementation success significantly depends on organizational fit of ERP and these implementation contingencies.
Realising Enhanced Value Due To Business Network Redesign Through Extended Er...Donovan Mulder
This document provides a case study of HLLNet, the business network system of Hindustan Lever Limited (HLL), one of the largest fast-moving consumer goods companies in India. It discusses how HLL extended its enterprise resource planning (ERP) system beyond internal operations to link its entire supply chain network. This allowed for improved information sharing and coordination across the network, leading to significant benefits such as reduced inventory levels, improved cash management, and increased value for consumers. The case study examines how redesigning business networks using ERP systems can optimize resources and create value across the entire supply chain.
Implementing Erp Systems In Small And Midsize Manufacturing FirmsDonovan Mulder
The document discusses a study on implementing enterprise resource planning (ERP) systems in small and midsize manufacturing firms. It presents profiles of 4 case study companies that implemented ERP systems. The companies ranged in annual revenue from $55-200 million.
It then outlines the ERP implementation process explored in the case studies, dividing it into 3 phases: planning activities, justification and selection activities, and installation activities. For each phase, it discusses the findings from the case studies and presents propositions related to the activities within each phase. The goal is to help understand what leads to successful ERP implementations.
15. Assessing Risk In Erp Projects Identify And Prioritize The FactorsDonovan Mulder
This document discusses assessing risk in ERP projects. It notes that ERP projects often face high failure rates and risks. The authors aim to identify and prioritize key risk factors for ERP projects using the Delphi method and analytic hierarchy process (AHP). They find that some important risk factors deserve more attention during ERP implementation. Identifying and managing risks is important to help mitigate failure and ensure successful ERP projects.
This document summarizes a research study on the effectiveness of implementing enterprise resource planning (ERP) systems in the printing industry in Pakistan. The study involved interviews with managers at five large printing firms in Pakistan. The research found that ERP systems could integrate key business functions and provide benefits to the printing industry, but not all ERP modules would be applicable. The study aims to increase awareness of ERP for the printing industry and determine which ERP modules should be prioritized for implementation. There are opportunities for the printing firms to gain competitive advantages through ERP implementations, but training and technical expertise will be required for successful adoption.
11. Requirements Of An Erp Enterprise Erp ModellerDonovan Mulder
The document discusses the requirements of an enterprise resource planning (ERP) enterprise modeler for optimally managing the fashion industry supply chain. It first provides background on the shift from push to pull supply chains and the benefits of supply chain integration. It then outlines that ERP systems have evolved as strategic tools for supply chain integration. The document also notes that while ERP adoption can enable integration, implementation is critical and can fail without proper business process reengineering. Finally, it describes forming an expert panel to identify the key features needed in an ERP enterprise modeler tailored specifically for the fashion industry.
This document summarizes a research paper that analyzed factors affecting ERP system adoption and compared adoption between small-to-medium enterprises (SMEs) and large companies. The summary identifies that business complexity and organizational change were examined as factors potentially influencing ERP adoption. An empirical study of 366 companies found that business complexity was a weak predictor of adoption, while company size was a strong predictor. SMEs cited structural and organizational reasons for non-adoption more than financial constraints, differing from large companies who cited organizational reasons most.
The document provides a conceptual model for understanding enterprise resource planning (ERP) systems. The model divides an ERP system into four major components: the software, the customer mindset, change management, and the flow of processes within the system. A fifth component, methodology, surrounds the other four and ensures they are integrated and implemented in an organized manner. The conceptual model is intended to help general and project managers understand ERP systems at a high level without technical details.
4. Expectation And Reality In Erp Implementation Consultant And Solution Prov...Donovan Mulder
The document discusses expectations and realities in ERP implementation from the perspective of consultants and solution providers. It summarizes literature on common ERP implementation challenges such as organizational readiness, customization vs standard systems, and managing change. It then presents results from a survey of 59 Finnish ERP consultants and vendors. The survey found that key challenges in ERP implementations relate to production planning, materials management, and meeting various requirements across different company levels and industries. Implementers emphasized the need for initial solutions and further research in production planning and complex products.
6. The Usefulness Of Erp Systems For Effective ManagementDonovan Mulder
This document summarizes a research paper that examines the usefulness of enterprise resource planning (ERP) systems for effective management. The paper studies, through a questionnaire, the reasons why companies adopt ERP systems, the impact on management processes including implementation problems, and the benefits derived from ERP systems. The empirical evidence confirms several benefits of ERP systems for management processes but also implementation problems encountered. The results provide a basis for future research on fully exploiting the potential of ERP systems for more effective business integration.
5. Change Management Strategies For Successful Erp ImplementationDonovan Mulder
1) The document discusses change management strategies for successful ERP implementation to overcome employee resistance. It proposes a 3-phase framework: knowledge formulation to understand employee attitudes, strategy implementation using marketing techniques to influence cognitive, affective, and conative components of attitudes, and status evaluation to assess effectiveness.
2) Key strategies discussed include communication of benefits, demonstration of system functionality, minimizing costs, differentiating the system through quality, training, gaining support of opinion leaders, and properly timing the system introduction. The strategies aim to reduce perceived risk and overcome habits to change employee attitudes toward accepting the new system.
3. Project Management A Case Study Of A Successful Erp ImplementationDonovan Mulder
This document provides a summary of a case study on a successful ERP implementation at an Irish subsidiary of a UK multinational company. The case study uses the Project Management Body of Knowledge (PMBOK) framework to analyze the project management approach. Interviews and questionnaires were conducted with the local implementation team over time to understand how their perceptions of the challenges evolved. The case study found that while the PMBOK framework broadly applies to ERP projects, some aspects require different emphasis due to the scope and complexity of ERP implementations.
2. Erp Innovation Implementation Model Incorporating Change ManagementDonovan Mulder
This document summarizes a research paper that proposes and provides preliminary validation of a model for how change management during an ERP implementation affects the effectiveness of that implementation. The paper presents a revised innovation implementation model that incorporates the impact of change management. Through a case study of an ERP implementation at a large Australian organization, the paper examines how specific change management activities influenced implementation climate and effectiveness. The findings provide qualitative support for the proposed model linking change management to implementation outcomes.
1. An Erp Performance Measurement Framework Using A Fuzzy Integral ApproachDonovan Mulder
This document proposes a seven-step framework for measuring the performance of an enterprise resource planning (ERP) system using a fuzzy integral approach. The framework aims to link performance indicators (PIs) to the objectives of the ERP implementation project. A fuzzy ERP performance index is used to account for ambiguities in evaluating ERP system performance. An example application in Taiwan demonstrates how the proposed framework can be applied in practice.
Using A Km Framework To Evaluate An Erp System ImplementationDonovan Mulder
This document summarizes a study that evaluated the implementation of an Enterprise Resource Planning (ERP) system from a knowledge management perspective. The researchers surveyed ERP users at a construction company to evaluate the system's effectiveness. They analyzed the results using a knowledge management capability maturity model framework. The framework indicated that the ERP system was seen as useful for cost management, and its effectiveness depended on human-to-human knowledge transfer about how to use the system. The framework also showed that leadership support is critical for successful ERP deployment. The study demonstrated that a capability maturity model can be used to assess an organization's knowledge management practices from an ERP adoption.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Managing Dirty Data In Organization Using Erp
1. Managing dirty data in organizations using ERP:
lessons from a case study
Jodi Vosburg
The University of Wisconsin-Whitewater, Wisconsin, USA
Anil Kumar
The University of Wisconsin-Whitewater, Wisconsin, USA
Keywords achieve a competitive advantage in the
Data, Data integrity, 1.0 Introduction marketplace (Sellar, 1999). On the other hand,
Enterprise resource planning,
Systems management Daily operations, planning, and decision- ``bad data can put a company at a competitive
making functions in organizations are disadvantage'' comments Greengard (1998). A
Abstract increasingly dependent on transaction data. recent study (Ferriss, 1998) found out that
The integrity of the data used to ``Canadian automotive insurers are taking a
This data is entered electronically and
operate and make decisions about
a business affects the relative manually and then organized, managed and major hit from organized and computer-
efficiency of operations and extracted for decision-making. The same data literate criminals who are staging crashes
quality of decisions made. entered and used to facilitate building, and taking advantage of dirty data in
Protecting that integrity can be corporate databases''. The study found out
shipping, and invoicing goods is also
difficult and becomes more
extracted and manipulated to evaluate that in one case several insurance firms lost
difficult as the size and complexity
of the business and its systems factory and sales force performance in the $56 million to one fraud ring.
increase. Recovering data short term. In the long term this data is used How does a company end up with dirty
integrity may be impossible once it data and what can be done to prevent this?
to chart the course of the business in terms of
is compromised. Stewards of
manufacturing facilities, products, and Disparate data stores (individual,
transactional and planning
systems must therefore employ a marketing. The integrity of the data used to departmental, and organizational) that have
combination of procedures operate and make decisions about a business been developed and used by organizational
including systematic safeguards users over the years lead to dirty data
affects the relative efficiency of operations
and user-training programs to
and quality of decisions made. Protecting problems. For example, dissimilar data
counteract and prevent dirty data
in those systems. Users of data integrity is a challenging task. Redman structures for the same customer data
transactional and planning (1995) comments that ``many managers are (spelling discrepancies, multiple account
systems must understand the
unaware of the quality of data they use and numbers, address variations), incomplete or
origins and effects of dirty data
perhaps assume that IT ensures that data are missing data, lack of legacy data standards,
and the importance of and means
of guarding against it. This perfect. Although poor quality appears to be actual data values being different from meta-
requires a shared understanding the norm, rather than the exception, they labels, use of free-form fields, etc. (Kay, 1997;
within the context of the business
have largely ignored the issue of quality''. Knowles, 1997; Weston, 1998). These problems
of the meaning, uses, and value of can be compounded by the volume of data
data across functional entities. In Other scholars (Greengard, 1998; Kilbane,
this paper, we discuss issues 1999; Tayi and Ballou, 1998; Wallace, 1999) that is stored and used in organizations. One
related to the origin of dirty data, also point out the importance of data quality way of overcoming this problem is to use
associated problems and costs of
for organizations. technologies that integrate the disparate data
using dirty data in an organization, stores for an organization and help
the process of dealing with dirty Maintaining the quality of the data that is
data in a migration to a new used in an organization is becoming an companies clean up their data. Enterprise
system: enterprise resource increasingly high priority for businesses. In resource planning (ERP) systems (SAP,
planning (ERP), and the benefits of
a recent survey of 300 IT executives Peoplesoft, Baan, J.D. Edwards, etc.) are
an ERP in managing dirty data. examples of such systems. ``A good ERP
These issues are explored in the conducted by Information Week (Wallace,
paper using a case study. 1999), majority of the respondents (81 per system offers an integrated option,
cent) said, ``improving customer data quality implementing browser and client-server
was the most important post-year 2000 modes while maintaining consistent data and
technology priority''. The respondents function within the enterprise and out to the
further stated that there would be supply chain'' (Stankovic, 1998). In recent
``significantly increased spending'' on data years, ERP vendors have gone beyond
quality in their organizations. Companies providing the traditional integrated
Industrial Management & that manage their data effectively are able to applications, such as manufacturing,
Data Systems financials, and human resources. Newer
101/1 [2001] 21±31 applications that have emerged include
The current issue and full text archive of this journal is available at
# MCB University Press supply chain management, customer-
[ISSN 0263-5577] http://www.emerald-library.com/ft
relationship management, data mining and
[ 21 ]
2. Jodi Vosburg and Anil Kumar data warehousing (Caldwell and Stein, 1998; the organization who were involved with this
Managing dirty data in Stankovic, 1998) and browser modes that project. These employees included the
organizations using ERP: enable organizations to reach out to manager of the CSC and marketing services,
lessons from a case study
customers and the supply chain. Caldwell an information analyst in the marketing
Industrial Management &
Data Systems and Stein (1998) also point out that ``most services group, and a customer support
101/1 [2001] 21±31 important, ERP forces discipline and representative (CSR). The manager of the
organization around processes, making the CSC is responsible for managing domestic
alignment of IT and business goals more order processing and sales and marketing
likely in the post-ERP era''. Aligning IT and reporting for the division. The information
business goals has always been a top priority analyst works with users and programmers
for senior management. Thus it might be to specify report requirements and does
helpful for a company to implement an ERP much of the testing and trouble-shooting for
system. those reports. The CSR is the data entry point
In this paper, we discuss the experiences of analyzing and translating customer purchase
a company, which implemented an ERP orders into ERP documents. This study will
system in their organization. The discussion look primarily at issues relating to the CSC.
is focussed primarily on the data aspect of the
implementation. The paper is organized as
follows. In the next section we describe the 3.0 Dirty data defined
case-study organization. Section 3 defines the At first, the abbreviation for black was blk.
concept of dirty data and its impact on the Then it was changed to bck. We didn't
integrity of organizational data. In Section 4 discover this change until someone said the
we list the costs incurred by organizations as color mix didn't look right (Horwitz, 1998).
a result of using dirty data. Section five
Dirty data exists when there are inaccuracies
highlights several lessons learnt from the
or inconsistencies within a collection of data
case-study organization and, finally, in
or when data extraction is inconsistent with
Section 6 we summarize the guidelines for
intent. Inclusion of dirty data in a data
companies planning to implement ERP
source may pollute the entire data source
solutions to overcome dirty data problems.
making it difficult or unwise to use the data
for analysis. Dirty data in a transactional
system can mean incorrect order taking,
2.0 The case study products not built to specification, or errors
The organization where this case study was in packaging, documentation, or billing. The
conducted is a $650 million division of a result is dissatisfied customers, loss of
Fortune 500 company located in the Midwest. shareholder confidence, unnecessary
This company is a manufacturer of electrical, material and labor costs, and the real and
lighting, and automotive equipment. The opportunity costs of time spent correcting
products of this company are marketed errors resulting from dirty data. Those
domestically and internationally. The interviewed define dirty data as follows:
The GIGO (garbage in, garbage out) theory
company employs approximately 1,600 people
applies to dirty data. If you don't have checks
in manufacturing and sales facilities located in the system that prevents human error, you
both domestically and internationally. There will have errors in your data. Data integrity
are 17 manufacturing facilities located in refers to data that is systematically edited or
North America and Asia. The case study was edited by ``experts'' after data entry to remove
used to understand the implications of dirty errors (Manager, CSC).
data at the company before and after the
Duplicate data or data that is incomplete or
implementation of an ERP system. The ERP extraneous (Information Analyst, Marketing
implementation in the company replaced a Services).
number of independent mainframe legacy
Anything that is entered incorrectly (CSR).
systems used for order and quotation
processing, manufacturing, transportation, The definitions used reflect each one's
billing, and finance applications. One of the experience with dirty data. Awareness of this
co-authors of the study works at the company problem is growing within the organization
as the system/support supervisor for the as users, systems people, and management
Customer Support Center (CSC). In this role, uncovers and deals with problems resulting
the author was directly involved in from dirty data.
identifying, trouble-shooting, and training Data integrity requires awareness and
for dirty data concerns in data entry and with control of dirty data. A collection of data has
specifying, testing, and distributing integrity if the data are logically consistent
customer and sales-force reports. In addition, and accurate. Data integrity requires that
we interviewed several other employees in data additions or changes be reflected in each
[ 22 ]
3. Jodi Vosburg and Anil Kumar of the locations where that data is stored and Each person's perspective is culled from that
Managing dirty data in that data is consistent across the storage person's training and experience. The CSR
organizations using ERP: medium(s) used. Data integrity also requires
lessons from a case study indicated that she had little understanding of
that the users of that data understand the the way in which the data she enters is used in
Industrial Management &
Data Systems meaning of the data within the context of the peripheral departments and how it becomes
101/1 [2001] 21±31 business. Maintaining data integrity part of reporting. For that reason, it is
requires a systematic approach to data important to examine the data and rationalize
processing, storage, sharing, manipulation, it. Data rationalization involves determining
and reporting. what data is important to which department
and prioritizing the value of those data sets.
Once this determination is made, plans to
4.0 Cost of using dirty data correct and prevent dirty data can be laid.
``Errors in data can cost a company millions
of dollars, alienate customers, and make
implementing new strategies difficult or 5.0 The ERP implementation:
impossible'' (Redman, 1995). The manager of lessons learned
CSC commented that: The start of data integrity problems is really
Any business that has to issue debits and a failure to treat data as a strategic business
credits or that throws out surplus, unusable resource. Scholars (Redman, 1995; Tayi and
inventory, understands the costs of dirty
Ballou, 1998) point out that data is a key
data. Each credit or debit is estimated to cost
organizational resource. However, as pointed
the company $75 for the clerical efforts of
analyzing, generating and disseminating the out by Kilbane (1999), ``Many companies who
document. Added to that are the following: use data contained in legacy systems are not
production errors from erroneous bills of leveraging it as a strategic company asset.''
material or misinterpretation of a customer's The primary challenge to maintaining data
specifications; freight costs for shipping and integrity is the lack of resources allocated to
returning product; inventory scrapping it. To maintain data integrity, people with an
charges where the product cannot be resold; understanding of the origins and results of
financial penalties charged by the customer
dirty data and the ways to prevent and
for our error; ordering of unneeded materials;
scrapping of raw materials; wasted labor correct it, must be dedicated to the task.
charges at the organization and its customer; Redman (1995) says that: ``Due largely to the
warranty charges to fix the product, if it can organizational politics, conflicts, and
be modified; and unknown cost of the passions that surround data, only a
customer not ordering additional product corporation's senior executives can address
from you because of your data problems. The many data quality issues. Only senior
managers and people involved in warranty, management can recognize data (and the
credit and collection and finance understand
processes that produce data) as a basic
the ramifications. The rest of the organization
understands what their managers or corporate asset and implement strategies to
supervisors have shared with them. Our proactively improve them.'' Where data
quality program emphasizes feedback to the integrity is one of many responsibilities of
person involved with a quality problem. It is people with no understanding of the concepts
up to the management team to insure that all surrounding data integrity, dirty data is the
people understand the problems dirty data result. Integrity, issues receive attention in
can cause as well as prevention. times of crisis, but as soon as the crisis is
The information analyst for marketing over, those with responsibilities other than
services was of the opinion that ``most of the data integrity turn to the pressing deadlines
costs associated with dirty data cannot be or daily tasks that they are responsible for. In
measured in terms of dollars. If these costs a complex ERP environment, this can result
could be quantified the management would in perpetual crisis management. In the
be shocked''. She stressed the cost of the following paragraphs we discuss the lessons
endless number of consultants required to learned from this case study.
configure the system to prevent a particular
data problem or to determine or correct the 5.1 Understanding and communicating
results of one. new demands of an ERP system
The CSR focused on ``costs associated with Before the move from legacy applications to
customer dissatisfaction ± lost confidence an ERP system takes place, considerable
and business are hard to measure and harder thought should be given to how the system
to win back''. She pointed out the frustration change will change the roles of the users. The
and time lost at the factory, in the marketing conversion to an ERP system is not just a
departments, and at the CSC in correcting data extraction, cleansing, transformation,
problems resulting from dirty data. and populate process to effectively
[ 23 ]
4. Jodi Vosburg and Anil Kumar implement an ERP system. An organization this way of working. The combination of
Managing dirty data in needs a strategy and a plan. Atre (1998) points these factors has increased the occurrence of
organizations using ERP: out that ``legacy data is invariably in worse
lessons from a case study inaccurate, inconsistent data being entered
condition than you realize''. Caldwell and on the ERP via sales orders, as CSRs attempt
Industrial Management &
Data Systems Stein (1998) comment that ``ultimately, by to complete their complex and time-
101/1 [2001] 21±31 feeling their way through the initial shock of consuming data entry work in the same
an ERP implementation ± new business amount of time they did prior to the ERP
processes, new job roles, new management implementation and without a clear
structures, and new technologies ± understanding of how that data is to be used
companies are transforming themselves''. by other functional areas in the business and
In this company there are 48 CSRs in the by upper management for analysis and
CSC. These CSRs are responsible for entering business decisions.
orders taken from domestic customers. Now, Lesson: Organizational users need to be
with the ERP, the items on these orders not educated and prepared for the changes that
only initiate the manufacturing, shipping, will take place as a result of ERP
and invoicing functions, but also are the raw implementation.
data used to generate the sales and marketing
reports. The sales and marketing reports feed 5.2 Developing shared understanding of
the decision-making processes that steer the data
business. The correct and consistent entering The lack of a shared understanding of the
of these orders is critical to preventing uses and value of data among those
dirty data. performing the same tasks and among those
Most CSRs believe that the order entry performing different tasks can lead to
process has increased in complexity with the creation of dirty data. Tayi and Ballou (1998)
implementation of the ERP. Some estimate point out that ``the data gatherer and initial
that the time required to enter an order has user may be fully aware of the nuances
increased two to four times. The reasons for regarding the meaning of the various data
this widely-held perception are threefold. items, but that will not be true for all of the
First, the ERP is still quite new ± system other users''. Where those performing the
glitches can mean several unsuccessful same tasks have a different understanding of
attempts at entering a single order and the the data being processed, inconsistencies are
eventual involvement of system support inevitable. For example, if the marketing
personnel in processing. Second, there are services department members differ on
more steps to the order entry processes, and whether abbreviations are to be used in
greater variation across product lines. customer master data, inconsistent entry is
Legacy systems were used for narrowly- the result. Locating this data becomes
defined transaction sets. For example, each of difficult for CSRs because they cannot be
the four product groups in the company had sure if they are using the wrong
their own manufacturing system. The abbreviation, or if the data has not been
homogeneity of the transactions and of the entered. The result of this lack of shared
users meant that the legacy systems could be understanding is duplicate records ± when
customized to accommodate those tasks the CSR cannot find the record that they are
without affecting the ability of other users to looking for, a new record is requested. Even
perform other tasks. Now that all users share if marketing services is able to locate the
a single system, transactions must be record and corrects the abbreviation before
generalized to fit all tasks. Where creating a duplicate record, both the CSR and
customization cannot be automated, it marketing services have spent unnecessary
becomes a manual part of user work time.
processes ± the order entry process varies A lack of a shared understanding is
greatly from product line to product line. common among data generators and report
Greater expertise is required on the part of writers. A CSR knows that the promised ship
the user, not only in the performance of their date on an order with a production block is
assigned tasks but also in those of others that not valid, but a consultant writing a backlog
are affected by their system transactions. The report probably does not. As a result, the
learning curve has been steeper than anyone invalid date is published on the report.
imagined. Third, data entry skills are no Geographical distances and functional
longer enough to successfully enter orders ± barriers exacerbate this complexity. The
the ERP requires system savvy and an further an employee is from another
analytical approach. It has become critical employee, and the less that employee
that CSRs understand the logic behind the understands what is required in the other's
processes and the ramifications of their position, the less likely they are to share a
actions on-line. Many are inexperienced in common understanding of the importance of
[ 24 ]
5. Jodi Vosburg and Anil Kumar the data each deals with. According to the ERP data structures that define transactional
Managing dirty data in CSC manager: data, and for authoring and generating sales
organizations using ERP: In the business right now, those entering the and marketing reports. Marketing services
lessons from a case study
data and those using the data are so confused has been successful in guarding against
Industrial Management & that there is little understanding of the data
Data Systems duplicated records, misspelled names,
101/1 [2001] 21±31 in the system. We are working with users inverted text, missing fields, outdated area
AND the IT departments to share the
codes and ZIP codes, and other kinds of dirty
knowledge about the entered, calculated AND
extracted data. Without this, we are, and have data in customer and salesforce master data
been, subject to interpretation of a field with a by employing combination of user training,
title meaning different things to those well-defined procedures, and tight control
entering versus using the data. We are and auditing of additions, changes, and
finding how difficult it is to deal with a deletes. Every CSR received at least four
program written in another language, as field hours of training on the use and import of
translations have always assisted users and this master data. During that training, CSRs
IT people in the past. In our ERP, there is no were asked to review the master data for their
such extra help available for those looking for
assigned customers and to advise marketing
field definitions and understandings.
services of any necessary changes. New
Lesson: Champions of the ERP master data requires the completion of a form
implementation project should ensure that all to ensure all necessary information is
users understand the organizational data in a provided. Only two people in the marketing
manner that is consistent throughout the services department do the actual addition of
organization.
the new data to ERP. In addition, an audit
report is run regularly to identify changes
5.3 Ownership of data and responsibilities made to the data. This report helps to catch
Responsibility for ensuring data integrity
mistakes and identify where additional
belongs to all employees. Tayi and Ballou
training is required. A data steward, who is
(1998) comment: ``The capability of judging the
responsible expressly for protecting data
reasonableness of the data is lost when users
integrity, should support the efforts of the
have no responsibility for the data's integrity
CSRs and the marketing services department.
and when they are removed from the
This data steward would be responsible for
gatherers.'' Atre (1998) points out: ``IT staff
raising awareness about data issues and
need help and cooperation from business
implementing systematic procedures for data
users to identify and cleanse operational data.
auditing and user training.
Users should be primarily responsible for Lesson: Ensuring that all stakeholders of an
determining the business value of data. Don't ERP system understand their responsibilities
rely on systems integrators ± they don't with respect to maintaining data integrity
understand the business value of the data.'' will lead to a better quality system. Data that
One has also to consider the ``politics'' which is a part of an ERP system belongs to an
play an important role. Often managers may organization and not to any individual
agree that they own the data, but may want department or user.
everybody to be involved in cleaning it. The
manager of the CSC commented: 5.4 Migrating legacy data
I believe that data integrity is the Ruber (1999) comments, ``Migrating
responsibility of every company employee. information from departmental databases
All positions, all departments are responsible and transaction-processing systems . . . is a
for insuring the data they are entering, daunting task.'' He goes on to say that the
reviewing or utilizing is error free. It is the ``hardest part is cleansing the data, yet people
responsibility of every manager to make sure tend to underestimate that part of the
the tools are in place to insure data integrity
process.'' Legacy systems in corporations,
for the data they are responsible for. . .. In the
past, users relied on the IT departments to
which were created in different generations,
make sure the edits were in place to make the create major stumbling blocks for migrating
data correct. With ERP systems and more data to integrated application systems. Quick
user controlled systems and input, it is a joint fixes that become embedded in the case of
responsibility. Users must understand legacy systems create complexities that are
systems better and IT personnel must difficult to overcome. Most of these systems
understand business problems better in order are usually lenient with the data that is
for them to work together to achieve the maintained, resulting in lack of data
highest level of data integrity. Too many IT standards or documentation in the form of
people are good programmers but not good
metadata. Before this data is migrated there
business analysts.
is a need to clean it. An effective strategy for
The marketing services staff is responsible companies planning to implement integrated
for maintaining customer and salesforce applications, such as ERP, may be to use
master data, for testing and maintaining the automated tools for cleaning legacy data
[ 25 ]
6. Jodi Vosburg and Anil Kumar before integrating it. Tools provided by Customer master data was loaded
Managing dirty data in vendors such as id.Centric, Vality programmatically initially. Customer master
organizations using ERP: data includes addressing for billing and
lessons from a case study Technology, HarteHanks, etc. (Knowles, 1997)
may benefit organizations significantly. shipping, tax identification numbers and
Industrial Management &
Data Systems Sales of such tools used for data extraction, designations of customer type and pricing
101/1 [2001] 21±31 refining and loading, was expected to reach levels. Migration from legacy systems to the
$210 million by the end of 1999 (Kay, 1997). ERP has allowed marketing services an
The initial ERP implementation involved a opportunity to scrutinize and clean data
programmatic load of legacy sales order maintained about our customers and sales
backlog onto the ERP. The order load force. More stringent master data
program was developed and tested over a requirements in the ERP, in fact, made this a
period of months by a programmer familiar necessity. For example, the legacy system
with the organization's business practices had used an address in Varnons, Georgia, for
and a team of users. The load was simplified one customer for years. This address was
by the fact that the legacy system was well kicked out in the programmatic load of
supported. That support meant not only that customer master data. On investigation, it
data to be converted was relatively clean, but was discovered that Varnons was not a city,
but a stop on a railway line. The ERP will
also that the data in the legacy system was
determine the zip code and county given a
well-defined and understood ± a program
city and state. This not only ensures that the
could be written to capture only relevant
city and state are entered accurately, but also
data. Unfortunately, the idiosyncrasies of the
that the customer has provided a valid city
order entry process for the various product
and state combination. Customer data moved
lines and of the ways in which CSRs entered
from the legacy to the ERP system were
the orders meant that no program could
relatively cleaner than they had been on the
convert the data without some errors.
well-maintained legacy system.
Because the data integrity of the orders was
so important, each converted order was Migrating poorly-maintained legacy data
reviewed on the ERP by the responsible CSR. Atre (1998) comments that ``you are likely to
Many data errors were caught and corrected run into problems such as incompatible data
during this review, including item quantity, formats, codes that no one can decipher, data
material number, and ship to errors. But that's embedded in long text fields,
some were missed. A tremendous amount of overlapping customer records from multiple
time has been spent and is still being spent to systems, some with redundant data and
correct these errors. One of the most common others with conflicting or outdated data and
errors involved contract release orders. The even chunks of mystery data of long-
program designed to convert the data forgotten provenance and uncertain
ownership''. Weston (1998) suggests using
somehow selected and input the wrong
flags for dirty data that is migrated. As a
material number into the converted order.
result, a decision-maker can decide if he/she
The CSRs, possibly tired after consecutive 12-
wants to use the information or leave it out
hour days of data verification, missed many
during data analysis activities. The customer
material errors. These kinds of errors,
and salesforce master load for the migration
though, are always found eventually ±
from a much less well-maintained system
usually by the customer. The results of these
was a more tedious and difficult procedure.
errors were: shipment of the wrong
Keeping that data clean on this legacy system
materials; angry customers; time spent was never a priority. The data was entered by
investigating the error; cost of processing the order entry group, as there was no
credit orders and replacement orders; position assigned to the management and
expedited production of the correct materials control of master data for this system.
(resulting in late shipments of other orders); Misspellings, duplicate records, and
transportation costs for returning the wrong inconsistencies were the result of a lack of
units; and/or cost of scrap or storage. No control over who could add, change, or delete
attempt has been made to assign a dollar customer master data, of instructions for
value, though, to the results of this dirty data. proper management of the data, and of
Overall, though, this data migration was a auditing procedures. The problems were
success ± the inevitable data errors were exacerbated by the fact that, when the
identified, some sooner, some later, and company purchased this facility, a
corrected. The success was due in large part completely new group of users began to enter
to the fact that the legacy system was well this data. A lack of shared definitions of the
supported, the migration process was well components of the master data and their uses
tested and documented, and those closest to increased the number of discrepancies and
the data verified the data after the migration. errors. Where the original group might
[ 26 ]
7. Jodi Vosburg and Anil Kumar define a salesperson as a customer or a and some in a closed status, some in an open
Managing dirty data in vendor or an agent, the company group status. Attempts to suppress this data on the
organizations using ERP: defined the salesperson as an agent only. conversion order might have, without
lessons from a case study
Subsequently, where there was no agent-type extensive testing, resulted in inadvertent
Industrial Management &
Data Systems record for a particular salesperson, one was suppression of materials that should be
101/1 [2001] 21±31 created, thereby creating the potential for converted ± the Miami order entry location
inaccurate reporting of sales data. Thus, uses freight items not to communicate
before any master data could be moved to the shipment information, but to charge the
ERP, each record had to be manually customer.
reviewed. The marketing services group This project, though, was also a success.
again handled this process. Spelling errors, While the manual conversion presented an
duplicate records, and incomplete data were opportunity for entry error, the process was
addressed before the data was loaded largely error free. This can be attributed to
to the ERP. the extensive testing of the backlog report
The sales order and production data on this serving as the basis for the conversion,
system had been subject to inexplicable simple comprehensive check-list style
changes. For example, in November of 1998, instructions for the CSRs in the use of the
the order entry group started to notice that backlog report, and, most importantly, a
some items on orders were being closed by group of CSRs now more comfortable and
the system for no apparent reason. Thus, experienced in the use of the current ERP.
they would never be built or shipped. The in- Again, migration to the new ERP was a boon,
house support could not identify the cause or because it drove the process of examining
propose a solution, nor could the and cleaning current data.
manufacturer of the software. The in-house Lesson: Migrating dirty data is a challenging
support group advised CSRs to address these task. Use of automated tools is a good strategy
for organizations planning to implement
system-generated cancellations as they
integrated application systems. The most
happened ± a virtually impossible task. After important factor is that the data needs to be
much discussion, the support team agreed to cleaned before it is migrated to an ERP
write a report to locate these items. system.
The data on this legacy system was not well
supported or understood. The data was in 5.5 Recognizing the complexity of
such poor condition that sales and marketing integrated data
reports generated from system data were The integration of several business functions
virtually useless. For example, the Canadian on a single system holds tremendous
order entry location might enter orders using potential for reporting. All transactional data
the same customer master record for is now available from one source. Reporting
different customer locations by overwriting that was difficult or not feasible in the past is
the sold-to-address text to reflect the different now possible. This consolidation of functions
location addresses. The domestic location onto one system has forced the various units
would add new customer master records for of the business to develop a greater
each customer location. Existing reports understanding of the work done by other
could not accurately reflect these units of the business and their interpretation
contradictory approaches. of the data. With this potential, though,
These factors combine to make a comes increased complexity. Tayi and Ballou
programmatic migration of the sales order (1998) point out ``personnel databases
data to the ERP infeasible. Instead, sales situated in different divisions of a company
orders were manually loaded onto the ERP by may be correct but unfit for use if the desire
CSRs using an expanded backlog report. The is to combine the two and they have
lack of understanding of the way the system incompatible formats''. Kilbane (1999) says
stores data, coupled with inaccuracies and ``the problem is that data is, too often, in
inconsistencies in order entry and different formats and companies don't know
processing, made the writing of this report how to properly bring it together and turn it
very difficult. For example, the initial run of into actionable information''.
the report included thousands of freight Locating data tables within the ERP system
items. Freight items are added to sales orders appropriate for the intended reporting has
by the shipping department to indicate turned out to be more tedious and difficult
carrier and shipment date of materials on the than anyone imagined. Reports used by the
order ± they are not backlogged. These were salesforce and in manufacturing to describe
difficult to suppress in the report because of sales order backlog have been found to be so
the inconsistent ways in which they have error-ridden that they have been totally
been added to the sales orders ± some were scrapped and rebuilt. Reports meant to
loaded as text items, some as freight items, describe incoming businesses took months to
[ 27 ]
8. Jodi Vosburg and Anil Kumar write. Several iterations of these reports suspicions were confirmed in March, when it
Managing dirty data in were developed before the set currently in was discovered why incoming business
organizations using ERP: circulation was completed. numbers seemed too high. CSRs had been
lessons from a case study
The information analyst describes an error entering the sales credit designation on
Industrial Management &
Data Systems that she stumbled across while researching orders more than once. Whenever a new item
101/1 [2001] 21±31 another reporting data discrepancy. It seems is added to an existing ERP sales order, the
that the same incoming business report was ERP returns an error indicating that sales
run for the month of March on April 1 and credit is missing. The correct action is to
then again on April 3. She noticed that the activate the existing sales credit designation
totals were different. This should never occur on the order for the new items. This problem
± once the month is closed, no updating was not anticipated or clearly understood.
should occur. She indicates that locating the Thus the correct handling was never made
cause of a problem like this is difficult and part of the ERP training for CSRs. So, CSRs
time consuming and sometimes proves to be generally entered an additional sales credit
impossible. In this case, though, they were designation with each addition to an order.
able to locate the source of the problem ± the Some orders showed a sales credit allocation
reporting structure was identifying the of 400 per cent or more of the net value of the
wrong date field as the determinant for order. The sales credit numbers are also used
which month a particular type of order would to report incoming business. In total, this
be allocated to. The correction of this data entry error resulted in an eight million-
structure error is perhaps more tedious than dollar overstatement of incoming business.
finding the cause of it ± the field reference Because this affected incoming business and
must be changed in more than 100 places in not shipments or production, the cost was
each of the several data structures. These and minimal financially. However, sales
other integrity problems detected in the early managers were forced to adjust sales
going have meant that several manual engineer bonuses downward as a result of the
adjustment schedules must be published with discovery.
each run of this report ± the data cannot be This data cannot be corrected on the ERP
cleaned. The information analyst attributes system. All adjustments had to be handled
these errors to a lack of comprehensive manually. Some preventative measures were
testing of the updating that occurs when immediately put in place. In the short term,
these orders are processed. She sights a lack additional training was provided to the
of communication between those that people who enter orders to make them aware
understand the way the company accrues of the impact of this error. A daily report is
and processes data and those responsible for being run to identify these errors as they are
building the data definition structures. As a made, allowing on-line corrections. In the
result, some basic assumptions were made in long term, the ERP configuration changes
the definition of data that were incorrect. have been requested to eliminate the
The complexity entailed by system misleading error message and to add
integration is compounded by the marketing messages when more than 100 per cent of the
services staff's inexperience with the selected value of the order is allocated as sales credit.
reporting bolt-on, the ERP data structures, According to the manager of the CSC: ``The
and the architecture of the data itself. Basic problem might have been prevented if we all
reporting requirements to operate the knew how to test wrong. In all the massive
business, coupled with this inexperience, testing done on order entry and reporting on
have resulted in an inordinate reliance on it, not enough was done to try to enter bad
consultants for report writing. While these data. Some of the edits seemed so self-evident,
consultants are skilled in report writing and that there lack was almost impossible to
the integration of ERP, their lack of comprehend. I think we are just now learning
understanding of company business and the how important understanding and testing for
transactional data and processes, and dirty data is in a truly integrated system.''
subsequent ERP configuration changes, has Lesson: Test, test and test again. Testing is a
impeded accurate reporting. crucial aspect of implementing ERP
Lesson: It takes time for users to comprehend solutions. There should be no short-cuts in
and use integrated data as a result of using testing. Different user groups should be
ERP packages. Care should be taken to ensure involved in the testing process to ensure that
that all users understand the concept of all possible scenarios are used for testing the
integrated corporate data and use it ERP system before the conversion to ERP is
accordingly. implemented.
5.6 Testing the new system 5.7 Training
The costs of insufficient testing prior to Lack of proper training can frustrate users
implementation can be very high. Months of when they begin using an ERP system in an
[ 28 ]
9. Jodi Vosburg and Anil Kumar organization. Caldwell and Stein (1998) point Even marketing services, though, does not
Managing dirty data in out the example of Amoco, where ``managers have a system in place to check data
organizations using ERP:
lessons from a case study found SAP so unfriendly they refused to use it. regularly for problems. The information
Few [of our] people use SAP directly because analyst indicates that the department spends
Industrial Management &
Data Systems you have to be an expert''. The authors further so much time ``putting out fires'' that there is
101/1 [2001] 21±31 comment that in the case of Owens Corning, little time left over for carrying out
the organization found out that ``the cultural systematic data checks. The problem is
and organizational impact on IT organizations exacerbated by a lack of tools for auditing.
is a little short of revolutionary''. The entry The information analyst attributes this to the
and extraction of dirty data can be prevented newness of the implementation.
with greater dedication to initial and on-going At present, data integrity is protected
training for those responsible for entering and through a combination of system safeguards,
extracting data. A lack of time is typically user training, and data entry procedures.
sighted as the reason for inadequate training. System safeguards are the result of building
The time required to investigate, understand, data integrity rules into ERP. For example,
correct, and prevent problems due to dirty ERP will prevent a CSR from entering a ship
data is considerably more, though, than that to address in a sold-to-field. This is a hard
required simply to understand and prevent error, preventing saving of the data. Soft
those problems. The additional cost of this error messages give the CSR the opportunity
reactive approach is the loss of shareholder to review potentially erroneous data.
confidence in the system, employees, and Additionally, many fields are populated from
data. A significant training effort was put into drop-down boxes, eliminating the chance for
teaching those that would be using and misspelled entries or entries outside the
entering data in the system. Each CSR acceptable domain for the field.
received in excess of 50 hours of training in Lesson: Organizations should emphasize that
meaning and population of the various fields maintaining data integrity is an on-going
comprising the order entry screens. Order process and everybody needs to play an active
entry procedures are documented in detail role. Maintaining data integrity does not stop
and available to all CSRs. The difficulty lies in with the implementation of the ERP system.
knowing how much training is enough ± a
difficult question to answer at conversion 5.9 Using consultants
Care must be taken to ensure that if
time, given the consultants' lack of
consultants are hired for the transition
understanding of the particular business and
the employees' lack of understanding of the project, the internal stewards of the system
new system and the potential problem areas. understand their work. For example, in this
There is no question, though, that additional company, consultants were responsible in
training will be required after large part for developing data structures for
implementation to address the numerous the new system, and form the system
unanticipated problems that will arise. metadata. These structures are used in
Lesson: On-going training is a prerequisite for conjunction with raw data to define the
success in implementing ERP systems. context of the data and to ensure that data
Organizations should plan ahead of time to reported is consistent with what is intended
train all users before and after the or required. For example, a structure may
implementation. Periodic exchange of ERP define incoming business as the value of the
experiences by users in an organization from selling prices on sales orders not including
their work environment will go a long way. items that have been cancelled. Thus reports
providing incoming business data will not
5.8 Prioritizing data maintenance include cancelled items. Direct involvement
According to the CSC manager: ``Data
by the manager of the CSC and the marketing
integrity is assigned a high priority at the
services staff throughout the development,
management and IT level. It is not as high a
ensured that data structures defined by the
priority at the middle manager and lower
consultants matched data the way users of
levels, as worrying about data integrity can
that data defined it. This prevents the
slow down production, order entry, shipping,
possibility that once the consultants leave
etc..'' The information analyst and the CSR
the project, the users of the system
expressed similar opinions when asked about
understand the data that is being processed
the prioritization of data integrity at CPS.
by the system.
The information analyst indicated that data Lesson: Hiring consultants to assist with the
integrity was critical in the marketing ERP implementation is an effective strategy if
services department, but prioritized much organizations ensure that all work done by
lower in departments dealing with the day-to- the consultants is understood and
day operations. documented. The ERP implementation
[ 29 ]
10. Jodi Vosburg and Anil Kumar knowledge should not leave the organization problems should be systematically
Managing dirty data in after the consultants work is completed. documented and stored so as to be easily
organizations using ERP:
lessons from a case study accessible to interested users. If a similar
5.10 Post-ERP implementation problem occurs, documentation of other
Industrial Management &
Data Systems Counteracting and preventing dirty data ± similar instances would be readily available.
101/1 [2001] 21±31 current perceptions and practices Where necessary, the communication should
Data entry procedures have been created to be followed up by training.
control the potential damage accruing to Regular training sessions should also be
dirty data. For example, CSRs are required to scheduled to ensure that users understand
place a production block on orders for some data integrity concepts and methods. These
material types. This gives the product group sessions would not only build a shared
marketing departments an opportunity to interpretation of data and preferred
review the order and correct any errors processing methods, but would also foster a
before production begins. Taken more global perspective on the part of the
individually, this procedure seems like a users ± instead of seeing only their own role,
reasonable safeguard. Taken together with users would see their role in the context of
all of the other exceptions and qualifiers to the business. This perspective would assist in
the basic order entry procedure based on paring down some of the current procedural
material type, or product line, and order complexity. Simpler procedures would
type, though, the procedures begin to seem further increase data accuracy.
like the source of the errors rather than the Equipped with an understanding of the
way to avoid them. As the procedures grow impact of their work on other areas of the
more complex, the likelihood of entering data
business, users can be analysts rather than
accurately and consistently drops.
data entry clerks. Analysts can make good
In almost every department within the
decisions in a complex and dynamic work
business, the increased complexity of
environment. Broadly-trained analysts
performing the job has meant more time
would also be in a position to work effectively
required to do the same work. This means
with consultants thus reducing our reliance
even less time and attention to data integrity
on them.
issues and more dirty data ± where someone
Performance measures should be taken
may have taken the time to find out what an
regularly to gauge the effectiveness of and to
error message means and to address the data
improve on the system and training
entry error prior to the implementation of
initiatives. All of these measures would
ERP, now they may pass the error without
directly improve data integrity and would
addressing it because of the work backlog.
serve to underline the importance of data
Out of necessity, though, where data
integrity to all users. These measures would
integrity is compromised, user involvement
reduce errors in carrying out tasks
into testing and reporting procedures is
throughout the business and all their
increasing. The correction will begin with an
associated costs and help to draw a sharper
investigation by system/support and/or
marketing services of the problem. Then, picture of the business to improve long- and
reporting tools will be generated to find all short-term decision-making.
instances of the error. Finally, users will be
enlisted to implement corrections. CSR
involvement in the corrections is critical 6.0 Conclusion
because of their intimacy with the data and Implementing ERP systems requires
as a training tool ± those repeating the error reinventing the business. Several legacy
most frequently will have the most systems are integrated in the process with a
corrections to make. The more corrections single integrated system for managing
that the CSR is required to make, the greater operations across the organization. Data that
the likelihood that they will be able to avoid resided in dozens of disparate sources is now
the mistake in the future. available through one integrated system for
Counteracting and preventing dirty data ± all users in an organization. To achieve
areas for improvement success in ERP systems implementation,
A systematic approach should replace the project champions should make sure that
more reactive crisis management approach they address the relevant issues. Some of the
to data integrity. Data audits should include key lessons from this study include, among
daily integrity checks within the system and others, the following issues:
regular audits performed by user groups. . The champion of the ERP implementation
Problems uncovered in those audits should project should ensure that the
be shared with all affected parties. The transformation is not viewed as an IT
causes, effects, and resolutions of those initiative, rather a business necessity.
[ 30 ]
11. Jodi Vosburg and Anil Kumar This requires educating the stakeholders from this case study would be valuable for
Managing dirty data in about the transition to an ERP. organizations planning to implement ERP
organizations using ERP: . The champion for the change to ERP systems.
lessons from a case study
should recognize the value of data as an
Industrial Management &
Data Systems organizational resource and educate users References
101/1 [2001] 21±31 about it. The issue of sharing corporate Atre, S. (1998), ``Beware dirty data'',
data and assigning responsibilities for Computerworld, Vol. 32 No. 38, pp. 67-9.
managing it should be done with a view to Caldwell, B. and Stein, T. (1998), ``New IT agenda'',
avoid any political issues arising from Information Week, No. 711, November, p. 30.
owners of disparate data sources. Ferriss, P. (1998), ``Insurers victims of DBMS
. The ERP implementation should be fraud'', Computing Canada, Vol. 24 No. 36,
planned to prepare users for the change. 28 September, pp. 13-15.
The expectations based on new Greengard, S. (1998), ``Don't let dirty data derail
responsibilities should be outlined you'', Workforce, Vol. 77 No. 11, November,
upfront to avoid any conflicts. pp. 107-8.
. The user community should be given time Horwitz, A.S. (1998), ``Ensuring the integrity of
your data'', Beyond Computing, Vol. 7 No. 4,
to accept the changes in their work
May.
environment to minimize the impact on
Kay, E. (1997), ``Dirty data challenges warehouses'',
organizational culture, such as
Software Magazine, October, pp. S5-S8.
overcoming comments like ``we've always
Kilbane, D. (1999), ``Are we overstocked with
done it this way''. Users should be
data'', Automatic I.D. News, Cleveland, OH.
encouraged to use the new system by
Vol. 15 No. 11, October, pp. 75-9.
providing incentives. Knowles, A. (1997), ``Dump your dirty data for
. All data that is migrated to an ERP system added profits'', Datamation, Vol. 43 No. 9,
should be cleaned before the migration. September, pp. 80-2.
Automated tools for data migration can be Redman, T.C. (1995), ``Improve data quality for
very useful for companies. competitive advantage'', Sloan Management
. Training users on a continual basis is Review, Cambridge, Vol. 36 No. 2, Winter.
very important. It is important that users Ruber, P. (1999), ``Migrating data to a warehouse'',
do not get bogged down by activities that Beyond Computing, November/December,
take up too much of their time. pp. 16-20.
. Extensive testing is required for Sellar, S. (1999), ``Dust off that data'', Sales and
implementing ERP systems. A good strategy Marketing Management, New York, NY,
would be to phase-in the implementation Vol. 151 No. 5, May, pp. 71-3.
rather than making a direct conversion. Stankovic, N. (1998), ``Dual access: lower costs,
. Consultants experienced with ERP tighten integration'', Computing Canada,
implementation can be very helpful. Care Vol. 24 No. 27, July, p. 30.
must be taken to ensure that all the work Tayi, G.K. and Ballou, D.P. (1998), ``Examining
done by consultants is documented for data quality'', Communications of the ACM,
future use. New York, NY, Vol. 41 No. 2, February,
pp. 54-7.
In this paper, we listed and discussed issues Wallace, B. (1999), ``Data quality moves to the
pertaining to ERP implementation. Though forefront'', Information Week Online,
implementation in different organizations 30 September.
can vary based on the organizational culture Weston, R. (1998), ``Using dirty data'',
and business needs we feel that the lessons Computerworld, Vol. 32 No. 22, 1 June, p. 54.
[ 31 ]