Businesses cannot compete without data. Every organization produces and consumes it. Data trends are hitting the mainstream and businesses are adopting buzzwords such as Big Data, data vault, data scientist, etc., to seek solutions for their fundamental data issues. Few realize that the importance of any solution, regardless of platform or technology, relies on the data model supporting it. Data modeling is not an optional task for an organization’s data remediation effort. Instead, it is a vital activity that supports the solution driving your business.
This webinar will address emerging trends around data model application methodology, as well as trends around the practice of data modeling itself. We will discuss abstract models and entity frameworks, as well as the general shift from data modeling being segmented to becoming more integrated with business practices.
Takeaways:
How are anchor modeling, data vault, etc. different and when should I apply them?
Integrating data models to business models and the value this creates
Application development (Data first, code first, object first)
Introduction
Big Data may well be the Next Big Thing in the IT world.
Big data burst upon the scene in the first decade of the 21st century.
The first organizations to embrace it were online and startup firms. Firms like Google, eBay, LinkedIn, and Face book were built around big data from the beginning.
Like many new information technologies, big data can bring about dramatic cost reductions, substantial improvements in the time required to perform a computing task, or new product and service offerings.
DAMA, Oregon Chapter, 2012 presentation - an introduction to Data Vault modeling. I will be covering parts of the methodology, comparison and contrast of issues in general for the EDW space. Followed by a brief technical introduction of the Data Vault modeling method.
After the presentation i I will be providing a demonstration of the ETL loading layers, LIVE!
You can find more on-line training at: http://LearnDataVault.com/training
Want to see a high-level overview of the products in the Microsoft data platform portfolio in Azure? I’ll cover products in the categories of OLTP, OLAP, data warehouse, storage, data transport, data prep, data lake, IaaS, PaaS, SMP/MPP, NoSQL, Hadoop, open source, reporting, machine learning, and AI. It’s a lot to digest but I’ll categorize the products and discuss their use cases to help you narrow down the best products for the solution you want to build.
Introduction
Big Data may well be the Next Big Thing in the IT world.
Big data burst upon the scene in the first decade of the 21st century.
The first organizations to embrace it were online and startup firms. Firms like Google, eBay, LinkedIn, and Face book were built around big data from the beginning.
Like many new information technologies, big data can bring about dramatic cost reductions, substantial improvements in the time required to perform a computing task, or new product and service offerings.
DAMA, Oregon Chapter, 2012 presentation - an introduction to Data Vault modeling. I will be covering parts of the methodology, comparison and contrast of issues in general for the EDW space. Followed by a brief technical introduction of the Data Vault modeling method.
After the presentation i I will be providing a demonstration of the ETL loading layers, LIVE!
You can find more on-line training at: http://LearnDataVault.com/training
Want to see a high-level overview of the products in the Microsoft data platform portfolio in Azure? I’ll cover products in the categories of OLTP, OLAP, data warehouse, storage, data transport, data prep, data lake, IaaS, PaaS, SMP/MPP, NoSQL, Hadoop, open source, reporting, machine learning, and AI. It’s a lot to digest but I’ll categorize the products and discuss their use cases to help you narrow down the best products for the solution you want to build.
Data Warehousing Trends, Best Practices, and Future OutlookJames Serra
Over the last decade, the 3Vs of data - Volume, Velocity & Variety has grown massively. The Big Data revolution has completely changed the way companies collect, analyze & store data. Advancements in cloud-based data warehousing technologies have empowered companies to fully leverage big data without heavy investments both in terms of time and resources. But, that doesn’t mean building and managing a cloud data warehouse isn’t accompanied by any challenges. From deciding on a service provider to the design architecture, deploying a data warehouse tailored to your business needs is a strenuous undertaking. Looking to deploy a data warehouse to scale your company’s data infrastructure or still on the fence? In this presentation you will gain insights into the current Data Warehousing trends, best practices, and future outlook. Learn how to build your data warehouse with the help of real-life use-cases and discussion on commonly faced challenges. In this session you will learn:
- Choosing the best solution - Data Lake vs. Data Warehouse vs. Data Mart
- Choosing the best Data Warehouse design methodologies: Data Vault vs. Kimball vs. Inmon
- Step by step approach to building an effective data warehouse architecture
- Common reasons for the failure of data warehouse implementations and how to avoid them
Data preprocessing techniques are applied before mining. These can improve the overall quality of the patterns mined and the time required for the actual mining.
Some important data preprocessing that must be needed before applying the data mining algorithm to any data sets are completely described in these slides.
Know different types of tips about Importance of dataware housing, Data Cleansing and Extracting etc . For more details visit: http://www.skylinecollege.com/business-analytics-course
Loan Prediction system is a system which provides you a interface for loan approval to the applicants application of loan. Applicants provides the system about their personal information and according to their information system gives his status of availability of loan.
Smart Health Prediction Using Data Mining.Data mining is a new powerful technology which is of high interest in computer world. It is a sub field of computer science that uses already existing data in different databases to transform it into new researches and results. It makes use of Artificial Intelligence, machine learning and database management to extract new patterns from large data sets and the knowledge associated with these patterns. The actual task is to extract data by automatic or semi-automatic means. The different parameters included in data mining includes clustering, forecasting, path analysis and predictive analysis.
Microsoft Data Platform - What's includedJames Serra
The pace of Microsoft product innovation is so fast that even though I spend half my days learning, I struggle to keep up. And as I work with customers I find they are often in the dark about many of the products that we have since they are focused on just keeping what they have running and putting out fires. So, let me cover what products you might have missed in the Microsoft data platform world. Be prepared to discover all the various Microsoft technologies and products for collecting data, transforming it, storing it, and visualizing it. My goal is to help you not only understand each product but understand how they all fit together and there proper use case, allowing you to build the appropriate solution that can incorporate any data in the future no matter the size, frequency, or type. Along the way we will touch on technologies covering NoSQL, Hadoop, and open source.
Building an Effective Data Warehouse ArchitectureJames Serra
Why use a data warehouse? What is the best methodology to use when creating a data warehouse? Should I use a normalized or dimensional approach? What is the difference between the Kimball and Inmon methodologies? Does the new Tabular model in SQL Server 2012 change things? What is the difference between a data warehouse and a data mart? Is there hardware that is optimized for a data warehouse? What if I have a ton of data? During this session James will help you to answer these questions.
This presenation explains basics of ETL (Extract-Transform-Load) concept in relation to such data solutions as data warehousing, data migration, or data integration. CloverETL is presented closely as an example of enterprise ETL tool. It also covers typical phases of data integration projects.
Big Data presence in the high volume in the data storages can help in various ways to learn more about the need and trends of the current market which will be useful for all type of organizations. Modern information technology used to analyze the relationship between social trends and market insights is a useful way to have indirectly interlinked to customers and their interests from unstructured and semi-structured data. Such analysis will give organizations a broader view towards the practical needs of customers and once banking industry or any industry could know the customers, they can serve better and with more flexibility. In this presentation, team has primarily created the platform and designed the architecture in big data technology for banking industry to maximize the users of credit card.
The Data Driven Enterprise - Roadmap to Big Data & Analytics SuccessBigInsights
The Data Driven Enterprise - Roadmap to Big Data & Analytics Success
Presentation used at the series of Breakfast seminar around Australia hosted by Lenovo/Intel/SAP/EY
This lecture gives various definitions of Data Mining. It also gives why Data Mining is required. Various examples on Classification , Cluster and Association rules are given.
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
Data modeling is a process used to define and analyze data requirements needed to support the business processes within the scope of corresponding information systems in organizations.
Data Warehousing Trends, Best Practices, and Future OutlookJames Serra
Over the last decade, the 3Vs of data - Volume, Velocity & Variety has grown massively. The Big Data revolution has completely changed the way companies collect, analyze & store data. Advancements in cloud-based data warehousing technologies have empowered companies to fully leverage big data without heavy investments both in terms of time and resources. But, that doesn’t mean building and managing a cloud data warehouse isn’t accompanied by any challenges. From deciding on a service provider to the design architecture, deploying a data warehouse tailored to your business needs is a strenuous undertaking. Looking to deploy a data warehouse to scale your company’s data infrastructure or still on the fence? In this presentation you will gain insights into the current Data Warehousing trends, best practices, and future outlook. Learn how to build your data warehouse with the help of real-life use-cases and discussion on commonly faced challenges. In this session you will learn:
- Choosing the best solution - Data Lake vs. Data Warehouse vs. Data Mart
- Choosing the best Data Warehouse design methodologies: Data Vault vs. Kimball vs. Inmon
- Step by step approach to building an effective data warehouse architecture
- Common reasons for the failure of data warehouse implementations and how to avoid them
Data preprocessing techniques are applied before mining. These can improve the overall quality of the patterns mined and the time required for the actual mining.
Some important data preprocessing that must be needed before applying the data mining algorithm to any data sets are completely described in these slides.
Know different types of tips about Importance of dataware housing, Data Cleansing and Extracting etc . For more details visit: http://www.skylinecollege.com/business-analytics-course
Loan Prediction system is a system which provides you a interface for loan approval to the applicants application of loan. Applicants provides the system about their personal information and according to their information system gives his status of availability of loan.
Smart Health Prediction Using Data Mining.Data mining is a new powerful technology which is of high interest in computer world. It is a sub field of computer science that uses already existing data in different databases to transform it into new researches and results. It makes use of Artificial Intelligence, machine learning and database management to extract new patterns from large data sets and the knowledge associated with these patterns. The actual task is to extract data by automatic or semi-automatic means. The different parameters included in data mining includes clustering, forecasting, path analysis and predictive analysis.
Microsoft Data Platform - What's includedJames Serra
The pace of Microsoft product innovation is so fast that even though I spend half my days learning, I struggle to keep up. And as I work with customers I find they are often in the dark about many of the products that we have since they are focused on just keeping what they have running and putting out fires. So, let me cover what products you might have missed in the Microsoft data platform world. Be prepared to discover all the various Microsoft technologies and products for collecting data, transforming it, storing it, and visualizing it. My goal is to help you not only understand each product but understand how they all fit together and there proper use case, allowing you to build the appropriate solution that can incorporate any data in the future no matter the size, frequency, or type. Along the way we will touch on technologies covering NoSQL, Hadoop, and open source.
Building an Effective Data Warehouse ArchitectureJames Serra
Why use a data warehouse? What is the best methodology to use when creating a data warehouse? Should I use a normalized or dimensional approach? What is the difference between the Kimball and Inmon methodologies? Does the new Tabular model in SQL Server 2012 change things? What is the difference between a data warehouse and a data mart? Is there hardware that is optimized for a data warehouse? What if I have a ton of data? During this session James will help you to answer these questions.
This presenation explains basics of ETL (Extract-Transform-Load) concept in relation to such data solutions as data warehousing, data migration, or data integration. CloverETL is presented closely as an example of enterprise ETL tool. It also covers typical phases of data integration projects.
Big Data presence in the high volume in the data storages can help in various ways to learn more about the need and trends of the current market which will be useful for all type of organizations. Modern information technology used to analyze the relationship between social trends and market insights is a useful way to have indirectly interlinked to customers and their interests from unstructured and semi-structured data. Such analysis will give organizations a broader view towards the practical needs of customers and once banking industry or any industry could know the customers, they can serve better and with more flexibility. In this presentation, team has primarily created the platform and designed the architecture in big data technology for banking industry to maximize the users of credit card.
The Data Driven Enterprise - Roadmap to Big Data & Analytics SuccessBigInsights
The Data Driven Enterprise - Roadmap to Big Data & Analytics Success
Presentation used at the series of Breakfast seminar around Australia hosted by Lenovo/Intel/SAP/EY
This lecture gives various definitions of Data Mining. It also gives why Data Mining is required. Various examples on Classification , Cluster and Association rules are given.
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
Data modeling is a process used to define and analyze data requirements needed to support the business processes within the scope of corresponding information systems in organizations.
C* Summit 2013: Data Modelers Still Have Jobs - Adjusting For the NoSQL Envir...DataStax Academy
Using concrete, real-world examples, the presenter will show the following: How abandoning modeling altogether is a recipe for disaster, even in—or especially in—NoSQL environments; How experienced relational modelers can leverage their skills for NoSQL projects; How the NoSQL context both simplifies and complicates the modeling endeavor.How lessons learned modeling for NoSQL projects can make you a more effective modeler for any kind of project.
Visualising Energistics WITSML XML Data Structures in Data Models. ECIM E&P conference, Haugesund Norway, September 2013.
chris.bradley@dmadvisors.co.uk
How to insert references and bibliography into your Word documentSylvia Matovu
This is a feature that many people ignore while working in MS Word even though it is available. Hopefully this presentation makes referencing and compiling a bibliography easier for the user.
Further discussion on Data Modeling with Apache Cassandra. Overview of formal data modeling techniques as well as practical. Real-world use cases and associated data models.
Database Normalization
The term Normalization is a process by which we can efficiently organize the data in a database. It associates relationship between individual tables according to policy designed both to care for the data and to create the database more flexible by eliminating redundancy and inconsistent dependency.
In other words, Database normalization is a process by which a presented database is tailored to bring its component tables into compliance with a sequence of progressive standard forms. It is an organized way of ensuring that a database construction is appropriate for general purpose querying and also includes the functions of insertion, deletion and updating.
Edgar Frank Codd was the person who introduced the process of database normalization firstly in his paper called A Relational Model of Data for Large Shared Data Banks. The two main objective of database normalization is eliminating redundant data and ensuring data dependencies make sense and make sure that every non-key column in every table is directly reliant on the key and the whole key.
Redundant data or unnecessary data will take more and more space in the database and later, creates the maintenance problem in the database. If data that exists in more than one place must be changed because it wastes disk space and the data must be changed in exactly the same way in all locations of the table.
The Information Technology have led us into an era where the production, sharing and use of information are now part of everyday life and of which we are often unaware actors almost: it is now almost inevitable not leave a digital trail of many of the actions we do every day; for example, by digital content such as photos, videos, blog posts and everything that revolves around the social networks (Facebook and Twitter in particular). Added to this is that with the "internet of things", we see an increase in devices such as watches, bracelets, thermostats and many other items that are able to connect to the network and therefore generate large data streams. This explosion of data justifies the birth, in the world of the term Big Data: it indicates the data produced in large quantities, with remarkable speed and in different formats, which requires processing technologies and resources that go far beyond the conventional systems management and storage of data. It is immediately clear that, 1) models of data storage based on the relational model, and 2) processing systems based on stored procedures and computations on grids are not applicable in these contexts. As regards the point 1, the RDBMS, widely used for a great variety of applications, have some problems when the amount of data grows beyond certain limits. The scalability and cost of implementation are only a part of the disadvantages: very often, in fact, when there is opposite to the management of big data, also the variability, or the lack of a fixed structure, represents a significant problem. This has given a boost to the development of the NoSQL database. The website NoSQL Databases defines NoSQL databases such as "Next Generation Databases mostly addressing some of the points: being non-relational, distributed, open source and horizontally scalable." These databases are: distributed, open source, scalable horizontally, without a predetermined pattern (key-value, column-oriented, document-based and graph-based), easily replicable, devoid of the ACID and can handle large amounts of data. These databases are integrated or integrated with processing tools based on the MapReduce paradigm proposed by Google in 2009. MapReduce with the open source Hadoop framework represent the new model for distributed processing of large amounts of data that goes to supplant techniques based on stored procedures and computational grids (step 2). The relational model taught courses in basic database design, has many limitations compared to the demands posed by new applications based on Big Data and NoSQL databases that use to store data and MapReduce to process large amounts of data.
Course Website http://pbdmng.datatoknowledge.it/
Contact me for other informations and to download the slides
Data Modelling 101 half day workshop presented by Chris Bradley at the Enterprise Data and Business Intelligence conference London on November 3rd 2014.
Chris Bradley is a leading independent information strategist.
Contact chris.bradley@dmadvisors.co.uk
Data-Ed Online: Trends in Data ModelingDATAVERSITY
Businesses cannot compete without data. Every organization produces and consumes it. Data trends are hitting the mainstream and businesses are adopting buzzwords such as Big Data, data vault, data scientist, etc., to seek solutions for their fundamental data issues. Few realize that the importance of any solution, regardless of platform or technology, relies on the data model supporting it. Data modeling is not an optional task for an organization’s data remediation effort. Instead, it is a vital activity that supports the solution driving your business.
This webinar will address emerging trends around data model application methodology, as well as trends around the practice of data modeling itself. We will discuss abstract models and entity frameworks, as well as the general shift from data modeling being segmented to becoming more integrated with business practices.
Takeaways:
How are anchor modeling, data vault, etc. different and when should I apply them?
Integrating data models to business models and the value this creates
Application development (Data first, code first, object first)
Data-Ed Webinar: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Takeaways:
Understanding how to contribute to organizational challenges beyond traditional data architecting
How to utilize data architectures in support of business strategy
Understanding foundational data architecture concepts based on the DAMA DMBOK
Data architecture guiding principles & best practices
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Find out more: http://www.datablueprint.com/resource-center/webinar-schedule/
Data-Ed Webinar: Data Modeling FundamentalsDATAVERSITY
Every organization produces and consumes data. Because data is so important to day to day operations, data trends are hitting the mainstream and businesses are adopting buzzwords such as Big Data, NoSQL, data scientist, etc., to seek solutions for their fundamental issues. Few realize that the importance of any solution, regardless of platform or technology, relies on the data model supporting it. Data modeling is not an optional task for an organization’s data effort. It is a vital activity that supports the solutions driving your business.
This webinar will address fundamental data modeling methodologies, as well as trends around the practice of data modeling itself. We will discuss abstract models and entity frameworks, as well as the general shift from data modeling being segmented to becoming more integrated with business practices.
Learning Objectives:
How are anchor modeling, data vault, etc. different and when should I apply them?
Integrating data models to business models and the value this creates
Application development (Data first, code first, object first)
Good systems development often depends on multiple data management disciplines. One of these is metadata. While much of the discussion around metadata focuses on understanding metadata itself along with associated technologies, this comprehensive issue often represents a typical tool-and-technology focus, which has not achieved significant results. A more relevant question when considering pockets of metadata is whether to include them in the scope of organizational metadata practices. By understanding metadata practices, you can begin to build systems that allow you to exercise sophisticated data management techniques and support business initiatives.
Learning Objectives:
How to leverage metadata in support of your business strategy
Understanding foundational metadata concepts based on the DAMA DMBOK
Guiding principles & lessons learned
Data Systems Integration & Business Value Pt. 1: MetadataDATAVERSITY
Certain systems are more data focused than others. Usually their primary focus is on accomplishing integration of disparate data. In these cases, failure is most often attributable to the adoption of a single pillar (silver bullet). The three webinars in the Data Systems Integration and Business Value series are designed to illustrate that good systems development more often depends on at least three DM disciplines (pie wedges) in order to provide a solid foundation.
Much of the discussion of metadata focuses on understanding it and the associated technologies. While these are important, they represent a typical tool/technology focus and this has not achieved significant results to date. A more relevant question when considering pockets of metadata is: Whether to include them in the scope organizational metadata practices. By understanding what it means to include items in the scope of your metadata practices, you can begin to build systems that allow you to practice sophisticated ways to advance their data management and supported business initiatives. After a bit of practice in this manner you can position your organization to better exploit any and all metadata technologies.
Data-Ed: Data Systems Integration & Business Value PT. 1: MetadataData Blueprint
Certain systems are more data focused than others. Usually their primary focus is on accomplishing integration of disparate data. In these cases, failure is most often attributable to the adoption of a single pillar (silver bullet). The three webinars in the Data Systems Integration and Business Value series are designed to illustrate that good systems development more often depends on at least three DM disciplines (pie wedges) in order to provide a solid foundation.
Much of the discussion of metadata focuses on understanding it and the associated technologies. While these are important, they represent a typical tool/technology focus and this has not achieved significant results to date. A more relevant question when considering pockets of metadata is: Whether to include them in the scope organizational metadata practices. By understanding what it means to include items in the scope of your metadata practices, you can begin to build systems that allow you to practice sophisticated ways to advance their data management and supported business initiatives. After a bit of practice in this manner you can position your organization to better exploit any and all metadata technologies.
You can sign up for future Data-Ed webinars here: http://www.datablueprint.com/resource-center/webinar-schedule/
Data Science Operationalization: The Journey of Enterprise AIDenodo
Watch full webinar here: https://bit.ly/3kVmYJl
As we move into a world driven by AI initiatives, we find ourselves facing new and diverse challenges when it comes to operationalization. Creating a solution and putting it into practice, is certainly not the same. The challenges span various organizational and data facades. In many instances, the data scientists may be working in silos and connecting to the live data may not always be possible. But how does one guarantee their developed model in a silo is still relevant to live data? How can we manage the data flow and data access across the entire AI operationalization cycle?
Watch on-demand to explore:
- The journey and challenges of the Data Scientist
- How Denodo data virtualization with data movement streamlines operationalization
- The best practices and techniques when dealing with siloed data
- How customers have used data virtualization in their data science initiatives
Organizations must realize what it means to utilize data quality management in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Data-Ed Webinar: Data Quality EngineeringDATAVERSITY
Organizations must realize what it means to utilize data quality management in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor data quality. Showing how data quality should be engineered provides a useful framework in which to develop an effective approach. This in turn allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Takeaways:
Understanding foundational data quality concepts based on the DAMA DMBOK
Utilizing data quality engineering in support of business strategy
Data Quality guiding principles & best practices
Steps for improving data quality at your organization
DataEd Slides: Data Modeling is FundamentalDATAVERSITY
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that any and all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, Data Modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important are the data models driving the engineering and architecture activities of your organization. This webinar illustrates Data Modeling as a key activity upon which so much technology depends.
Data Modeling Best Practices - Business & Technical ApproachesDATAVERSITY
Data Modeling is hotter than ever, according to a number of recent surveys. Part of the appeal of data models lies in their ability to translate complex data concepts in an intuitive, visual way to both business and technical stakeholders. This webinar provides real-world best practices in using Data Modeling for both business and technical teams.
DataEd Slides: Unlock Business Value Using Reference and Master Data Manageme...DATAVERSITY
Data tends to pile up and can be rendered unusable or obsolete without careful maintenance processes. Reference and Master Data Management (MDM) has been a popular Data Management approach to effectively gain mastery over not just the data but the supporting architecture for processing it from a master/transaction perspective. This webinar presents MDM as a strategic approach to improving and formalizing practices around those data items that provide context for organizational transactions – its master data. Too often, MDM has been implemented technology-first and achieved the same very poor track record (1/3 succeeding on-time, within budget, achieving planned functionality). MDM success depends on a coordinated approach involving typically Data Governance and Data Quality activities. Program learning objectives include:
• Understanding foundational reference and MDM concepts
• Why they are an important component of your Data Architecture
• Awareness of Reference and MDM Frameworks and building blocks
• What consists of MDM guiding principles and best practices
• How to utilize Reference and MDM in support of business strategy
DAS Slides: Metadata Management From Technical Architecture & Business Techni...DATAVERSITY
Metadata provides context for the “who, what, when, where, and why” of data, and is of critical interest in today’s data-driven business environment. Since metadata is created and used by both business and IT, architectural and organizational techniques need to encompass a holistic approach across the organization to address all audiences. This webinar provides practical ways to manage metadata in your organization using both technical architecture and business techniques.
Architecting Data For The Modern Enterprise - Data Summit 2017, Closing KeynoteCaserta
The “Big Data era” has ushered in an avalanche of new technologies and approaches for delivering information and insights to business users. What is the role of the cloud in your analytical environment? How can you make your migration as seamless as possible? This closing keynote, delivered by Joe Caserta, a prominent consultant who has helped many global enterprises adopt Big Data, provided the audience with the inside scoop needed to supplement data warehousing environments with data intelligence—the amalgamation of Big Data and business intelligence.
This presentation was given as the closing keynote at DBTA's annual Data Summit in NYC.
• History of Data Management
• Business Drivers for implementation of data governance • Building Data Strategy & Governance Framework
• Data Management Maturity Models
• Data Quality Management
• Metadata and Governance
• Metadata Management
• Data Governance Stakeholder Communication Strategy
How to use the EU e-Competence Framework together with the Body of Knowledge Edison when implementing a new workforce model. In this use case Workforce Transformation, Curricula, Career paths and assessments. Human Resource Development tooling.
The Importance of MDM - Eternal Management of the Data MindDATAVERSITY
Despite its immaterial nature, data has a tendency to pile up as time goes on, and can quickly be rendered unusable or obsolete without careful maintenance and streamlining of processes for its management. This presentation will provide you with an understanding of reference and master data management (MDM), one such method for keeping mass amounts of business data organized and functional towards achieving business goals.
MDM’s guiding principles include the establishment and implementation of authoritative data sources and effective means of delivering data to various business processes, as well as increases to the quality of information used in organizational analytical functions (such as BI).
To that end, attendees of this webinar will learn how to:
- Structure their data management processes around these principles
- Incorporate data quality engineering into the planning of reference and MDM
- Understand why MDM is so critical to their organization’s overall data strategy
Data-Ed Online Webinar: Business Value from MDMDATAVERSITY
This presentation provides you with an understanding of the goals of reference and master data management (MDM), including establishing and implementing authoritative data sources, establishing and implementing more effective means of delivery data to various business processes, as well as increasing the quality of information used in organizational analytical functions (such as BI). You will understand the parallel importance of incorporating data quality engineering into the planning of reference and MDM.
Takeaways:
What is reference and MDM?
Why are reference and MDM important?
Reference and MDM Frameworks
Guiding principles & best practices
Architecture, Products, and Total Cost of Ownership of the Leading Machine Le...DATAVERSITY
Organizations today need a broad set of enterprise data cloud services with key data functionality to modernize applications and utilize machine learning. They need a comprehensive platform designed to address multi-faceted needs by offering multi-function data management and analytics to solve the enterprise’s most pressing data and analytic challenges in a streamlined fashion.
In this research-based session, I’ll discuss what the components are in multiple modern enterprise analytics stacks (i.e., dedicated compute, storage, data integration, streaming, etc.) and focus on total cost of ownership.
A complete machine learning infrastructure cost for the first modern use case at a midsize to large enterprise will be anywhere from $3 million to $22 million. Get this data point as you take the next steps on your journey into the highest spend and return item for most companies in the next several years.
Data at the Speed of Business with Data Mastering and GovernanceDATAVERSITY
Do you ever wonder how data-driven organizations fuel analytics, improve customer experience, and accelerate business productivity? They are successful by governing and mastering data effectively so they can get trusted data to those who need it faster. Efficient data discovery, mastering and democratization is critical for swiftly linking accurate data with business consumers. When business teams can quickly and easily locate, interpret, trust, and apply data assets to support sound business judgment, it takes less time to see value.
Join data mastering and data governance experts from Informatica—plus a real-world organization empowering trusted data for analytics—for a lively panel discussion. You’ll hear more about how a single cloud-native approach can help global businesses in any economy create more value—faster, more reliably, and with more confidence—by making data management and governance easier to implement.
What is data literacy? Which organizations, and which workers in those organizations, need to be data-literate? There are seemingly hundreds of definitions of data literacy, along with almost as many opinions about how to achieve it.
In a broader perspective, companies must consider whether data literacy is an isolated goal or one component of a broader learning strategy to address skill deficits. How does data literacy compare to other types of skills or “literacy” such as business acumen?
This session will position data literacy in the context of other worker skills as a framework for understanding how and where it fits and how to advocate for its importance.
Building a Data Strategy – Practical Steps for Aligning with Business GoalsDATAVERSITY
Developing a Data Strategy for your organization can seem like a daunting task – but it’s worth the effort. Getting your Data Strategy right can provide significant value, as data drives many of the key initiatives in today’s marketplace – from digital transformation, to marketing, to customer centricity, to population health, and more. This webinar will help demystify Data Strategy and its relationship to Data Architecture and will provide concrete, practical ways to get started.
Uncover how your business can save money and find new revenue streams.
Driving profitability is a top priority for companies globally, especially in uncertain economic times. It's imperative that companies reimagine growth strategies and improve process efficiencies to help cut costs and drive revenue – but how?
By leveraging data-driven strategies layered with artificial intelligence, companies can achieve untapped potential and help their businesses save money and drive profitability.
In this webinar, you'll learn:
- How your company can leverage data and AI to reduce spending and costs
- Ways you can monetize data and AI and uncover new growth strategies
- How different companies have implemented these strategies to achieve cost optimization benefits
Data Catalogs Are the Answer – What is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
Data Catalogs Are the Answer – What Is the Question?DATAVERSITY
Organizations with governed metadata made available through their data catalog can answer questions their people have about the organization’s data. These organizations get more value from their data, protect their data better, gain improved ROI from data-centric projects and programs, and have more confidence in their most strategic data.
Join Bob Seiner for this lively webinar where he will talk about the value of a data catalog and how to build the use of the catalog into your stewards’ daily routines. Bob will share how the tool must be positioned for success and viewed as a must-have resource that is a steppingstone and catalyst to governed data across the organization.
In this webinar, Bob will focus on:
-Selecting the appropriate metadata to govern
-The business and technical value of a data catalog
-Building the catalog into people’s routines
-Positioning the data catalog for success
-Questions the data catalog can answer
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, data modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization. This webinar illustrates data modeling as a key activity upon which so much technology and business investment depends.
Specific learning objectives include:
- Understanding what types of challenges require data modeling to be part of the solution
- How automation requires standardization on derivable via data modeling techniques
- Why only a working partnership between data and the business can produce useful outcomes
Analytics play a critical role in supporting strategic business initiatives. Despite the obvious value to analytic professionals of providing the analytics for these initiatives, many executives question the economic return of analytics as well as data lakes, machine learning, master data management, and the like.
Technology professionals need to calculate and present business value in terms business executives can understand. Unfortunately, most IT professionals lack the knowledge required to develop comprehensive cost-benefit analyses and return on investment (ROI) measurements.
This session provides a framework to help technology professionals research, measure, and present the economic value of a proposed or existing analytics initiative, no matter the form that the business benefit arises. The session will provide practical advice about how to calculate ROI and the formulas, and how to collect the necessary information.
How a Semantic Layer Makes Data Mesh Work at ScaleDATAVERSITY
Data Mesh is a trending approach to building a decentralized data architecture by leveraging a domain-oriented, self-service design. However, the pure definition of Data Mesh lacks a center of excellence or central data team and doesn’t address the need for a common approach for sharing data products across teams. The semantic layer is emerging as a key component to supporting a Hub and Spoke style of organizing data teams by introducing data model sharing, collaboration, and distributed ownership controls.
This session will explain how data teams can define common models and definitions with a semantic layer to decentralize analytics product creation using a Hub and Spoke architecture.
Attend this session to learn about:
- The role of a Data Mesh in the modern cloud architecture.
- How a semantic layer can serve as the binding agent to support decentralization.
- How to drive self service with consistency and control.
Enterprise data literacy. A worthy objective? Certainly! A realistic goal? That remains to be seen. As companies consider investing in data literacy education, questions arise about its value and purpose. While the destination – having a data-fluent workforce – is attractive, we wonder how (and if) we can get there.
Kicking off this webinar series, we begin with a panel discussion to explore the landscape of literacy, including expert positions and results from focus groups:
- why it matters,
- what it means,
- what gets in the way,
- who needs it (and how much they need),
- what companies believe it will accomplish.
In this engaging discussion about literacy, we will set the stage for future webinars to answer specific questions and feature successful literacy efforts.
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...DATAVERSITY
Change is hard, especially in response to negative stimuli or what is perceived as negative stimuli. So organizations need to reframe how they think about data privacy, security and governance, treating them as value centers to 1) ensure enterprise data can flow where it needs to, 2) prevent – not just react – to internal and external threats, and 3) comply with data privacy and security regulations.
Working together, these roles can accelerate faster access to approved, relevant and higher quality data – and that means more successful use cases, faster speed to insights, and better business outcomes. However, both new information and tools are required to make the shift from defense to offense, reducing data drama while increasing its value.
Join us for this panel discussion with experts in these fields as they discuss:
- Recent research about where data privacy, security and governance stand
- The most valuable enterprise data use cases
- The common obstacles to data value creation
- New approaches to data privacy, security and governance
- Their advice on how to shift from a reactive to resilient mindset/culture/organization
You’ll be educated, entertained and inspired by this panel and their expertise in using the data trifecta to innovate more often, operate more efficiently, and differentiate more strategically.
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Data Governance Trends - A Look Backwards and ForwardsDATAVERSITY
As DATAVERSITY’s RWDG series hurdles into our 12th year, this webinar takes a quick look behind us, evaluates the present, and predicts the future of Data Governance. Based on webinar numbers, hot Data Governance topics have evolved over the years from policies and best practices, roles and tools, data catalogs and frameworks, to supporting data mesh and fabric, artificial intelligence, virtualization, literacy, and metadata governance.
Join Bob Seiner as he reflects on the past and what has and has not worked, while sharing examples of enterprise successes and struggles. In this webinar, Bob will challenge the audience to stay a step ahead by learning from the past and blazing a new trail into the future of Data Governance.
In this webinar, Bob will focus on:
- Data Governance’s past, present, and future
- How trials and tribulations evolve to success
- Leveraging lessons learned to improve productivity
- The great Data Governance tool explosion
- The future of Data Governance
Data Governance Trends and Best Practices To Implement TodayDATAVERSITY
Would you share your bank account information on social media? How about shouting your social security number on the New York City subway? We didn’t think so either – that’s why data governance is consistently top of mind.
In this webinar, we’ll discuss the common Cloud data governance best practices – and how to apply them today. Join us to uncover Google Cloud’s investment in data governance and learn practical and doable methods around key management and confidential computing. Hear real customer experiences and leave with insights that you can share with your team. Let’s get solving.
Topics that you will hear addressed in this webinar:
- Understanding the basics of Cloud Incident Response (IR) and anticipated data governance trends
- Best practices for key management and apply data governance to your day-to-day
- The next wave of Confidential Computing and how to get started, including a demo
It is a fascinating, explosive time for enterprise analytics.
It is from the position of analytics leadership that the enterprise mission will be executed and company leadership will emerge. The data professional is absolutely sitting on the performance of the company in this information economy and has an obligation to demonstrate the possibilities and originate the architecture, data, and projects that will deliver analytics. After all, no matter what business you’re in, you’re in the business of analytics.
The coming years will be full of big changes in enterprise analytics and data architecture. William will kick off the fifth year of the Advanced Analytics series with a discussion of the trends winning organizations should build into their plans, expectations, vision, and awareness now.
Too often I hear the question “Can you help me with our data strategy?” Unfortunately, for most, this is the wrong request because it focuses on the least valuable component: the data strategy itself. A more useful request is: “Can you help me apply data strategically?” Yes, at early maturity phases the process of developing strategic thinking about data is more important than the actual product! Trying to write a good (must less perfect) data strategy on the first attempt is generally not productive –particularly given the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” This program refocuses efforts on learning how to iteratively improve the way data is strategically applied. This will permit data-based strategy components to keep up with agile, evolving organizational strategies. It also contributes to three primary organizational data goals. Learn how to improve the following:
- Your organization’s data
- The way your people use data
- The way your people use data to achieve your organizational strategy
This will help in ways never imagined. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges programmatically includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs as organizations identify prioritized areas where better assets, literacy, and support (data strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are also covered, including:
- A cohesive argument for why data strategy is necessary for effective data governance
- An overview of prerequisites for effective strategic use of data strategy, as well as common pitfalls
- A repeatable process for identifying and removing data constraints
- The importance of balancing business operation and innovation
Who Should Own Data Governance – IT or Business?DATAVERSITY
The question is asked all the time: “What part of the organization should own your Data Governance program?” The typical answers are “the business” and “IT (information technology).” Another answer to that question is “Yes.” The program must be owned and reside somewhere in the organization. You may ask yourself if there is a correct answer to the question.
Join this new RWDG webinar with Bob Seiner where Bob will answer the question that is the title of this webinar. Determining ownership of Data Governance is a vital first step. Figuring out the appropriate part of the organization to manage the program is an important second step. This webinar will help you address these questions and more.
In this session Bob will share:
- What is meant by “the business” when it comes to owning Data Governance
- Why some people say that Data Governance in IT is destined to fail
- Examples of IT positioned Data Governance success
- Considerations for answering the question in your organization
- The final answer to the question of who should own Data Governance
It is clear that Data Management best practices exist and so does a useful process for improving existing Data Management practices. The question arises: Since we understand the goal, how does one design a process for Data Management goal achievement? This program describes what must be done at the programmatic level to achieve better data use and a way to implement this as part of your data program. The approach combines DMBoK content and CMMI/DMM processes – permitting organizations with the opportunity to benefit from the best of both. It also permits organizations to understand:
- Their current Data Management practices
- Strengths that should be leveraged
- Remediation opportunities
MLOps – Applying DevOps to Competitive AdvantageDATAVERSITY
MLOps is a practice for collaboration between Data Science and operations to manage the production machine learning (ML) lifecycles. As an amalgamation of “machine learning” and “operations,” MLOps applies DevOps principles to ML delivery, enabling the delivery of ML-based innovation at scale to result in:
Faster time to market of ML-based solutions
More rapid rate of experimentation, driving innovation
Assurance of quality, trustworthiness, and ethical AI
MLOps is essential for scaling ML. Without it, enterprises risk struggling with costly overhead and stalled progress. Several vendors have emerged with offerings to support MLOps: the major offerings are Microsoft Azure ML and Google Vertex AI. We looked at these offerings from the perspective of enterprise features and time-to-value.
Putting the SPARK into Virtual Training.pptxCynthia Clay
This 60-minute webinar, sponsored by Adobe, was delivered for the Training Mag Network. It explored the five elements of SPARK: Storytelling, Purpose, Action, Relationships, and Kudos. Knowing how to tell a well-structured story is key to building long-term memory. Stating a clear purpose that doesn't take away from the discovery learning process is critical. Ensuring that people move from theory to practical application is imperative. Creating strong social learning is the key to commitment and engagement. Validating and affirming participants' comments is the way to create a positive learning environment.
Tata Group Dials Taiwan for Its Chipmaking Ambition in Gujarat’s DholeraAvirahi City Dholera
The Tata Group, a titan of Indian industry, is making waves with its advanced talks with Taiwanese chipmakers Powerchip Semiconductor Manufacturing Corporation (PSMC) and UMC Group. The goal? Establishing a cutting-edge semiconductor fabrication unit (fab) in Dholera, Gujarat. This isn’t just any project; it’s a potential game changer for India’s chipmaking aspirations and a boon for investors seeking promising residential projects in dholera sir.
Visit : https://www.avirahi.com/blog/tata-group-dials-taiwan-for-its-chipmaking-ambition-in-gujarats-dholera/
3.0 Project 2_ Developing My Brand Identity Kit.pptxtanyjahb
A personal brand exploration presentation summarizes an individual's unique qualities and goals, covering strengths, values, passions, and target audience. It helps individuals understand what makes them stand out, their desired image, and how they aim to achieve it.
RMD24 | Debunking the non-endemic revenue myth Marvin Vacquier Droop | First ...BBPMedia1
Marvin neemt je in deze presentatie mee in de voordelen van non-endemic advertising op retail media netwerken. Hij brengt ook de uitdagingen in beeld die de markt op dit moment heeft op het gebied van retail media voor niet-leveranciers.
Retail media wordt gezien als het nieuwe advertising-medium en ook mediabureaus richten massaal retail media-afdelingen op. Merken die niet in de betreffende winkel liggen staan ook nog niet in de rij om op de retail media netwerken te adverteren. Marvin belicht de uitdagingen die er zijn om echt aansluiting te vinden op die markt van non-endemic advertising.
Personal Brand Statement:
As an Army veteran dedicated to lifelong learning, I bring a disciplined, strategic mindset to my pursuits. I am constantly expanding my knowledge to innovate and lead effectively. My journey is driven by a commitment to excellence, and to make a meaningful impact in the world.
Enterprise Excellence is Inclusive Excellence.pdfKaiNexus
Enterprise excellence and inclusive excellence are closely linked, and real-world challenges have shown that both are essential to the success of any organization. To achieve enterprise excellence, organizations must focus on improving their operations and processes while creating an inclusive environment that engages everyone. In this interactive session, the facilitator will highlight commonly established business practices and how they limit our ability to engage everyone every day. More importantly, though, participants will likely gain increased awareness of what we can do differently to maximize enterprise excellence through deliberate inclusion.
What is Enterprise Excellence?
Enterprise Excellence is a holistic approach that's aimed at achieving world-class performance across all aspects of the organization.
What might I learn?
A way to engage all in creating Inclusive Excellence. Lessons from the US military and their parallels to the story of Harry Potter. How belt systems and CI teams can destroy inclusive practices. How leadership language invites people to the party. There are three things leaders can do to engage everyone every day: maximizing psychological safety to create environments where folks learn, contribute, and challenge the status quo.
Who might benefit? Anyone and everyone leading folks from the shop floor to top floor.
Dr. William Harvey is a seasoned Operations Leader with extensive experience in chemical processing, manufacturing, and operations management. At Michelman, he currently oversees multiple sites, leading teams in strategic planning and coaching/practicing continuous improvement. William is set to start his eighth year of teaching at the University of Cincinnati where he teaches marketing, finance, and management. William holds various certifications in change management, quality, leadership, operational excellence, team building, and DiSC, among others.
Skye Residences | Extended Stay Residences Near Toronto Airportmarketingjdass
Experience unparalleled EXTENDED STAY and comfort at Skye Residences located just minutes from Toronto Airport. Discover sophisticated accommodations tailored for discerning travelers.
Website Link :
https://skyeresidences.com/
https://skyeresidences.com/about-us/
https://skyeresidences.com/gallery/
https://skyeresidences.com/rooms/
https://skyeresidences.com/near-by-attractions/
https://skyeresidences.com/commute/
https://skyeresidences.com/contact/
https://skyeresidences.com/queen-suite-with-sofa-bed/
https://skyeresidences.com/queen-suite-with-sofa-bed-and-balcony/
https://skyeresidences.com/queen-suite-with-sofa-bed-accessible/
https://skyeresidences.com/2-bedroom-deluxe-queen-suite-with-sofa-bed/
https://skyeresidences.com/2-bedroom-deluxe-king-queen-suite-with-sofa-bed/
https://skyeresidences.com/2-bedroom-deluxe-queen-suite-with-sofa-bed-accessible/
#Skye Residences Etobicoke, #Skye Residences Near Toronto Airport, #Skye Residences Toronto, #Skye Hotel Toronto, #Skye Hotel Near Toronto Airport, #Hotel Near Toronto Airport, #Near Toronto Airport Accommodation, #Suites Near Toronto Airport, #Etobicoke Suites Near Airport, #Hotel Near Toronto Pearson International Airport, #Toronto Airport Suite Rentals, #Pearson Airport Hotel Suites
Discover the innovative and creative projects that highlight my journey throu...dylandmeas
Discover the innovative and creative projects that highlight my journey through Full Sail University. Below, you’ll find a collection of my work showcasing my skills and expertise in digital marketing, event planning, and media production.
Discover the innovative and creative projects that highlight my journey throu...
Trends in Data Modeling
1. Trends in Data Modeling
Presented by James Michael Lee and Peter Aiken, Ph.D.
2. Welcome: Trends in Data Modeling
Date: August 11, 2015
Time: 2:00 PM ET
Presented by: Peter Aiken, PhD
Steven MacLauchlan
Michael Lee
2Copyright 2015 by Data Blueprint Slide #
Businesses cannot compete without data. Every organization produces and
consumes it. Data trends are hitting the mainstream and businesses are adopting
buzzwords such as Big data, Data Vault, Data Scientist, etc., to seek solutions to
their fundamental data issues. Few realize that the importance of any solution,
regardless of platform or technology relies on the data model supporting it. Data
modeling is not an optional task for an organization’s data remediation effort.
Instead, it is a vital activity that supports the solution driving your business.
This webinar will address emerging trends around data model application
technology, as well as trends around the practice of data modeling itself. We will
discuss abstract models and entity frameworks, as well as the general shift from
data modeling being segmented to becoming more integrated with business
practices.
Takeaways:
• NoSQL, data vault, etc., different and when should I apply them?
• How Data Modeling relates to business process
• Application development (data first, code first, object first?)
3. Steven MacLauchlan
• 10 years of experience in Application
Development and Data Modeling with a
focus on Healthcare solutions.
• Delivers tailored data management solutions
that provide focus on data’s business value
while enhancing clients’ overall capability to
manage data
• Certified Data Management Professional (CDMP)
• Computer Science degree from Virginia Commonwealth
University
• Most recent focus: Understanding emerging
data modeling trends and how these can
best be leveraged for the Enterprise.
3Copyright 2015 by Data Blueprint Slide #
4. Peter Aiken, Ph.D.
• 30+ years in data management
• Repeated international recognition
• Founder, Data Blueprint (datablueprint.com)
• Associate Professor of IS (vcu.edu)
• DAMA International (dama.org)
• 9 books and dozens of articles
• Experienced w/ 500+ data
management practices
• Multi-year immersions:
– US DoD (DISA/Army/Marines/DLA)
– Nokia
– Deutsche Bank
– Wells Fargo
– Walmart
– …
• DAMA International President 2009-2013
• DAMA International Achievement Award 2001 (with
Dr. E. F. "Ted" Codd
• DAMA International Community Award 2005
PETER AIKEN WITH JUANITA BILLINGS
FOREWORD BY JOHN BOTTEGA
MONETIZING
DATA MANAGEMENT
Unlocking the Value in Your Organization’s
Most Important Asset.
The Case for the
Chief Data Officer
Recasting the C-Suite to Leverage
Your MostValuable Asset
Peter Aiken and
Michael Gorman
4Copyright 2015 by Data Blueprint Slide #
5. James “Michael” Lee
• Data Consultant certified in a number of areas, including Data
Vault 2.0 Practitioner, Kimball ETL Architecture and Certified
Data Management Professional (CDMP).
• Over 7 years of experience with
– Designing data quality solutions
– improving data management practices
– implementing Data Governance frameworks
– architecting data warehouses
– implementation of system upgrades and migrations
• In the following industries:
– telecommunications
– banking
– insurance
– government (defense)
– commercial manufacturing
– international shipping
5Copyright 2015 by Data Blueprint Slide #
6. We believe ...
Data
Assets
Financial
Assets
Real
Estate Assets
Inventory
Assets
Non-
depletable
Available for
subsequent
use
Can be
used up
Can be
used up
Non-
degrading √ √ Can degrade
over time
Can degrade
over time
Durable Non-taxed √ √
Strategic
Asset √ √ √ √
• Today, data is the most powerful, yet underutilized and poorly
managed organizational asset
• Data is your
– Sole
– Non-depleteable
– Non-degrading
– Durable
– Strategic
• Asset
– Data is the new oil!
– Data is the new (s)oil!
– Data is the new bacon!
• Our mission is to unlock business value by
– Strengthening your data management capabilities
– Providing tailored solutions, and
– Building lasting partnerships
6Copyright 2015 by Data Blueprint Slide #
Asset: A resource controlled by the organization as a result of past events or transactions and from which
future economic benefits are expected to flow [Wikipedia]
7. Trends in Data Modeling
Copyright 2015 by Data Blueprint
• Business to Data: the Relationship
• What is a Data Model?
• Conceptual, Logical, Physical
• What issues can poor data modeling
introduce?
• Different Models, Different Uses
• Traditional (3NF, Star Schema, Data Vault)
• NoSQL Technologies (Key-Value/Document,
Graph, Column Family)
• Trends
- Move to the business
- Self Service and Virtualization
- Agile
- Data Sharing World (The API’s)
- Patterns and Reuse
- Metadata Modeling
7
8. What is a Data Model*?
• A data model organizes data
elements and standardizes how the
data elements relate to one another.
• In “Data Modeling Made Simple” by
Steve Hoberman, he says: "A data
model is a wayfinding tool for both
business and IT professionals,
which uses a set of symbols and
text to precisely explain a subset of
real information to improve
communication within the
organization and thereby lead to a
more flexible and stable application
environment."
8Copyright 2015 by Data Blueprint Slide #
*According to ANSI.
9. Why should we care about poor data models?
• Poor data modeling up front can cause Data Quality issues “downstream”
• If the model isn’t a true representation of the business concepts, this will impact
confidence in the data, inhibit business insights and innovation
• Potential for poor DB/Application performance for reads/writes. Example: Over-
normalization
• Lack of flexibility can cause difficulty aligning with evolving business requirements
• Difficulty integrating data in the future
• Constrains business agility by complicating reengineering
• Creates operational inefficiencies (ex: poor application performance)
• Limits workflow transparency
• Proliferates system work-arounds,
including shadow systems
developed by end users
• Impact Analysis
9Copyright 2015 by Data Blueprint Slide #
10. How are Data Models Expressed as Architectures?
• Attributes are organized into entities/objects
– Attributes are characteristics of "things"
– Entitles/objects are "things" whose information is
managed in support of strategy
• Entities/objects are organized into models
– Combinations of attributes and entities are structured
to represent information requirements
– Poorly structured data, constrains organizational
information delivery capabilities
• Models are organized into architectures
– When building new systems, architectures are used to
plan development
– More often, data managers do not know what existing
architectures are and - therefore - cannot make use of
them in support of strategy implementation
10Copyright 2015 by Data Blueprint Slide #
More Granular
More Abstract
11. The Conceptual Data Model
• Represents entities and relationships
• Should Identify the domain and scope of data
• Should be easily understood by business users in order to
communicate core data concepts, and drive application
requirements
11Copyright 2015 by Data Blueprint Slide #
Example:
We need to model customer
address data. A customer may have
many addresses, and many
customers may share one address.
“many to many”
13. Data map of DISPOSITION
• At least one but possibly more system USERS enter the DISPOSITION facts into the system.
• An ADMISSION is associated with one and only one DISCHARGE.
• An ADMISSION is associated with zero or more FACILITIES.
• An ADMISSION is associated with zero or more PROVIDERS.
• An ADMISSION is associated with one or more ENCOUNTERS.
• An ENCOUNTER may be recorded by a system USER.
• An ENCOUNTER may be associated with a PROVIDER.
• An ENCOUNTER may be associated with one or more DIAGNOSES.
13Copyright 2015 by Data Blueprint Slide #
ADMISSION Contains information about patient admission history
related to one or more inpatient episodes
DIAGNOSIS Contains the International Disease Classification (IDC) of
code representation and/or description of a patient's health
related to an inpatient code
DISCHARGE A table of codes describing disposition types available for
an inpatient at a FACILITY
ENCOUNTER Tracking information related to inpatient episodes
FACILITY File containing a list of all facilities in regional health care
system
PROVIDER Full name of a member of the FACILITY team providing
services to the patient
USER Any user with access to create, read, update, and delete
DISPOSITION data
14. A sample data entity and associated metadata
Entity: BED
Data Asset Type: Principal Data Entity
Purpose: This is a substructure within the Room
substructure of the Facility Location. It contains
information about beds within rooms.
Source: Maintenance Manual for File and Table
Data (Software Version 3.0, Release 3.1)
Attributes: Bed.Description
Bed.Status
Bed.Sex.To.Be.Assigned
Bed.Reserve.Reason
Associations: >0-+ Room
Status: Validated
14Copyright 2015 by Data Blueprint Slide #
• A purpose statement describing why the organization is maintaining information
about this business concept;
• Sources of information about it;
• A partial list of the attributes or characteristics of the entity; and
• Associations with other data items; this one is read as "One room contains zero or
many beds."
15. The Logical Data Model
• Should represent the Conceptual Data model more
thoroughly, but be otherwise very similar
• Will include attributes, names, relationships, and other
metadata
• Will be developed using Data Modeling notation (ex: UML)
15Copyright 2015 by Data Blueprint Slide #
16. The Physical Data Model
• Describes the specific database implementation of the
data
• Attributes will be named according to naming conventions
• Displays data types, accurate table names, Key
information, etc
16Copyright 2015 by Data Blueprint Slide #
17. CM2 Component Evolution is technology derived but technology independent
17Copyright 2015 by Data Blueprint Slide #
18. Data Reengineering for More Shareable Data
18Copyright 2015 by Data Blueprint Slide #
Other logical as-is
data architecture
components
20. Trends in Data Modeling
Copyright 2015 by Data Blueprint
• Business to Data: the Relationship
• What is a Data Model?
• Conceptual, Logical, Physical
• What issues can poor data modeling
introduce?
• Different Models, Different Uses
• Traditional (3NF, Star Schema, Data Vault)
• NoSQL Technologies (Key-Value/Document,
Graph, Column Family)
• Trends
- Move to the business
- Self Service and Virtualization
- Agile
- Data Sharing World (The API’s)
- Patterns and Reuse
- Metadata Modeling
20
21. Normalization Rules Overview
• 1st Normal Form - no repeating non-
key attributes for a given primary key
• 2nd Normal Form - no non-key
attributes that depend on only a
portion of the primary key
• 3rd Normal Form - no attributes
depend on something other than the
primary key
• 4th Normal Form - attributes depend
on not only key but the value of the
key
• 5th Normal Form - an entity is in
5NBF if its dependencies on
occurrences of the same entity of
entity type have been moved into a
structured entity
21Copyright 2015 by Data Blueprint Slide #
The row in every table is
dependent on the key, the whole
key and northern but the key
22. Third Normal Form
• Each attribute in the relationship is a fact about a key
• Highly normalized structure
22Copyright 2015 by Data Blueprint Slide #
• Use Cases:
– Transactional Systems.
– Operational Data Stores.
23. Third Normal Form: Pros and Cons
• Pros
– Easily understood by business and end users
– Reduced data redundancy
– Enforced referential integrity
– Indexed attributes/flexible querying
• Cons
– Joins can be expensive
– Does not scale
23Copyright 2015 by Data Blueprint Slide #
Neo4j.com
24. Star Schema
24Copyright 2015 by Data Blueprint Slide #
• Comprised of “fact tables” that contain quantitative data,
and any number of adjoining “dimension” tables
• Optimized for business reporting
• Use Cases:
– OLAP (Online Analytic Processing)
– BI
Wikipedia
25. Star Schema Pros and Cons
• Pros
– Simple Design
– Fast Queries
– Most major DBMS
are optimized for
Star Schema
Designs
• Cons
– Questions must be
built into the design
– Data marts are often
centralized on one
fact table
25Copyright 2015 by Data Blueprint Slide #
26. Data Vault
• Designed to facilitate long-term historical storage, focusing on ease
of implementation
• Retains data lineage information (source/date)
• “All the data, all the time”. Hybrid approach of Inmon and Kimball.
• Comprised of Hubs (which contain a list of business keys that do not
change often), Links (Associations/transactions between hubs), and
Satellites (descriptive attributes associated with hubs and links)
26Copyright 2015 by Data Blueprint Slide #
• Use Cases:
– Data Warehousing
– Complete Auditability
Bukhantsov.org
27. Data Vault Pros and Cons
• Pros
– Simple integration
– Houses immense
amounts of data with
excellent performance
– Full data lineage
captured
• Cons
– Complication is pushed
to the “back end”
– Can be difficult to setup
for many data workers
– No widespread support
for ETL tools yet
27Copyright 2015 by Data Blueprint Slide #
28. Model Comparison Matrix
28Copyright 2015 by Data Blueprint Slide #
3NF Dimensional Vault
Scalability ☑ ☑ ☑
Flexibility ☒ ☒ ☑
Reengineering ☒ ☒ ☑
Auditability ☑
Business Interpretable ☑ ☑ ☒
Presentation Layer ☒ ☑ ☒
Performance ☒ ☑ ☑
Support ☑ ☑
29. 29Copyright 2015 by Data Blueprint Slide #
Technology Trigger: A potential technology breakthrough kicks things off. Early proof-of-concept stories and media interest
trigger significant publicity. Often no usable products exist and commercial viability is unproven.
Trough of Disillusionment: Interest wanes as experiments and implementations fail to deliver. Producers of the
technology shake out or fail. Investments continue only if the surviving providers improve their products to the
satisfaction of early adopters.
Peak of Inflated Expectations: Early publicity produces a number of
success stories—often accompanied by scores of failures. Some
companies take action; many do not.
Slope of Enlightenment: More instances of how the technology can benefit the
enterprise start to crystallize and become more widely understood. Second- and third-
generation products appear from technology providers. More enterprises fund pilots;
conservative companies remain cautious.
Plateau of Productivity: Mainstream adoption starts to
take off. Criteria for assessing provider viability are more
clearly defined. The technology’s broad market
applicability and relevance are clearly paying off.
Gartner Five-phase Hype Cycle
31. 2012 Big Data in Hype Cycle
31Copyright 2015 by Data Blueprint Slide #
32. 2013 Big Data in Hype Cycle
32Copyright 2015 by Data Blueprint Slide #
33. 2014 Big Data in Hype Cycle
33Copyright 2015 by Data Blueprint Slide #
"A focus on big data is not a substitute for the
fundamentals of information management."
34. NoSQL Solutions*
• Document/Key Value
– “Schema-less” design empowers developers*
– Scalable
– High availability
– Economically viable (scale out not up!)
• RDF/Triple Store
– Purpose-built to store triples (“bob likes football”)
– SPARQL is a query language specific to RDF.
– One of the pillars of “Semantic Web”
• Graph
– Structure comprised of “nodes”, “edges”, and “properties”
– Focused on the interconnection between entities
– Fast queries to find associative data
• Column Family
– Columns are stored individually (but clustered by “family” unlike traditional
columnar databases)
– By only querying specific column families, we can have nearly unlimited
numbers of columns without causing expensive queries
34Copyright 2015 by Data Blueprint Slide #
*not exhaustive!
35. NoSQL Data Models
35Copyright 2015 by Data Blueprint Slide #
RDF/Triple Store
Graph (Source: Neo4J)
Document Store (Source: MongoDB)
Column Store (Source: Toadworld)
38. Trends in Data Modeling
Copyright 2015 by Data Blueprint
• Business to Data: the Relationship
• What is a Data Model?
• Conceptual, Logical, Physical
• What issues can poor data modeling
introduce?
• Different Models, Different Uses
• Traditional (3NF, Star Schema, Data Vault)
• NoSQL Technologies (Key-Value/Document,
Graph, Column Family)
• Trends
- Move to the business
- Self Service and Virtualization
- Agile
- Data Sharing World (The API’s)
- Patterns and Reuse
- Metadata Modeling
38
39. Move it to the Business
• Models need to add value
• Models need to be part of the process
– (Not a documentation of the process)
• Models need to assist in improving capabilities, not
hindering them
– Self Service BI
39Copyright 2015 by Data Blueprint Slide #
40. Self Service and Virtualization
• Self Service BI requires end user understanding of
the system
• Presentation Data Models
40Copyright 2015 by Data Blueprint Slide #
41. Agile
• Incremental build of models
– Not an excuse to create bad models
• 80/20 Rule
• The problem with code first
– Rules exist in code
– Reengineering concerns
– Governance concerns
– Lack Business Insights
• Database First
– Creates value in modeling
– Enforced integrity and lineage of the data
– Integrates the model into the process
– Used to generate code
41Copyright 2015 by Data Blueprint Slide #
42. A Data Sharing World
• Adding structure to information allows us to obtain
exactly what we want, when we want it.
• Allows applications to serve up data to external
sources in a structured way- “Post-schema”.
42Copyright 2015 by Data Blueprint Slide #
43. Design Patterns
• Why are the restrooms generally in the same place in each building?
• What about the electrical wiring?
• HVAC? Floorplans? ...
• Architecture design patterns (spoke and hub,
hub of hubs, warehouse, cloud, MDM,
changing tires, portal)
43Copyright 2015 by Data Blueprint Slide #
44. Patterns and Reuse
• Common rule of thumb:
– One third of a data model
contains fields common to all
business.
– One third contains fields common
to the industry, and the
– Other third is specific to the
organization.
• Patterns should theoretically provide
an organization with a base-line to
quickly develop data infrastructure.
• Off-the-shelf solutions may require
in-depth customization or
specialization.
44Copyright 2015 by Data Blueprint Slide #
46. Marco & Jennings's Metadata Model
Source:http://dmreview.com/article_sub.cfm?articleID=1000941 used with permission
46Copyright 2015 by Data Blueprint Slide #
47. Trends in Data Modeling
Copyright 2015 by Data Blueprint
• Business to Data: the Relationship
• What is a Data Model?
• Conceptual, Logical, Physical
• What issues can poor data modeling
introduce?
• Different Models, Different Uses
• Traditional (3NF, Star Schema, Data Vault)
• NoSQL Technologies (Key-Value/Document,
Graph, Column Family)
• Trends
- Move to the business
- Self Service and Virtualization
- Agile
- Data Sharing World (The API’s)
- Patterns and Reuse
- Metadata Modeling
47
48. Conclusions
• Data Modeling is
important to get right.
• Getting it “right” is
hugely dependent on
the business case,
maturity of the
organization,
flexibility for future
growth, and so much
more.
• There are many
technologies and
ideas available to
help solve a number
of problems.
• Don't try any of this
without considering
the various
architectures involved
48Copyright 2015 by Data Blueprint Slide #
49. Questions?
49Copyright 2015 by Data Blueprint Slide #
It’s your turn!
Use the chat feature or Twitter (#dataed) to submit
your questions to Peter, Michael and Steven now.
50. Upcoming Events
Data Quality Success Stories
September 8, 2015
@ 2:00 PM ET/11:00 AM PT
Design & Manage Data Structures
October 13, 2015
@ 2:00 PM ET/11:00 AM PT
Sign up here:
• www.datablueprint.com/webinar-schedule
• or www.dataversity.net
50Copyright 2015 by Data Blueprint Slide #
51. Sources
• Data model. (2014, October 7). In Wikipedia, The Free
Encyclopedia. Retrieved October 7, 2014, from http://
en.wikipedia.org/w/index.php?
title=Data_model&oldid=628639882
• Data Modeling 101. (2006). In Agile Data. Retrieved
October 7, 2014, from http://www.agiledata.org/essays/
dataModeling101.html
51Copyright 2015 by Data Blueprint Slide #