This presentation is focused on the Data Integration and Interoperability section of the DMBOK2. It focuses on Data virtualization as a key tool of Data Integration.
The Analytical HR Professional: A Look at Data-Driven Talent ManagementHuman Capital Media
Faced with unprecedented generational shifts and evolving business imperatives for smart growth, the need for a complete and comprehensive view of the workforce - from skills, competencies and performance to hiring and retiring - has never been higher. However, many organizations lack the critical insight needed to identify high performers, develop high potentials, ensure that the right employees receive the right training and effectively deploy one of their most valuable resources - their people - to execute the organization's strategy. Join Lisa Rowan of IDC Research as she discusses the role of workforce for HR leaders seeking a more proactive role in driving business strategy. She will discuss how organizations can use a data-driven approach to HR to advance the workforce and seize market opportunities, and share tips for getting started.
Successfully Kickstarting Data Governance's Social Dynamics: Define, Collabor...Stijn (Stan) Christiaens
Learn how to launch your data governance program, by answering three questions:
- What does my data mean: collect and manage business definitions and relations, taxonomies and classifications, business rules and ontologies;
- How can I involve all stakeholders: engage them across business units and geographies, with stewards, data owners, … in a guiding workflow;
- How do I operationalize data governance: link MDM, DQ and BI to the business, use business-driven semantic modelling, achieve end-to end traceabilitiy. During this session we will use examples from different verticals: Finance, Government, Utilities,… .
We discuss their main drivers for starting a Data Governance initiative, as well as their pragmatic approach in moving from gradual roll out to support and sustain their Data Governance program.
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Find out more: http://www.datablueprint.com/resource-center/webinar-schedule/
Data-Ed Online: Trends in Data ModelingDATAVERSITY
Businesses cannot compete without data. Every organization produces and consumes it. Data trends are hitting the mainstream and businesses are adopting buzzwords such as Big Data, data vault, data scientist, etc., to seek solutions for their fundamental data issues. Few realize that the importance of any solution, regardless of platform or technology, relies on the data model supporting it. Data modeling is not an optional task for an organization’s data remediation effort. Instead, it is a vital activity that supports the solution driving your business.
This webinar will address emerging trends around data model application methodology, as well as trends around the practice of data modeling itself. We will discuss abstract models and entity frameworks, as well as the general shift from data modeling being segmented to becoming more integrated with business practices.
Takeaways:
How are anchor modeling, data vault, etc. different and when should I apply them?
Integrating data models to business models and the value this creates
Application development (Data first, code first, object first)
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “big data,” “NoSQL,” “data scientist,” and so on. Few realize that any and all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, Data Modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization become. This webinar illustrates Data Modeling as a key activity upon which so much technology depends.
The Analytical HR Professional: A Look at Data-Driven Talent ManagementHuman Capital Media
Faced with unprecedented generational shifts and evolving business imperatives for smart growth, the need for a complete and comprehensive view of the workforce - from skills, competencies and performance to hiring and retiring - has never been higher. However, many organizations lack the critical insight needed to identify high performers, develop high potentials, ensure that the right employees receive the right training and effectively deploy one of their most valuable resources - their people - to execute the organization's strategy. Join Lisa Rowan of IDC Research as she discusses the role of workforce for HR leaders seeking a more proactive role in driving business strategy. She will discuss how organizations can use a data-driven approach to HR to advance the workforce and seize market opportunities, and share tips for getting started.
Successfully Kickstarting Data Governance's Social Dynamics: Define, Collabor...Stijn (Stan) Christiaens
Learn how to launch your data governance program, by answering three questions:
- What does my data mean: collect and manage business definitions and relations, taxonomies and classifications, business rules and ontologies;
- How can I involve all stakeholders: engage them across business units and geographies, with stewards, data owners, … in a guiding workflow;
- How do I operationalize data governance: link MDM, DQ and BI to the business, use business-driven semantic modelling, achieve end-to end traceabilitiy. During this session we will use examples from different verticals: Finance, Government, Utilities,… .
We discuss their main drivers for starting a Data Governance initiative, as well as their pragmatic approach in moving from gradual roll out to support and sustain their Data Governance program.
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Find out more: http://www.datablueprint.com/resource-center/webinar-schedule/
Data-Ed Online: Trends in Data ModelingDATAVERSITY
Businesses cannot compete without data. Every organization produces and consumes it. Data trends are hitting the mainstream and businesses are adopting buzzwords such as Big Data, data vault, data scientist, etc., to seek solutions for their fundamental data issues. Few realize that the importance of any solution, regardless of platform or technology, relies on the data model supporting it. Data modeling is not an optional task for an organization’s data remediation effort. Instead, it is a vital activity that supports the solution driving your business.
This webinar will address emerging trends around data model application methodology, as well as trends around the practice of data modeling itself. We will discuss abstract models and entity frameworks, as well as the general shift from data modeling being segmented to becoming more integrated with business practices.
Takeaways:
How are anchor modeling, data vault, etc. different and when should I apply them?
Integrating data models to business models and the value this creates
Application development (Data first, code first, object first)
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “big data,” “NoSQL,” “data scientist,” and so on. Few realize that any and all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, Data Modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business. Since quality engineering/architecture work products do not happen accidentally, the more your organization depends on automation, the more important the data models driving the engineering and architecture activities of your organization become. This webinar illustrates Data Modeling as a key activity upon which so much technology depends.
DataOps - The Foundation for Your Agile Data ArchitectureDATAVERSITY
Achieving agility in data and analytics is hard. It’s no secret that most data organizations struggle to deliver the on-demand data products that their business customers demand. Recently, there has been much hype around new design patterns that promise to deliver this much sought-after agility.
In this webinar, Chris Bergh, CEO and Head Chef of DataKitchen will cut through the noise and describe several elegant and effective data architecture design patterns that deliver low errors, rapid development, and high levels of collaboration. He’ll cover:
• DataOps, Data Mesh, Functional Design, and Hub & Spoke design patterns;
• Where Data Fabric fits into your architecture;
• How different patterns can work together to maximize agility; and
• How a DataOps platform serves as the foundational superstructure for your agile architecture.
Data-Ed Online: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Takeaways:
Understanding how to contribute to organizational challenges beyond traditional data architecting
How to utilize data architectures in support of business strategy
Understanding foundational data architecture concepts based on the DAMA DMBOK
Data architecture guiding principles & best practices
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality management effectively in support of business strategy. This, in turn, allows for speedy identification of business problems, the delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This, in turn, allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Emerging Trends in Data Architecture – What’s the Next Big ThingDATAVERSITY
Digital Transformation is a top priority for many organizations, and a successful digital journey requires a strong data foundation. Creating this digital transformation requires a number of core data management capabilities such as MDM, With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Data-Ed Webinar: Data Modeling FundamentalsDATAVERSITY
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that any and all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, Data Modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business.
Instead of the technical minutiae of Data Modeling, this webinar will focus on its value and practicality for your organization. In doing so, we will:
Address fundamental Data Modeling methodologies, their differences and various practical applications, and trends around the practice of Data Modeling itself
Discuss abstract models and entity frameworks, as well as some basic tenets for application development
Examine the general shift from segmented Data Modeling to more business-integrated practices
Discuss fundamental Data Modeling concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
Digital Transformation is a top priority for many organizations, and a successful digital journey requires a strong data foundation. Creating this digital transformation requires a number of core data management capabilities such as MDM, With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Too often I hear the question “Can you help me with our Data Strategy?” Unfortunately, for most, this is the wrong request because it focuses on the least valuable component – the Data Strategy itself. A more useful request is this: “Can you help me apply data strategically?”Yes, at early maturity phases the process of developing strategic thinking about data is more important than the actual product! Trying to write a good (must less perfect) Data Strategy on the first attempt is generally not productive –particularly given the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” Refocus on learning how to iteratively improve the way data is strategically applied. This will permit data-based strategy components to keep up with agile, evolving organizational strategies. This approach can also contribute to three primary organizational data goals.
In this webinar, you will learn how improving your organization’s data, the way your people use data, and the way your people use data to achieve your organizational strategy will help in ways never imagined. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges programmatically includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs, as organizations identify prioritized areas where better assets, literacy, and support (Data Strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are also covered, including:
- A cohesive argument for why Data Strategy is necessary for effective Data Governance
- An overview of prerequisites for effective strategic use of Data Strategy, as well as common pitfalls
- A repeatable process for identifying and removing data constraints
- The importance of balancing business operation and innovation
How Enterprises are Using NoSQL for Mission-Critical ApplicationsDATAVERSITY
NoSQL databases including Couchbase are increasingly being selected as the backend technology for web and mobile apps. Document databases in particular are well suited for a large number of different use cases as an operational datastore.
In this webinar, Perry Krug, Principal Solutions Architect at Couchbase, will give a brief overview of Couchbase Server, a document database and its underlying distributed architecture. In addition, Perry will share how some of the biggest brands in the world use Couchbase, including:
Paypal A scalable NoSQL and big data architecture with real time analytics
Concur A highly available cache solution that supports 1B operations/day
Amadeus A backend data store that supports 1.6B transactions/day
How you can gain rapid insights and create more flexibility by capturing and storing data from a variety of sources and structures into a NoSQL database.
Information management plays a critical role in supporting strategic business initiatives. Despite the apparent value of providing the data infrastructure for these initiatives, many executives question the economic feasibility of business intelligence and analytics. This requires information professionals to calculate and present the business value in terms business executives can understand.
Unfortunately, most IT professionals lack the knowledge required to develop comprehensive cost-benefit analyses and return on investment (ROI) measurements.
This session provides a framework to help IT professionals research, measure, and present the economic value of a proposed or existing information initiative. The session will provide practical advice about how to calculate ROI, which formula to use, and how to collect the necessary information.
Building a Collaborative Data ArchitectureDATAVERSITY
Data isn’t stored in a vault where no one can touch it. Both business and technical professionals need the right access to the data, and need to work together to ensure that they all have the same understanding as to what the data means. Ron Huizenga, Senior Product Manager for ER/Studio, will discuss and demonstrate how data models, business glossaries, and collaboration can enhance your data architecture.
A presentation to explain why selling of Information Architecture is important and how the architect has to include strategy points even before the IA is sold.
Slides: How AI Makes Analytics More HumanDATAVERSITY
People think AI makes analytics less human, replacing human decision making. But the truth is, AI actually makes analytics more human. Augmented analytics are helping organizations finally break through the low levels of adoption and limitations typical of 2nd generation visualization tools.
Most business problems cannot be solved purely by algorithms or machine learning — they require human interaction and perspective. Uniting precedent-based machine learning systems with natural human intuition and curiosity is the foundation of 3rd generation BI and democratizing data across an enterprise.
It is a natural flow to enhance your data eco-system by deploying a platform with augmented intelligence to work alongside users in the pursuit of surfacing new insights, automating tasks, and supporting natural language interaction. All work as accelerators for achieving active intelligence and Data Literacy.
In that session we will discuss about Data Governance, mainly around that fantastic platform Power BI (but also around on-prem concerns).
How to avoid dataset-hell ? What are the best practices for sharing queries ? Who is the famous Data Steward and what is its role in a department or in the whole company ? How do you choose the right person ?
Keywords : Power Query, Data Management Gateway, Power BI Admin Center, Datastewardship, SharePoint 2013, eDiscovery
Level 200
DataEd Slides: Data Management + Data Strategy = InteroperabilityDATAVERSITY
Few organizations operate without having to exchange data. (Many do it professionally and well!) The larger the data exchange burden (DEB), the greater the organizational overhead incurred. This death by 1,000 cuts must be factored into each organization’s calculations. Unfortunately, most organizations do not know if their organization’s DEB is great or small. A somewhat greater number of organizations have organized Data Management practices. Focusing Data Management efforts on increasing interoperability by decreasing the DEB friction is a good area to “practice.”
Learning Objectives:
• Gaining a good understanding of both important topics
• Understanding that data only operates at a very intricate, specifically dependent intent and what this means
• Understand state-of-the-practice
• Coordination is key, requiring necessary but insufficient interdependencies and sequencing
• Practice makes perfect
Data Lakehouse, Data Mesh, and Data Fabric (r1)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. I’ll include use cases so you can see what approach will work best for your big data needs.
Traditionally, data integration has meant compromise. No matter how rapidly data architects and developers could complete a project before its deadline, speed would always come at the expense of quality. On the other hand, if they focused on delivering a quality project, it would generally drag on for months thus exceeding its deadline. Finally, if the teams concentrated on both quality and rapid delivery, the costs would invariably exceed the budget. Regardless of which path you chose, the end result would be less than desirable. This led some experts to revisit the scope of data integration. This write up shall focus on the same issue.
DataOps - The Foundation for Your Agile Data ArchitectureDATAVERSITY
Achieving agility in data and analytics is hard. It’s no secret that most data organizations struggle to deliver the on-demand data products that their business customers demand. Recently, there has been much hype around new design patterns that promise to deliver this much sought-after agility.
In this webinar, Chris Bergh, CEO and Head Chef of DataKitchen will cut through the noise and describe several elegant and effective data architecture design patterns that deliver low errors, rapid development, and high levels of collaboration. He’ll cover:
• DataOps, Data Mesh, Functional Design, and Hub & Spoke design patterns;
• Where Data Fabric fits into your architecture;
• How different patterns can work together to maximize agility; and
• How a DataOps platform serves as the foundational superstructure for your agile architecture.
Data-Ed Online: Data Architecture RequirementsDATAVERSITY
Data architecture is foundational to an information-based operational environment. It is your data architecture that organizes your data assets so they can be leveraged in your business strategy to create real business value. Even though this is important, not all data architectures are used effectively. This webinar describes the use of data architecture as a basic analysis method. Various uses of data architecture to inform, clarify, understand, and resolve aspects of a variety of business problems will be demonstrated. As opposed to showing how to architect data, your presenter Dr. Peter Aiken will show how to use data architecting to solve business problems. The goal is for you to be able to envision a number of uses for data architectures that will raise the perceived utility of this analysis method in the eyes of the business.
Takeaways:
Understanding how to contribute to organizational challenges beyond traditional data architecting
How to utilize data architectures in support of business strategy
Understanding foundational data architecture concepts based on the DAMA DMBOK
Data architecture guiding principles & best practices
Good data is like good water: best served fresh, and ideally well-filtered. Data Management strategies can produce tremendous procedural improvements and increased profit margins across the board, but only if the data being managed is of high quality. Determining how Data Quality should be engineered provides a useful framework for utilizing Data Quality management effectively in support of business strategy. This, in turn, allows for speedy identification of business problems, the delineation between structural and practice-oriented defects in Data Management, and proactive prevention of future issues. Organizations must realize what it means to utilize Data Quality engineering in support of business strategy. This webinar will illustrate how organizations with chronic business challenges often can trace the root of the problem to poor Data Quality. Showing how Data Quality should be engineered provides a useful framework in which to develop an effective approach. This, in turn, allows organizations to more quickly identify business problems as well as data problems caused by structural issues versus practice-oriented defects and prevent these from re-occurring.
Emerging Trends in Data Architecture – What’s the Next Big ThingDATAVERSITY
Digital Transformation is a top priority for many organizations, and a successful digital journey requires a strong data foundation. Creating this digital transformation requires a number of core data management capabilities such as MDM, With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Data-Ed Webinar: Data Modeling FundamentalsDATAVERSITY
Because every organization produces and propagates data as part of their day-to-day operations, data trends are becoming more and more important in the mainstream business world’s consciousness. For many organizations in various industries, though, comprehension of this development begins and ends with buzzwords: “Big Data,” “NoSQL,” “Data Scientist,” and so on. Few realize that any and all solutions to their business problems, regardless of platform or relevant technology, rely to a critical extent on the data model supporting them. As such, Data Modeling is not an optional task for an organization’s data effort, but rather a vital activity that facilitates the solutions driving your business.
Instead of the technical minutiae of Data Modeling, this webinar will focus on its value and practicality for your organization. In doing so, we will:
Address fundamental Data Modeling methodologies, their differences and various practical applications, and trends around the practice of Data Modeling itself
Discuss abstract models and entity frameworks, as well as some basic tenets for application development
Examine the general shift from segmented Data Modeling to more business-integrated practices
Discuss fundamental Data Modeling concepts based on “The DAMA Guide to the Data Management Body of Knowledge” (DAMA DMBOK)
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
Digital Transformation is a top priority for many organizations, and a successful digital journey requires a strong data foundation. Creating this digital transformation requires a number of core data management capabilities such as MDM, With technological innovation and change occurring at an ever-increasing rate, it’s hard to keep track of what’s hype and what can provide practical value for your organization. Join this webinar to see the results of a recent DATAVERSITY survey on emerging trends in Data Architecture, along with practical commentary and advice from industry expert Donna Burbank.
Too often I hear the question “Can you help me with our Data Strategy?” Unfortunately, for most, this is the wrong request because it focuses on the least valuable component – the Data Strategy itself. A more useful request is this: “Can you help me apply data strategically?”Yes, at early maturity phases the process of developing strategic thinking about data is more important than the actual product! Trying to write a good (must less perfect) Data Strategy on the first attempt is generally not productive –particularly given the widespread acceptance of Mike Tyson’s truism: “Everybody has a plan until they get punched in the face.” Refocus on learning how to iteratively improve the way data is strategically applied. This will permit data-based strategy components to keep up with agile, evolving organizational strategies. This approach can also contribute to three primary organizational data goals.
In this webinar, you will learn how improving your organization’s data, the way your people use data, and the way your people use data to achieve your organizational strategy will help in ways never imagined. Data are your sole non-depletable, non-degradable, durable strategic assets, and they are pervasively shared across every organizational area. Addressing existing challenges programmatically includes overcoming necessary but insufficient prerequisites and developing a disciplined, repeatable means of improving business objectives. This process (based on the theory of constraints) is where the strategic data work really occurs, as organizations identify prioritized areas where better assets, literacy, and support (Data Strategy components) can help an organization better achieve specific strategic objectives. Then the process becomes lather, rinse, and repeat. Several complementary concepts are also covered, including:
- A cohesive argument for why Data Strategy is necessary for effective Data Governance
- An overview of prerequisites for effective strategic use of Data Strategy, as well as common pitfalls
- A repeatable process for identifying and removing data constraints
- The importance of balancing business operation and innovation
How Enterprises are Using NoSQL for Mission-Critical ApplicationsDATAVERSITY
NoSQL databases including Couchbase are increasingly being selected as the backend technology for web and mobile apps. Document databases in particular are well suited for a large number of different use cases as an operational datastore.
In this webinar, Perry Krug, Principal Solutions Architect at Couchbase, will give a brief overview of Couchbase Server, a document database and its underlying distributed architecture. In addition, Perry will share how some of the biggest brands in the world use Couchbase, including:
Paypal A scalable NoSQL and big data architecture with real time analytics
Concur A highly available cache solution that supports 1B operations/day
Amadeus A backend data store that supports 1.6B transactions/day
How you can gain rapid insights and create more flexibility by capturing and storing data from a variety of sources and structures into a NoSQL database.
Information management plays a critical role in supporting strategic business initiatives. Despite the apparent value of providing the data infrastructure for these initiatives, many executives question the economic feasibility of business intelligence and analytics. This requires information professionals to calculate and present the business value in terms business executives can understand.
Unfortunately, most IT professionals lack the knowledge required to develop comprehensive cost-benefit analyses and return on investment (ROI) measurements.
This session provides a framework to help IT professionals research, measure, and present the economic value of a proposed or existing information initiative. The session will provide practical advice about how to calculate ROI, which formula to use, and how to collect the necessary information.
Building a Collaborative Data ArchitectureDATAVERSITY
Data isn’t stored in a vault where no one can touch it. Both business and technical professionals need the right access to the data, and need to work together to ensure that they all have the same understanding as to what the data means. Ron Huizenga, Senior Product Manager for ER/Studio, will discuss and demonstrate how data models, business glossaries, and collaboration can enhance your data architecture.
A presentation to explain why selling of Information Architecture is important and how the architect has to include strategy points even before the IA is sold.
Slides: How AI Makes Analytics More HumanDATAVERSITY
People think AI makes analytics less human, replacing human decision making. But the truth is, AI actually makes analytics more human. Augmented analytics are helping organizations finally break through the low levels of adoption and limitations typical of 2nd generation visualization tools.
Most business problems cannot be solved purely by algorithms or machine learning — they require human interaction and perspective. Uniting precedent-based machine learning systems with natural human intuition and curiosity is the foundation of 3rd generation BI and democratizing data across an enterprise.
It is a natural flow to enhance your data eco-system by deploying a platform with augmented intelligence to work alongside users in the pursuit of surfacing new insights, automating tasks, and supporting natural language interaction. All work as accelerators for achieving active intelligence and Data Literacy.
In that session we will discuss about Data Governance, mainly around that fantastic platform Power BI (but also around on-prem concerns).
How to avoid dataset-hell ? What are the best practices for sharing queries ? Who is the famous Data Steward and what is its role in a department or in the whole company ? How do you choose the right person ?
Keywords : Power Query, Data Management Gateway, Power BI Admin Center, Datastewardship, SharePoint 2013, eDiscovery
Level 200
DataEd Slides: Data Management + Data Strategy = InteroperabilityDATAVERSITY
Few organizations operate without having to exchange data. (Many do it professionally and well!) The larger the data exchange burden (DEB), the greater the organizational overhead incurred. This death by 1,000 cuts must be factored into each organization’s calculations. Unfortunately, most organizations do not know if their organization’s DEB is great or small. A somewhat greater number of organizations have organized Data Management practices. Focusing Data Management efforts on increasing interoperability by decreasing the DEB friction is a good area to “practice.”
Learning Objectives:
• Gaining a good understanding of both important topics
• Understanding that data only operates at a very intricate, specifically dependent intent and what this means
• Understand state-of-the-practice
• Coordination is key, requiring necessary but insufficient interdependencies and sequencing
• Practice makes perfect
Data Lakehouse, Data Mesh, and Data Fabric (r1)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. I’ll include use cases so you can see what approach will work best for your big data needs.
Traditionally, data integration has meant compromise. No matter how rapidly data architects and developers could complete a project before its deadline, speed would always come at the expense of quality. On the other hand, if they focused on delivering a quality project, it would generally drag on for months thus exceeding its deadline. Finally, if the teams concentrated on both quality and rapid delivery, the costs would invariably exceed the budget. Regardless of which path you chose, the end result would be less than desirable. This led some experts to revisit the scope of data integration. This write up shall focus on the same issue.
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
An Introduction to Data Virtualization in 2018Denodo
Watch full webinar on demand here: https://goo.gl/Rdrc1w
"Through 2020, 50% of enterprises will implement some form of data virtualization as one enterprise production option for data integration" according to Gartner. It is clear that data virtualization has become a driving force for companies to implement an agile, real-time and flexible enterprise data architecture.
Attend this session to learn:
• What data virtualization actually means and how it differs from traditional data integration approaches
• The all important use cases and key patterns of data virtualization
• What to expect in the upcoming sessions in the Packed Lunch Webinar Series, which will take a deeper dive into various challenges solved by data virtualization in big data analytics, cloud migration and various other scenarios
Agenda:
• Introduction & benefits of DV
• Summary & next steps
• Q&A
The Pivotal Business Data Lake provides a flexible blueprint to meet your business's future information and analytics needs while avoiding the pitfalls of typical EDW implementations. Pivotal’s products will help you overcome challenges like reconciling corporate and local needs, providing real-time access to all types of data, integrating data from multiple sources and in multiple formats, and supporting ad hoc analysis.
Watch Paul's session from Fast Data Strategy on-demand here: https://goo.gl/3veKqw
"Through 2020, 50% of enterprises will implement some form of data virtualization as one enterprise production option for data integration" according to Gartner. It is clear that data virtualization has become a driving force for companies to implement an agile, real-time and flexible enterprise data architecture.
Attend this session to learn:
• What data virtualization actually means and how it differs from traditional data integration approaches
• The most important use cases and key patterns of data virtualization
• The benefits of data virtualization
Watch full webinar here: https://bit.ly/2vN59VK
What started to evolve as the most agile and real-time enterprise data fabric, data virtualization is proving to go beyond its initial promise and is becoming one of the most important enterprise big data fabrics.
Attend this session to learn:
- What data virtualization really is.
- How it differs from other enterprise data integration technologies.
- Why data virtualization is finding enterprise-wide deployment inside some of the largest organizations.
Traditional BI vs. Business Data Lake – A ComparisonCapgemini
Traditional Business Intelligence (BI) systems provide various levels and kinds of analyses on structured data but they are not designed to handle unstructured data.
For these systems Big Data brings big problems because the data that flows in may be either structured or unstructured. That makes them hugely limited when it comes to delivering Big Data benefits.
The way forward is a complete rethink of the way we use BI - in terms of how the data is ingested, stored and analyzed.
More information: http://www.capgemini.com/big-data-analytics/pivotal
Building an Effective Data Warehouse ArchitectureJames Serra
Why use a data warehouse? What is the best methodology to use when creating a data warehouse? Should I use a normalized or dimensional approach? What is the difference between the Kimball and Inmon methodologies? Does the new Tabular model in SQL Server 2012 change things? What is the difference between a data warehouse and a data mart? Is there hardware that is optimized for a data warehouse? What if I have a ton of data? During this session James will help you to answer these questions.
Presentation on Data Mesh: The paradigm shift is a new type of eco-system architecture, which is a shift left towards a modern distributed architecture in which it allows domain-specific data and views “data-as-a-product,” enabling each domain to handle its own data pipelines.
Modern Integrated Data Environment - Whitepaper | QuboleVasu S
A whit-paper is about building a modern data platform for data driven organisations with using cloud data warehouse with modern data platform architecture
https://www.qubole.com/resources/white-papers/modern-integrated-data-environment
Data Ninja Webinar Series: Realizing the Promise of Data LakesDenodo
Watch the full webinar: Data Ninja Webinar Series by Denodo: https://goo.gl/QDVCjV
The expanding volume and variety of data originating from sources that are both internal and external to the enterprise are challenging businesses in harnessing their big data for actionable insights. In their attempts to overcome big data challenges, organizations are exploring data lakes as consolidated repositories of massive volumes of raw, detailed data of various types and formats. But creating a physical data lake presents its own hurdles.
Attend this session to learn how to effectively manage data lakes for improved agility in data access and enhanced governance.
This is session 5 of the Data Ninja Webinar Series organized by Denodo. If you want to learn more about some of the solutions enabled by data virtualization, click here to watch the entire series: https://goo.gl/8XFd1O
The Shifting Landscape of Data IntegrationDATAVERSITY
Enterprises and organizations from every industry and scale are working to leverage data to achieve their strategic objectives — whether they are to be more profitable, effective, risk-tolerant, prepared, sustainable, and/or adaptable in an ever-changing world. Data has exploded in volume during the last decade as humans and machines alike produce data at an exponential pace. Also, exciting technologies have emerged around that data to improve our abilities and capabilities around what we can do with data.
Behind this data revolution, there are forces at work, causing enterprises to shift the way they leverage data and accelerate the demand for leverageable data. Organizations (and the climates in which they operate) are becoming more and more complex. They are also becoming increasingly digital and, thus, dependent on how data informs, transforms, and automates their operations and decisions. With increased digitization comes an increased need for both scale and agility at scale.
In this session, we have undertaken an ambitious goal of evaluating the current vendor landscape and assessing which platforms have made, or are in the process of making, the leap to this new generation of Data Management and integration capabilities.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
2. DII – The new kid in the
block
Data Integration and Interoperability (DII) describes processes
related to the movement and consolidation of data within
and between data stores, applications and organizations.
Why we’re geeking out for DII
1. SOA/Microservices are becoming more popular.
2. Integration of structured and unstructured data
3. Deliver value faster… avoid ROGUE users
4. Although it’s not new, DII in DMBOK provides clear
guidelines to organizations aiming to become more efficient
through IT.
5. History
Data virtualization exists
since Bill Inmon
popularized data
warehouse in the 1990s.
But virtual models back
then were not very
popular due to the lack
of computer power
available (or
accessible).
Today, change in data
types and business
expectations on
information velocity have
made virtualization a
more popular concept.
Did you know?
Last time Bill Inmon wrote about
Data virtualization he compared
it to a frustrating whack a mole
game, where no matter how
much you hit the mole… it
keeps coming back!
http://www.b-eye-
network.com/view/9956
8. Why data virtualization?
Fast and Easy
• Rapid data integration which enables a faster time to solution
• Integrations and changes are easy (No need to update Extractions, tables, DataMart's)
Integrate more!
• Opportunity to integrate structured and unstructured data
Cheaper and more secure
• Less expensive to maintain
• No need to replicate data
• Reduces overhead of management of data integration systems (Easier + Faster = Less
reqauired resources)
Agile
• Enables iterative development with quick deliverables (Note: very important one since in most
cases, users don’t know what they want… too many iterations)
• Developers are more focused on business instead of understanding the mechanics of data
manipulation (Why? Because Data virtualization tools automatically connect to many data
sources )
9.
10. Use Cases
Data Warehouse augmentation
Problem
• Bringing in new data sources to a data warehouse
takes a significant amount of effort, but even more
so, if the data sources include unstructured data.
Fix
• Data virtualization can be applied to augment
existing data warehouse with virtual views that
incorporate unstructured data.
Support ETL process
Problem
• It is sometimes too complicated to access web
services data, extract it and make it part of the ETL,
specially if you need to develop access methods for
external or new types of data.
Fix
• Data virtualization tools have access methods which
can be used to easily extract data from web
services, pre-process this data and have it ready to
Data Warehouse Federation/Canonical
Problem
• Some organizations have multiple separate data
warehouses which may take too much effort to
integrate.
Fix
• Data virtualization allows to quickly generate
federated views of all these data warehouses and
integrate this data for different services. Individual
warehouses continue to operate with no
interruptions. (Same thing for DWH migrations!)
11. Use Cases
Data Warehouse prototyping
Problem
• Organizations are moving to agile development,
where iterations and short term sprints are key to
delivering value on a weekly ot bi weekly basis.
Fix
• When data prototypes are built fast and are
validated by users, this then generates a proven
product that can then be materialized saving time
and therefore money.
Data Mashups
Problem
• Web mashups are enabled by APIs and most
corporate data sources do not have accessible APIs
to support this mashup process.
Fix
• Data virtualization tools are enables of mashups
since they use same protocols and data delivery
formats as APIs.
Master Data on Steroids– Past, present and future
data
Problem
• Master Data Hubs traditionally only hold identity and
descriptive information, but transactional data is
usually not stored in MDHs.
Fix
• With data virtualization, you could make a canonical
layer where you would input data from the MDH and
other sources and enrich master data with
summarized transactional data. (E.g. adding value
of customer over time, purchasing forecast etc…)
12. So is ETL going away?
This does not mean ETL is not needed, its more around identifying when ETL is not
enough, and use virtualization to enhance Data integration! When ETL is too slow, or
data sources are difficult to access or data types are challenging.
Maybe in the future it’ll be the other way around, where we’ll look at ETL for cases
when data virtualization is not enough. For instance, when you need to perform highly
complex transformations that could impact performance in a virtual database.
Today, It is common to virtualize in development and materialize in production.
Misconceptions
1. VDB DOES NOT replace a DWH. VDB enhances DWH by:
• Combine structured and unstructured data into a single data layer