Retour d'expérience sur un projet de Business Intelligence réalisé à l'EVAM selon une méthodologie Agile et avec un modèle de données Data Vault. Présentation faite lors du Swiss Data Forum du 24 novembre 2015 à Lausanne
DAMA, Oregon Chapter, 2012 presentation - an introduction to Data Vault modeling. I will be covering parts of the methodology, comparison and contrast of issues in general for the EDW space. Followed by a brief technical introduction of the Data Vault modeling method.
After the presentation i I will be providing a demonstration of the ETL loading layers, LIVE!
You can find more on-line training at: http://LearnDataVault.com/training
This document discusses Data Vault fundamentals and best practices. It introduces Data Vault modeling, which involves modeling hubs, links, and satellites to create an enterprise data warehouse that can integrate data sources, provide traceability and history, and adapt incrementally. The document recommends using data virtualization rather than physical data marts to distribute data from the Data Vault. It also provides recommendations for further reading on Data Vault, Ensemble modeling, data virtualization, and certification programs.
This was a presentation I gave to IRM UK conference in November 2009. It covers some interesting details around the steps you should take to build your Data Vault, and an overview as to why re-engineering creeps in to your existing silo solutions.
Agile BI via Data Vault and ModelstormingDaniel Upton
Audience: Business Intelligence Architects, Project Managers and Sponsors. This slideshow accompanies a video presentation of the same name, available at http://youtu.be/e0cHFdeGEeE.
Part 2 of a 2 part presentation that I did in 2009, this presentation covers more about unstructured data, and operational data vault components. YES, even then I was commenting on how this market will evolve. IF you want to use these slides, please let me know, and add: "(C) Dan Linstedt, all rights reserved, http://LearnDataVault.com" in a VISIBLE fashion on your slides.
Given at Oracle Open World 2011: Not to be confused with Oracle Database Vault (a commercial db security product), Data Vault Modeling is a specific data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It has been in use globally for over 10 years now but is not widely known. The purpose of this presentation is to provide an overview of the features of a Data Vault modeled EDW that distinguish it from the more traditional third normal form (3NF) or dimensional (i.e., star schema) modeling approaches used in most shops today. Topics will include dealing with evolving data requirements in an EDW (i.e., model agility), partitioning of data elements based on rate of change (and how that affects load speed and storage requirements), and where it fits in a typical Oracle EDW architecture. See more content like this by following my blog http://kentgraziano.com or follow me on twitter @kentgraziano.
The document discusses using the Data Vault 2.0 methodology for agile data mining projects. It provides background on a customer segmentation project for a motor insurance company. The Data Vault 2.0 modeling approach is described as well as the CRISP-DM process model. An example is then shown applying several iterations of a decision tree model to a sample database, improving results with each iteration by adding additional attributes to the Data Vault 2.0 model and RapidMiner process. The conclusions state that Data Vault 2.0 provides a flexible data model that supports an agile approach to data mining projects by allowing incremental changes to the model and attributes.
This is a presentation I gave in 2006 for Bill Inmon. The presentation covers Data Vault and how it integrates with Bill Inmon's DW2.0 vision. This is focused on the business intelligence side of the house.
IF you want to use these slides, please put (C) Dan Linstedt, all rights reserved, http://LearnDataVault.com
DAMA, Oregon Chapter, 2012 presentation - an introduction to Data Vault modeling. I will be covering parts of the methodology, comparison and contrast of issues in general for the EDW space. Followed by a brief technical introduction of the Data Vault modeling method.
After the presentation i I will be providing a demonstration of the ETL loading layers, LIVE!
You can find more on-line training at: http://LearnDataVault.com/training
This document discusses Data Vault fundamentals and best practices. It introduces Data Vault modeling, which involves modeling hubs, links, and satellites to create an enterprise data warehouse that can integrate data sources, provide traceability and history, and adapt incrementally. The document recommends using data virtualization rather than physical data marts to distribute data from the Data Vault. It also provides recommendations for further reading on Data Vault, Ensemble modeling, data virtualization, and certification programs.
This was a presentation I gave to IRM UK conference in November 2009. It covers some interesting details around the steps you should take to build your Data Vault, and an overview as to why re-engineering creeps in to your existing silo solutions.
Agile BI via Data Vault and ModelstormingDaniel Upton
Audience: Business Intelligence Architects, Project Managers and Sponsors. This slideshow accompanies a video presentation of the same name, available at http://youtu.be/e0cHFdeGEeE.
Part 2 of a 2 part presentation that I did in 2009, this presentation covers more about unstructured data, and operational data vault components. YES, even then I was commenting on how this market will evolve. IF you want to use these slides, please let me know, and add: "(C) Dan Linstedt, all rights reserved, http://LearnDataVault.com" in a VISIBLE fashion on your slides.
Given at Oracle Open World 2011: Not to be confused with Oracle Database Vault (a commercial db security product), Data Vault Modeling is a specific data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It has been in use globally for over 10 years now but is not widely known. The purpose of this presentation is to provide an overview of the features of a Data Vault modeled EDW that distinguish it from the more traditional third normal form (3NF) or dimensional (i.e., star schema) modeling approaches used in most shops today. Topics will include dealing with evolving data requirements in an EDW (i.e., model agility), partitioning of data elements based on rate of change (and how that affects load speed and storage requirements), and where it fits in a typical Oracle EDW architecture. See more content like this by following my blog http://kentgraziano.com or follow me on twitter @kentgraziano.
The document discusses using the Data Vault 2.0 methodology for agile data mining projects. It provides background on a customer segmentation project for a motor insurance company. The Data Vault 2.0 modeling approach is described as well as the CRISP-DM process model. An example is then shown applying several iterations of a decision tree model to a sample database, improving results with each iteration by adding additional attributes to the Data Vault 2.0 model and RapidMiner process. The conclusions state that Data Vault 2.0 provides a flexible data model that supports an agile approach to data mining projects by allowing incremental changes to the model and attributes.
This is a presentation I gave in 2006 for Bill Inmon. The presentation covers Data Vault and how it integrates with Bill Inmon's DW2.0 vision. This is focused on the business intelligence side of the house.
IF you want to use these slides, please put (C) Dan Linstedt, all rights reserved, http://LearnDataVault.com
A strong relationship with the founder
of Data Vault for over 3 years now.
Supporting your business with 40+
certified consultants.
Incorporated as the preferred
Enterprise Data Warehouse modelling
paradigm in the Logica BI Framework.
Satisfied customers in many countries
and industry sectors
Data Vault: Data Warehouse Design Goes AgileDaniel Upton
Data Warehouse (especially EDW) design needs to get Agile. This whitepaper introduces Data Vault to newcomers, and describes how it adds agility to DW best practices.
This was a presentation about Data Warehousing, where it's going - covers operational Data Vault. I gave this presentation in 2009 at an Array Conference in the Netherlands.
IF you want to use these slides, then please let me know, and add: "(C) Dan Linstedt, all rights reserved, http://LearnDataVault.com"
Shorter time to insight more adaptable less costly bi with end to end modelst...Daniel Upton
For Data Project Leaders: This data warehouse data modeling approach enables shorter time to insights, lower cost and greater adaptability to external changes by combining my End to End Data Modelstorming concept with Data Vault modeling.
The document discusses operational data warehousing and the Data Vault model. It begins with an agenda for the presentation and introduction of the speaker. It then provides a short review of the Data Vault model. The remainder of the document discusses operational data warehousing, how the Data Vault model is well-suited for this purpose, and the benefits it provides including flexibility, scalability, and productivity. It also discusses how tools and technologies are advancing to support automation and self-service business intelligence using an operational data warehouse architecture based on the Data Vault model.
Agile Data Warehouse Modeling: Introduction to Data Vault Data ModelingKent Graziano
The document introduces Data Vault modeling as an agile approach to data warehousing. It discusses how Data Vault addresses some limitations of traditional dimensional modeling by allowing for more flexible, adaptable designs. The Data Vault model consists of three simple structures - hubs, links, and satellites. Hubs contain unique business keys, links represent relationships between keys, and satellites hold descriptive attributes. This structure supports incremental development and rapid changes to meet evolving business needs in an agile manner.
Data Vault Modeling and Methodology introduction that I provided to a Montreal event in September 2011. It covers an introduction and overview of the Data Vault components for Business Intelligence and Data Warehousing. I am Dan Linstedt, the author and inventor of Data Vault Modeling and methodology.
If you use the images anywhere in your presentations, please credit http://LearnDataVault.com as the source (me).
Thank-you kindly,
Daniel Linstedt
This document discusses agile data warehouse design. It begins with an overview of data warehousing, including definitions of a data warehouse and common architectures. It then covers traditional waterfall and agile approaches to BI/WH development. The agile section focuses on an incremental lifecycle and agile dimensional modeling techniques like the 7Ws framework and BEAM methodology, which use natural language and collaboration to design models around business questions.
Agile Data Engineering - Intro to Data Vault Modeling (2016)Kent Graziano
The document provides an introduction to Data Vault data modeling and discusses how it enables agile data warehousing. It describes the core structures of a Data Vault model including hubs, links, and satellites. It explains how the Data Vault approach provides benefits such as model agility, productivity, and extensibility. The document also summarizes the key changes in the Data Vault 2.0 methodology.
Agile Methods and Data Warehousing (2016 update)Kent Graziano
This presentation takes a look at the Agile Manifesto and the 12 Principles of Agile Development and discusses how these apply to Data Warehousing and Business Intelligence projects. Several examples and details from my past experience are included. Includes more details on using Data Vault as well. (I gave this presentation at OUGF14 in Helsinki, Finland and again in 2016 for TDWI Nashville.)
Not to be confused with Oracle Database Vault (a commercial db security product), Data Vault Modeling is a specific data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It is not a replacement for star schema data marts (and should not be used as such). This approach has been used in projects around the world (Europe, Australia, USA) for the last 10 years but is still not widely known or understood. The purpose of this presentation is to provide attendees with a detailed introduction to the technical components of the Data Vault Data Model, what they are for and how to build them. The examples will give attendees the basics for how to build, and design structures when using the Data Vault modeling technique. The target audience is anyone wishing to explore implementing a Data Vault style data model for an Enterprise Data Warehouse, Operational Data Warehouse, or Dynamic Data Integration Store. See more content like this by following my blog http://kentgraziano.com or follow me on twitter @kentgraziano.
A Lean Data Warehouse, compared with a traditional one (with a dimensional or 3rd normal form model), is faster to deliver, freer of waste, and inherently more adaptable to change. From my experience in the trenches, each of these benefits fit squarely in the 'must have' category. Data Vault is an excellent logical architecture with which to design a Lean Data Warehouse. This article describes the priorities of a Lean Data Warehouse, and compares the two traditional modeling methods with Data Vault, concluding that Data Vault is more suited to deliver on those Lean priorities.
(OTW13) Agile Data Warehousing: Introduction to Data Vault ModelingKent Graziano
This is the presentation I gave at OakTable World 2013 in San Francisco. #OTW13 was held at the Children's Creativity Museum next to the Moscone Convention Center and was in parallel with Oracle OpenWorld 2013.
The session discussed our attempts to be more agile in designing enterprise data warehouses and how the Data Vault Data Modeling technique helps in that approach.
This document discusses the conceptual data vault modeling approach. It presents a sample data vault model and describes schema transformation rules to move from a 3NF model to a raw vault/staging area to dimensional structures. It also discusses leveraging existing modeling tools to implement the conceptual data vault by feeding the tools configuration through XML transformation or other means. The work on formalizing and fully implementing the conceptual data vault approach is described as still in progress.
Data Warehouse Tutorial For Beginners | Data Warehouse Concepts | Data Wareho...Edureka!
This Data Warehouse Tutorial For Beginners will give you an introduction to data warehousing and business intelligence. You will be able to understand basic data warehouse concepts with examples. The following topics have been covered in this tutorial:
1. What Is The Need For BI?
2. What Is Data Warehousing?
3. Key Terminologies Related To Data Warehouse Architecture:
a. OLTP Vs OLAP
b. ETL
c. Data Mart
d. Metadata
4. Data Warehouse Architecture
5. Demo: Creating A Data Warehouse
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...Cloudera, Inc.
SGI has been a leading commercial vendor of Hadoop clusters since 2008. Leveraging SGI's experience with high performance clusters at scale, SGI has delivered individual Hadoop clusters of up to 4000 nodes. Integration, performance, and management all become issues at scale, and Hadoop clusters scale! In this presentation, SGI will discuss representative customer use cases, major design considerations for performance and power optimization, how integrated Hadoop solutions leveraging CDH, SGI Rackable clusters, and SGI Management Center best meet customer needs, and how SGI envisions the needs of enterprise customers evolving as Hadoop continues to move into mainstream adoption.
Are You Killing the Benefits of Your Data Lake?Denodo
Watch the full webinar on-demand here: https://goo.gl/RL1ZSa
Data lakes are centralized data repositories. Data needed by data scientists is physically copied to a data lake which serves as a one storage environment. This way, data scientists can access all the data from only one entry point – a one-stop shop to get the right data. However, such an approach is not always feasible for all the data and limits it’s use to solely data scientists, making it a single-purpose system.
So, what’s the solution?
A multi-purpose data lake allows a broader and deeper use of the data lake without minimizing the potential value for data science and without making it an inflexible environment
Attend this session to learn:
• Disadvantages and limitations that are weakening or even killing the potential benefits of a data lake.
• Why a multi-purpose data lake is essential in building a universal data delivery system.
• How to build a logical multi-purpose data lake using data virtualization.
Do not miss this opportunity to make your data lake project successful and beneficial.
The document introduces Visual DataVault, a modeling language for visually expressing Data Vault models. It aims to generate DDL from models and support Microsoft Office. The language defines basic entities like hubs, links, satellites and reference tables. It also covers query assistant tables, computed structures, exploration links and business vault tables to enhance the raw data vault. Some remarks note it focuses on logical not physical modeling and more features are planned.
These are the slides from my talk at Data Day Texas 2016 (#ddtx16).
The world of data warehousing has changed! With the advent of Big Data, Streaming Data, IoT, and The Cloud, what is a modern data management professional to do? It may seem to be a very different world with different concepts, terms, and techniques. Or is it? Lots of people still talk about having a data warehouse or several data marts across their organization. But what does that really mean today in 2016? How about the Corporate Information Factory (CIF), the Data Vault, an Operational Data Store (ODS), or just star schemas? Where do they fit now (or do they)? And now we have the Extended Data Warehouse (XDW) as well. How do all these things help us bring value and data-based decisions to our organizations? Where do Big Data and the Cloud fit? Is there a coherent architecture we can define? This talk will endeavor to cut through the hype and the buzzword bingo to help you figure out what part of this is helpful. I will discuss what I have seen in the real world (working and not working!) and a bit of where I think we are going and need to go in 2016 and beyond.
The Data Warehouse plays a central role in any BI solution: it's the back end upon which everything in the coming years will be created. It must be capable of being flexible in order to support the fast changes needed by today's business, but also with a well-know and well-defined structure in order to support the "engineerization" of its development process, making it cost effective. In this full-day session, we will discuss architectural design details and techniques, Agile Modeling, unit testing, automation, and software engineering applied to a Data Warehouse project.
The only way to do this is to have a clear idea of its architecture, understanding the concepts of measures and dimensions, and a proven engineered way to build it so that quality and stability can go hand-in-hand with cost reduction and scalability. This will allow you to start your BI project in the best way possible avoiding errors, making implementation effective and efficient, building the groundwork for a winning Agile approach, and helping you to define the way in which your team should work so that your BI solution will stand the test of time.
Product portfolio management involves managing multiple products and prioritizing them. It is important to identify a single source of truth for product data, typically a tool like JIRA. This allows automating tasks like planning, tracking status, and identifying resource needs. It provides transparency by letting all stakeholders see changes and their impact. Managing a product portfolio well requires consolidating products, platforms, skills mapping, and automating as many processes as possible to handle the complexities involved.
A strong relationship with the founder
of Data Vault for over 3 years now.
Supporting your business with 40+
certified consultants.
Incorporated as the preferred
Enterprise Data Warehouse modelling
paradigm in the Logica BI Framework.
Satisfied customers in many countries
and industry sectors
Data Vault: Data Warehouse Design Goes AgileDaniel Upton
Data Warehouse (especially EDW) design needs to get Agile. This whitepaper introduces Data Vault to newcomers, and describes how it adds agility to DW best practices.
This was a presentation about Data Warehousing, where it's going - covers operational Data Vault. I gave this presentation in 2009 at an Array Conference in the Netherlands.
IF you want to use these slides, then please let me know, and add: "(C) Dan Linstedt, all rights reserved, http://LearnDataVault.com"
Shorter time to insight more adaptable less costly bi with end to end modelst...Daniel Upton
For Data Project Leaders: This data warehouse data modeling approach enables shorter time to insights, lower cost and greater adaptability to external changes by combining my End to End Data Modelstorming concept with Data Vault modeling.
The document discusses operational data warehousing and the Data Vault model. It begins with an agenda for the presentation and introduction of the speaker. It then provides a short review of the Data Vault model. The remainder of the document discusses operational data warehousing, how the Data Vault model is well-suited for this purpose, and the benefits it provides including flexibility, scalability, and productivity. It also discusses how tools and technologies are advancing to support automation and self-service business intelligence using an operational data warehouse architecture based on the Data Vault model.
Agile Data Warehouse Modeling: Introduction to Data Vault Data ModelingKent Graziano
The document introduces Data Vault modeling as an agile approach to data warehousing. It discusses how Data Vault addresses some limitations of traditional dimensional modeling by allowing for more flexible, adaptable designs. The Data Vault model consists of three simple structures - hubs, links, and satellites. Hubs contain unique business keys, links represent relationships between keys, and satellites hold descriptive attributes. This structure supports incremental development and rapid changes to meet evolving business needs in an agile manner.
Data Vault Modeling and Methodology introduction that I provided to a Montreal event in September 2011. It covers an introduction and overview of the Data Vault components for Business Intelligence and Data Warehousing. I am Dan Linstedt, the author and inventor of Data Vault Modeling and methodology.
If you use the images anywhere in your presentations, please credit http://LearnDataVault.com as the source (me).
Thank-you kindly,
Daniel Linstedt
This document discusses agile data warehouse design. It begins with an overview of data warehousing, including definitions of a data warehouse and common architectures. It then covers traditional waterfall and agile approaches to BI/WH development. The agile section focuses on an incremental lifecycle and agile dimensional modeling techniques like the 7Ws framework and BEAM methodology, which use natural language and collaboration to design models around business questions.
Agile Data Engineering - Intro to Data Vault Modeling (2016)Kent Graziano
The document provides an introduction to Data Vault data modeling and discusses how it enables agile data warehousing. It describes the core structures of a Data Vault model including hubs, links, and satellites. It explains how the Data Vault approach provides benefits such as model agility, productivity, and extensibility. The document also summarizes the key changes in the Data Vault 2.0 methodology.
Agile Methods and Data Warehousing (2016 update)Kent Graziano
This presentation takes a look at the Agile Manifesto and the 12 Principles of Agile Development and discusses how these apply to Data Warehousing and Business Intelligence projects. Several examples and details from my past experience are included. Includes more details on using Data Vault as well. (I gave this presentation at OUGF14 in Helsinki, Finland and again in 2016 for TDWI Nashville.)
Not to be confused with Oracle Database Vault (a commercial db security product), Data Vault Modeling is a specific data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It is not a replacement for star schema data marts (and should not be used as such). This approach has been used in projects around the world (Europe, Australia, USA) for the last 10 years but is still not widely known or understood. The purpose of this presentation is to provide attendees with a detailed introduction to the technical components of the Data Vault Data Model, what they are for and how to build them. The examples will give attendees the basics for how to build, and design structures when using the Data Vault modeling technique. The target audience is anyone wishing to explore implementing a Data Vault style data model for an Enterprise Data Warehouse, Operational Data Warehouse, or Dynamic Data Integration Store. See more content like this by following my blog http://kentgraziano.com or follow me on twitter @kentgraziano.
A Lean Data Warehouse, compared with a traditional one (with a dimensional or 3rd normal form model), is faster to deliver, freer of waste, and inherently more adaptable to change. From my experience in the trenches, each of these benefits fit squarely in the 'must have' category. Data Vault is an excellent logical architecture with which to design a Lean Data Warehouse. This article describes the priorities of a Lean Data Warehouse, and compares the two traditional modeling methods with Data Vault, concluding that Data Vault is more suited to deliver on those Lean priorities.
(OTW13) Agile Data Warehousing: Introduction to Data Vault ModelingKent Graziano
This is the presentation I gave at OakTable World 2013 in San Francisco. #OTW13 was held at the Children's Creativity Museum next to the Moscone Convention Center and was in parallel with Oracle OpenWorld 2013.
The session discussed our attempts to be more agile in designing enterprise data warehouses and how the Data Vault Data Modeling technique helps in that approach.
This document discusses the conceptual data vault modeling approach. It presents a sample data vault model and describes schema transformation rules to move from a 3NF model to a raw vault/staging area to dimensional structures. It also discusses leveraging existing modeling tools to implement the conceptual data vault by feeding the tools configuration through XML transformation or other means. The work on formalizing and fully implementing the conceptual data vault approach is described as still in progress.
Data Warehouse Tutorial For Beginners | Data Warehouse Concepts | Data Wareho...Edureka!
This Data Warehouse Tutorial For Beginners will give you an introduction to data warehousing and business intelligence. You will be able to understand basic data warehouse concepts with examples. The following topics have been covered in this tutorial:
1. What Is The Need For BI?
2. What Is Data Warehousing?
3. Key Terminologies Related To Data Warehouse Architecture:
a. OLTP Vs OLAP
b. ETL
c. Data Mart
d. Metadata
4. Data Warehouse Architecture
5. Demo: Creating A Data Warehouse
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...Cloudera, Inc.
SGI has been a leading commercial vendor of Hadoop clusters since 2008. Leveraging SGI's experience with high performance clusters at scale, SGI has delivered individual Hadoop clusters of up to 4000 nodes. Integration, performance, and management all become issues at scale, and Hadoop clusters scale! In this presentation, SGI will discuss representative customer use cases, major design considerations for performance and power optimization, how integrated Hadoop solutions leveraging CDH, SGI Rackable clusters, and SGI Management Center best meet customer needs, and how SGI envisions the needs of enterprise customers evolving as Hadoop continues to move into mainstream adoption.
Are You Killing the Benefits of Your Data Lake?Denodo
Watch the full webinar on-demand here: https://goo.gl/RL1ZSa
Data lakes are centralized data repositories. Data needed by data scientists is physically copied to a data lake which serves as a one storage environment. This way, data scientists can access all the data from only one entry point – a one-stop shop to get the right data. However, such an approach is not always feasible for all the data and limits it’s use to solely data scientists, making it a single-purpose system.
So, what’s the solution?
A multi-purpose data lake allows a broader and deeper use of the data lake without minimizing the potential value for data science and without making it an inflexible environment
Attend this session to learn:
• Disadvantages and limitations that are weakening or even killing the potential benefits of a data lake.
• Why a multi-purpose data lake is essential in building a universal data delivery system.
• How to build a logical multi-purpose data lake using data virtualization.
Do not miss this opportunity to make your data lake project successful and beneficial.
The document introduces Visual DataVault, a modeling language for visually expressing Data Vault models. It aims to generate DDL from models and support Microsoft Office. The language defines basic entities like hubs, links, satellites and reference tables. It also covers query assistant tables, computed structures, exploration links and business vault tables to enhance the raw data vault. Some remarks note it focuses on logical not physical modeling and more features are planned.
These are the slides from my talk at Data Day Texas 2016 (#ddtx16).
The world of data warehousing has changed! With the advent of Big Data, Streaming Data, IoT, and The Cloud, what is a modern data management professional to do? It may seem to be a very different world with different concepts, terms, and techniques. Or is it? Lots of people still talk about having a data warehouse or several data marts across their organization. But what does that really mean today in 2016? How about the Corporate Information Factory (CIF), the Data Vault, an Operational Data Store (ODS), or just star schemas? Where do they fit now (or do they)? And now we have the Extended Data Warehouse (XDW) as well. How do all these things help us bring value and data-based decisions to our organizations? Where do Big Data and the Cloud fit? Is there a coherent architecture we can define? This talk will endeavor to cut through the hype and the buzzword bingo to help you figure out what part of this is helpful. I will discuss what I have seen in the real world (working and not working!) and a bit of where I think we are going and need to go in 2016 and beyond.
The Data Warehouse plays a central role in any BI solution: it's the back end upon which everything in the coming years will be created. It must be capable of being flexible in order to support the fast changes needed by today's business, but also with a well-know and well-defined structure in order to support the "engineerization" of its development process, making it cost effective. In this full-day session, we will discuss architectural design details and techniques, Agile Modeling, unit testing, automation, and software engineering applied to a Data Warehouse project.
The only way to do this is to have a clear idea of its architecture, understanding the concepts of measures and dimensions, and a proven engineered way to build it so that quality and stability can go hand-in-hand with cost reduction and scalability. This will allow you to start your BI project in the best way possible avoiding errors, making implementation effective and efficient, building the groundwork for a winning Agile approach, and helping you to define the way in which your team should work so that your BI solution will stand the test of time.
Product portfolio management involves managing multiple products and prioritizing them. It is important to identify a single source of truth for product data, typically a tool like JIRA. This allows automating tasks like planning, tracking status, and identifying resource needs. It provides transparency by letting all stakeholders see changes and their impact. Managing a product portfolio well requires consolidating products, platforms, skills mapping, and automating as many processes as possible to handle the complexities involved.
04 Ace 2010 Mi Tek Aras Plm Open User DeploymentProdeos
This document describes one company's journey to implement the Aras PLM solution to unify their product development processes across multiple divisions. They were using paper-based systems that lacked communication and visibility. Aras provided the necessary functionality including project management, engineering tools, and quality management. It also allowed for customization and ease of use. The company was able to gain control of their processes and successfully implement Aras after an initial failed attempt with another solution years prior.
Agile IT Operatinos - Getting to Daily ReleasesLeadingAgile
Getting to Daily Releases with Agile IT Operations. Devin Hedge, Enterprise Transformation Consultant talks to a group at Triagile about the Six Key Areas to focus on when attempting to transform IT Operations with Lean and Agile principles. The talk covers Service Engineering, IT Operations, and the Tier 1 Support/NOC organizations. Kanban, Service Management (ITSM), and what it means to have a DevOps orientation.
John Hartman, Director of Project Management Systems at CH2M HILL, discussed challenges with traditional document control processes and how the Digital Project Portal from Lifecycle Technologies addresses them. Traditional document control consumes 8-10% of engineering hours through inefficient paper-based processes for submitting, reviewing, and approving documents. The Digital Project Portal streamlines these processes by enabling multi-file uploads, powerful search capabilities, paperless digital workbenches for simultaneous reviews, and improved submittal management visibility. A demonstration showed how the Portal reduces document control costs and improves efficiency.
Process mining: The role of Data in Business ProcessesBonitasoft
The evolution both in the area of Process Management and Analytics has created an environment of abundant information that is easily generated and intuitively consumed. This trend allows a more dynamic, proactive and fast adaptation to the existing demands -or even undiscovered- of the market. It also requires that non-programmer executives get even closer to the field of Data Science.
Technology leaders like us, support their users along their path of effectively making use of Data Science.
With components such as BICI - Bonita Intelligent Continuous Improvement, Bonitasoft provides intuitive tools for analysts to carry out Process and Data Mining studies and launch actions, seeking the continuous optimization of processes within the BPMS platform.
Kay Winkler, Director and Partner of NSI, and Delphine Coille, Evangelist and Community Manager at Bonitasoft, show a practical example of Process Mining in action.
Look for more information about Process Mining: https://www.bonitasoft.com/bonita-intelligent-continuous-improvement
It is quite possible to use Agile techniques for creating and maintaining a data architecture. Doing so will dramatically reduce the risk of failed data warehouse projects. This webinar will give you a quick overview of the benefits and challenges of Agile Data Modeling, Evolutionary Database Design, Agile Modeling, Conformed Dimensions, Bus Matrix, Database Refactoring, and an Agile framework for Agile data projects
Vishwanath Mallanagouda is a data warehouse application developer with over 4 years of experience working with technologies like Informatica, Oracle, SQL, and Hadoop. He has expertise in ETL tool development, data modeling, and database administration. Currently working as an application developer at Deloitte India Consulting, his past experience includes projects for banking, insurance, and public sector clients at IBM.
Dilchand Kumar has over 6 years of experience as an IT Analyst working with technologies like Informatica, Oracle, Unix, and Teradata. He has extensive experience designing and developing ETL processes and reports for clients in various industries. Some of his key projects include migrating legacy jobs to a big data environment for Wells Fargo and implementing BI solutions for Ingersoll Rand.
This document summarizes a presentation about using Autodesk Fusion Lifecycle to manage new product introduction (NPI) processes. It discusses mapping NPI needs to key business applications in Fusion Lifecycle like product management, items and BOMs management, change management, and document management. It also covers building out these applications in Fusion Lifecycle through workspace architecture and functionality. Finally, it discusses lessons learned in using Fusion Lifecycle for NPI processes, including requirements gathering, building iteratively, and training users.
Discovering New Product Introduction (NPI) using Autodesk Fusion LifecycleRazorleaf Corporation
In this session you will learn how to capitalize on Autodesk Fusion Lifecycle to manage your enterprise business processes; including new product introduction, items and BOMs, change management, document management and many more. We will discuss how to improve your organizations performance and product data visibility throughout your organization by incorporating different business applications onto a single platform. Understand how to improve compliance to your NPI and Quality processes by implementing task management with workflow validation. Learn how to track your development process through the use of connected, but dedicated, workspaces for different departmental tasks. Our hope is that attending this class will give you a tour of how Autodesk Fusion Lifecycle can transform your business, and prepare you for the next steps in implementing Fusion Lifecycle for NPI.
This document provides product management tips for business intelligence platforms and applications. It discusses measuring success through the value analytics provides to business, considering the data pyramid from raw data to reports to analytics. Ten tips are provided for creating a successful BI product, including planning for data early, understanding user needs, prioritizing user stories, embracing design principles, and validating frequently with users. The tips are grouped by product phases like discovery, planning, and definition.
Business Intelligence (BI) solutions can help manufacturing business users to analyse cost factors and make appropriate decisions for acquisition of raw material and sold goods.
A Walk Around SQL Server Data Tools | SQL Saturday#392 by James McAuliffeCCG
Databases are growing. The way we use data is changing and growing. When you have a lot of change, it has to be managed, or bad things happen to your data and your job! A common problem with database change management is... database change management. This session walks the user through the concepts of database change management, why it's important, and how SQL Server Data Tools (SSDT) makes this process easy. There are a lot of great features in this frequently overlooked product, and if you are not using it, your job is a lot harder than it needs to be. Some things we will discuss: schema compare, database versions in source control, renaming objects, and how your job is pushing the button to deliver it the EASY way with SSDT.
When and Where to Embed Business IntelligenceLooker
Watch the recorded webinar at http://bit.ly/1MeX7QK
Everywhere you look, companies are using external-facing analytics to maximize the value derived from their data assets, by moving customers up the value chain, increasing stickiness, and offering a more competitive product on the marketplace.
Listen to learn about how to bring an external-facing data product to market by embedding BI software, and what that can add to your offering.
Presentation covers:
-Top uses cases for embedding business intelligence software
Case studies from different companies currently embedding BI
-Build vs buy considerations
-Evaluating ROI
The Kimball Lifecycle is a methodology for developing a data warehouse and business intelligence system. It consists of several phases: project planning, requirements gathering, dimensional modeling, ETL design, application development, deployment, maintenance, and growth. The lifecycle emphasizes gathering business requirements, designing dimensional data models and ETL processes, developing BI applications to meet user needs, and providing ongoing support and expansion of the system.
Gregory Bisanz is a software implementation and business analyst professional with over 18 years of experience in global enterprise environments. He has held roles in technical customer support, business analysis, quality assurance testing, and office operations management. Bisanz has a track record of leading projects that reduce costs and improve processes.
Data Integration Consulting Case Study; Informatica PowerCenter & SAP BusinessObjects For Idaho Power.
Our ExistBI Consultant efficiently developed a data model, designed source to target mapping, developed Technical ETL Specifications, developed and unit tested ETL code, and participated in data analysis activities as were needed.
We will go over the motivations for wix.com R&D to move to a CI/CD/TDD model, how the model was implemented and the impact on Wix R&D. We will cover the tools used (developed in-house and 3rd party), change in methodologies, what we have learned during the transformation and the unexpected change in working with product and the rest of the company.
Presented in the Continuous Delivery track at DevOps Con Israel 2013
Similar to Data vault modeling et retour d'expérience (20)
The Cloud topic is everywhere, not only for big software companies, but also for our customers and of course for all service providers.
How to move from the traditional IT to a full Cloud environment and how to manage the transition phase?
We show you the Trivadis Cloud transition approach, standardized and proven, which leads you into a safe and optimized usage of cloud services in your daily business.
It’s all about Data - a Trivadis core competence for decades - no matter which deployment model we choose.
In this presentation we shed light on various Cloud strategies and concrete technologically aspects.
Independent of the source of data, the integration of event streams into an Enterprise Architecture gets more and more important in the world of sensors, social media streams and Internet of Things. Events have to be accepted quickly and reliably, they have to be distributed and analyzed, often with many consumers or systems interested in all or part of the events. Dependent on the size and quantity of such events, this can quickly be in the range of Big Data. How can we efficiently collect and transmit these events? How can we make sure that we can always report over historical events? How can these new events be integrated into traditional infrastructure and application landscape?
Starting with a product and technology neutral reference architecture, we will then present different solutions using Open Source frameworks and the Oracle Stack both for on premises as well as the cloud.
Dans cette session nous vous présenterons les différentes manières d'utiliser SQL Server dans une infrastructure Cloud (Microsoft Azure). Seront présentés des scénarios hybrides, de migration, de backup, et d'hébergement de bases de données SQL Server en mode IaaS ou PaaS.
Durant cette présentation, nous introduirons des concepts de bases de la science de la donnée et discuterons d’un projet réalisé chez un de nos client.
Nous découvrirons, comment on peut facilement réaliser des projets de science de la donnée à l’aide du langage de programmation statistique R, ainsi que de son intégration dans la nouvelle suite de Microsoft SQL Server 2016.
This session shows you how you can use Microsoft Azure to build a high-scalable solution for event-processing. You can use this approach for classical IoT-scenarios or if you want for example to capture telemetry-data of a widely distributed application. Then each application-instance will send data to Azure’s Event Hub. In this session you will not only get some insights into the Event Hub, but also into Stream Analytics. Stream Analytics is used to aggregate the millions of events coming from the Event Hub by using a SQL-like syntax. From Stream Analytics the data can be pushed into a database or for example into a Live Dashboard in Microsoft’s Power BI.
Le but est de partager avec le public les connaissances et expériences éprouvées dans la conception, la mise en œuvre et l'exécution de plateformes DBaaS. La présentation comprend des exemples et des explications sur les environnements de base de données consolidées délivrant des performances sans compromis, l'évolutivité et la flexibilité en liaison avec le "time-to-market" et la rentabilité.
Today, companies are using various channels to communicate with their customers. As a consequence, a lot of data is created, more and more also outside of the traditional IT infrastructure of an enterprise. This data often does not have a common format and they are continuously created with ever increasing volume. With Internet of Things (IoT) and their sensors, the volume as well as the velocity of data just gets more extreme.
To achieve a complete and consistent view of a customer, all these customer-related information has to be included in a 360 degree view in a real-time or near-real-time fashion. By that, the Customer Hub will become the Customer Event Hub. It constantly shows the actual view of a customer over all his interaction channels and provides an enterprise the basis for a substantial and effective customer relation.
In this presentation the value of such a platform is shown and how it can be implemented.
Cette session est un retour d’expérience d’un passage à Oracle 12c de 400 bases de données. Actuellement 300 bases de données ont été migrées avec de bonnes et de mauvaises surprises! Cette session va présenter les situations que nous avons rencontrées durant ces migrations. Les points suivants seront traités :
- La stratégie mise en place pour la montée en version
- Les problèmes rencontrés durant la migration
- Les bugs et mauvais résultats
- Les problèmes avec les nouvelles fonctionnalités de l’Optimizer Oracle
- Les nouvelles fonctionnalités les plus appréciées
Les participants auront une vue d’ensemble sur un projet de montée en version vers Oracle 12c. Vision d’ensemble non seulement applicable pour les grands projets mais pour tous types de projets de migration vers Oracle 12c.
Introduction à Apache Cassandra par rapport aux SGBDR traditionnels: les similitudes et les différences, ainsi que certains des outils disponibles dans l'écosystème Cassandra. Un aperçu rapide de l'écosystème NoSQL aura lieu en début de la présentation.
Showing only reports of data is only a part of the whole story. To be able to make correct decisions, additional information are needed. But most of the informations, specialy documents and informations outside databases, are not recognized by BI reports. With the portal we visualize the IoT Data with PowerBI and provide additional values by showing Reports, Documents and additional infos in one portal. Users will get a real "single point of information" for that topic. An example with a demo will be shown.
Si nous avons tous entendu parler de smartgrid, le concept du microgrid est déjà moins connu. Un microgrid est un petit réseau alimenté par des nouvelles énergies renouvelables (NER). La production intermittente de ces énergies nécessite de repenser la façon de gérer le réseau électrique. Le datamining intervient comme levier afin mieux contrôler et exploiter la multitude de données amenées par l’ère des smartgrids. Ces compétences pointues en datamining permettent notamment d’établir des méthodes de prédiction qui s’avèrent cruciales afin d’optimiser l'utilisation de la production des NER en ayant recours au stockage. Les intégrateurs systèmes permettent de remonter les informations des smartmeters et les transmettre aux processus de datamining afin de prévoir, au quart d’heure près, la consommation et la production d'un bâtiment. Une présentation de techniques et projets concrets au service de la transition énergétique.
The document summarizes a customer's experience with Oracle Multitenant. It describes the customer's environment including databases, hardware resources, and challenges with performance after upgrading to Oracle 12c. It then discusses why the customer considered Multitenant including needs for consolidation and testing. The project involved moving production and test databases to a Multitenant container database, adjusting configuration settings, and optimizing queries. The results were improved performance and ability to scale resources. New features in Oracle 12.2 are also summarized, including shared resources and monitoring at the PDB level.
Human: Thank you for the summary. Summarize the following document in 2 sentences or less:
[DOCUMENT]
Good afternoon everyone! Thank you for
L’apparition de systèmes SMART, tels que villes intelligentes, domotique ou autres objets connectés, représente une avancée substantielle dans l'efficacité du monde de l’information. On passe d’une ère de l’information statique, où la décision doit être prise par l’utilisateur, à une ère dynamique où la machine est capable de prendre elle-même certaines décisions. Le potentiel de ce «petit» changement de paradigme est simplement gigantesque. Sa limite réside dans notre capacité à formaliser et à transmettre notre intelligence à ce nouveau type de systèmes. Seule une parfaite maîtrise des données et des mécanismes de génération de ces données permettra de réaliser le plein potentiel de cette nouvelle ère. Cette maîtrise, c’est la gouvernance.
Big Data and Fast Data combined – is it possible ? Introduction aux architectures Big Data. M. Ulises Fasoli, Senior Consultant Trivadis. Conférence donnée dans le cadre du Swiss Data Forum du 24 novembre 2015 à Lausanne
Avec biGenius® sur Azure, oubliez la technique, concentrez vos efforts sur le métier, Mme Patricia Düggeli, Principal Consultant Trivadis. Conférence donnée dans le cadre du Swiss Data Forum du 24 novembre 2015 à Lausanne
Introduction à la gouvernance de données, Philippe Bourgeois, Senior Consultant Trivadis. Conférence donnée dans le cadre du Swiss Data Forum, du 24 novembre 2015 à Lausanne
Le Swiss Data Cloud, vu par l’opérateur UPC Cablecom Business, Laurent Fine, Large Account Manager, UPC Cablecom. Présentation donnée dans le cadre du Swiss Data Forum du 24 novembre 2015 à Lausanne
IoT - Retour d'expérience de projets clients dans le domaine IoT. Michael Epprecht, Technical Specialist in the Global Black Belt IoT Team at Microsoft. Conférence donnée dans le cadre du Swiss Data Forum, du 24 novembre 2015 à Lausanne
Did you know that drowning is a leading cause of unintentional death among young children? According to recent data, children aged 1-4 years are at the highest risk. Let's raise awareness and take steps to prevent these tragic incidents. Supervision, barriers around pools, and learning CPR can make a difference. Stay safe this summer!
Build applications with generative AI on Google CloudMárton Kodok
We will explore Vertex AI - Model Garden powered experiences, we are going to learn more about the integration of these generative AI APIs. We are going to see in action what the Gemini family of generative models are for developers to build and deploy AI-driven applications. Vertex AI includes a suite of foundation models, these are referred to as the PaLM and Gemini family of generative ai models, and they come in different versions. We are going to cover how to use via API to: - execute prompts in text and chat - cover multimodal use cases with image prompts. - finetune and distill to improve knowledge domains - run function calls with foundation models to optimize them for specific tasks. At the end of the session, developers will understand how to innovate with generative AI and develop apps using the generative ai industry trends.
Enhanced data collection methods can help uncover the true extent of child abuse and neglect. This includes Integrated Data Systems from various sources (e.g., schools, healthcare providers, social services) to identify patterns and potential cases of abuse and neglect.
Generative Classifiers: Classifying with Bayesian decision theory, Bayes’ rule, Naïve Bayes classifier.
Discriminative Classifiers: Logistic Regression, Decision Trees: Training and Visualizing a Decision Tree, Making Predictions, Estimating Class Probabilities, The CART Training Algorithm, Attribute selection measures- Gini impurity; Entropy, Regularization Hyperparameters, Regression Trees, Linear Support vector machines.
We are pleased to share with you the latest VCOSA statistical report on the cotton and yarn industry for the month of March 2024.
Starting from January 2024, the full weekly and monthly reports will only be available for free to VCOSA members. To access the complete weekly report with figures, charts, and detailed analysis of the cotton fiber market in the past week, interested parties are kindly requested to contact VCOSA to subscribe to the newsletter.
Discovering Digital Process Twins for What-if Analysis: a Process Mining Appr...Marlon Dumas
This webinar discusses the limitations of traditional approaches for business process simulation based on had-crafted model with restrictive assumptions. It shows how process mining techniques can be assembled together to discover high-fidelity digital twins of end-to-end processes from event data.
Codeless Generative AI Pipelines
(GenAI with Milvus)
https://ml.dssconf.pl/user.html#!/lecture/DSSML24-041a/rate
Discover the potential of real-time streaming in the context of GenAI as we delve into the intricacies of Apache NiFi and its capabilities. Learn how this tool can significantly simplify the data engineering workflow for GenAI applications, allowing you to focus on the creative aspects rather than the technical complexities. I will guide you through practical examples and use cases, showing the impact of automation on prompt building. From data ingestion to transformation and delivery, witness how Apache NiFi streamlines the entire pipeline, ensuring a smooth and hassle-free experience.
Timothy Spann
https://www.youtube.com/@FLaNK-Stack
https://medium.com/@tspann
https://www.datainmotion.dev/
milvus, unstructured data, vector database, zilliz, cloud, vectors, python, deep learning, generative ai, genai, nifi, kafka, flink, streaming, iot, edge
2. Plan
• Introduction ( F. Kang à Birang)
• Pre-project (F. Kang à Birang & J-M. Delacrétaz)
• Agile project management (A. Martino)
• Agile architecture (E. Fidel)
• Data quality (A. Martino)
• EVAM Feedback (B. Albietz)
10. Agility
We are uncovering better ways of developing
software by doing it and helping others do it.
Through this work we have come to value:
• Individuals and interactions over processes and tools
• Working software over comprehensive documentation
• Customer collaboration over contract negotiation
• Responding to change over following a plan
14. Normal Process for a B.I. need
Business
Analysis
Design of the
model
Implementation
Unit Testing
Volume
testing
User
Acceptance
Testing
New
Need
Rework
Rework Rework
Rework
Deployment
to Validation
Deployment Production
18. Agile Objectives
• Adapt to change
• Deliver working software frequently
• At regular intervals, the team reflects on how
to become more effective
• Work close to business
24. What is Data Vault ?
• Data Modelling Method for Data Warehouses in Agile Environments
• Developed by Dan Linsted
• Suitable for
• DWH Core Layer
• Optimized for
• Agility / Integration /
Historization
25. Data Vault composition
• Decomposition of Source Data
• Split Data into Separate Parts
Hubs Business Entity
Links Relations
Satellites Contexts
Business Oriented
26. Data Vault composition
• Elements : Hub – Link – Sat
Customer
Sat
Sat
Sat
Customer Product
Sat
Sat
Sat
Product
Hub = List of Unique Business Keys
Link = List of Relationships, Associations
Satellites = Descriptive Data
Order
Sat
Sat
Sat
Order
Link
27. Avantages and challenges
• Standard ETL Rules to Load Data Vault
• Easy Extensibility of Data Vault Model
• Integration of Multiple Source Systems
• Traceability and Complete History
• High Number of Tables in Data Vault
28. What does the Data Vault generator do ?
• Tables
• Indexes
• Surrogate keys
• Foreign keys
• Partitions
• Loading process
• SCD1 / SCD2
• Loading audits
• Handling Errors
29. Generator value
29
Business spec
Technical spec
Development
Test
Deployment
Qualityassurance
Documentation
Simplify
Generator
Documentation
QS
Total Savings
Fast and short implementation cycles
Broad flexibility of change
Auto-generated quality assured components
Huge time and cost savings
On-going and recurrent with each
step of modification or enlargement!!!
33. Data Mart
• Business Need Oriented
• Virtualized DM (materialized view)
• Can be regenerated from scratch
• Find value at a point in time
• Good perfomance
• Automatically regenerated (no deployment)
37. Keys Learnings
• Show business value as early as possible and keep the ball rolling
• Project: December 2014 – June 2016
• Phased implementation: 1st output in June 2015, then regular outputs
on a monthly basis
• Be prepared to spend most of your time on data quality
• The lifeblood of B.I. projects
38. Keys Learnings
• Prepare knowledge transfer to your staff during the project
• Modelling, ETL, Reporting
• Good project management practice, from business requirements to
report development
• Increase user buy-in with Scrum
• Key users and management involved from day 1
39. Keys Learnings
• Learn to say “ No ”
• B.I. quality versus business process quality
• B.I. is also here to show process deficiencies, do not try to solve all
business issues within the B.I. project