Watch Here: https://bit.ly/2NcqU6F
We take on the 2nd myth about data virtualization and it’s one that suggests a BI tool can substitute a data virtualization software.
You might be thinking: If I can have multi-source queries and define a logical model in my reporting tool, why would I need a data virtualization software?
Reporting tools, no doubt important and necessary, focus on the visualization of data and it’s presentation to the business user. Data virtualization is a governed data access layer designed to connect to and provide transparency of all enterprise data.
Yet the myth suggests that these technologies are interchangeable. So we’re going to take it on!
Watch this webinar as we compare and contrast BI tools and data virtualization to draw a final conclusion.
Enabling digital transformation api ecosystems and data virtualizationDenodo
Watch the full webinar here: https://buff.ly/2KBKzLJ
Digital transformation, as cliché as it sounds, is on top of every decision maker’s strategic initiative list. And at the heart of any digital transformation, no matter the industry or the size of the company, there is an application programming interface (API) strategy. While API platforms enable companies to manage large numbers of APIs working in tandem, monitor their usage, and establish security between them, they are not optimized for data integration, so they cannot easily or quickly integrate large volumes of data between different systems. Data virtualization, however, can greatly enhance the capabilities of an API platform, increasing the benefits of an API-based architecture. With data virtualization as part of an API strategy, companies can streamline digital transformations of any size and scope.
Join us for this webinar to see these technologies in action in a demo and to get the answers to the following questions:
*How can data virtualization enhance the deployment and exposure of APIs?
*How does data virtualization work as a service container, as a source for microservices and as an API gateway?
*How can data virtualization create managed data services ecosystems in a thriving API economy?
*How are GetSmarter and others are leveraging data virtualization to facilitate API-based initiatives?
Enabling Data as a Service with the JBoss Enterprise Data Services Platformprajods
This presentation was given at JUDCon 2013, Jan 17,18 at Bangalore. Presented by Prajod Vettiyattil and Gnanaguru Sattanathan. The presentation deals with the Why, What and How of Data Services and Data Services Platforms. It also explains the features of the JBoss Enterprise Data Services Platform.
The need for Data Services is explained with 3 Business use cases:
1. Post purchase customer experience improvement for an Auto manufacturer
2. Enterprise Data Access Layer
3. Data Services for Regulatory Reporting requirements like Dodd Frank
Data virtualization, Data Federation & IaaS with Jboss TeiidAnil Allewar
Enterprise have always grappled with the problem of information silos that needed to be merged using multiple data warehouses(DWs) and business intelligence(BI) tools so that enterprises could mine this disparate data for businessdecisions and strategy. Traditionally this data integration was done with ETL by consolidating multiple DBMS into a single data storage facility.
Data virtualization enables abstraction, transformation, federation, and delivery of data taken from variety of heterogeneous data sources as if it is a single virtual data source without the need to physically copy the data for integration. It allows consuming applications or users to access data from these various sources via a request to a single access point and delivers information-as-a-service (IaaS).
In this presentation, we will explore what data virtualization is and how it differs from the traditional data integration architecture. We’ll also look at validating the data virtualization and federation concepts by working through an example(see videos at the GitHub repo) to federate data across 2 heterogeneous data sources; mySQL and MongoDB using the JBoss Teiid data virtualization platform.
(OTW13) Agile Data Warehousing: Introduction to Data Vault ModelingKent Graziano
This is the presentation I gave at OakTable World 2013 in San Francisco. #OTW13 was held at the Children's Creativity Museum next to the Moscone Convention Center and was in parallel with Oracle OpenWorld 2013.
The session discussed our attempts to be more agile in designing enterprise data warehouses and how the Data Vault Data Modeling technique helps in that approach.
Wonder what this data mesh stuff is all about? What are the principles of data mesh? Can you or should you consider data mesh as the approach for your analytics platform? And most important - how can Snowflake help?
Given in Montreal on 14-Dec-2021
Agile Data Warehousing: Using SDDM to Build a Virtualized ODSKent Graziano
(This is the talk I gave at Houston DAMA and Agile Denver BI meetups)
At a past client, in order to meet timelines to fulfill urgent, unmet reporting needs, I found it necessary to build a virtualized Operational Data Store as the first phase of a new Data Vault 2.0 project. This allowed me to deliver new objects, quickly and incrementally to the report developer so we could quickly show the business users their data. In order to limit the need for refactoring in later stages of the data warehouse development, I chose to build this virtualization layer on top of a Type 2 persistent staging layer. All of this was done using Oracle SQL Developer Data Modeler (SDDM) against (gasp!) a MS SQL Server Database. In this talk I will show you the architecture for this approach, the rationale, and then the tricks I used in SDDM to build all the stage tables and views very quickly. In the end you will see actual SQL code for a virtual ODS that can easily be translated to an Oracle database.
Enabling digital transformation api ecosystems and data virtualizationDenodo
Watch the full webinar here: https://buff.ly/2KBKzLJ
Digital transformation, as cliché as it sounds, is on top of every decision maker’s strategic initiative list. And at the heart of any digital transformation, no matter the industry or the size of the company, there is an application programming interface (API) strategy. While API platforms enable companies to manage large numbers of APIs working in tandem, monitor their usage, and establish security between them, they are not optimized for data integration, so they cannot easily or quickly integrate large volumes of data between different systems. Data virtualization, however, can greatly enhance the capabilities of an API platform, increasing the benefits of an API-based architecture. With data virtualization as part of an API strategy, companies can streamline digital transformations of any size and scope.
Join us for this webinar to see these technologies in action in a demo and to get the answers to the following questions:
*How can data virtualization enhance the deployment and exposure of APIs?
*How does data virtualization work as a service container, as a source for microservices and as an API gateway?
*How can data virtualization create managed data services ecosystems in a thriving API economy?
*How are GetSmarter and others are leveraging data virtualization to facilitate API-based initiatives?
Enabling Data as a Service with the JBoss Enterprise Data Services Platformprajods
This presentation was given at JUDCon 2013, Jan 17,18 at Bangalore. Presented by Prajod Vettiyattil and Gnanaguru Sattanathan. The presentation deals with the Why, What and How of Data Services and Data Services Platforms. It also explains the features of the JBoss Enterprise Data Services Platform.
The need for Data Services is explained with 3 Business use cases:
1. Post purchase customer experience improvement for an Auto manufacturer
2. Enterprise Data Access Layer
3. Data Services for Regulatory Reporting requirements like Dodd Frank
Data virtualization, Data Federation & IaaS with Jboss TeiidAnil Allewar
Enterprise have always grappled with the problem of information silos that needed to be merged using multiple data warehouses(DWs) and business intelligence(BI) tools so that enterprises could mine this disparate data for businessdecisions and strategy. Traditionally this data integration was done with ETL by consolidating multiple DBMS into a single data storage facility.
Data virtualization enables abstraction, transformation, federation, and delivery of data taken from variety of heterogeneous data sources as if it is a single virtual data source without the need to physically copy the data for integration. It allows consuming applications or users to access data from these various sources via a request to a single access point and delivers information-as-a-service (IaaS).
In this presentation, we will explore what data virtualization is and how it differs from the traditional data integration architecture. We’ll also look at validating the data virtualization and federation concepts by working through an example(see videos at the GitHub repo) to federate data across 2 heterogeneous data sources; mySQL and MongoDB using the JBoss Teiid data virtualization platform.
(OTW13) Agile Data Warehousing: Introduction to Data Vault ModelingKent Graziano
This is the presentation I gave at OakTable World 2013 in San Francisco. #OTW13 was held at the Children's Creativity Museum next to the Moscone Convention Center and was in parallel with Oracle OpenWorld 2013.
The session discussed our attempts to be more agile in designing enterprise data warehouses and how the Data Vault Data Modeling technique helps in that approach.
Wonder what this data mesh stuff is all about? What are the principles of data mesh? Can you or should you consider data mesh as the approach for your analytics platform? And most important - how can Snowflake help?
Given in Montreal on 14-Dec-2021
Agile Data Warehousing: Using SDDM to Build a Virtualized ODSKent Graziano
(This is the talk I gave at Houston DAMA and Agile Denver BI meetups)
At a past client, in order to meet timelines to fulfill urgent, unmet reporting needs, I found it necessary to build a virtualized Operational Data Store as the first phase of a new Data Vault 2.0 project. This allowed me to deliver new objects, quickly and incrementally to the report developer so we could quickly show the business users their data. In order to limit the need for refactoring in later stages of the data warehouse development, I chose to build this virtualization layer on top of a Type 2 persistent staging layer. All of this was done using Oracle SQL Developer Data Modeler (SDDM) against (gasp!) a MS SQL Server Database. In this talk I will show you the architecture for this approach, the rationale, and then the tricks I used in SDDM to build all the stage tables and views very quickly. In the end you will see actual SQL code for a virtual ODS that can easily be translated to an Oracle database.
Power BI for Big Data and the New Look of Big Data SolutionsJames Serra
New features in Power BI give it enterprise tools, but that does not mean it automatically creates an enterprise solution. In this talk we will cover these new features (composite models, aggregations tables, dataflow) as well as Azure Data Lake Store Gen2, and describe the use cases and products of an individual, departmental, and enterprise big data solution. We will also talk about why a data warehouse and cubes still should be part of an enterprise solution, and how a data lake should be organized.
Data Integration through Data Virtualization (SQL Server Konferenz 2019)Cathrine Wilhelmsen
Data Integration through Data Virtualization - PolyBase and new SQL Server 2019 Features (Presented at SQL Server Konferenz 2019 on February 21st, 2019)
Every business today wants to leverage data to drive strategic initiatives with machine learning, data science and analytics — but runs into challenges from siloed teams, proprietary technologies and unreliable data.
That’s why enterprises are turning to the lakehouse because it offers a single platform to unify all your data, analytics and AI workloads.
Join our How to Build a Lakehouse technical training, where we’ll explore how to use Apache SparkTM, Delta Lake, and other open source technologies to build a better lakehouse. This virtual session will include concepts, architectures and demos.
Here’s what you’ll learn in this 2-hour session:
How Delta Lake combines the best of data warehouses and data lakes for improved data reliability, performance and security
How to use Apache Spark and Delta Lake to perform ETL processing, manage late-arriving data, and repair corrupted data directly on your lakehouse
Microsoft and Hortonworks Delivers the Modern Data Architecture for Big DataHortonworks
Joint webinar with Microsoft and Hortonworns on the power of combining the Hortonworks Data Platform with Microsoft’s ubiquitous Windows, Office, SQL Server, Parallel Data Warehouse, and Azure platform to build the Modern Data Architecture for Big Data.
The data lake has become extremely popular, but there is still confusion on how it should be used. In this presentation I will cover common big data architectures that use the data lake, the characteristics and benefits of a data lake, and how it works in conjunction with a relational data warehouse. Then I’ll go into details on using Azure Data Lake Store Gen2 as your data lake, and various typical use cases of the data lake. As a bonus I’ll talk about how to organize a data lake and discuss the various products that can be used in a modern data warehouse.
Not to be confused with Oracle Database Vault (a commercial db security product), Data Vault Modeling is a specific data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It is not a replacement for star schema data marts (and should not be used as such). This approach has been used in projects around the world (Europe, Australia, USA) for the last 10 years but is still not widely known or understood. The purpose of this presentation is to provide attendees with a detailed introduction to the technical components of the Data Vault Data Model, what they are for and how to build them. The examples will give attendees the basics for how to build, and design structures when using the Data Vault modeling technique. The target audience is anyone wishing to explore implementing a Data Vault style data model for an Enterprise Data Warehouse, Operational Data Warehouse, or Dynamic Data Integration Store. See more content like this by following my blog http://kentgraziano.com or follow me on twitter @kentgraziano.
Agile Data Engineering: Introduction to Data Vault 2.0 (2018)Kent Graziano
(updated slides used for North Texas DAMA meetup Oct 2018) As we move more and more towards the need for everyone to do Agile Data Warehousing, we need a data modeling method that can be agile with us. Data Vault Data Modeling is an agile data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It is a hybrid approach using the best of 3NF and dimensional modeling. It is not a replacement for star schema data marts (and should not be used as such). This approach has been used in projects around the world (Europe, Australia, USA) for over 15 years and is now growing in popularity. The purpose of this presentation is to provide attendees with an introduction to the components of the Data Vault Data Model, what they are for and how to build them. The examples will give attendees the basics:
• What the basic components of a DV model are
• How to build, and design structures incrementally, without constant refactoring
As a follow-on to the presentation "Building an Effective Data Warehouse Architecture", this presentation will explain exactly what Big Data is and its benefits, including use cases. We will discuss how Hadoop, the cloud and massively parallel processing (MPP) is changing the way data warehouses are being built. We will talk about hybrid architectures that combine on-premise data with data in the cloud as well as relational data and non-relational (unstructured) data. We will look at the benefits of MPP over SMP and how to integrate data from Internet of Things (IoT) devices. You will learn what a modern data warehouse should look like and how the role of a Data Lake and Hadoop fit in. In the end you will have guidance on the best solution for your data warehouse going forward.
Many companies today move mountains of data using ETL (extract, transform, load) technology. But data volumes are growing too large to move, customers are now expecting real-time data, and ETL costs now account for 10-15% of computing capacity. In this slide presentation, you can see how data virtualization enables data structures that were designed independently to be leveraged together, in real time, and without data movement, reducing complexity, lowering IT costs, and minimizing risk.
Introduction to Microsoft’s Master Data Services (MDS)James Serra
Master Data Services is bundled with SQL Server 2012 to help resolve many of the Master Data Management issues that companies are faced with when integrating data. In this session, James will show an overview of Master Data Services 2012, including the out of the box Web UI, the highly developed Excel Add-in, and how to get started with loading MDS with your data.
Power BI Overview, Deployment and GovernanceJames Serra
Deploying Power BI in a large enterprise is a complex task, and one that requires a lot of thought and planning. The purpose of this presentation is to help you make your Power BI deployment a success. After a quick Power BI overview, I’ll discuss deployment strategies, common usage scenarios, how to store and refresh data, prototyping options, how to share externally, and then finish with how to administer and secure Power BI. I’ll outline considerations and best practices for achieving an optimal, well-performing, enterprise level Power BI deployment.
An introduction to data virtualization in business intelligenceDavid Walker
A brief description of what Data Virtualisation is and how it can be used to support business intelligence applications and development. Originally presented to the ETIS Conference in Riga, Latvia in October 2013
Big data insights with Red Hat JBoss Data VirtualizationKenneth Peeples
You’re hearing a lot about big data these days. And big data and the technologies that store and process it, like Hadoop, aren’t just new data silos. You might be looking to integrate big data with existing enterprise information systems to gain better understanding of your business. You want to take informed action.
During this session, we’ll demonstrate how Red Hat JBoss Data Virtualization can integrate with Hadoop through Hive and provide users easy access to data. You’ll learn how Red Hat JBoss Data Virtualization:
Can help you integrate your existing and growing data infrastructure.
Integrates big data with your existing enterprise data infrastructure.
Lets non-technical users access big data result sets.
We’ll also provide typical uses cases and examples and a demonstration of the integration of Hadoop sentiment analysis with sales data.
Top Five Cool Features in Oracle SQL Developer Data ModelerKent Graziano
This is the presentation I gave at OUGF14 in Helsinki, Finland in June 2014.
Oracle SQL Developer Data Modeler (SDDM) has been around for a few years now and is up to version 4.x. It really is an industrial strength data modeling tool that can be used for any data modeling task you need to tackle. Over the years I have found quite a few features and utilities in the tool that I rely on to make me more efficient (and agile) in developing my models. This presentation will demonstrate at least five of these features, tips, and tricks for you. I will walk through things like modifying the delivered reporting templates, how to create and applying object naming templates, how to use a table template and transformation script to add audit columns to every table, and using the new meta data export tool and several other cool things you might not know are there. Since there will likely be patches and new releases before the conference, there is a good chance there will be some new things for me to show you as well. This might be a bit of a whirlwind demo, so get SDDM installed on your device and bring it to the session so you can follow along.
SAP Analytics Cloud: Haben Sie schon alle Datenquellen im Live-Zugriff?Denodo
Watch full webinar here: https://bit.ly/3hfEO6d
Die SAP Analytics Cloud (kurz "SAC" genannt) ist ein Service in der Cloud, der umfangreiche Analysefunktionen für Benutzer in einem Produkt bereit stellt. Wie immer bei der SAP ist auch die SAC technologisch gut integriert in die Welt der SAP Systeme.
Doch die Daten, die Unternehmen heutzutage analysieren möchten, befinden sich sehr häufig in den unterschiedlichsten Datenquellen: In relationalen Datenbanken, in Data Lakes, in Webservices, in Dateien, in NoSQL Datenbanken,... Und so stellt sich zwangsläufig die Frage, wie Sie aus der SAC heraus alle Daten konnektieren, transformieren und kombinieren können. Und das möglichst live, d.h. mit Abfragen auf Echtzeit-Daten! Hier kommt die Datenvirtualisierung ins Spiel: Sie bietet Anwendungen (so auch der SAC) einen einheitlichen, integrierten und performanten Zugriff auf SAP Daten und non-SAP Daten.
Erfahren Sie in diesem Webcast:
- Wie die Datenvirtualisierung funktioniert (in a Nutshell)
- Wie Sie aus der SAC heraus auf alle ihre Daten in Echtzeit zugreifen können ("Live Data Connection" genannt)
- Wie die Datenvirtualisierung die Performance auch für Abfragen auf grossen Datenmengen optimiert
Best practices to deliver data analytics to the business with power biSatya Shyam K Jayanty
Get your data to life with Power BI visualization and insights!
With the changing landscape of Power BI features it is essential to get hold of configuration and deployment practices within your data platform that will ensure you are on-par with compliance & security practices. In this session we will overview from the basics leading into advanced tricks on this landscape:
How to deploy Power BI?
How to implement configuration parameters and package BI features as a part of Office 365 roll out in your organisation?
What are newest features and enhancements on this Power BI landscape?
How to manage on-premise vs on-cloud connectivity?
How can you help and support the Power BI community as well?
Having said that within the objectives of this session, cloud computing is another aspect of this technology made is possible to get data within few clicks and ticks to the end-user. Let us review how to manage & connect on-premise data to cloud capabilities that can offer full advantage of data catalogue capabilities by keeping data secure as per Information Governance standards. Not just with nuts and bolts, performance is another aspect that every Admin is keeping up, let us look into few settings on how to maximize performance to optimize access to data as required. Gain understanding and insight into number of tools that are available for your Business Intelligence needs. There will be a showcase of events to demonstrate where to begin and how to proceed in BI world.
- D BI A Consulting
consulting@dbia.uk
Power BI for Big Data and the New Look of Big Data SolutionsJames Serra
New features in Power BI give it enterprise tools, but that does not mean it automatically creates an enterprise solution. In this talk we will cover these new features (composite models, aggregations tables, dataflow) as well as Azure Data Lake Store Gen2, and describe the use cases and products of an individual, departmental, and enterprise big data solution. We will also talk about why a data warehouse and cubes still should be part of an enterprise solution, and how a data lake should be organized.
Data Integration through Data Virtualization (SQL Server Konferenz 2019)Cathrine Wilhelmsen
Data Integration through Data Virtualization - PolyBase and new SQL Server 2019 Features (Presented at SQL Server Konferenz 2019 on February 21st, 2019)
Every business today wants to leverage data to drive strategic initiatives with machine learning, data science and analytics — but runs into challenges from siloed teams, proprietary technologies and unreliable data.
That’s why enterprises are turning to the lakehouse because it offers a single platform to unify all your data, analytics and AI workloads.
Join our How to Build a Lakehouse technical training, where we’ll explore how to use Apache SparkTM, Delta Lake, and other open source technologies to build a better lakehouse. This virtual session will include concepts, architectures and demos.
Here’s what you’ll learn in this 2-hour session:
How Delta Lake combines the best of data warehouses and data lakes for improved data reliability, performance and security
How to use Apache Spark and Delta Lake to perform ETL processing, manage late-arriving data, and repair corrupted data directly on your lakehouse
Microsoft and Hortonworks Delivers the Modern Data Architecture for Big DataHortonworks
Joint webinar with Microsoft and Hortonworns on the power of combining the Hortonworks Data Platform with Microsoft’s ubiquitous Windows, Office, SQL Server, Parallel Data Warehouse, and Azure platform to build the Modern Data Architecture for Big Data.
The data lake has become extremely popular, but there is still confusion on how it should be used. In this presentation I will cover common big data architectures that use the data lake, the characteristics and benefits of a data lake, and how it works in conjunction with a relational data warehouse. Then I’ll go into details on using Azure Data Lake Store Gen2 as your data lake, and various typical use cases of the data lake. As a bonus I’ll talk about how to organize a data lake and discuss the various products that can be used in a modern data warehouse.
Not to be confused with Oracle Database Vault (a commercial db security product), Data Vault Modeling is a specific data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It is not a replacement for star schema data marts (and should not be used as such). This approach has been used in projects around the world (Europe, Australia, USA) for the last 10 years but is still not widely known or understood. The purpose of this presentation is to provide attendees with a detailed introduction to the technical components of the Data Vault Data Model, what they are for and how to build them. The examples will give attendees the basics for how to build, and design structures when using the Data Vault modeling technique. The target audience is anyone wishing to explore implementing a Data Vault style data model for an Enterprise Data Warehouse, Operational Data Warehouse, or Dynamic Data Integration Store. See more content like this by following my blog http://kentgraziano.com or follow me on twitter @kentgraziano.
Agile Data Engineering: Introduction to Data Vault 2.0 (2018)Kent Graziano
(updated slides used for North Texas DAMA meetup Oct 2018) As we move more and more towards the need for everyone to do Agile Data Warehousing, we need a data modeling method that can be agile with us. Data Vault Data Modeling is an agile data modeling technique for designing highly flexible, scalable, and adaptable data structures for enterprise data warehouse repositories. It is a hybrid approach using the best of 3NF and dimensional modeling. It is not a replacement for star schema data marts (and should not be used as such). This approach has been used in projects around the world (Europe, Australia, USA) for over 15 years and is now growing in popularity. The purpose of this presentation is to provide attendees with an introduction to the components of the Data Vault Data Model, what they are for and how to build them. The examples will give attendees the basics:
• What the basic components of a DV model are
• How to build, and design structures incrementally, without constant refactoring
As a follow-on to the presentation "Building an Effective Data Warehouse Architecture", this presentation will explain exactly what Big Data is and its benefits, including use cases. We will discuss how Hadoop, the cloud and massively parallel processing (MPP) is changing the way data warehouses are being built. We will talk about hybrid architectures that combine on-premise data with data in the cloud as well as relational data and non-relational (unstructured) data. We will look at the benefits of MPP over SMP and how to integrate data from Internet of Things (IoT) devices. You will learn what a modern data warehouse should look like and how the role of a Data Lake and Hadoop fit in. In the end you will have guidance on the best solution for your data warehouse going forward.
Many companies today move mountains of data using ETL (extract, transform, load) technology. But data volumes are growing too large to move, customers are now expecting real-time data, and ETL costs now account for 10-15% of computing capacity. In this slide presentation, you can see how data virtualization enables data structures that were designed independently to be leveraged together, in real time, and without data movement, reducing complexity, lowering IT costs, and minimizing risk.
Introduction to Microsoft’s Master Data Services (MDS)James Serra
Master Data Services is bundled with SQL Server 2012 to help resolve many of the Master Data Management issues that companies are faced with when integrating data. In this session, James will show an overview of Master Data Services 2012, including the out of the box Web UI, the highly developed Excel Add-in, and how to get started with loading MDS with your data.
Power BI Overview, Deployment and GovernanceJames Serra
Deploying Power BI in a large enterprise is a complex task, and one that requires a lot of thought and planning. The purpose of this presentation is to help you make your Power BI deployment a success. After a quick Power BI overview, I’ll discuss deployment strategies, common usage scenarios, how to store and refresh data, prototyping options, how to share externally, and then finish with how to administer and secure Power BI. I’ll outline considerations and best practices for achieving an optimal, well-performing, enterprise level Power BI deployment.
An introduction to data virtualization in business intelligenceDavid Walker
A brief description of what Data Virtualisation is and how it can be used to support business intelligence applications and development. Originally presented to the ETIS Conference in Riga, Latvia in October 2013
Big data insights with Red Hat JBoss Data VirtualizationKenneth Peeples
You’re hearing a lot about big data these days. And big data and the technologies that store and process it, like Hadoop, aren’t just new data silos. You might be looking to integrate big data with existing enterprise information systems to gain better understanding of your business. You want to take informed action.
During this session, we’ll demonstrate how Red Hat JBoss Data Virtualization can integrate with Hadoop through Hive and provide users easy access to data. You’ll learn how Red Hat JBoss Data Virtualization:
Can help you integrate your existing and growing data infrastructure.
Integrates big data with your existing enterprise data infrastructure.
Lets non-technical users access big data result sets.
We’ll also provide typical uses cases and examples and a demonstration of the integration of Hadoop sentiment analysis with sales data.
Top Five Cool Features in Oracle SQL Developer Data ModelerKent Graziano
This is the presentation I gave at OUGF14 in Helsinki, Finland in June 2014.
Oracle SQL Developer Data Modeler (SDDM) has been around for a few years now and is up to version 4.x. It really is an industrial strength data modeling tool that can be used for any data modeling task you need to tackle. Over the years I have found quite a few features and utilities in the tool that I rely on to make me more efficient (and agile) in developing my models. This presentation will demonstrate at least five of these features, tips, and tricks for you. I will walk through things like modifying the delivered reporting templates, how to create and applying object naming templates, how to use a table template and transformation script to add audit columns to every table, and using the new meta data export tool and several other cool things you might not know are there. Since there will likely be patches and new releases before the conference, there is a good chance there will be some new things for me to show you as well. This might be a bit of a whirlwind demo, so get SDDM installed on your device and bring it to the session so you can follow along.
SAP Analytics Cloud: Haben Sie schon alle Datenquellen im Live-Zugriff?Denodo
Watch full webinar here: https://bit.ly/3hfEO6d
Die SAP Analytics Cloud (kurz "SAC" genannt) ist ein Service in der Cloud, der umfangreiche Analysefunktionen für Benutzer in einem Produkt bereit stellt. Wie immer bei der SAP ist auch die SAC technologisch gut integriert in die Welt der SAP Systeme.
Doch die Daten, die Unternehmen heutzutage analysieren möchten, befinden sich sehr häufig in den unterschiedlichsten Datenquellen: In relationalen Datenbanken, in Data Lakes, in Webservices, in Dateien, in NoSQL Datenbanken,... Und so stellt sich zwangsläufig die Frage, wie Sie aus der SAC heraus alle Daten konnektieren, transformieren und kombinieren können. Und das möglichst live, d.h. mit Abfragen auf Echtzeit-Daten! Hier kommt die Datenvirtualisierung ins Spiel: Sie bietet Anwendungen (so auch der SAC) einen einheitlichen, integrierten und performanten Zugriff auf SAP Daten und non-SAP Daten.
Erfahren Sie in diesem Webcast:
- Wie die Datenvirtualisierung funktioniert (in a Nutshell)
- Wie Sie aus der SAC heraus auf alle ihre Daten in Echtzeit zugreifen können ("Live Data Connection" genannt)
- Wie die Datenvirtualisierung die Performance auch für Abfragen auf grossen Datenmengen optimiert
Best practices to deliver data analytics to the business with power biSatya Shyam K Jayanty
Get your data to life with Power BI visualization and insights!
With the changing landscape of Power BI features it is essential to get hold of configuration and deployment practices within your data platform that will ensure you are on-par with compliance & security practices. In this session we will overview from the basics leading into advanced tricks on this landscape:
How to deploy Power BI?
How to implement configuration parameters and package BI features as a part of Office 365 roll out in your organisation?
What are newest features and enhancements on this Power BI landscape?
How to manage on-premise vs on-cloud connectivity?
How can you help and support the Power BI community as well?
Having said that within the objectives of this session, cloud computing is another aspect of this technology made is possible to get data within few clicks and ticks to the end-user. Let us review how to manage & connect on-premise data to cloud capabilities that can offer full advantage of data catalogue capabilities by keeping data secure as per Information Governance standards. Not just with nuts and bolts, performance is another aspect that every Admin is keeping up, let us look into few settings on how to maximize performance to optimize access to data as required. Gain understanding and insight into number of tools that are available for your Business Intelligence needs. There will be a showcase of events to demonstrate where to begin and how to proceed in BI world.
- D BI A Consulting
consulting@dbia.uk
Self Service Analytics and a Modern Data Architecture with Data Virtualizatio...Denodo
Watch full webinar here: https://bit.ly/32TT2Uu
Data virtualization is not just for self-service, it’s also a first-class citizen when it comes to modern data platform architectures. Technology has forced many businesses to rethink their delivery models. Startups emerged, leveraging the internet and mobile technology to better meet customer needs (like Amazon and Lyft), disrupting entire categories of business, and grew to dominate their categories.
Schedule a complimentary Data Virtualization Discovery Session with g2o.
Traditional companies are still struggling to meet rising customer expectations. During this webinar with the experts from g2o and Denodo we covered the following:
- How modern data platforms enable businesses to address these new customer expectation
- How you can drive value from your investment in a data platform now
- How you can use data virtualization to enable multi-cloud strategies
Leveraging the strategy insights of g2o and the power of the Denodo platform, companies do not need to undergo the costly removal and replacement of legacy systems to modernize their systems. g2o and Denodo can provide a strategy to create a modern data architecture within a company’s existing infrastructure.
Today, data lakes are widely used and have become extremely affordable as data volumes have grown. However, they are only meant for storage and by themselves provide no direct value. With up to 80% of data stored in the data lake today, how do you unlock the value of the data lake? The value lies in the compute engine that runs on top of a data lake.
Join us for this webinar where Ahana co-founder and Chief Product Officer Dipti Borkar will discuss how to unlock the value of your data lake with the emerging Open Data Lake analytics architecture.
Dipti will cover:
-Open Data Lake analytics - what it is and what use cases it supports
-Why companies are moving to an open data lake analytics approach
-Why the open source data lake query engine Presto is critical to this approach
Modern Data Management for Federal ModernizationDenodo
Watch full webinar here: https://bit.ly/2QaVfE7
Faster, more agile data management is at the heart of government modernization. However, Traditional data delivery systems are limited in realizing a modernized and future-proof data architecture.
This webinar will address how data virtualization can modernize existing systems and enable new data strategies. Join this session to learn how government agencies can use data virtualization to:
- Enable governed, inter-agency data sharing
- Simplify data acquisition, search and tagging
- Streamline data delivery for transition to cloud, data science initiatives, and more
So you got a handle on what Big Data is and how you can use it to find business value in your data. Now you need an understanding of the Microsoft products that can be used to create a Big Data solution. Microsoft has many pieces of the puzzle and in this presentation I will show how they fit together. How does Microsoft enhance and add value to Big Data? From collecting data, transforming it, storing it, to visualizing it, I will show you Microsoft’s solutions for every step of the way
Denodo DataFest 2016: Comparing and Contrasting Data Virtualization With Data...Denodo
Watch the full session: Denodo DataFest 2016 sessions: https://goo.gl/Bvmvc9
Data prep and data blending are terms that have come to prominence over the last year or two. On the surface, they appear to offer functionality similar to data virtualization…but there are important differences!
In this session, you will learn:
• How data virtualization complements or contrasts technologies such as data prep and data blending
• Pros and cons of functionality provided by data prep, data catalog and data blending tools
• When and how to use these different technologies to be most effective
This session is part of the Denodo DataFest 2016 event. You can also watch more Denodo DataFest sessions on demand here: https://goo.gl/VXb6M6
Denodo Partner Connect: Technical Webinar - Ask Me AnythingDenodo
Watch full webinar here: https://buff.ly/47jH4lk
In this session, Denodo experts will cover a deeper dive into the top 5 differentiated use cases for Denodo by answering any questions since the previous session.
Additionally, we invite partners to bring any general questions related to Denodo, the Denodo Platform, or data management.
This is Part 4 of the GoldenGate series on Data Mesh - a series of webinars helping customers understand how to move off of old-fashioned monolithic data integration architecture and get ready for more agile, cost-effective, event-driven solutions. The Data Mesh is a kind of Data Fabric that emphasizes business-led data products running on event-driven streaming architectures, serverless, and microservices based platforms. These emerging solutions are essential for enterprises that run data-driven services on multi-cloud, multi-vendor ecosystems.
Join this session to get a fresh look at Data Mesh; we'll start with core architecture principles (vendor agnostic) and transition into detailed examples of how Oracle's GoldenGate platform is providing capabilities today. We will discuss essential technical characteristics of a Data Mesh solution, and the benefits that business owners can expect by moving IT in this direction. For more background on Data Mesh, Part 1, 2, and 3 are on the GoldenGate YouTube channel: https://www.youtube.com/playlist?list=PLbqmhpwYrlZJ-583p3KQGDAd6038i1ywe
Webinar Speaker: Jeff Pollock, VP Product (https://www.linkedin.com/in/jtpollock/)
Mr. Pollock is an expert technology leader for data platforms, big data, data integration and governance. Jeff has been CTO at California startups and a senior exec at Fortune 100 tech vendors. He is currently Oracle VP of Products and Cloud Services for Data Replication, Streaming Data and Database Migrations. While at IBM, he was head of all Information Integration, Replication and Governance products, and previously Jeff was an independent architect for US Defense Department, VP of Technology at Cerebra and CTO of Modulant – he has been engineering artificial intelligence based data platforms since 2001. As a business consultant, Mr. Pollock was a Head Architect at Ernst & Young’s Center for Technology Enablement. Jeff is also the author of “Semantic Web for Dummies” and "Adaptive Information,” a frequent keynote at industry conferences, author for books and industry journals, formerly a contributing member of W3C and OASIS, and an engineering instructor with UC Berkeley’s Extension for object-oriented systems, software development process and enterprise architecture.
Data Virtualization: Introduction and Business Value (UK)Denodo
Watch full webinar here: https://bit.ly/30mHuYH
What started to evolve as the most agile and real-time enterprise data fabric, data virtualization is proving to go beyond its initial promise and is becoming one of the most important enterprise big data fabrics. Denodo’s vision is to provide a unified data delivery layer as a logical data fabric, to bridge the gap between the IT and the business, hiding the underlying complexity and creating a semantic layer to expose data in a business friendly manner.
Attend this webinar to learn:
- What data virtualization really is
- How it differs from other enterprise data integration technologies
- Why data virtualization is finding enterprise-wide deployment inside some of the largest organizations
- Business Value of data virtualization and customer use cases
- Highlights of the newly launched Denodo Platform 8.0
Myth Busters III: I’m Building a Data Lake, So I Don’t Need Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/2XXAzU3
So you’re building a data lake to solve your big data challenges. A data lake will allow you to keep all of your raw, detailed data in a single, consolidated repository; therefore, your problem is solved. Or is it? Is it really that easy?
Data lakes have their use and purpose, and we’re not here to argue that. However, data lakes on their own are constrained by factors such as duplication of data and therefore higher costs, governance limitations, and the risk of becoming another data silo.
With the addition of data virtualization, a physical data lake, can turn into a virtual or logical data like through an abstraction layer. Data virtualization can facilitate and expedite accessing and exploring critical data in a cost-effective manner and assist in deriving a greater return on the data lake investment.
You might still not be convinced. Give us an opportunity and join us as we try to bust this myth!
Watch this webinar as we explore the promises of a data lake as well as its downfalls to draw a final conclusion.
Choosing technologies for a big data solution in the cloudJames Serra
Has your company been building data warehouses for years using SQL Server? And are you now tasked with creating or moving your data warehouse to the cloud and modernizing it to support “Big Data”? What technologies and tools should use? That is what this presentation will help you answer. First we will cover what questions to ask concerning data (type, size, frequency), reporting, performance needs, on-prem vs cloud, staff technology skills, OSS requirements, cost, and MDM needs. Then we will show you common big data architecture solutions and help you to answer questions such as: Where do I store the data? Should I use a data lake? Do I still need a cube? What about Hadoop/NoSQL? Do I need the power of MPP? Should I build a "logical data warehouse"? What is this lambda architecture? Can I use Hadoop for my DW? Finally, we’ll show some architectures of real-world customer big data solutions. Come to this session to get started down the path to making the proper technology choices in moving to the cloud.
Embarking on building a modern data warehouse in the cloud can be an overwhelming experience due to the sheer number of products that can be used, especially when the use cases for many products overlap others. In this talk I will cover the use cases of many of the Microsoft products that you can use when building a modern data warehouse, broken down into four areas: ingest, store, prep, and model & serve. It’s a complicated story that I will try to simplify, giving blunt opinions of when to use what products and the pros/cons of each.
Think of big data as all data, no matter what the volume, velocity, or variety. The simple truth is a traditional on-prem data warehouse will not handle big data. So what is Microsoft’s strategy for building a big data solution? And why is it best to have this solution in the cloud? That is what this presentation will cover. Be prepared to discover all the various Microsoft technologies and products from collecting data, transforming it, storing it, to visualizing it. My goal is to help you not only understand each product but understand how they all fit together, so you can be the hero who builds your companies big data solution.
Data - and the things we want to do with data - exist in many different forms. Getting those formats and tasks to play nicely together can sometimes be a painstaking grind. The difficulty escalates if we need to switch between specialized tools, designed to address only a small subset of what we need to accomplish.
Enter, the Composable DataFlow.
Composable DataFlows are event-driven pipelines that consist of functional modules, strung together to form full analytical workflows. For developers, DataFlows can represent independently-deployable microservices, and can be used as part of a broader Microservice Architecture.
In this session, we will use a Composable DataFlow to extract data via API, transform JSON into a tabular structure, and load that data into a database of our own creation (using Composable DataPortal). We will also explore the DataFlow's Module Library to see what other options we have to help make our data... flow.
Meetup: https://www.meetup.com/boston-data-engineering/events/289525162/
Cloud Modernization and Data as a Service OptionDenodo
Watch here: https://bit.ly/36tEThx
The current data landscape is fragmented, not just in location but also in terms of shape and processing paradigms. Cloud has become a key component of modern architecture design. Data lakes, IoT, NoSQL, SaaS, etc. coexist with relational databases to fuel the needs of modern analytics, ML and AI. Exploring and understanding the data available within your organization is a time-consuming task. Dealing with bureaucracy, different languages and protocols, and the definition of ingestion pipelines to load that data into your data lake can be complex. And all of this without even knowing if that data will be useful at all.
Attend this session to learn:
- How dynamic data challenges and the speed of change requires a new approach to data architecture – one that is real-time, agile and doesn’t rely on physical data movement.
- Learn how logical data architecture can enable organizations to transition data faster to the cloud with zero downtime and ultimately deliver faster time to insight.
- Explore how data as a service and other API management capabilities is a must in a hybrid cloud environment.
Similar to Myth Busters II: BI Tools and Data Virtualization are Interchangeable (20)
Enterprise Monitoring and Auditing in DenodoDenodo
Watch full webinar here: https://buff.ly/3P3l4oK
Proper monitoring of an enterprise system is critical to understanding its capacity and growth, anticipating potential issues, and even understanding key ROI metrics. This also facilitates the implementation of policies and user access audits which are key to optimizing the resource utilization in an organization. Do you want to learn more about the new Denodo features for monitoring, auditing, and visualizing enterprise monitoring data?
Join us for the session with Vijayalakshmi Mani, Data Engineer at Denodo, to understand how the new features and components help in monitoring your Denodo Servers and the resource utilizations and how to extract the most out of the logs that the Denodo Platform generates including FinOps information.
Watch on-demand and Learn:
- What is a Denodo Monitor and what’s new in it?
- How to visualize the Denodo Monitor Information and use of Diagnostics & Monitoring Tool
- Introduction to the new Denodo Dashboard
- Demonstration on the Denodo Dashboard
Lunch and Learn ANZ: Mastering Cloud Data Cost Control: A FinOps ApproachDenodo
Watch full webinar here: https://buff.ly/4bYOOgb
With the rise of cloud-first initiatives and pay-per-use systems, forecasting IT costs has become a challenge. It's easy to start small, but it's equally easy to get skyrocketing bills with little warning. FinOps is a discipline that tries to tackle these issues, by providing the framework to understand and optimize cloud costs in a more controlled manner. The Denodo Platform, being a middleware layer in charge of global data delivery, sits in a privileged position not only to help us understand where costs are coming from, but also to take action, manage, and reduce them.
Attend this session to learn:
- The importance of FinOps in a cloud architecture.
- How the Denodo Platform can help you collect and visualize key FinOps metrics to understand where your costs are coming from?
- What actions and controls the Denodo Platform offers to keep costs at bay.
Achieving Self-Service Analytics with a Governed Data Services LayerDenodo
Watch full webinar here: https://buff.ly/3wBhxYb
In an increasingly distributed and complex data landscape, it is becoming increasingly difficult to govern and secure data effectively throughout the enterprise. Whether it be securing data across different repositories or monitoring access across different business units, the proliferation of data technologies and repositories across both on-premises and in the cloud is making the task unattainable. The challenge is only made greater by the ongoing pressure to offer self-service data access to business users.
Watch on-demand and learn:
- How to use a logical data fabric to build an enterprise-wide data access role model.
- Centralise security when data is spread across multiple systems residing both on-premises and in the cloud.
- Control and audit data access across different regions.
What you need to know about Generative AI and Data Management?Denodo
Watch full webinar here: https://buff.ly/3UXy0A2
It should be no surprise that Generative AI will have a profound impact to data management in years to come. Much like other areas of the technology sector, the opportunities presented by GenAI will accelerate our efforts around all aspects of data management, including self-service, automation, data governance and security. On the other hand, it is also becoming clearer that to unleash the true potential of AI assistants powered by GenAI, we need novel implementation strategies and a reimagined data architecture. This presents an exhilarating yet challenging future, demanding innovative thinking and methodologies in data management.
Join us on this webinar to learn about:
- The opportunities and challenges presented by GenAI today.
- Exploiting GenAI to democratize data management.
- How to augment GenAI applications with corporate data and knowledge.
- How to get started.
Mastering Data Compliance in a Dynamic Business LandscapeDenodo
Watch full webinar here: https://buff.ly/48rpLQ3
Join us for an enlightening webinar, "Mastering Data Compliance in a Dynamic Business Landscape," presented by Denodo Technologies and W5 Consulting. This session is tailored for business leaders and decision-makers who are navigating the complexities of data compliance in an ever-evolving business environment.
This webinar will focus on why data compliance is crucial for your business. Discover how to turn compliance into a competitive advantage, enhancing operational efficiency and market trust. We'll also address the risks of non-compliance, including financial penalties and the loss of customer trust, and provide strategies to proactively overcome these challenges.
Key Takeaways:
- How can your business leverage data management practices to stay agile and compliant in a rapidly changing regulatory landscape?
- Keys to balancing data accessibility with security and privacy in today's data-driven environment.
- What are the common pitfalls in achieving compliance with regulations like GDPR, CCPA, and HIPAA, and how can your business avoid them?
We will go beyond the technical aspects and delve into how you can strategically position your organization in the realm of data management and compliance. Learn how to craft a data compliance strategy that aligns with your business goals, enhances operational efficiency, and builds stakeholder trust.
Denodo Partner Connect: Business Value Demo with Denodo Demo LiteDenodo
Watch full webinar here: https://buff.ly/3OCQvGk
In this session, Denodo Sales Engineer, Yik Chuan Tan, will guide you through the art of delivering a compelling demo of the Denodo Platform with Denodo Demo Lite. Watch to uncover the significant functionalities that set Denodo apart and learn how to effectively win over potential customers.
In this session, we will cover:
Understanding the Denodo Platform & Tailoring Your Demo to Prospect Needs: By gaining a comprehensive understanding of the Denodo Platform, its architecture, and how it addresses data management challenges, you can customize your demo to align with the specific needs and pain points of your prospects, including:
- seamless data integration with real-time access
- data security and governance
- self-service data discovery
- advanced analytics and reporting
- performance optimization scalability and deployment
Watch this Denodo demo session and acquire the skills and knowledge necessary to captivate your prospects. Whether you're a seasoned technical professional or new to the field, this session will equip you with the skills to deliver compelling demos that lead to successful conversions.
Expert Panel: Overcoming Challenges with Distributed Data to Maximize Busines...Denodo
Watch full webinar here: https://buff.ly/3wdI1il
As organizations compete in new markets and new channels, business data requirements include new data platforms and applications. Migration to the cloud typically adds more distributed data when operations set up their own data platforms. This spreads important data across on-premises and cloud-based data platforms. As a result, data silos proliferate and become difficult to access, integrate, manage, and govern. Many organizations are using cloud data platforms to consolidate data, but distributed environments are unlikely to go away.
Organizations need holistic data strategies for unifying distributed data environments to improve data access and data governance, optimize costs and performance, and take advantage of modern technologies as they arrive. This TDWI Expert Panel will focus on overcoming challenges with distributed data to maximize business value.
Key topics this panel will address include:
- Developing the right strategy for your use cases and workloads in distributed data environments, such as data fabrics, data virtualization, and data mesh
- Deciding whether to consolidate data silos or bridge them with distributed data technologies
- Enabling easier self-service access and analytics across a distributed data environment
- Maximizing the value of data catalogs and other data intelligence technologies for distributed data environments
- Monitoring and data observability for spotting problems and ensuring business satisfaction
Watch full webinar here: https://buff.ly/3UE5K5l
The ability to recognize and flag sensitive information within corporate datasets is essential for compliance with emerging privacy laws, for completing a privacy impact assessment (PIA) or data subject access request (DSAR), and also for cyber-insurance compliance. During this session, we will discuss data privacy laws, the challenges they present, and how they can be applied with modern tools.
Join us for the session driven by Mark Rowan, CEO at Data Sentinel, and Bhavita Jaiswal, SE at Denodo, who will show how a data classification engine augments Data Catalog to support data governance and compliance objectives.
Watch on-demand & Learn:
- Changing landscape of data privacy laws and compliance requirements
- How to create a data classification framework
- How Data Sentinel classifies data and this can be integrated into Denodo
- Using the enhanced data classifications via consuming tools such as Data Catalog and Power BI
Знакомство с виртуализацией данных для профессионалов в области данныхDenodo
Watch full webinar here: https://buff.ly/3OETC08
По данным аналитической компании Gartner, "к 2022 году 60% предприятий включат виртуализацию данных в качестве основного метода доставки данных в свою интеграционную архитектуру". Компания Gartner назвала Denodo лидером в Магическом квадранте 2020 года по инструментам интеграции данных.
В ходе этого 1,5-часового занятия вы узнаете, как виртуализация данных революционизирует бизнес и ИТ-подход к доступу, доставке, потреблению, управлению и защите данных, независимо от возраста вашей технологии, формата данных или их местонахождения. Эта зрелая технология устраняет разрыв между ИТ и бизнес-пользователями и обеспечивает значительную экономию средств и времени.
**ФОРМАТ
Онлайн-семинар продолжительностью 1 час 30 минут.
Благодаря записи вы можете выполнять упражнения в своем собственном темпе.
**ДЛЯ КОГО ЭТОТ СЕМИНАР?
ИТ-менеджеры / архитекторы
Специалисты по анализу данных / аналитики
CDO
**СОДЕРЖАНИЕ
В программе: введение в суть виртуализации данных, примеры использования, реальные примеры из практики клиентов и демонстрация возможностей платформы Denodo Platform:
Интеграция и предоставление данных быстро и легко с помощью платформы Denodo Platform 8.0
Оптимизатор запросов Denodo предоставляет данные в режиме реального времени, по запросу, даже для очень больших наборов данных
Выставлять данные в качестве "сервисов данных" для потребления различными пользователями и инструментами
Каталог данных: Открывайте и документируйте данные с помощью нашего Каталога данных
пространства для самостоятельного доступа к данным.
Виртуализация данных играет ключевую роль в управлении и обеспечении безопасности данных в вашей организации
**ПОВЕСТКА
Введение в виртуализацию данных
Примеры использования и примеры из практики клиентов
Архитектура - Управление и безопасность
Производительность
Демо
Следующие шаги: как самостоятельно протестировать и внедрить платформу
Интерактивная сессия вопросов и ответов
Data Democratization: A Secret Sauce to Say Goodbye to Data FragmentationDenodo
Watch full webinar here: https://buff.ly/41Zf31D
Despite recent and evolving technological advances, the vast amounts of data that exist in a typical enterprise is not always available to all stakeholders when they need it. In modern enterprises, there are broad sets of users, with varying levels of skill sets, who strive to make data-driven decisions daily but struggle to gain access to the data needed in a timely manner.
Join our webinar to learn how to:
- Unlock the Power of Your Data: Discover how data democratization can transform your organization by giving every user access to the data they need, when they need it.
- Say 'Goodbye' to Data Fragmentation: Learn practical strategies to break down data silos and foster a more collaborative and efficient data environment.
- Realize the Full Potential of Your Data: Hear success stories about industry leaders who have embraced data democratization and witnessed tangible results.
Denodo Partner Connect - Technical Webinar - Ask Me AnythingDenodo
Watch full webinar here: https://buff.ly/48ZpEf1
In this session, we will cover a deeper dive into the Denodo Platform 8.0 Certified Architect Associate (DEN80EDUCAA) exam by answering any questions that have developed since the previous session.
Additionally, we invite partners to bring any general questions related to Denodo, the Denodo Platform, or data management.
Lunch and Learn ANZ: Key Takeaways for 2023!Denodo
Watch full webinar here: https://buff.ly/3SnH5QY
2023 is coming to an end where organisations dependency on trusted, accurate, secure and contextual data only grows more challenging. The perpetual aspect in seeking new architectures, processes, organisational team structures to "get the business their data" and reduce the operating costs continues unabated. While confidence from the business in what "value" is being derived or "to be" delivered from these investments in data, is being heavily scrutinised. 2023 saw significant new releases from vendors, focusing on the Data Fabric.
At this session we will look at these topics and key takeaways for 2023, including;
- Data management and data integration market highlights for 2023
- Key achievements for Denodo in their journey as a leader in this market
- A few case studies from Australian organisations in how they are delivering strategic business value through Denodo's Data Fabric platform and what they have been doing differently
It’s a Wrap! 2023 – A Groundbreaking Year for AI and The Way ForwardDenodo
Watch full webinar here: https://buff.ly/3S4Y49o
A little over a year ago, we would not have expected the disruptions caused by the rise of Generative AI. If 2023 was a groundbreaking year for AI, what will 2024 bring? More importantly, what can you do now to take advantage of these trends and ensure you are future-proof?
For example:
- Generative AI will become more powerful and user-friendly, enabling novel and realistic content creation and automation.
- Data Architectures will need to adapt to feed these powerful new models.
- Data ecosystems are moving to the cloud, but there is a growing need to maintain control of costs and optimize workloads better.
Join us for a discussion on the most significant trends in the Data & AI space, and how you can prepare to ride this wave!
Quels sont les facteurs-clés de succès pour appliquer au mieux le RGPD à votr...Denodo
Watch full webinar here: https://buff.ly/3O7rd2R
Afin d’être conformes au RGPD, les entreprises ont besoin d'avoir une vue d'ensemble sur toutes leurs données et d'établir des contrôles de sécurité sur toute l'infrastructure. La virtualisation des données de Denodo permet de rassembler les multiples sources de données, de les rendre accessibles à partir d'une seule couche, et offre des capacités de monitoring pour surveiller les changements.
Pour cela, Square IT Services a développé pour l’un de ses grands clients français prestigieux dans le secteur du luxe une interface utilisateur ergonomique qui lui permet de consulter les informations personnelles de ses clients, vérifier leur éligibilité à pratiquer leur droit à l'oubli, et de désactiver leurs différents canaux de notification. Elle dispose aussi d'une fonctionnalité d'audit qui permet de tracer l'historique des opérations effectuées, et lui permet donc de retrouver notamment la date à laquelle la personne a été anonymisée.
L'ensemble des informations remontées au niveau de l'application sont récupérées à partir des APIs REST exposées par Denodo.
Dans ce webinar, nous allons détailler l’ensemble des fonctionnalités de l’application DPO-Cockpit autour d’une démo, et expliquer à chaque étape le rôle central de Denodo pour réussir à simplifier la gestion du RGPD tout en étant compliant.
Les points clés abordés:
- Contexte client face aux enjeux du RGPD
- Défis et challenges rencontrés
- Options et choix retenu (Denodo)
- Démarche: architecture de la solution proposée
- Démo de l'outil: fonctionnalités principales
Lunch and Learn ANZ: Achieving Self-Service Analytics with a Governed Data Se...Denodo
Watch full webinar here: https://buff.ly/48zzN2h
In an increasingly distributed and complex data landscape, it is becoming increasingly difficult to govern and secure data effectively throughout the enterprise. Whether it be securing data across different repositories or monitoring access across different business units, the proliferation of data technologies and repositories across both on-premises and in the cloud is making the task unattainable. The challenge is only made greater by the ongoing pressure to offer self-service data access to business users.
Tune in and learn:
- How to use a logical data fabric to build an enterprise-wide data access role model.
- Centralise security when data is spread across multiple systems residing both on-premises and in the cloud.
- Control and audit data access across different regions.
How to Build Your Data Marketplace with Data Virtualization?Denodo
Watch full webinar here: https://buff.ly/4aAi0cS
Organizations continue to collect mounds of data and it is spread over different locations and in different formats. The challenge is navigating the vastness and complexity of the modern data ecosystem to find the right data to suit your specific business purpose. Data is an important corporate asset and it needs to be leveraged but also protected.
By adopting an alternate approach to data management and adapting a logical data architecture, data can be democratized while providing centralized control within a distributed data landscape. The web-based Data Catalog tool acts as a single access point for secure enterprise-wide data access and governance. This corporate data marketplace provides visibility into your data ecosystem and allows data to be shared without compromising data security policies.
Catch this live webinar to understand how this approach can transform how you leverage data across the business:
- Empower the knowledge worker with data and increase productivity
- Promote data accuracy and trust to encourage re-use of important data assets
- Apply consistent security and governance policies across the enterprise data landscape
Webinar #2 - Transforming Challenges into Opportunities for Credit UnionsDenodo
Watch full webinar here: https://buff.ly/3vhzqL5
Join our exclusive webinar series designed to empower credit unions with transformative insights into the untapped potential of data. Explore how data can be a strategic asset, enabling credit unions to overcome challenges and foster substantial growth.
This webinar will delve into how data can serve as a catalyst for addressing key challenges faced by credit unions, propelling them towards a future of enhanced efficiency and growth.
Enabling Data Catalog users with advanced usabilityDenodo
Watch full webinar here: https://buff.ly/48A4Yu1
Data catalogs are increasingly important in any modern data-driven organization. They are essential to manage and make the most of the huge amount of data that any organization uses. As this information is continuously growing in size and complexity, data catalogs are key to providing Data Discovery, Data Governance, and Data Lineage capabilities.
Join us for the session driven by David Fernandez, Senior Technical Account Manager at Denodo, to review the latest features aimed at improving the usability of the Denodo Data Catalog.
Watch on-demand & Learn:
- Enhanced search capabilities using multiple terms.
- How to create workflows to manage internal requests.
- How to leverage the AI capabilities of Data Catalog to generate SQL queries from natural language.
Watch full webinar here: https://buff.ly/3vjrn0s
The purpose of the Denodo Platform 8.0 Certified Architect Associate (DEN80EDUCAA) exam is to provide organizations that use Denodo Platform 8.0 with a means of identifying suitably qualified data architects who understand the role and position of the Denodo Platform within their broader information architecture.
This exam covers the following technical topics and subject areas:
- Denodo Platform functionality, including
- Governance and metadata management
- Security
- Performance optimization
- Caching
- Defining Denodo Platform use scenarios
Along with some sample questions, a Denodo Sales Engineer will help you prepare for exam topics and ace the exam.
Join us now to start your journey toward becoming a Certified Denodo Architect Associate!
GenAI y el futuro de la gestión de datos: mitos y realidadesDenodo
Watch full webinar here: https://buff.ly/3NLMSNM
El Generative AI y los Large Language Models (LLMs), encabezados por GPT de OpenAI, han supuesto la mayor revolución en el mundo de la computación de los últimos años. Pero ¿Cómo afectan realmente a la gestión de datos? ¿Reemplazarán los LLMs al profesional de la gestion de datos? ¿Cuánto hay de mito y cuánto de realidad?
En esta sesión revisaremos:
- Que es la Generative AI y por qué es importante para la gestión de datos
- Presente y futuro de aplicación de genAI en el mundo de los datos
- Cómo preparar tu organización para la adopción de genAI
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
StarCompliance is a leading firm specializing in the recovery of stolen cryptocurrency. Our comprehensive services are designed to assist individuals and organizations in navigating the complex process of fraud reporting, investigation, and fund recovery. We combine cutting-edge technology with expert legal support to provide a robust solution for victims of crypto theft.
Our Services Include:
Reporting to Tracking Authorities:
We immediately notify all relevant centralized exchanges (CEX), decentralized exchanges (DEX), and wallet providers about the stolen cryptocurrency. This ensures that the stolen assets are flagged as scam transactions, making it impossible for the thief to use them.
Assistance with Filing Police Reports:
We guide you through the process of filing a valid police report. Our support team provides detailed instructions on which police department to contact and helps you complete the necessary paperwork within the critical 72-hour window.
Launching the Refund Process:
Our team of experienced lawyers can initiate lawsuits on your behalf and represent you in various jurisdictions around the world. They work diligently to recover your stolen funds and ensure that justice is served.
At StarCompliance, we understand the urgency and stress involved in dealing with cryptocurrency theft. Our dedicated team works quickly and efficiently to provide you with the support and expertise needed to recover your assets. Trust us to be your partner in navigating the complexities of the crypto world and safeguarding your investments.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Myth Busters II: BI Tools and Data Virtualization are Interchangeable
1. W E B I N A R S E R I E S
BI Tools and Data
Virtualization are
Interchangeable
2. W E B I N A R S E R I E S
BI Tools and Data
Virtualization are
Interchangeable
Paul Moxon
SVP Data Architectures & Chief Evangelist
Denodo
17nd June 2020
7. 7
Welcome to my Universe
• BusinessObjects added Universe as semantic layer to
BI tool
• Special tools to design business-oriented data
objects
• Hide technical nature of physical data storage
• Initially use Data Federator to access multiple data
sources
• Multi-source Universe capability subsumed Data
Federator tool
• Made BusinessObjects the leading BI Tool vendor
• Increased usability and appeal to ‘citizen analysts’
8. 8
Follow the Leader
• Other vendors followed this approach
• MicroStrategy, Cognos, etc.
• New entrants initially focused on visualization
and analysis of data
• Tableau, Qlik, Power BI
• Quickly added ‘data blending’ capabilities
• Support multiple data source integration
• With limitations
9. 9
Data Blending Everywhere
• Most reporting tools now offer capabilities to create reports with data coming from
multiple data sources
• Some in real time, with their own federation engines (e.g. Tableau, MicroStrategy,
Business Objects, etc.)
• Some based on replication in the reporting tool engine (Qlik, SiSense, ThoughtSpot,
etc.)
• Some of them also provide data modeling capabilities (Looker, Business Objects,
MicroStrategy, PowerBI, etc.)
So if I can have multi-source queries and define a logical model in my
reporting tool, why would I need Data Virtualization?
11. 11
Source: “Gartner Market Guide for Data Virtualization, November 16, 2018”
Data virtualization can be used to create virtualized and
integrated views of data in-memory rather than executing data
movement and physically storing integrated views in a target
data structure. It provides a layer of abstraction above the
physical implementation of data, to simplify query logic.
12. 12
What is Data Virtualization?
Consume
in business applications
Combine
related data into views
Connect
to disparate data sources
2
3
1
DATA CONSUMERS
DISPARATE DATA SOURCES
Enterprise Applications, Reporting, BI, Portals, ESB, Mobile, Web, Users
Databases & Warehouses, Cloud/Saas Applications, Big Data, NoSQL, Web, XML, Excel, PDF, Word...
Analytical Operational
Less StructuredMore Structured
CONNECT COMBINE PUBLISH
Multiple Protocols,
Formats
Query, Search,
Browse
Request/Reply,
Event Driven
Secure
Delivery
SQL,
MDX
Web
Services
Big Data
APIs
Web Automation
and Indexing
CONNECT COMBINE CONSUME
Share, Deliver,
Publish, Govern,
Collaborate
Discover, Transform,
Prepare, Improve
Quality, Integrate
Normalized views of
disparate data
“Data virtualization
integrates disparate
data sources in real
time or near-real
time to meet
demands for
analytics and
transactional data.”
– Create a Road Map For A
Real-time, Agile, Self-
Service Data Platform,
Forrester Research, Dec 16,
2015
13. 13
What is Data Virtualization?
1. Single Access Point to all
Data at any location
2. Semantic Layer – Expose
Data in Business-Friendly
form, adapted to the
needs of each consumer
3. Abstract changes in the
underlying infrastructure
4. Single entry point to apply
security and governance
policies
5. Avoid data replication: Up
to 80% reduction in
integration costs, in terms
of resources and
technology data
14. 14
(Almost) Any-to-Many Connectivity
Relational Databases
• MS SQL*Server (JDBC, ODBC): 2000, 2005, 2008,
2008R2, 2012, 2014, 2016
• Oracle (JDBC): 8i, 9i, 10g, 11g, 12c, 18
• Oracle E-Business Suite (JDBC): 12
• IBM DB2 (JDBC): 8, 9, 10, 11, 12 for LUW; 9,10 for z/OS
• Informix (JDBC): 7, 12
• Sybase Adaptive Server Enterprise (JDBC): 12, 15
• MySQL (JDBC): 4, 5
• PostgreSQL (JDBC): 8, 9
• Denodo Platform (JDBC): 5.5, 6.0, 7.0
- For multi-location architecture deployments
• MS Access (ODBC)
• Apache Derby (JDBC): 10
• Generic (JDBC)
In-Memory Databases
• SAP HANA (JDBC): 1
• Oracle TimesTen (JDBC): 11g
• Oracle 12c In-Memory
Parallel databases and appliances
• GreenPlum (JDBC): 4.2
• HP Vertica (JDBC): 7, 8
• Oracle Exadata (JDBC): X5-2
• ParAccel 8.0.2 (using ParAccel 2.5.0.0 JDBC3g/SSL
driver)
• Netezza (JDBC): 4.6, 5.0, 6.0, 7.0
• SybaseIQ (JDBC) 12.x, 15.x
• Teradata (JDBC): 12, 13, 14, 15
Multi-Dimensional Sources
• SAP BW (BAPI/XMLA): 3.x
• SAP BI 7.x (BAPI): 7.x
• Mondrian (XMLA): 3.x
• MS SQL Server Analysis Services 200x
• Essbase (XMLA): 9, 11
Cloud Data Warehouse
• Amazon Redshift (JDBC)
• Amazon Athena (JDBC)
• Amazon Aurora (JDBC)
• Snowflake (JDBC)
• Amazon DynamoDB
• Azure SQL Data Warehouse
• Azure CosmosDB (SQL API and MongoDB API)
Big Data/NoSQL
• Apache Hive (JDBC): 0.12, 1.1.0, 1.1.0 for Cloudera
1.2.1 for Hortonworks 2.0.0
• MapR-XD, MapR-DB, MapR-ES, Hive, and Drill for
MapR 6.1
• Impala (JDBC): 2.3
• Spark SQL (JDBC): 1.5, 1.6
• Google BigQuery (JDBC)
• Presto (JDBC)
Web Automation
• Denodo’s ITPilot automates extraction from web
pages
Indexes and unstructured content
• CMS, file systems, pdf, word, text, email servers,
knowledge bases, indexes
• Elastic Search
Web Services
• SOAP
• REST (XML, RSS, ATOM, JSON)
• OData v2 and v4
Packaged Applications
• SAP ERP/ECC (BAPIs and RFC tables)
• Oracle E-Business Suite 12
• Siebel
• SAS (SAS JDBC Driver): 7 and higher
Semantic Repositories
• Semantic repositories in Triple Stores / RDF
accessed through SPARQL endpoints.
Flat and Binary Files
• CSV, pipe-delimited, Regular expression-parsed
• MS Excel xls 97-2003
• MS Excel xlsx 2007 or later
• MS Access
• XML
• JSON
All files can be locally accessible or in remote
filesystems, through FTP/ SFTP/FTPS, and in clear,
zipped and/or encrypted format.
Active Directory as source or leveraging security
• LDAP v3
• Microsoft Active Directory 2003, 2008
Cloud, SaaS, Web Sources with Simplified OAuth
Security
• Amazon
• Google
• Facebook
• LinkedIn
• MS Azure Data Lake
• MS SharePoint (by using the OData connector)
• MS Dynamics
• ServiceNow
• Marketo
• Salesforce
• Twitter via APIs with simplified Oauth integration
(1.0, 1.0a and 2.0)
• Workday
MS Queues as data source and Delivery
• MQSeries
• SonicMQ
• ActiveMQ
• Tibco EMS
Denodo SDK for Custom Connectors
• CouchDB
• Lotus Domino
• MongoDB and Mongo Atlas DBaaS
Mainframe
• IMS
• IBM IMS native drivers: 8, 9
• IMS Universal Drivers: 11
Hierarchical databases
• Adabas (SOA Gateway and Denodo’s SOAP
connector): 5, 6
Legacy
• Microsoft FoxPro (ODBC)
The following data sources have been successfully
tested with Denodo using JDBC and ODBC drivers,
WS/SOAP and WS/REST, and DenodoConnect
adapters (not exhaustive list):
• Apache Solr
• Kafka Messages
• SAS Files
• Hadoop HBase
• Hadoop HCatalog
• Hadoop HDFS (Avro, CSV, Parquet)
• Files in Amazon S3 (incl. Parquet files)
• IBM BigInsights
• Pivotal HAWQ
15. 15
(Almost) Any-to-Many Connectivity
Many Consumers
Protocols and Formats
• SQL Based access via JDBC, ODBC and ADO.NET
• Web Services
• SOAP (XML/JSON)
• REST (JSON/XML)
• OData
• Open API (a.k.a Swagger)
• Web Parts (for SharePoint), Portlets
• Kafka and JMS listeners for message queues
• Denodo Scheduler for batch process and ‘ETL lite’
Security Options
• Authentication using LDAP or Active Directory
• Kerberos for Single Sign-On (SSO)
• OAuth, OAuth 2.0 (JWT)
• SAML
• SSL/TLS
• WS-Security, X.509 certificates
BI/Reporting tools
• Microstrategy, Cognos, Business Objects, Oracle OBIEE
• Tableau, Qlikview, Spotfire, Microsoft PowerBI
• Excel
Analytical Tools/Languages
• SAS, Statistica, SPSS, MatLab
• R, Python, Java, Scala, etc.
• Azure ML Studio, Amazon Machine Learning
Portals
• SharePoint, Enterprise portals, Web/mobile apps
Enterprise Service Bus
• Oracle Service Bus, Azure Service Bus, TIBCO Active Matrix
Bus
ETL tools
• SAP Data Services, Informatica Powercenter, IBM Data
Stage, Talend ETL
API Management tools
• CA (Layer 7), TIBCO Mashery, Apigee
17. 17
Data Blending Silos
Q: Is SAP planning to release SAP Universe connections for Power BI and Tableau?
A: The answer is no. No. There are no plans for this.
Gregory Botticchio, Director of Product Management, SAP BusinessObjects
Suite 360 webinar for SAP BusinessObjects 4.3 Release Preview
Beside SAP BusinessObjects, are you
using other analytics solution(s)?
18. 18
Data Blending Limitations
Shared Dataset
(Import Mode)
Shared Dataset
(Direct Mode)
Direct mode is limited
to 1 data source
and 1 million rows
19. 19
Francois Ajenstat, Chief Product Officer, Tableau Software
There are two flows; the ad-hoc and the operational…where we are
coming from is…I just want to integrate these two sources. It's not
formalized, per se, it's not a project. I just want to connect this and this
and I want to analyze it. How do we go from data to analysis as quickly as
possible? And when you want to formalize it, operationalize it, make it
repeatable, then [you use other tools].
21. 21
Denodo’s Coronavirus Data Portal
File
Denodo Express
COVID-19 Edition
Data
Catalog
Data
Portal
JDBC
ODBC
API
GraphQL
GeoJSON
Sandbox
Sandbox
Sandbox
22. 22
Connected Data Sources
Australian Bureau of Statistics Labor Force
Survey
ACAPS
Air Quality Open Data Platform
Allen Institute for AI
ArcGIS Hub
Becker Friedman Institute for Research in
Economics, University of Chicago
California Health and Human Services (CHHS)
Carnegie Mellon University
Centraal Bureau voor de Statistiek (CBS),
Netherlands
COVID19-India (covid19india.org)
Data Science for Social Impact Research Group
(DSFSI), University of Pretoria
Dipartimento della Protezione Civile, Italy
Europa Press
European Centre for Disease Prevention and
Control (ECDC)
Federal Ministry of Social Affairs, Health, Care
and Consumer Protection (BMSGPK), Austria
France GEOJSON
French Government Open Data (data.gouv.fr)
GlobalHealth 50/50
Google - COVID-19 Community Mobility
Reports
Hong Kong Department of Health
Humanitarian Data Exchange
Institute for Health Metrics and Evaluation
(IHME)
Instituto de Salud Carlos III
International Monetary Fund (IMF)
Istituto Nazionale di Statistica, Italy
Johns Hopkins University (JHU) Center for
Systems Science and Engineering (CSSE)
Junta de Castilla y Léon
Kaiser Family Foundation (KFF)
Ministerio de Sanidad, Spain
Ministry of Health of New Zealand
Ministry of Health, Brazil
Ministry of Health, Consumer Affairs and
Social Welfare, Spain
Ministry of Health, Labor and Welfare, Japan
National Institute for Health (NIH) - National
Library of Medicine (NLM)
Netherlands National Institute for Public
Health and the Environment (RIVM)
New York City Department of Health and
Mental Hygiene (DOHMH)
Office for National Statistics, UK
Organisation for Economic Co-operation and
Development (OECD)
Our World in Data
Public Health England
Robert Koch Institute (RKI)
RSS News Feeds
San Francisco Department of Public Health
(SFDPH)
Servicio Publico de Empleo Estatal (SEPE),
Spain
Statista.com
Statistics Austria
Statistics Canada
Statistics Norway
Statistics Sweden
Taiwan Centers for Disease Control
Texas Department of State, Health Services
Thailand Department for Disease Control
The COVID Tracking Project
The Economist
The Government of the Hong Kong Special
Administrative Region - Census and Statistics
Department
The New York Times
The World Bank
United Kingdom Government Open Data
(gov.uk)
United Nations Educational, Scientific and
Cultural Organization (UNESCO)
United Nations Population Division, Department
of Economic and Social Affairs
US Department of Labor
Wharton School of Business, University of
Pennsylvania
World Health Organization (WHO)
25. 25
Comparing Apples to Oranges
• Data Virtualization and ‘Data Blending’ serve two different purposes
• Data Blending is focused on a single vendor’s toolset
• It makes it easier for ‘citizen analysts’ to use a specific BI Tool
• It provides a semantic layer for that specific toolset
• It has limitations on real-time use
• Data Virtualization provides an enterprise-wide data fabric layer
• Supports many different consuming tools
• Creates a general purpose semantic layer for all users
• Can mix data delivery modes without limitations
• Use the right tool for the right task