This is a lecture from the advanced service engineering course from the Vienna University of Technology. See http://dsg.tuwien.ac.at/teaching/courses/ase/
In their webinar "Big Data Fabric 2.0 Drives Data Democratization" Ben Szekley, Cambridge Semantics’ SVP of Field Operations, and guest speaker, Forrester’s Noel Yuhanna, author of the Forrester report: “Big Data Fabric 2.0 Drives Data Democratization”, explored why data-driven businesses are making a big data fabric part of their data strategy to minimize data complexity, integrate siloed data, deliver real-time trusted insights, and to create new business opportunities. These are the slides from that webinar.
In this webinar, data analytics gurus Sathish Thyagarajan and Steve Sarsfield introduce AnzoGraph™, our graph OLAP database, demonstrate the different types of analyses you can perform with it and how it complements Neo4j, AWS Neptune and other OLTP systems. Finally, they’ll show how you can get it up and running on your laptop in about 5 minutes.
DataGraft: Data-as-a-Service for Open Datadapaasproject
Tutorial at "The Data Matters Series – Transforming Service Industry via Big Data Analytics", May 4, 2016, Cyberjaya, Malaysia
http://www.eventbrite.com/e/the-data-matters-series-transforming-service-industry-via-big-data-analytics-tickets-24617911837
In their webinar "Big Data Fabric 2.0 Drives Data Democratization" Ben Szekley, Cambridge Semantics’ SVP of Field Operations, and guest speaker, Forrester’s Noel Yuhanna, author of the Forrester report: “Big Data Fabric 2.0 Drives Data Democratization”, explored why data-driven businesses are making a big data fabric part of their data strategy to minimize data complexity, integrate siloed data, deliver real-time trusted insights, and to create new business opportunities. These are the slides from that webinar.
In this webinar, data analytics gurus Sathish Thyagarajan and Steve Sarsfield introduce AnzoGraph™, our graph OLAP database, demonstrate the different types of analyses you can perform with it and how it complements Neo4j, AWS Neptune and other OLTP systems. Finally, they’ll show how you can get it up and running on your laptop in about 5 minutes.
DataGraft: Data-as-a-Service for Open Datadapaasproject
Tutorial at "The Data Matters Series – Transforming Service Industry via Big Data Analytics", May 4, 2016, Cyberjaya, Malaysia
http://www.eventbrite.com/e/the-data-matters-series-transforming-service-industry-via-big-data-analytics-tickets-24617911837
Scaling up business value with real-time operational graph analyticsConnected Data World
Graph-based solutions have been in the market for over a decade with deployments in financial services, healthcare, retail, and manufacturing. The graph technology of the past limited them to simple queries (1 or 2 hops), modest data sizes, or slow response times, which limited their value.
A new generation of fast, scalable graph databases, led by TigerGraph, is opening up a new world of business insight and performance. Join us, as we explore some new exciting use cases powered by native parallel graph database with storage and computation capability for each node:
A large financial services payment provider is using graph-based pattern detection (7 to 11 hop queries) to detect more fraud and money laundering in real time, handling peak volume of 256,000 transactions per second.
IceKredit, an innovative FinTech is transforming the near-prime and sub-prime credit market in United States, China and South Asian countries with customer 360 analytics for credit approval and ongoing monitoring.
A biotech and pharmaceutical giant is building a prescriber and patient 360 graph and using multi-hop exploratory and analytic queries to understand the most efficient ways of launching a new drug for maximum return.
Wish.com is delivering real-time personalized recommendations to increase eCommerce revenue.
As we enter the digital economy, it becomes increasingly transparent that the information and data ecosphere will continue to be a complex environment for the foreseeable future, with information being provided from a variety of internal and external sources in the form of files, messages, queries and streams. It would be foolish for any organization to place their bets on any one platform to be their platform of choice because it is incongruent to the thought patterns of the consumers, suppliers, regulators, partners and financiers who will participate in their information ecosphere through data feeds, information requests and a host of other interfaces.
Rather, there is a role of each of these platforms which serve as the conduit for data and the transformation of data into information aligned with the value propositions of the organization. This writing is focused on the big data platform because there are some unique characteristics of the big data environment that require an approach different than many of the legacy environments that exist in organizations. Furthermore, while big data is the one environment that is new and requires these special handling characteristics, there will be future platforms with the same requirements as big data requires today, and hopefully lessons learned will be left to not revisit each of the challenges as the next transformational information ecosphere is made available.
Figure 1 The Fourth Industrial Revolution, World Economic Forum, InfoSight Partners, 2016
This time is different, in that information is the catalyst to achieving value and the platform ideally suited to house information not optimal for storage in the form of rows and columns is the big data environment. Understanding which information is delivered with intended consequences and having the management prowess to tune information shared with customers, prospects, suppliers, partners, regulators and financiers is critical for the digital economy. Additionally, it is specific to understand the challenges each platform housing information bring to the equation. This writing will focus on big data.
The most profitable insurance organizations will outperform competitors in key areas as personalized customer service, claims processing, subrogation recovery, fraud detection and product innovation. This requires thinking beyond the traditional data warehouse to the data fabric - an emerging data management architecture.
In this webinar Andy Sohn, Senior Advisor at NewVantage Partners, and Bob Parker, Senior Director for Insurance at Cambridge Semantics, explore the role of the data discovery and integration layer in an enterprise data fabric for the Insurance industry. These are their slides.
Risk Analytics Using Knowledge Graphs / FIBO with Deep LearningCambridge Semantics
This EDM Council webinar, sponsored by Cambridge Semantics Inc. and featuring FI Consulting, explores the challenges common to a risk analytics pipeline, application of graph analytics to mortgage loan data and use cases in adjacent areas including customer service, collections, fraud and AML.
Annual Big Data Landscape prepared by FIrstMark. Check out full blog post: "Is Big Data Still a Thing"? at http://mattturck.com/2016/02/01/big-data-landscape/
IDS: Update on Reference Architecture and Ecosystem DesignBoris Otto
This presentation motivates the Industrial Data Space and gives an update on the IDS Reference Architecture Model as well as the related ecosystem. It sets data in the context of business model innovation and points out how the IDS Reference Architecture relates to alternative data architecture styles such as data lakes and blockchain technology, for example. The presentation was given at the IDSA Summit on March 22, 2018.
Delivering Quality Open Data by Chelsea UrsanerData Con LA
Abstract:- The value of data is exponentially related to the number of people and applications that have access to it. The City of Los Angeles embraces this philosophy and is committed to opening as much of its data as it can in order to stimulate innovation, collaboration, and informed discourse. This presentation will be a review of what you can find and do on our open data portals as well as our strategy for delivering the best open data program in the nation.
How Financial Institutions Are Leveraging Data Virtualization to Overcome the...Denodo
Watch full webinar here: https://bit.ly/2KkJ08B
Financial institutions need to implement new strategies and services that will drive them securely to their digital objectives over their entire infrastructure.
- How to securely move legacy systems and data to new technologies such as the Big Data and Cloud?
- How to break down silos and ensure a global, centralized, secure and agile access to meaningful data?
- How to facilitate data sharing while applying strict and coherent governance and security rules?
- How to avoid downtime and to guarantee the success of IT initiaves while optimizing costs and resources?
- How to produce and to maintain efficient reports and financial aggregations for the holdings and CxO managers?
We are pleased to invite you to this online session to discover how data virtualization can answer these questions and contribute to the digital transformation of financial institutions.
WHAT IS IT ABOUT?
This virtual event will be organized in two parts. First, we will conduct a conference focusing on the impact of digital transformation in the financial sector, in addition to the general concepts of Data Virtualization and how it has supported the new business goals of financial companies in terms of IT modernization, risk management, governance and security. Then, we will conduct will conduct a hands-on session with a guided live demo to help you discover the main features and benefits of Denodo Platform for Data Virtualization.
Presentation at Data/Graph Day Texas Conference.
Austin, Texas
January 14, 2017
This talk grew out Juan Sequeda's office hours following the Seattle Graph Meetup. Some of the questions posed were: How do I recognize problem best solved with a graph solution? How do I determine the best type of graph to solve the problem? How do I manage the data where both graph and relational operations will be performed? Juan did such a great job of explaining the options, we asked him to develop his responses into a formal talk.
International Data Spaces: Data Sovereignty and Interoperability for Business...Boris Otto
This presentation was held in a workshop session on IoT Business Models and Data Interoperability at the Max Planck Institute for Innovation and Competition in Munich on 8 October 2018. The presenation introduces the concept of business ecosystems and the role of data within the latter, then outlines the state of the art in terms of interoperability and sovereignty and finally sketches the IDS contribution.
Watch Alberto's presentation from Fast Data Strategy on-demand here: https://goo.gl/CRjYuD
In this session, we will review Denodo Platform 7.0 key capabilities.
Watch this session to learn more about:
• The vision behind the Denodo Platform
• The new data catalog and self-service features of Denodo Platform 7.0
• The new connectivity, data transformation, and enterprise-wide deployment features
Using Cloud Automation Technologies to Deliver an Enterprise Data FabricCambridge Semantics
The world of database management is changing. Cloud adoption is accelerating, offering a path for companies to increase their database capabilities while keeping costs in line. To help IT decision-makers survive and thrive in the cloud era, DBTA hosted this special roundtable webinar.
Powering Self Service Business Intelligence with Hadoop and Data VirtualizationDenodo
A Webinar with Hortonworks and Denodo (watch on demand here: https://goo.gl/xuP1Ak)
Vizient needed a unified view of their accounting and financial data marts to enable business users to discover the information they need in a self-service manner and to be able to provide excellent service to their members. Vizient selected Hortonworks Big Data Platform and Denodo Data Virtualization Platform so that they can unify their distributed data sets in a data lake, and at the same time provide an abstraction for end users for easy self-serviceable information access.
During this webinar, you will learn:
1) The role, use, and benefits of Hortonworks Data Platform in the Modern Data Architecture.
2) How Hadoop and data virtualisation simplify data management and self-service data discovery.
3) What data virtualisation is and how it can simplify big data projects. Best practices of using Hadoop with data virtualisation
About Vizient
Vizient, Inc. is the largest nationwide network of community-owned health care systems and their physicians in the US. Vizient™ combines the strengths of VHA, University HealthSystem Consortium (UHC), Novation and MedAssets SCM and Sg2, trusted leaders focused on solving health care's most pressing challenges. Vizient delivers brilliant resources and powerful data driven insights to healthcare organizations.
Designing an Agile Fast Data Architecture for Big Data Ecosystem using Logica...Denodo
Autodesk designed a modern data architecture that heavily uses data virtualization to integrate both legacy data sources and contemporary big data analytics like Spark into a single unified logical data warehouse. In this presentation, you will learn how to build a logical data warehouse using data virtualization and create a single, unified enterprise-wide access and governance point for any data used within the company.
This presentation is part of the Fast Data Strategy Conference, and you can watch the video here goo.gl/Ab4PDB.
Using a Semantic and Graph-based Data Catalog in a Modern Data FabricCambridge Semantics
Watch this webinar to learn about the benefits of using semantic and graph database technology to create a Data Catalog of all of an enterprise's data, regardless of source or format, as part of a modern IT or data management stack and an important step toward building an Enterprise Data Fabric.
Unified views of business-critical information across all customer-facing processes and HR-related tasks are most relevant for decision makers.
In this talk we present a SharePoint extension that supports the automatic linking of unstructured content like Word documents with structured information from other databases, such as statistical data. As a result, decision makers have knowledge portals based on linked data at their fingertips.
While the importance of managed metadata and Term Store is clear to most SharePoint architects, the significance of a semantic layer outside of the content silos has not yet been explored systematically.
We will present a four-layered content architecture and will take a close look on some of the aspects of the semantic layer and its integration with SharePoint:
- Keeping Term Store and the semantic layer in sync
- Automatic tagging of SharePoint content
- Use of graph databases to store tags
- Entity-centric search & analytics applications
Metadata is most often stored per data source, and therefore it is meaningless outside of the silo. In this presentation, we will give a live demo of a SharePoint extension that makes use of an explicit semantic layer based on standards. This approach builds the basis to start linking data across the silos in a most agile way.
The resulting knowledge graph can start on a small scale, to develop continuously and to grow with the requirements. In this presentation we will give an example to illustrate how initially disconnected HR-related data (CVs in SharePoint; statistical data from labour market; skills and competencies taxonomies; salary spreadsheets) gets linked automatically, and is then made available through an extensive search & analytics application.
How Semantics Solves Big Data ChallengesDATAVERSITY
Today, organizations want both IT simplicity and innovation, but reliance on traditional databases only leads to more complexity, longer development cycles, and more silos. In fact, organizations report that the #1 impediment to big data success is having too many silos. In this webinar, we will discuss how a new database technology, semantics, solves this problem by providing a new approach to modeling data that focuses on relationships and context, making it easier for data to be understood, searched, and shared. With semantics, world-leading organizations are integrating disparate data faster and easier and building smarter applications with richer analytic capabilities—benefits that we look forward to diving into during the webinar.
Scaling up business value with real-time operational graph analyticsConnected Data World
Graph-based solutions have been in the market for over a decade with deployments in financial services, healthcare, retail, and manufacturing. The graph technology of the past limited them to simple queries (1 or 2 hops), modest data sizes, or slow response times, which limited their value.
A new generation of fast, scalable graph databases, led by TigerGraph, is opening up a new world of business insight and performance. Join us, as we explore some new exciting use cases powered by native parallel graph database with storage and computation capability for each node:
A large financial services payment provider is using graph-based pattern detection (7 to 11 hop queries) to detect more fraud and money laundering in real time, handling peak volume of 256,000 transactions per second.
IceKredit, an innovative FinTech is transforming the near-prime and sub-prime credit market in United States, China and South Asian countries with customer 360 analytics for credit approval and ongoing monitoring.
A biotech and pharmaceutical giant is building a prescriber and patient 360 graph and using multi-hop exploratory and analytic queries to understand the most efficient ways of launching a new drug for maximum return.
Wish.com is delivering real-time personalized recommendations to increase eCommerce revenue.
As we enter the digital economy, it becomes increasingly transparent that the information and data ecosphere will continue to be a complex environment for the foreseeable future, with information being provided from a variety of internal and external sources in the form of files, messages, queries and streams. It would be foolish for any organization to place their bets on any one platform to be their platform of choice because it is incongruent to the thought patterns of the consumers, suppliers, regulators, partners and financiers who will participate in their information ecosphere through data feeds, information requests and a host of other interfaces.
Rather, there is a role of each of these platforms which serve as the conduit for data and the transformation of data into information aligned with the value propositions of the organization. This writing is focused on the big data platform because there are some unique characteristics of the big data environment that require an approach different than many of the legacy environments that exist in organizations. Furthermore, while big data is the one environment that is new and requires these special handling characteristics, there will be future platforms with the same requirements as big data requires today, and hopefully lessons learned will be left to not revisit each of the challenges as the next transformational information ecosphere is made available.
Figure 1 The Fourth Industrial Revolution, World Economic Forum, InfoSight Partners, 2016
This time is different, in that information is the catalyst to achieving value and the platform ideally suited to house information not optimal for storage in the form of rows and columns is the big data environment. Understanding which information is delivered with intended consequences and having the management prowess to tune information shared with customers, prospects, suppliers, partners, regulators and financiers is critical for the digital economy. Additionally, it is specific to understand the challenges each platform housing information bring to the equation. This writing will focus on big data.
The most profitable insurance organizations will outperform competitors in key areas as personalized customer service, claims processing, subrogation recovery, fraud detection and product innovation. This requires thinking beyond the traditional data warehouse to the data fabric - an emerging data management architecture.
In this webinar Andy Sohn, Senior Advisor at NewVantage Partners, and Bob Parker, Senior Director for Insurance at Cambridge Semantics, explore the role of the data discovery and integration layer in an enterprise data fabric for the Insurance industry. These are their slides.
Risk Analytics Using Knowledge Graphs / FIBO with Deep LearningCambridge Semantics
This EDM Council webinar, sponsored by Cambridge Semantics Inc. and featuring FI Consulting, explores the challenges common to a risk analytics pipeline, application of graph analytics to mortgage loan data and use cases in adjacent areas including customer service, collections, fraud and AML.
Annual Big Data Landscape prepared by FIrstMark. Check out full blog post: "Is Big Data Still a Thing"? at http://mattturck.com/2016/02/01/big-data-landscape/
IDS: Update on Reference Architecture and Ecosystem DesignBoris Otto
This presentation motivates the Industrial Data Space and gives an update on the IDS Reference Architecture Model as well as the related ecosystem. It sets data in the context of business model innovation and points out how the IDS Reference Architecture relates to alternative data architecture styles such as data lakes and blockchain technology, for example. The presentation was given at the IDSA Summit on March 22, 2018.
Delivering Quality Open Data by Chelsea UrsanerData Con LA
Abstract:- The value of data is exponentially related to the number of people and applications that have access to it. The City of Los Angeles embraces this philosophy and is committed to opening as much of its data as it can in order to stimulate innovation, collaboration, and informed discourse. This presentation will be a review of what you can find and do on our open data portals as well as our strategy for delivering the best open data program in the nation.
How Financial Institutions Are Leveraging Data Virtualization to Overcome the...Denodo
Watch full webinar here: https://bit.ly/2KkJ08B
Financial institutions need to implement new strategies and services that will drive them securely to their digital objectives over their entire infrastructure.
- How to securely move legacy systems and data to new technologies such as the Big Data and Cloud?
- How to break down silos and ensure a global, centralized, secure and agile access to meaningful data?
- How to facilitate data sharing while applying strict and coherent governance and security rules?
- How to avoid downtime and to guarantee the success of IT initiaves while optimizing costs and resources?
- How to produce and to maintain efficient reports and financial aggregations for the holdings and CxO managers?
We are pleased to invite you to this online session to discover how data virtualization can answer these questions and contribute to the digital transformation of financial institutions.
WHAT IS IT ABOUT?
This virtual event will be organized in two parts. First, we will conduct a conference focusing on the impact of digital transformation in the financial sector, in addition to the general concepts of Data Virtualization and how it has supported the new business goals of financial companies in terms of IT modernization, risk management, governance and security. Then, we will conduct will conduct a hands-on session with a guided live demo to help you discover the main features and benefits of Denodo Platform for Data Virtualization.
Presentation at Data/Graph Day Texas Conference.
Austin, Texas
January 14, 2017
This talk grew out Juan Sequeda's office hours following the Seattle Graph Meetup. Some of the questions posed were: How do I recognize problem best solved with a graph solution? How do I determine the best type of graph to solve the problem? How do I manage the data where both graph and relational operations will be performed? Juan did such a great job of explaining the options, we asked him to develop his responses into a formal talk.
International Data Spaces: Data Sovereignty and Interoperability for Business...Boris Otto
This presentation was held in a workshop session on IoT Business Models and Data Interoperability at the Max Planck Institute for Innovation and Competition in Munich on 8 October 2018. The presenation introduces the concept of business ecosystems and the role of data within the latter, then outlines the state of the art in terms of interoperability and sovereignty and finally sketches the IDS contribution.
Watch Alberto's presentation from Fast Data Strategy on-demand here: https://goo.gl/CRjYuD
In this session, we will review Denodo Platform 7.0 key capabilities.
Watch this session to learn more about:
• The vision behind the Denodo Platform
• The new data catalog and self-service features of Denodo Platform 7.0
• The new connectivity, data transformation, and enterprise-wide deployment features
Using Cloud Automation Technologies to Deliver an Enterprise Data FabricCambridge Semantics
The world of database management is changing. Cloud adoption is accelerating, offering a path for companies to increase their database capabilities while keeping costs in line. To help IT decision-makers survive and thrive in the cloud era, DBTA hosted this special roundtable webinar.
Powering Self Service Business Intelligence with Hadoop and Data VirtualizationDenodo
A Webinar with Hortonworks and Denodo (watch on demand here: https://goo.gl/xuP1Ak)
Vizient needed a unified view of their accounting and financial data marts to enable business users to discover the information they need in a self-service manner and to be able to provide excellent service to their members. Vizient selected Hortonworks Big Data Platform and Denodo Data Virtualization Platform so that they can unify their distributed data sets in a data lake, and at the same time provide an abstraction for end users for easy self-serviceable information access.
During this webinar, you will learn:
1) The role, use, and benefits of Hortonworks Data Platform in the Modern Data Architecture.
2) How Hadoop and data virtualisation simplify data management and self-service data discovery.
3) What data virtualisation is and how it can simplify big data projects. Best practices of using Hadoop with data virtualisation
About Vizient
Vizient, Inc. is the largest nationwide network of community-owned health care systems and their physicians in the US. Vizient™ combines the strengths of VHA, University HealthSystem Consortium (UHC), Novation and MedAssets SCM and Sg2, trusted leaders focused on solving health care's most pressing challenges. Vizient delivers brilliant resources and powerful data driven insights to healthcare organizations.
Designing an Agile Fast Data Architecture for Big Data Ecosystem using Logica...Denodo
Autodesk designed a modern data architecture that heavily uses data virtualization to integrate both legacy data sources and contemporary big data analytics like Spark into a single unified logical data warehouse. In this presentation, you will learn how to build a logical data warehouse using data virtualization and create a single, unified enterprise-wide access and governance point for any data used within the company.
This presentation is part of the Fast Data Strategy Conference, and you can watch the video here goo.gl/Ab4PDB.
Using a Semantic and Graph-based Data Catalog in a Modern Data FabricCambridge Semantics
Watch this webinar to learn about the benefits of using semantic and graph database technology to create a Data Catalog of all of an enterprise's data, regardless of source or format, as part of a modern IT or data management stack and an important step toward building an Enterprise Data Fabric.
Unified views of business-critical information across all customer-facing processes and HR-related tasks are most relevant for decision makers.
In this talk we present a SharePoint extension that supports the automatic linking of unstructured content like Word documents with structured information from other databases, such as statistical data. As a result, decision makers have knowledge portals based on linked data at their fingertips.
While the importance of managed metadata and Term Store is clear to most SharePoint architects, the significance of a semantic layer outside of the content silos has not yet been explored systematically.
We will present a four-layered content architecture and will take a close look on some of the aspects of the semantic layer and its integration with SharePoint:
- Keeping Term Store and the semantic layer in sync
- Automatic tagging of SharePoint content
- Use of graph databases to store tags
- Entity-centric search & analytics applications
Metadata is most often stored per data source, and therefore it is meaningless outside of the silo. In this presentation, we will give a live demo of a SharePoint extension that makes use of an explicit semantic layer based on standards. This approach builds the basis to start linking data across the silos in a most agile way.
The resulting knowledge graph can start on a small scale, to develop continuously and to grow with the requirements. In this presentation we will give an example to illustrate how initially disconnected HR-related data (CVs in SharePoint; statistical data from labour market; skills and competencies taxonomies; salary spreadsheets) gets linked automatically, and is then made available through an extensive search & analytics application.
How Semantics Solves Big Data ChallengesDATAVERSITY
Today, organizations want both IT simplicity and innovation, but reliance on traditional databases only leads to more complexity, longer development cycles, and more silos. In fact, organizations report that the #1 impediment to big data success is having too many silos. In this webinar, we will discuss how a new database technology, semantics, solves this problem by providing a new approach to modeling data that focuses on relationships and context, making it easier for data to be understood, searched, and shared. With semantics, world-leading organizations are integrating disparate data faster and easier and building smarter applications with richer analytic capabilities—benefits that we look forward to diving into during the webinar.
Practical DoDAF Presentation to International Council on Systems Engineering Washington Metro Area by Steven H. Dam Ph.D., ESEP, founder of SPEC Innovations
Big Data: Architecture and Performance Considerations in Logical Data LakesDenodo
This presentation explains in detail what a Data Lake Architecture looks like, how data virtualization fits into the Logical Data Lake, and goes over some performance tips. Also it includes an example demonstrating this model's performance.
This presentation is part of the Fast Data Strategy Conference, and you can watch the video here goo.gl/9Jwfu6.
Data Modelling 101 half day workshop presented by Chris Bradley at the Enterprise Data and Business Intelligence conference London on November 3rd 2014.
Chris Bradley is a leading independent information strategist.
Contact chris.bradley@dmadvisors.co.uk
Data modelling for the business half day workshop presented at the Enterprise Data & Business Intelligence conference in London on November 3rd 2014
chris.bradley@dmadvisors.co.uk
TUW - 184.742 Data marketplaces: models and conceptsHong-Linh Truong
This presentation is part of the course "184.742 Advanced Services Engineering" at The Vienna University of Technology, in Winter Semester 2012. Check the course at: http://www.infosys.tuwien.ac.at/teaching/courses/ase/
Fast Data Strategy Houston Roadshow PresentationDenodo
Fast Data Strategy Houston Roadshow focused on the next industrial revolution on the horizon, driven by the application of big data, IoT and Cloud technologies.
• Denodo’s innovative customer, Anadarko, elaborated on how data virtualization serves as the key component in their prescriptive and predictive analytics initiatives, driven by multi-structured data ranging from customer data to equipment data.
• Denodo’s session, Unleashing the Power of Data, described the complexity of the modern data ecosystem and how to overcome challenges and successfully harness insights.
• Our Partner Noah Consulting, an expert analytics solutions provider in the energy industry, explained how your peers are innovating using new business models and reducing cost in areas such as Asset Management and Operations by leveraging Data Virtualization and Prescriptive and Predictive Analytics.
For more information on upcoming roadshows near you, follow this link: https://goo.gl/WBDHiE
TUW-ASE Summer 2015: Advanced service-based data analytics: Models, Elasticit...Hong-Linh Truong
This is a lecture from the advanced service engineering course from the Vienna University of Technology. See http://dsg.tuwien.ac.at/teaching/courses/ase
Big Data LDN 2018: CONNECTING SILOS IN REAL-TIME WITH DATA VIRTUALIZATIONMatt Stubbs
Date: 14th November 2018
Location: Keynote Theatre
Time: 13:50 - 14:20
Speaker: Becky Smith
Organisation: Denodo
About: How many users inside and outside of your organization access your organization’s data? Dozens? Hundreds is probably more like it, each with their own structure and content requirements as well as different access rights. As a result, many organizations have witnessed the formation of “data delivery mills,” in various shapes and sizes. How does one create order and reliability in this world of chaotic data streams? Quite easily, if it’s done with data virtualization.
According to Gartner, "through 2020, 50% of enterprises will implement some form of data virtualization as one enterprise production option for data integration.” Data virtualization enables organizations to gain data insights from multiple, distributed data sources without the time-consuming processes of data extraction and loading. This allows for faster insights and fact-based decisions, which help business realize value sooner.
Join us to find out more about:
• What data virtualization actually means and how it differs from traditional data integration approaches.
• How you can connect and combine all your data in real-time, without compromising on scalability, security or governance.
• The benefits of data virtualization and its most important use cases.
Emerging Dynamic TUW-ASE Summer 2015 - Distributed Systems and Challenges for...Hong-Linh Truong
This is a lecture from the advanced service engineering course from the Vienna University of Technology. See http://dsg.tuwien.ac.at/teaching/courses/ase/
Your Data is Waiting. What are the Top 5 Trends for Data in 2022? (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3saONRK
COVID-19 has pushed every industry and organization to embrace digital transformation at scale, upending the way many businesses will operate for the foreseeable future. Organizations no longer tolerate monolithic and centralized data architecture; they are embracing flexibility, modularity, and distributed data architecture to help drive innovation and modernize processes.
The pandemic has compelled organizations to accelerate their digital transformation initiatives and look for smarter and more agile ways to manage and leverage their corporate data assets. Data governance has become challenging in the ever-increasing complexity and distributed nature of the data ecosystem. Interoperability, collaboration and trust in data are imperative for a business to succeed. Data needs to be easily accessible and fit for purpose.
In this session, Denodo experts will discuss 5 key trends that are expected to be top of mind for CIOs and CDOs;
- Distributed Data Environments
- Decision Intelligence
- Modern Data Architecture
- Composable Data & Analytics
- Hyper-personalized Experiences
A Taxonomy of the Data Resource in the Networked IndustryBoris Otto
This presentation reports on the design of a taxonomy of the data resource in the networked industry. It was held on the 7th International Scientific Symposium on Logistics on June 6, 2014, in Cologne, Germany. The presentation motivates the topic, analyzes four networking industry cases and discusses a first version of the taxonomy. The presentation argues that for companies aiming at designing a future-proof data architecture leveraging the potentials of the industrial internet, collaborative forms of organizations etc. transparency about data sources, data ownership, criticality, compliance of standards of data, data quality are key for success. In addition, the presentation introduces a first sketch of a method supporting businesses in applying the taxonomy.
Denodo Partner Connect: A Review of the Top 5 Differentiated Use Cases for th...Denodo
Watch full webinar here: https://buff.ly/46pRfV7
This Denodo session explores the power of data virtualization, shedding light on its architecture, customer value, and a diverse range of use cases. Attendees will discover how the Denodo Platform enables seamless connectivity to various data sources while effortlessly combining, cleansing, and delivering data through 5 differentiated use cases.
Architecture: Delve into the core architecture of the Denodo Platform and learn how it empowers organizations to create a unified virtual data layer. Understand how data is accessed, integrated, and delivered in a real-time, agile manner.
Value for the Customer: Explore the tangible benefits that Denodo offers to its customers. From cost savings to improved decision-making, discover how the Denodo Platform helps organizations derive maximum value from their data assets.
Five Different Use Cases: Uncover five real-world use cases where Denodo's data virtualization platform has made a significant impact. From data governance to analytics, Denodo proves its versatility across a variety of domains.
- Logical Data Fabric
- Self Service Analytics
- Data Governance
- 360 degree of Entities
- Hybrid/Multi-Cloud Integration
Watch this illuminating session to gain insights into the transformative capabilities of the Denodo Platform.
Watch full webinar here: https://bit.ly/2vN59VK
What started to evolve as the most agile and real-time enterprise data fabric, data virtualization is proving to go beyond its initial promise and is becoming one of the most important enterprise big data fabrics.
Attend this session to learn:
- What data virtualization really is.
- How it differs from other enterprise data integration technologies.
- Why data virtualization is finding enterprise-wide deployment inside some of the largest organizations.
Webinar presented live on August 11, 2017
Today, the majority of big data and analytics use cases are built on hybrid cloud infrastructure. A hybrid cloud is a combination of on-premises and local cloud resources integrated with one or more dedicated cloud(s) and one or more public cloud(s). Hybrid cloud computing has matured to support data security and privacy requirements as well as increased scalability and computational power needed for big data and analytics solutions.
This webinar summarizes what hybrid cloud is, explains why it is important in the context of big data and analytics, and discusses implementation considerations unique to hybrid cloud computing.
The presentation draws from the CSCC's deliverable, Hybrid Cloud Considerations for Big Data and Analytics:
http://www.cloud-council.org/deliverables/hybrid-cloud-considerations-for-big-data-and-analytics.htm
Download the presentation deck here:
http://www.cloud-council.org/webinars/hybrid-cloud-considerations-for-big-data-and-analytics.htm
Data Virtualization – Gateway to a Digital Business - Barry DevlinDenodo
Next-Generation Data Management Afternoon
with InfoRoad and Denodo. Presentation by Dr Barry Devlin, Founder and Principal 9sight Consulting on data virtualization.
Data Virtualization: Introduction and Business Value (UK)Denodo
Watch full webinar here: https://bit.ly/30mHuYH
What started to evolve as the most agile and real-time enterprise data fabric, data virtualization is proving to go beyond its initial promise and is becoming one of the most important enterprise big data fabrics. Denodo’s vision is to provide a unified data delivery layer as a logical data fabric, to bridge the gap between the IT and the business, hiding the underlying complexity and creating a semantic layer to expose data in a business friendly manner.
Attend this webinar to learn:
- What data virtualization really is
- How it differs from other enterprise data integration technologies
- Why data virtualization is finding enterprise-wide deployment inside some of the largest organizations
- Business Value of data virtualization and customer use cases
- Highlights of the newly launched Denodo Platform 8.0
Similar to TUW-ASE Summer 2015: Data marketplaces: core models and concepts (20)
Integrated Analytics for IIoT Predictive Maintenance using IoT Big Data Cloud...Hong-Linh Truong
For predictive maintenance of equipment with In-
dustrial Internet of Things (IIoT) technologies, existing IoT Cloud
systems provide strong monitoring and data analysis capabilities
for detecting and predicting status of equipment. However, we
need to support complex interactions among different software
components and human activities to provide an integrated analyt-
ics, as software algorithms alone cannot deal with the complexity
and scale of data collection and analysis and the diversity of
equipment, due to the difficulties of capturing and modeling
uncertainties and domain knowledge in predictive maintenance.
In this paper, we describe how we design and augment complex
IoT big data cloud systems for integrated analytics of IIoT
predictive maintenance. Our approach is to identify various
complex interactions for solving system incidents together with
relevant critical analytics results about equipment. We incorpo-
rate humans into various parts of complex IoT Cloud systems
to enable situational data collection, services management, and
data analytics. We leverage serverless functions, cloud services,
and domain knowledge to support dynamic interactions between
human and software for maintaining equipment. We use a real-
world maintenance of Base Transceiver Stations to illustrate our
engineering approach which we have prototyped with state-of-
the art cloud and IoT technologies, such as Apache Nifi, Hadoop,
Spark and Google Cloud Functions.
Modeling and Provisioning IoT Cloud Systems for Testing UncertaintiesHong-Linh Truong
Modern Cyber-Physical Systems (CPS) and Internet of Things (IoT)
systems consist of both loosely and tightly interactions among
various resources in IoT networks, edge servers and cloud data
centers. These elements are being built atop virtualization layers
and deployed in both edge and cloud infrastructures. They also deal
with a lot of data through the interconnection of different types of
networks and services. Therefore, several new types of uncertainties
are emerging, such as data, actuation, and elasticity uncertainties.
This triggers several challenges for testing uncertainty in such
systems. However, there is a lack of novel ways to model and
prepare the right infrastructural elements covering requirements
for testing emerging uncertainties. In this paper, first we present
techniques for modeling CPS/IoT Systems and their uncertainties
to be tested. Second, we introduce techniques for determining and
generating deployment configuration for testing in different IoT
and cloud infrastructures. We illustrate our work with a real-world
use case for monitoring and analysis of Base Transceiver Stations.
Testing Uncertainty of Cyber-Physical Systems in IoT Cloud Infrastructures: C...Hong-Linh Truong
Today’s cyber-physical systems (CPS) span IoT and cloud-based
datacenter infrastructures, which are highly heterogeneous with
various types of uncertainty. Thus, testing uncertainties in these
CPS is a challenging and multidisciplinary activity. We need several
tools for modeling, deployment, control, and analytics to test and
evaluate uncertainties for different configurations of the same CPS.
In this paper, we explain why using state-of-the art model-driven
engineering (MDE) and model-based testing (MBT) tools is not
adequate for testing uncertainties of CPS in IoT Cloud infrastruc-
tures. We discus how to combine them with techniques for elastic
execution to dynamically provision both CPS under test and testing
utilities to perform tests in various IoT Cloud infrastructures.
Towards a Resource Slice Interoperability Hub for IoTHong-Linh Truong
Interoperability for IoT is a challenging problem
because it requires us to tackle (i) cross-system interoperability
issues at the IoT platform sides as well as relevant network
functions and clouds in the edge systems and data centers
and (ii) cross-layer interoperability, e.g., w.r.t. data formats,
communication protocols, data delivery mechanisms, and perfor-
mance. However, existing solutions are quite static w.r.t software
deployment and provisioning for interoperability. Many middle-
ware, services and platforms have been built and deployed as
interoperability bridges but they are not dynamically provisioned
and reconfigured for interoperability at runtime. Furthermore,
they are often not considered together with other services as a
whole in application-specific contexts. In this paper, we focus
on dynamic aspects by introducing the concept of Resource
Slice Interoperability Hub (rsiHub). Our approach leverages
existing software artifacts and services for interoperability to
create and provision dynamic resource slices, including IoT,
network functions and clouds, for addressing application-specific
interoperability requirements. We will present our key concepts,
architectures and examples toward the realization of rsiHub.
On Supporting Contract-aware IoT Dataspace ServicesHong-Linh Truong
Advances in the Internet of Things (IoT) enable a
huge number of connected devices that produce large amounts
of data. Such data is increasingly shared among various
stakeholders to support advanced (predictive) analytics and
precision decision making in different application domains like
smart cities and industrial internet. Currently there are several
platforms that facilitate sharing, buying and selling IoT data.
However, these platforms do not support the establishment and
monitoring of usage contracts for IoT data. In this paper we
address this research issue by introducing a new extensible
platform for enabling contract-aware IoT dataspace services,
which supports data contract specification and IoT data flow
monitoring based on established data contracts. We present
a general architecture of contract monitoring services for
IoT dataspaces and evaluate our platform through illustrative
examples with real-world datasets and through performance
analysis.
Towards the Realization of Multi-dimensional Elasticity for Distributed Cloud...Hong-Linh Truong
As multiple types of distributed, heterogeneous cloud computing environments have proliferated, cloud software can leverage
diverse types of infrastructural, platform and data resources with di
erent cost and quality models. This introduces a multi-
dimensional elasticity perspective for cloud software that would greatly meet changing demands from the user. However, we argue
that current techniques are not enough for dealing with multi-dimensional elasticity in distributed cloud environments. We present
our approach to the realization of multi-dimensional elasticity by introducing novel concepts and a roadmap to achieve them.
On Engineering Analytics of Elastic IoT Cloud SystemsHong-Linh Truong
Developing IoT cloud platforms is very challenging, as IoT
cloud platforms consist of a mix of cloud services and IoT elements, e.g.,
for sensor management, near-realtime events handling, and data analyt-
ics. Developers need several tools for deployment, control, governance
and analytics actions to test and evaluate designs of software compo-
nents and optimize the operation of di erent design con gurations. In
this paper, we describe requirements and our techniques on support-
ing the development and testing of IoT cloud platforms. We present our
choices of tools and engineering actions that help the developer to design,
test and evaluate IoT cloud platforms in multi-cloud environments.
HINC – Harmonizing Diverse Resource Information Across IoT, Network Functions...Hong-Linh Truong
Effective resource management in IoT systems must
represent IoT resources, edge-to-cloud network capabilities, and
cloud resources at a high-level, while being able to link to diverse
low-level types of IoT devices, network functions, and cloud
computing infrastructures. Hence resource management in such
a context demands a highly distributed and extensible approach,
which allows us to integrate and provision IoT, network functions,
and cloud resources from various providers. In this paper, we
address this crucial research issue. We first present a high-
level information model for virtualized IoT, network functions
and cloud resource modeling, which also incorporates software-
defined gateways, network slicing and data centers. This model
is used to glue various low-level resource models from different
types of infrastructures in a distributed manner to capture
sets of resources spanning across different sub-networks. We
then develop a set of utilities and a middleware to support
the integration of information about distributed resources from
various sources. We present a proof of concept prototype with
various experiments to illustrate how various tasks in IoT cloud
systems can be simplified as well as to evaluate the performance
of our framework.
SINC – An Information-Centric Approach for End-to-End IoT Cloud Resource Prov...Hong-Linh Truong
We present SINC –
Slicing IoT, Network Functions, and Clouds – which enables designers to dynamically create/update end-to-end slices of the overall IoT network in order to simultaneously meet multiple user needs.
Governing Elastic IoT Cloud Systems under UncertaintiesHong-Linh Truong
we introduce U-GovOps – a novel framework for
dynamic, on-demand governance of elastic IoT cloud systems under
uncertainty. We introduce a declarative policy language to simplify
the development of uncertainty- and elasticity-aware governance
strategies. Based on that we develop runtime mechanisms, which
enable mitigating the uncertainties by monitoring and governing
the IoT cloud systems through specified strategies.
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
How to Make a Field invisible in Odoo 17Celine George
It is possible to hide or invisible some fields in odoo. Commonly using “invisible” attribute in the field definition to invisible the fields. This slide will show how to make a field invisible in odoo 17.
Unit 8 - Information and Communication Technology (Paper I).pdfThiyagu K
This slides describes the basic concepts of ICT, basics of Email, Emerging Technology and Digital Initiatives in Education. This presentations aligns with the UGC Paper I syllabus.
Operation “Blue Star” is the only event in the history of Independent India where the state went into war with its own people. Even after about 40 years it is not clear if it was culmination of states anger over people of the region, a political game of power or start of dictatorial chapter in the democratic setup.
The people of Punjab felt alienated from main stream due to denial of their just demands during a long democratic struggle since independence. As it happen all over the word, it led to militant struggle with great loss of lives of military, police and civilian personnel. Killing of Indira Gandhi and massacre of innocent Sikhs in Delhi and other India cities was also associated with this movement.
Palestine last event orientationfvgnh .pptxRaedMohamed3
An EFL lesson about the current events in Palestine. It is intended to be for intermediate students who wish to increase their listening skills through a short lesson in power point.
Francesca Gottschalk - How can education support child empowerment.pptxEduSkills OECD
Francesca Gottschalk from the OECD’s Centre for Educational Research and Innovation presents at the Ask an Expert Webinar: How can education support child empowerment?
Embracing GenAI - A Strategic ImperativePeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
The French Revolution Class 9 Study Material pdf free download
TUW-ASE Summer 2015: Data marketplaces: core models and concepts
1. Data marketplaces: core models and
concepts
Hong-Linh Truong
Distributed Systems Group,
Vienna University of Technology
truong@dsg.tuwien.ac.at
http://dsg.tuwien.ac.at/staff/truong
1ASE Summer 2015
Advanced Services Engineering,
Summer 2015, Lecture 6
Advanced Services Engineering,
Summer 2015, Lecture 6
2. Outline
Data marketplaces
Description models
Data agreement exchange models and
architectures
Data contract model and evaluation
ASE Summer 2015 2
3. Data service unitData service unit
3
Recall – data service units in
clouds/internet
datadata
Internet/CloudInternet/Cloud
Data service unitData service unit
People
data
Data service unitData service unit
Things
ASE Summer 2015
data data
4. Data-as-a-Service – service modelsData-as-a-Service – service models
Recall – data as a service
ASE Summer 2015 4
Storage-as-a-Service
(Basic storage functions)
Storage-as-a-Service
(Basic storage functions)
Database-as-a-Service
(Structured/non-structured
querying systems)
Database-as-a-Service
(Structured/non-structured
querying systems)
Data publish/subcription
middleware as a service
Data publish/subcription
middleware as a service
Sensor-as-a-ServiceSensor-as-a-Service
Private/Public/Hybrid/Community CloudsPrivate/Public/Hybrid/Community Clouds
deploy
5. Data platform or marketplace?
ASE Summer 2015 5
http://www.guavus.com/platform/
http://datamarket.azure.com/browse/data
6. Data marketplaces
More than just DaaS
DaaS focuses on data provisioning features
Stakeholders in data marketplaces
Multiple data providers and consumers
Marketplace providers
Marketplace authorities
Analytics providers
Data transportation providers
Billing and payment providers
ASE Summer 2015 6
7. Example of stakeholders
ASE Summer 2015 7
Questions: specific data market (Tokyo Tsukiji) or generic data
market (Donau Zentrum)
Tien-Dung Cao, Quang-Hieu
Vu, Duc-Hung Le, Hong-Linh
Truong, Schahram Dustdar:
MARSA: A Marketplace for
Realtime Human-Sensing Data.
On submission.
http://dungcao.github.io/marsa/
Tien-Dung Cao, Quang-Hieu
Vu, Duc-Hung Le, Hong-Linh
Truong, Schahram Dustdar:
MARSA: A Marketplace for
Realtime Human-Sensing Data.
On submission.
http://dungcao.github.io/marsa/
8. Technical services, protocols,
mechanisms in data marketplaces
Multiple DaaS provisioning
Access models and interfaces
Complex interactions among DaaS providers,
data providers, data consumers, marketplace
providers, etc.
Data exchange as well as payment
Complex billing and pricing models
Market dynamics
Service and data contracts
ASE Summer 2015 8
10. Description Model for DaaS (1)
Which levels must be covered?
ASE Summer 2015 10
Data
items
Data
items
Data
items
Data resourceData resource
Data
assets
Data resourceData resource Data resourceData resource
Data resourceData resourceData resourceData resource
Consumer
Consumer
DaaS
Here
11. Description model for DaaS – types
of information
Which types of information must be covered?
ASE Summer 2015 11
Quality of
data
Ownership
Price
License ....
Service
interface
Service
license
Quality of
service ....
12. DEMOS – a description model for
Data-as-a-Service
ASE Summer 2015 12
See prototype:
http://www.infosys.tuwien.ac.at/
prototype/SOD1/demods/
Quang Hieu Vu, Tran Vu Pham, Hong
Linh Truong,, Schahram Dustdar,
Rasool Asal: DEMODS: A Description
Model for Data-as-a-Service. AINA
2012: 605-612
Quang Hieu Vu, Tran Vu Pham, Hong
Linh Truong,, Schahram Dustdar,
Rasool Asal: DEMODS: A Description
Model for Data-as-a-Service. AINA
2012: 605-612
14. Exchange data agreement (1)
ASE Summer 2015 14
DaaS
Consumer
DaaS
Sensor
DaaS
Consumer DaaS provider Data
provider
How do they interact w.r.t. data concerns?
How do their data agreements look like?
15. Exchange data agreement (2)
Lack of models and protocols for data
agreement in data marketplaces
Constraints for data usage are not clear
Inadequate data/service description → hindering
automatic (near realtime) data selection and
integration
Existing techniques are not adequate for
dynamic data agreement exchange in data
marketplaces
Need generic exchange models suitable for different
ways of data provisioning in data marketplaces
Need generic exchange models suitable for different
ways of data provisioning in data marketplaces
ASE Summer 2015 15
16. Data Agreement Exchange as a
Service (DAES)
Metamodel for data agreement exchange
Techniques for enriching and associating data
assets with agreement terms
Interaction models for data agreement exchange
Hong Linh Truong, Schahram Dustdar, Joachim Götze, Tino Fleuren, Paul Müller, Salah-Eddine Tbahriti, Michael Mrissa,
Chirine Ghedira: Exchanging Data Agreements in the DaaS Model. APSCC 2011: 153-160
Hong Linh Truong, Schahram Dustdar, Joachim Götze, Tino Fleuren, Paul Müller, Salah-Eddine Tbahriti, Michael Mrissa,
Chirine Ghedira: Exchanging Data Agreements in the DaaS Model. APSCC 2011: 153-160
ASE Summer 2015 16
17. Metamodel for data agreements
Different
category of
agreements
Licensing,
privacy, quality
of data
Extensions
Languages
Different types
of agreements
Different
specifications
ASE Summer 2015 17
18. Associating data with data
agreements
Solutions
(a) directly inserting agreements into data assets
(b) providing two-step access to agreements and data
assets
(c) linking data agreements to the description of DaaS
(d) linking data agreements to the message sent by
DaaS
ASE Summer 2015 18
20. DAES – conceptual architecture
Using URIs to identify agreements
ASE Summer 2015 20
21. DAES – managed information
Specific applications: agreement creation, agreement validation,
agreement compatibility analysis, agreement management
Specific applications: agreement creation, agreement validation,
agreement compatibility analysis, agreement management
ASE Summer 2015 21
22. Illustrating examples – insert
agreement into data asset
A pay-per-use consumer uses dataAPI of DaaS
search for data
The consumer pays the use APIs
Each call can return different types of data
Example of
searching people
But a strong consequence
for data service engineering
techniques: dealing with
elastic requirements!
But a strong consequence
for data service engineering
techniques: dealing with
elastic requirements!
ASE Summer 2015 22
23. Illustrating examples – link
agreements to geospatial data
Domain-specific DaaS: different agreements for different data requests
Vector data of geographic features via Web-Feature-Service (WFS)
Terrain elevation data via Web-Coverage Services (WCS)
Domain-specific DaaS: different agreements for different data requests
Vector data of geographic features via Web-Feature-Service (WFS)
Terrain elevation data via Web-Coverage Services (WCS)
ASE Summer 2015 23
24. Illustrating examples – link
agreements to geospatial data
Software can interpret and
reason if the data can be
used for specific purposes
Software can interpret and
reason if the data can be
used for specific purposes
ASE Summer 2015 24
26. Illustrative examples – develop an
app for policy compliance (2)
Configuration
Results
ASE Summer 2015 26
27. HOW DOES NEAR-REALTIME DATA IMPACT
ON DATA AGREEMENT EXCHANGE?
Discussion time
ASE Summer 2015 27
28. Data contract
How to specific data contract?
ASE Summer 2015 28
Data
items
Data
items
Data
items
Data resourceData resource
Data
assets
Data resourceData resource Data resourceData resource
Data resourceData resourceData resourceData resource
Consumer
Consumer
DaaS
29. Data contracts
Give a clear information about data usage
Have a remedy against the consumer where the
circumstances are such that the acts complained
of do not
Limit the liability of data providers in case of
failure of the provided data;
Specify information on data delivery,
acceptance, and payment
29ASE Summer 2015
30. 30
Data contracts
Well-researched contracts for services but not
for DaaS and data marketplaces
But service APIs != data APIs =! data assets
Several open questions
Right to use data? Quality of data in the data
agreement? Search based on data contract? Etc.
➔ Require extensible models
➔ Capture contractual terms for data contracts
➔ Support (semi-)automatic data service/data selection
techniques.
➔ Require extensible models
➔ Capture contractual terms for data contracts
➔ Support (semi-)automatic data service/data selection
techniques.
Hong-Linh Truong, Marco Comerio, Flavio De Paoli, G.R. Gangadharan, Schahram Dustdar, "Data Contracts for
Cloud-based Data Marketplaces ", International Journal of Computational Science and Engineering, 2012 Vol.7, No.4,
pp.280 - 295
Hong-Linh Truong, Marco Comerio, Flavio De Paoli, G.R. Gangadharan, Schahram Dustdar, "Data Contracts for
Cloud-based Data Marketplaces ", International Journal of Computational Science and Engineering, 2012 Vol.7, No.4,
pp.280 - 295
ASE Summer 2015
31. Study of main data contract terms
Data rights
Derivation, Collection, Reproduction, Attribution
Quality of Data (QoD)
Not mentioned, Not clear how to establish QoD metrics
Regulatory Compliance
Sarbanes-Oxley, EU data protection directive, etc.
Pricing model
Different models, pricing for data APIs and for data assets
Control and Relationship
Evolution terms, support terms, limitation of liability, etc
31
Most information is in human-readable formMost information is in human-readable form
ASE Summer 2015
32. 32
Data contract study
ASE Summer 2015
Hong-Linh Truong, Marco Comerio, Flavio De Paoli, G.R. Gangadharan, Schahram Dustdar, "Data Contracts for
Cloud-based Data Marketplaces ", International Journal of Computational Science and Engineering, 2012 Vol.7, No.4,
pp.280 - 295
Hong-Linh Truong, Marco Comerio, Flavio De Paoli, G.R. Gangadharan, Schahram Dustdar, "Data Contracts for
Cloud-based Data Marketplaces ", International Journal of Computational Science and Engineering, 2012 Vol.7, No.4,
pp.280 - 295
33. 33
Developing data contracts in cloud-
based data marketplaces
Follow community-based approach for data
contract
Propose generic structures to represent data
contract terms and abstract data contracts
Develop frameworks for data contract applications
Incorporate data contracts into data-as-a-service
description
Develop data contract applications
ASE Summer 2015
34. 34
Community view on data contract
development
Community users can develop:
Term categories, term names, values, and units
Rules for data contracts
Common contract and contract fragments
Community users =!
novice users
ASE Summer 2015
35. 35
Representing data contract terms
Contract term: (termName,termValue)
Term name: common terms or user-specific terms
Term value: a single value, a set, or a range
ASE Summer 2015
36. 36
Structuring abstract data contracts
Concrete data contracts can be in
RDF, XML or JSON
generates
Use Identifiers and
Tags for identifying
and searches
ASE Summer 2015
37. 37
Development of contract
applications
Main applications:
Data contract compatibility evaluation, data contract
composition
Some common steps
Extract DCTermType in TermCategoryType
Extact comprable terms from all contracts,
- e.g., dataRight: Derivation, Composition and Reproduction
Use evaluation rules associated with DCTermType
from rule repositories
Execute rules by passing comparable terms to rules
Aggregate results
ASE Summer 2015
38. Evaluating Data Contracts
Goal
Check the quality and reputation of a data contract
We can check data contracts using quality of
data metrics
Timeliness, Completeness, Reputation, Consistency
metrics
Examples
Free-per-use but cost = 100EUR
Missing „data accuracy“ concern
ASE Summer 2015 38
39. Data Contract Compatibility
Goal
If multiple data contracts are compatible with the
consumer needs
The consumer requires multiple data associated with
different contracts
Contract compatibility
Matching contract terms
Evaluating contract term compatibility and
completeness w.r.t. application needs
Making decision in using data
ASE Summer 2015 39
41. Conceptual architecture for contract
management and evaluation
Prototype
RDF for representing term categories,
term names, term values, units
Allegro Graph for storing contract
knowledge
ASE Summer 2015 41
42. 42
Illustrating examples
A large sustainability monitoring data platform
shows how green buildings are
Real-time total and per capita of CO2 emission
of monitored building
Open government data about CO2 per capita at
national level
We created contracts from
Open Data Commons Attribution License
Open Government License
ASE Summer 2015
44. 44
Step 2: provide OpenBuildingCO2
OpenBuildingCO2 by
modifying quality of
data and data right
OpenBuildingCO2 by
modifying quality of
data and data right
OpenGov for
government data
OpenGov for
government data
Data contract for green building dataData contract for green building data
ASE Summer 2015
48. MARSA Description for Human-
sensing data marketplace
ASE Summer 2015 48
Tien-Dung Cao, Quang-Hieu
Vu, Duc-Hung Le, Hong-Linh
Truong, Schahram Dustdar:
MARSA: A Marketplace for
Realtime Human-Sensing Data.
On submission.
http://dungcao.github.io/marsa/
Tien-Dung Cao, Quang-Hieu
Vu, Duc-Hung Le, Hong-Linh
Truong, Schahram Dustdar:
MARSA: A Marketplace for
Realtime Human-Sensing Data.
On submission.
http://dungcao.github.io/marsa/
49. MARSA
ASE Summer 2015 49
Tien-Dung Cao, Quang-Hieu
Vu, Duc-Hung Le, Hong-Linh
Truong, Schahram Dustdar:
MARSA: A Marketplace for
Realtime Human-Sensing Data.
On submission.
http://dungcao.github.io/marsa/
Tien-Dung Cao, Quang-Hieu
Vu, Duc-Hung Le, Hong-Linh
Truong, Schahram Dustdar:
MARSA: A Marketplace for
Realtime Human-Sensing Data.
On submission.
http://dungcao.github.io/marsa/
51. Data Market without Marketplace?
ASE Summer 2015 51
Dominic Wörner and Thomas von
Bomhard. 2014. When your sensor
earns money: exchanging data for
cash with Bitcoin. In Proceedings of the
2014 ACM International Joint Conference
on Pervasive and Ubiquitous Computing:
Adjunct Publication (UbiComp '14
Adjunct). ACM, New York, NY, USA, 295-
298.
Dominic Wörner and Thomas von
Bomhard. 2014. When your sensor
earns money: exchanging data for
cash with Bitcoin. In Proceedings of the
2014 ACM International Joint Conference
on Pervasive and Ubiquitous Computing:
Adjunct Publication (UbiComp '14
Adjunct). ACM, New York, NY, USA, 295-
298.
Kay Noyen, Dirk Volland, Dominic
Wörner, Elgar Fleisch:
When Money Learns to Fly: Towards
Sensing as a Service Applications Using
Bitcoin.
Kay Noyen, Dirk Volland, Dominic
Wörner, Elgar Fleisch:
When Money Learns to Fly: Towards
Sensing as a Service Applications Using
Bitcoin.
But what about data contract?
52. Exercises
Read mentioned papers
Examine existing data marketplaces and write
DEMODS-based specification for some of them
Develop some specific data contracts for open
government data
Work on some algorithms for checking data
contract compatiblity
ASE Summer 2015 52
53. 53
Thanks for
your attention
Hong-Linh Truong
Distributed Systems Group
Vienna University of Technology
truong@dsg.tuwien.ac.at
http://dsg.tuwien.ac.at/staff/truong
ASE Summer 2015