suggesting a way to manage linked data platform to be used domain specific applications
Best parper awarded - http://www.iaria.org/conferences2015/AwardsALLDATA15.html
- About the importance of Linked Data technologies to make real/practical of the benefits of Open Data.
- Linked Data, it's all about open data quality and Implementation of the applications/services
OpenAIRE-Advance: Advancing Open Scholarship (Presentation at RDA 11th Plenary)OpenAIRE
Presentation by Natalia Manola, OpenAIRE Director, at RDA 11th Plenary BoF meeting - EOSC-related European Projects getting Global: Engaging with the RDA
Conference Opening Science to Meet Future Challenges, Warsaw, March 11, 2014, organized by Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw.
Aleksandar Kapisoda: The semantic approach for tracking scientific publicationsSemantic Web Company
The document discusses Boehringer Ingelheim Pharma's development of a publication tracking system using semantic technologies. It aims to automatically import publication data, perform data curation, and enable advanced visualization and analysis. Some key challenges include cleaning noisy author and institution data, adding internal BI data, and linking to external impact factors. The system utilizes tools like PoolParty, Virtuoso, and SPARQL to semantically enrich and link publication data. It is meant to provide advanced analytics beyond what was possible in their previous manually curated system.
Presentation given at EuropeanaTech 2018 in Rotterdam, The Netherlands. Provides a summary of insights gained from working for about a decade on challenges related to temporal aspects of the web, persistence.
Perseverance on persistence by Herbert Van de Sompel - EuropeanaTech Conferen...Europeana
The document discusses strategies for maintaining links and content over time on the web (link rot and content drift). It proposes using persistent identifiers (PIDs) assigned to web resources and versions to combat these issues. PIDs allow the "web of the present" to remain the same as the "web of the past" by redirecting links over time. The document provides examples of communities like scholarly communication that care about maintaining the integrity of web archives and references over long periods. It also discusses the use of relation types like "cite-as" to convey a preferred URI for references that may change.
The document introduces ALIADA, an open source tool that automatically converts and publishes library and museum metadata as Linked Data. ALIADA takes metadata from library and museum collection management systems in MARCXML or LIDOXML format, converts it to RDF according to the ALIADA ontology based on FRBRoo, and links it to other datasets. It then publishes the Linked Data and makes it available through a SPARQL endpoint. The goal of ALIADA is to help libraries and museums easily share their metadata as Linked Open Data without having to learn complex Semantic Web technologies.
- About the importance of Linked Data technologies to make real/practical of the benefits of Open Data.
- Linked Data, it's all about open data quality and Implementation of the applications/services
OpenAIRE-Advance: Advancing Open Scholarship (Presentation at RDA 11th Plenary)OpenAIRE
Presentation by Natalia Manola, OpenAIRE Director, at RDA 11th Plenary BoF meeting - EOSC-related European Projects getting Global: Engaging with the RDA
Conference Opening Science to Meet Future Challenges, Warsaw, March 11, 2014, organized by Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw.
Aleksandar Kapisoda: The semantic approach for tracking scientific publicationsSemantic Web Company
The document discusses Boehringer Ingelheim Pharma's development of a publication tracking system using semantic technologies. It aims to automatically import publication data, perform data curation, and enable advanced visualization and analysis. Some key challenges include cleaning noisy author and institution data, adding internal BI data, and linking to external impact factors. The system utilizes tools like PoolParty, Virtuoso, and SPARQL to semantically enrich and link publication data. It is meant to provide advanced analytics beyond what was possible in their previous manually curated system.
Presentation given at EuropeanaTech 2018 in Rotterdam, The Netherlands. Provides a summary of insights gained from working for about a decade on challenges related to temporal aspects of the web, persistence.
Perseverance on persistence by Herbert Van de Sompel - EuropeanaTech Conferen...Europeana
The document discusses strategies for maintaining links and content over time on the web (link rot and content drift). It proposes using persistent identifiers (PIDs) assigned to web resources and versions to combat these issues. PIDs allow the "web of the present" to remain the same as the "web of the past" by redirecting links over time. The document provides examples of communities like scholarly communication that care about maintaining the integrity of web archives and references over long periods. It also discusses the use of relation types like "cite-as" to convey a preferred URI for references that may change.
The document introduces ALIADA, an open source tool that automatically converts and publishes library and museum metadata as Linked Data. ALIADA takes metadata from library and museum collection management systems in MARCXML or LIDOXML format, converts it to RDF according to the ALIADA ontology based on FRBRoo, and links it to other datasets. It then publishes the Linked Data and makes it available through a SPARQL endpoint. The goal of ALIADA is to help libraries and museums easily share their metadata as Linked Open Data without having to learn complex Semantic Web technologies.
Smart Data Applications powered by the Wikidata Knowledge GraphPeter Haase
This document discusses Wikidata and how it can power smart data applications. Wikidata is a large, structured, collaborative knowledge graph containing over 15 million entities. It collects data in a structured form from Wikipedia pages and can be queried like a database using the Wikidata Query Service. The document promotes metaphacts, an enterprise knowledge graph platform that can be used to build applications using Wikidata, enrich Wikidata with private data, and enable companies to build and leverage their own knowledge graphs for various domains such as cultural heritage and pharma.
The document summarizes a CrossRef workshop held in Gauteng, South Africa in September 2015. It introduced CrossRef and its role in managing digital object identifiers (DOIs) to uniquely identify scholarly works, enabling linking between references and cited works. It outlined CrossRef's history and services, including DOI registration and metadata deposit, and encouraging long-term archiving of scholarly works. The document also reviewed CrossRef participation benefits and growing statistics on registered DOIs and annual clicks to publisher sites.
Large corporations have to master vast amounts of heterogeneous data in order to stay competitive. While existing approaches have attempted to consolidate and manage the data by forcing it into a single shared data model, data lakes recently emerged that instead provide a central storage point for holding all data sets in their original form. In this talk, we present eccenca Corporate Memory, which extends the data lake paradigm with a semantic integration layer for managing diverse, but semantically enriched data. In addition to that, we depict our vision for public / private data co-evolution and how we research this topic in the joint project Linked Enterprise Data Services (LEDS) together with the University of Leipzig and other partners.
from René Pietzsch | Head of Product Management, Eccenca
and Dr. Michael Martin | AKSW, Universität Leipzig, LEDS Project
Presentation at Sachsentag der Angewandten informatik 2016 in Leipzig in the context with the results of the LEDS project
2015 GIS in Colorado Track: Challenges Standardizing and Implementing Metadat...GIS in the Rockies
Having metadata as well as a catalog of all GIS related data, is fundamental to the success of any organization. There is not a right or wrong approach, but any approach is challenging. This is how Denver Water approached the problem and found a solution.
Enabling combined Software and Data engineering at Web-scaleMonika Solanki
The ALIGNED suite of ontologies was developed to integrate software and data engineering processes for web-scale systems. It provides semantic models to describe system specifications, lifecycles, and quality reports. The suite aims to improve productivity, agility, and quality for data-intensive applications. It was deployed and evaluated using four large-scale use cases, including a legal information platform that combines over 1 million documents.
II-SDV 2016 Patrick Beaucamp - Data Science with R and Vanilla AirDr. Haxel Consult
Companies are facing challenges to analyze Big data databases & data Lakes in uncertain technologies environment, in order to provide accurate analysis and build forecast model.
In a context of budgdet constraints, the R project is a reliable alternative to legacy commmercial software to develop and deploy business analytics data model. R has a worldwide recognation and fast adoption from companies everywhere in the world. Together with Vanilla Air, everybody can start now a Data Science project and share his analysis in an instant.
This session comes along with a presentation of the Vanilla Air, a new "cloud / on premise" Data Science platform to develop & deploy R data model at enterprise level
Advanced Topics in OpenAPI: Added Value Services and Protection in the OpenTr...🧑💻 Manuel Coppotelli
The objectives of this work were to study a series of advanced aspects that an organization can consider when expose data through an OpenService.
I studied the problems relative to the implementation of Added Value Services using the information exposed through an OpenAPI, in particular a complex route planner that combines both timetables and real-time data on the public transport.
The exposed information can also be used by a byzantine user to infer whether a service provider is respecting the terms of its SLA.
Obviously an organization do not want to expose data that would allow to infer this kind of information; therefore arises the problem of studying what is the right tradeoff that allows to have a sort of protection but, a the same time, maintain the openness of the data.
The solution studied for this work have been applied to the real case of OpenTrasporti (a project by the Italian Ministry for Transportation and Infrastructures)
The document discusses infrastructure for learning analytics. It notes that organizations with centralized student data will have a competitive advantage over those without through improved learning analytics services. It outlines the University of Oxford's aim to become a world-leading center for learning analytics research and ensure effective translation of research into business improvements. Finally, it discusses standards, tools and initiatives that can help build scalable learning analytics infrastructure, including the xAPI, LTI, OLA and JISC frameworks.
The document describes an ontology-driven approach to integrating software engineering data from different tools like Confluence and JIRA. It presents a case study applying this approach to requirements data from the PoolParty Thesaurus project. An evaluation with software engineers found the unified semantic search took half the time to answer questions compared to the original tools. It also revealed limitations in the full text search of the governance dashboard. The approach aims to consolidate project experiences and requirements to improve software development.
Adopting linked data principles for accelerating business transformation proc...Quentin Reul
The document discusses Wolters Kluwer's adoption of linked data principles to accelerate their business transformation processes. It describes Wolters Kluwer as an information services company providing tools for legal, tax, finance, and healthcare professionals. It also outlines their shift from print to digital content delivery and their development of a content standard and semantic model using linked data to power use cases like cross-source querying and a semantic integrator.
Advanced Analytics and Machine Learning with Data VirtualizationDenodo
Watch here: https://bit.ly/3719Bi7
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spent most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Attend this webinar and learn:
-How data virtualization can accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- How popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc. integrate with Denodo
- How you can use the Denodo Platform with large data volumes in an efficient way
-About the success McCormick has had as a result of seasoning the Machine Learning and Blockchain Landscape with data virtualization
- Springer, a leading scientific publisher, sought to build a new content platform in 7 months to better serve its corporate customers and double sales to that market within 5 years.
- They used MarkLogic to build the platform due to its ability to handle Springer's large XML assets, develop quickly through MarkLogic performing more tasks, and solve the challenge of granting customized access to content.
- The new platform organized content for corporate customers, offered security and reporting features, and allowed for flexible business models like pay-for-access subscriptions to custom search queries.
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...semanticsconference
The NXTM Project is a research project between a university and IT company aimed at developing technology to analyze unstructured data streams and extract structured information. It involves processing documents through various analysis engines to identify semantics and link related data. The extracted structured data is stored in a database and made searchable through a semantic search engine. Search results are interactively represented as a graph to discover related information. The goal is to help small businesses extract valuable insights from unstructured data sources.
Project Description of the Linked Open Data (LOD) PILOT Austria - presented at the PiLOD event at VU Amsterdam (Netherlands) on 29.01. 2014 (see: http://www.pilod.nl/) by Martin Kaltenböck of Semantic Web Company.
PoolParty Semantic Suite - Solutions for Sustainable DevelopmentMartin Kaltenböck
Presentation of the webinar: PoolParty for Sustainable Development - the Climate Tagger - taking place on 5 November 2015. More information and other presentations to be found here: http://bit.ly/1NpTcGT.
Recording of the webinar: https://www.youtube.com/watch?v=3GxtFfLL1ps.
The document describes an architecture for semantically integrating enterprise data lakes. It proposes a knowledge graph that links metadata, data models and key performance indicators to provide a common meaning for data. Raw data is stored in a data lake and ingested from various sources. A metadata layer captures dataset metadata, ontologies and integration rules to link disparate data. An interface allows users to access consolidated views generated by executing queries on Hadoop. The process involves cataloging, discovering, lifting, linking and validating datasets to integrate them based on rules into the knowledge graph.
Technical Overview and RoadMap of WordLift 2.0, a Semantic Web plug-in for WordPress.
Table of Contents:
* Architecture Overview
* Functional Overview
* WordLift Plug-in
* WordLift Plug-in Dependencies
* WordLift End-Point
* WordLift Client-Side Protocol
* WordLift Server and Procotol
* Development Environment
* RoadMap
* Current State
* Product Backlog
* Hands-on
* Have Fun!
* Want to know more?
The Dendro research data management platform: Applying ontologies to long-ter...João Rocha da Silva
It has been shown that data management should start as early as possible in the research workflow to minimize the risks of data loss. Given the large numbers of datasets produced every day, curators may be unable to describe them all, so researchers should take an active part in the process. However, since they are not data management experts, they must be provided with user-friendly but powerful tools to capture the context information necessary for others to interpret and reuse their datasets. In this paper, we present Dendro, a fully ontology-based collaborative platform for research data management. Its graph data model innovates in the sense that it allows domain-specific lightweight ontologies to be used in resource description, acting as a staging area for later deposit in long-term preservation solutions.
This document outlines challenges and a proposed architecture for connecting government systems across central, state and local levels. It discusses problems with current paper-based and siloed systems, and proposes a shared architecture with common services, applications, identity management and data. Key elements include citizen and employee portals, centralized workflow, policies and master data, with local customization options. The goal is to move processes from manual to automated while integrating previously disconnected systems and stakeholders in a centralized way.
Falcon - Data Management Platform on Hadoop (Beyond ETL)DataWorks Summit
Hadoop and its ecosystem of products have made storing and processing massive amounts of data common place. This has enabled numerous businesses to gain valuable foresights that they never could have in the past. While it is easy to leverage Hadoop for crunching large volumes of data, organizing data, managing life cycle of data and processing data is fairly involved. This is solved adequately well in a traditional data platform involving data warehouses and standard ETL (extract-transform-load) tools, but remains largely unsolved today. Besides data processing complexities, Hadoop presents new set of challenges relating to management of data. Data Management on Hadoop encompasses data motion (import/export), process orchestration (data pipelines, late/re-processing, scheduling), lifecycle management (retention, replication, DR, anonymization, archival), data discovery (data classification, Lineage), etc. among other concerns that are beyond ETL. The presentation focuses on a new data processing and management platform for Hadoop, Falcon that attempts to solve this problem by leveraging existing stacks in the Hadoop ecosystem. Falcon has been in production for nearly a year at InMobi and has been managing hundreds of feeds and processes.
This document provides an overview of a presentation given by WSO2 on their platform. The agenda includes discussing WSO2's company overview, platform, Carbon architecture, cloud computing, big data analytics, API management, mobile, IoT, and customer use cases. It describes WSO2's vision of being a 100% open source middleware platform and global corporation. It also summarizes some of WSO2's products, Carbon middleware platform, private PaaS architecture, App Factory, data analytics capabilities, IoT/device management, and API management platform. It highlights three customer use cases - eBay using WSO2 for a scalable middleware platform, Boeing using it for an integrated platform, and StubHub using it for
Smart Data Applications powered by the Wikidata Knowledge GraphPeter Haase
This document discusses Wikidata and how it can power smart data applications. Wikidata is a large, structured, collaborative knowledge graph containing over 15 million entities. It collects data in a structured form from Wikipedia pages and can be queried like a database using the Wikidata Query Service. The document promotes metaphacts, an enterprise knowledge graph platform that can be used to build applications using Wikidata, enrich Wikidata with private data, and enable companies to build and leverage their own knowledge graphs for various domains such as cultural heritage and pharma.
The document summarizes a CrossRef workshop held in Gauteng, South Africa in September 2015. It introduced CrossRef and its role in managing digital object identifiers (DOIs) to uniquely identify scholarly works, enabling linking between references and cited works. It outlined CrossRef's history and services, including DOI registration and metadata deposit, and encouraging long-term archiving of scholarly works. The document also reviewed CrossRef participation benefits and growing statistics on registered DOIs and annual clicks to publisher sites.
Large corporations have to master vast amounts of heterogeneous data in order to stay competitive. While existing approaches have attempted to consolidate and manage the data by forcing it into a single shared data model, data lakes recently emerged that instead provide a central storage point for holding all data sets in their original form. In this talk, we present eccenca Corporate Memory, which extends the data lake paradigm with a semantic integration layer for managing diverse, but semantically enriched data. In addition to that, we depict our vision for public / private data co-evolution and how we research this topic in the joint project Linked Enterprise Data Services (LEDS) together with the University of Leipzig and other partners.
from René Pietzsch | Head of Product Management, Eccenca
and Dr. Michael Martin | AKSW, Universität Leipzig, LEDS Project
Presentation at Sachsentag der Angewandten informatik 2016 in Leipzig in the context with the results of the LEDS project
2015 GIS in Colorado Track: Challenges Standardizing and Implementing Metadat...GIS in the Rockies
Having metadata as well as a catalog of all GIS related data, is fundamental to the success of any organization. There is not a right or wrong approach, but any approach is challenging. This is how Denver Water approached the problem and found a solution.
Enabling combined Software and Data engineering at Web-scaleMonika Solanki
The ALIGNED suite of ontologies was developed to integrate software and data engineering processes for web-scale systems. It provides semantic models to describe system specifications, lifecycles, and quality reports. The suite aims to improve productivity, agility, and quality for data-intensive applications. It was deployed and evaluated using four large-scale use cases, including a legal information platform that combines over 1 million documents.
II-SDV 2016 Patrick Beaucamp - Data Science with R and Vanilla AirDr. Haxel Consult
Companies are facing challenges to analyze Big data databases & data Lakes in uncertain technologies environment, in order to provide accurate analysis and build forecast model.
In a context of budgdet constraints, the R project is a reliable alternative to legacy commmercial software to develop and deploy business analytics data model. R has a worldwide recognation and fast adoption from companies everywhere in the world. Together with Vanilla Air, everybody can start now a Data Science project and share his analysis in an instant.
This session comes along with a presentation of the Vanilla Air, a new "cloud / on premise" Data Science platform to develop & deploy R data model at enterprise level
Advanced Topics in OpenAPI: Added Value Services and Protection in the OpenTr...🧑💻 Manuel Coppotelli
The objectives of this work were to study a series of advanced aspects that an organization can consider when expose data through an OpenService.
I studied the problems relative to the implementation of Added Value Services using the information exposed through an OpenAPI, in particular a complex route planner that combines both timetables and real-time data on the public transport.
The exposed information can also be used by a byzantine user to infer whether a service provider is respecting the terms of its SLA.
Obviously an organization do not want to expose data that would allow to infer this kind of information; therefore arises the problem of studying what is the right tradeoff that allows to have a sort of protection but, a the same time, maintain the openness of the data.
The solution studied for this work have been applied to the real case of OpenTrasporti (a project by the Italian Ministry for Transportation and Infrastructures)
The document discusses infrastructure for learning analytics. It notes that organizations with centralized student data will have a competitive advantage over those without through improved learning analytics services. It outlines the University of Oxford's aim to become a world-leading center for learning analytics research and ensure effective translation of research into business improvements. Finally, it discusses standards, tools and initiatives that can help build scalable learning analytics infrastructure, including the xAPI, LTI, OLA and JISC frameworks.
The document describes an ontology-driven approach to integrating software engineering data from different tools like Confluence and JIRA. It presents a case study applying this approach to requirements data from the PoolParty Thesaurus project. An evaluation with software engineers found the unified semantic search took half the time to answer questions compared to the original tools. It also revealed limitations in the full text search of the governance dashboard. The approach aims to consolidate project experiences and requirements to improve software development.
Adopting linked data principles for accelerating business transformation proc...Quentin Reul
The document discusses Wolters Kluwer's adoption of linked data principles to accelerate their business transformation processes. It describes Wolters Kluwer as an information services company providing tools for legal, tax, finance, and healthcare professionals. It also outlines their shift from print to digital content delivery and their development of a content standard and semantic model using linked data to power use cases like cross-source querying and a semantic integrator.
Advanced Analytics and Machine Learning with Data VirtualizationDenodo
Watch here: https://bit.ly/3719Bi7
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spent most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Attend this webinar and learn:
-How data virtualization can accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- How popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc. integrate with Denodo
- How you can use the Denodo Platform with large data volumes in an efficient way
-About the success McCormick has had as a result of seasoning the Machine Learning and Blockchain Landscape with data virtualization
- Springer, a leading scientific publisher, sought to build a new content platform in 7 months to better serve its corporate customers and double sales to that market within 5 years.
- They used MarkLogic to build the platform due to its ability to handle Springer's large XML assets, develop quickly through MarkLogic performing more tasks, and solve the challenge of granting customized access to content.
- The new platform organized content for corporate customers, offered security and reporting features, and allowed for flexible business models like pay-for-access subscriptions to custom search queries.
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...semanticsconference
The NXTM Project is a research project between a university and IT company aimed at developing technology to analyze unstructured data streams and extract structured information. It involves processing documents through various analysis engines to identify semantics and link related data. The extracted structured data is stored in a database and made searchable through a semantic search engine. Search results are interactively represented as a graph to discover related information. The goal is to help small businesses extract valuable insights from unstructured data sources.
Project Description of the Linked Open Data (LOD) PILOT Austria - presented at the PiLOD event at VU Amsterdam (Netherlands) on 29.01. 2014 (see: http://www.pilod.nl/) by Martin Kaltenböck of Semantic Web Company.
PoolParty Semantic Suite - Solutions for Sustainable DevelopmentMartin Kaltenböck
Presentation of the webinar: PoolParty for Sustainable Development - the Climate Tagger - taking place on 5 November 2015. More information and other presentations to be found here: http://bit.ly/1NpTcGT.
Recording of the webinar: https://www.youtube.com/watch?v=3GxtFfLL1ps.
The document describes an architecture for semantically integrating enterprise data lakes. It proposes a knowledge graph that links metadata, data models and key performance indicators to provide a common meaning for data. Raw data is stored in a data lake and ingested from various sources. A metadata layer captures dataset metadata, ontologies and integration rules to link disparate data. An interface allows users to access consolidated views generated by executing queries on Hadoop. The process involves cataloging, discovering, lifting, linking and validating datasets to integrate them based on rules into the knowledge graph.
Technical Overview and RoadMap of WordLift 2.0, a Semantic Web plug-in for WordPress.
Table of Contents:
* Architecture Overview
* Functional Overview
* WordLift Plug-in
* WordLift Plug-in Dependencies
* WordLift End-Point
* WordLift Client-Side Protocol
* WordLift Server and Procotol
* Development Environment
* RoadMap
* Current State
* Product Backlog
* Hands-on
* Have Fun!
* Want to know more?
The Dendro research data management platform: Applying ontologies to long-ter...João Rocha da Silva
It has been shown that data management should start as early as possible in the research workflow to minimize the risks of data loss. Given the large numbers of datasets produced every day, curators may be unable to describe them all, so researchers should take an active part in the process. However, since they are not data management experts, they must be provided with user-friendly but powerful tools to capture the context information necessary for others to interpret and reuse their datasets. In this paper, we present Dendro, a fully ontology-based collaborative platform for research data management. Its graph data model innovates in the sense that it allows domain-specific lightweight ontologies to be used in resource description, acting as a staging area for later deposit in long-term preservation solutions.
This document outlines challenges and a proposed architecture for connecting government systems across central, state and local levels. It discusses problems with current paper-based and siloed systems, and proposes a shared architecture with common services, applications, identity management and data. Key elements include citizen and employee portals, centralized workflow, policies and master data, with local customization options. The goal is to move processes from manual to automated while integrating previously disconnected systems and stakeholders in a centralized way.
Falcon - Data Management Platform on Hadoop (Beyond ETL)DataWorks Summit
Hadoop and its ecosystem of products have made storing and processing massive amounts of data common place. This has enabled numerous businesses to gain valuable foresights that they never could have in the past. While it is easy to leverage Hadoop for crunching large volumes of data, organizing data, managing life cycle of data and processing data is fairly involved. This is solved adequately well in a traditional data platform involving data warehouses and standard ETL (extract-transform-load) tools, but remains largely unsolved today. Besides data processing complexities, Hadoop presents new set of challenges relating to management of data. Data Management on Hadoop encompasses data motion (import/export), process orchestration (data pipelines, late/re-processing, scheduling), lifecycle management (retention, replication, DR, anonymization, archival), data discovery (data classification, Lineage), etc. among other concerns that are beyond ETL. The presentation focuses on a new data processing and management platform for Hadoop, Falcon that attempts to solve this problem by leveraging existing stacks in the Hadoop ecosystem. Falcon has been in production for nearly a year at InMobi and has been managing hundreds of feeds and processes.
This document provides an overview of a presentation given by WSO2 on their platform. The agenda includes discussing WSO2's company overview, platform, Carbon architecture, cloud computing, big data analytics, API management, mobile, IoT, and customer use cases. It describes WSO2's vision of being a 100% open source middleware platform and global corporation. It also summarizes some of WSO2's products, Carbon middleware platform, private PaaS architecture, App Factory, data analytics capabilities, IoT/device management, and API management platform. It highlights three customer use cases - eBay using WSO2 for a scalable middleware platform, Boeing using it for an integrated platform, and StubHub using it for
This document discusses digital tools in retail spaces and what constitutes innovation. It provides examples of digital in-store technologies that lacked benefits or understanding of retail needs. The document advocates focusing on enhancing the sensory and human aspects of in-store shopping. It outlines five main goals for digital in-store solutions: to show, guide, help sell, entertain customers and provide an interactive directory. The overall message is that digital should complement but not replace the in-store experience.
This document provides definitions and explanations of a data management platform (DMP). It summarizes that a DMP is a centralized system that collects proprietary and third-party data, integrates that data, and applies it to advertising strategies. It allows users to create custom segments, predict volumes, and synchronize segments across sources for advanced analytics and integration with demand-side platforms. The core elements of a DMP include data integration, management capabilities for segmentation, data labeling for flexibility, a self-service interface, and connections to relevant marketing channels. A DMP's main functions are to record, save, analyze, and identify targets from collected data to provide customized ad delivery and enhance customer relationship management through more accurate advertising.
EA Intensive Course "Building Enterprise Architecture" by mr.danairatSoftware Park Thailand
This document outlines the agenda for a two-day course on building enterprise architecture. Day one covers introductions, current architecture challenges, the need for enterprise architecture, definitions of enterprise architecture, reference architecture frameworks, and group workshops. Day two covers maturity models, technology platforms, the TOGAF standard, cloud computing roadmaps, governance, and building a target architecture.
I created this presentation with the intent to inform and clarify what a data management platform is and show how it an be used to enhance marketing initiatives.
Converteo et Makazi ont le plaisir d’annoncer la parution de leur baromètre sur les usages des DMP.
Cette étude menée auprès de 260 décideurs du digital se divise en 4 grands axes d’analyse :
- La notoriété et le niveau d’adoption des DMP
- L’organisation et la gouvernance des projets de déploiement des DMP
- Les usages souhaités et effectifs des DMP
- Le niveau de satisfaction et de ROI des acteurs ayant déployé une DMP
How to Win Machine Learning Competitions ? HackerEarth
This presentation was given by Marios Michailidis (a.k.a Kazanova), Current Kaggle Rank #3 to help community learn machine learning better. It comprises of useful ML tips and techniques to perform better in machine learning competitions. Read the full blog: http://blog.hackerearth.com/winning-tips-machine-learning-competitions-kazanova-current-kaggle-3
The document discusses a father taking his 11-year-old son on a school trip. It describes how the trip strengthened their relationship and the son's self-esteem. The son opened up to his father during the trip, asking thoughtful questions about life and sharing private thoughts. The father realized the trip provided an opportunity to better understand his son and how experiences in everyday life can teach important lessons.
Le Digital Store devient réalité, mais…
…Quels sont les processus à digitaliser en point de vente ou en agence ?
…Comment bâtir une plate-forme marketing Retail, adaptée au Digital In Store, génératrice de ROI et d'engagement client dans un monde ultra connecté ?
Hier vous vendiez des produits, aujourd'hui vous devez offrir une expérience Digitale à vos clients, y compris sur le point de vente (In Store), ou dans un réseau d’agences banque/assurance où le libre service connecté côtoie l’accueil personnalisé.
Pour atteindre cet objectif, vous devez collecter, maîtriser et valoriser vos données structurées ou non-structurées !
Spécialiste de la relation client et du marketing data driven, Soft Computing met en œuvre des plates-formes marketing pour comprendre et prédire les comportements des consommateurs, générer des expériences positives sur les canaux physiques, digitaux ou cross canal, et piloter la relation client & prospect dans la durée.
Ce séminaire vous apportera des éclairages sur les opportunités de ré-enchanter des parcours en points de vente ou agence, la façon de construire une plate-forme, et la déployer sur le terrain.
Pour répondre à vos interrogations, nous vous invitons à un séminaire gratuit, au cours duquel nous partagerons nos meilleures pratiques au travers de cas concrets.
This document provides an agenda for the CITA'15 Workshop held in August 2015. The workshop schedule includes 4 sessions taking place between 8:30 am and 5:00 pm with morning and afternoon breaks. The workshop agenda covers topics such as big data analytics, open data, semantic data description using ontologies and RDF, and a case study on converting a dataset to linked open data. The format of the workshop will be interactive with exercises and discussion encouraged.
Linked Statistical Data: does it actually pay off?Oscar Corcho
Invited keynote at the ISWC2015 Workshop on Semantics and Statistics (SemStats 2015). http://semstats.github.io/2015/
The release of the W3C RDF Data Cube recommendation was a significant milestone towards improving the maturity of the area of Linked Statistical Data. Many Data Cube-based datasets have been released since then. Tools for the generation and exploitation of such datasets have also appeared. While the benefits for the usage of RDF Data Cube and the generation of Linked Data in this area seem to be clear, there are still many challenges associated to the generation and exploitation of such data. In this talk we will reflect about them, based on our experience on generating and exploiting such type of data, and hopefully provoke some discussion about what the next steps should be.
Repository Power: How Repositories can support Open Access Mandates (OR2015 O...OpenAIRE
OpenAIRE presentation at the Open Repositories Conference (OR2015), in Indianapolis, 10/Jun/2015 - Session - P4B: Supporting Open Scholarship and Open Science. Presented by Wolfram Horstmann (Univ. Goettingen) on behalf of the paper authors: Najla Rettberg, Jochen Schirrwagen, Pedro Principe, Eloy Rodrigues, José Carvalho, Paolo Manghi, Natalia Manola.
Knowledge extraction and incorporation is currently considered to be beneficial for efficient Big Data analytics. Knowledge can take part in workflow design, constraint definition, parameter selection and configuration, human interactive and decision-making strategies. Here we present BIGOWL, an ontology to support knowledge management in Big Data analytics. BIGOWL is designed to cover a wide vocabulary of terms concerning Big Data analytics workflows, including their components and how they are connected, from data sources to the analytics visualization. It also takes into consideration aspects such as parameters, restrictions and formats. This ontology defines not only the taxonomic relationships between the different concepts, but also instances representing specific individuals to guide the users in the design of Big Data analytics workflows. For testing purposes, two case studies are developed, which consists in: first, real-world streaming processing with Spark of traffic Open Data, for route optimization in urban environment of New York city; and second, data mining classification of an academic dataset on local/cloud platforms. The analytics workflows resulting from the BIGOWL semantic model are validated and successfully evaluated.
WEBINAR: Open Research Data in Horizon 2020OpenAIRE
The document discusses a webinar about open research data in the Horizon 2020 program. It provides information on the European Commission's open research data pilot, including details on the flexible nature of the pilot, which areas are participating, and how projects can partially or fully opt-out. It also covers what a data management plan is, how to write one, and what it should include regarding FAIR data principles and making data findable, accessible, interoperable and reusable. Recommendations are provided on issues like metadata, file formats, storage, and where to deposit data for long-term preservation and access.
The document provides guidelines for publishing data as Linked Data. It discusses identifying appropriate data sources, reusing existing vocabularies and non-ontological resources, generating RDF data from relational databases or geometrical data using tools like R2O, ODEMapster and geometry2rdf, and publishing the data on the web by resolving URIs. The Ontology Engineering Group at Universidad Politécnica de Madrid has published Spanish geospatial and statistical data as part of projects like GeoLinkedData following these guidelines.
Values & Vision - Cloud Sandboxes for BIG Earth Sciencesterradue
Terradue is an Italian SME focused on providing cloud services for earth science research. They have developed an open platform to help scientists access and analyze large datasets through web and cloud technologies. Their goal is to stimulate new scientific applications and help researchers adapt to increasing data volumes. The platform allows scientists to share data access points, processing chains, and collaborate across distributed systems delivered as a service. Terradue is focusing on new services like data and software as a service to create marketplaces and leverage linked open data. They are also exploring how to use analytics and human resources like data scientists to help optimize the platform.
OpenAIRE services and tools - 6th National Open Access Conference and OpenAIR...OpenAIRE
OpenAIRE provides services and tools to support open science, including:
1. An e-infrastructure that connects publications repositories, research data repositories, CRIS systems, registries, open access journals, and software repositories.
2. Services for content providers, funders, and researchers, including validation, enrichment, and discovery of publications, data, projects, and software.
3. APIs, search interfaces, and tools to link research outputs like publications and data to projects and funders.
OpenAIRE services and tools - 6th National Open Access Conference and OpenAIR...Pedro Príncipe
OpenAIRE provides services and tools to support open science, including:
1. An e-infrastructure that connects publications repositories, research data repositories, CRIS systems, registries, open access journals, and software repositories.
2. Services for content providers, funders, and researchers, including validation, enrichment, and discovery of publications, data, projects, and software.
3. APIs, search interfaces, and tools to link research outputs to projects and funders and support reporting and analysis of open access trends and research impact.
The document provides an introduction to data visualization for open access advocacy. It discusses the PASTEUR4OA project and ROARMAP registry, which contain policy data that can be visualized. Examples of visualizations include maps and charts showing policy data by country, funder type, and mandate status. The document outlines considerations for manipulating, delivering, and working with live data visualizations, and recommends testing different tools to choose the best for the tasks.
Presentation done by Ander García, Maria Teresa Linaza, Javier Franco and Miriam Juaristi, during "Data management" workshop, of the ENTER2015 eTourism conference.
Introducing ORCID at Imperial College LondonTorsten Reimer
Presentation given at the 6th German Library Congress (6. Bibliothekskongress Deutschland) in Leipzig, 15th March 2016. I discuss the position of ORCID in the scholarly communications system - in particular with reference to developments in the UK - and the ORCID implementation project at Imperial College London.
OpenAIRE infrastructure and Services (OpenAIRE Workshop Malta)Pedro Príncipe
OpenAIRE provides infrastructure and services to support open scholarship across Europe. It operates 34 National Open Access Desks, offers over 15 technical services, and supports stakeholders across the research process from funding to publication. OpenAIRE links research results like publications, data, and software to provide a comprehensive view of projects and their outputs and outcomes.
This document discusses how libraries can leverage data from their collections to support new research and discovery. It outlines several initiatives that treat library collections as data, including the Library of Congress labs and a project exploring computationally-driven research. The document also discusses OCLC's work analyzing institutional repository data through its Repository Analytics and Metrics Portal (RAMP) and making data more interoperable through support of the IIIF standard for sharing images and metadata.
Open Data Initiatives – Empowering Students to Make More Informed Choices? - ...Terminalfour
Are the developments in the HEFCE and JISC XCRI XML standards finally giving students the information they need to compare one university against the other? By standardising the output of content about courses, results and post graduation success in the job market, are students better positioned to evaluate the right University for them. This presentation will answer this question and discuss how universities can utilize Web Content Management to deliver on the demands of these standards and maximise student engagement.
Easy SPARQLing for the Building Performance ProfessionalMartin Kaltenböck
Slides of Martin Kaltenböcks (SWC) presentation at SEMANTiCS2014 conference in Leipzig on 5th of September 2014 about the 'Tool for Building Energy Performance Scenarios' of GBPN (Global Buildings Performance Network, http://gbpn.org) that provides a prediction tool for buildings performance worldwide by making use of Linked Open Data (LOD).
Text Analytics & Linked Data Management As-a-ServiceMarin Dimitrov
slides from the talk on "Text Analytics & Linked Data Management As-a-Service with S4" from the ESWC'2015 workshop on Semantic Web Enterprise Adoption & Best Practices
full paper available at http://2015.wasabi-ws.org/papers/wasabi15_1.pdf
Towards a better measure of business proximity: Topic modeling for industry i...Gene Moo Lee
The document presents a new approach for measuring business proximity between firms using topic modeling. It aims to overcome limitations of existing approaches by developing a data-driven, scalable method that provides finer-grained analysis with limited data requirements. The approach applies latent Dirichlet allocation to uncover topics from company descriptions in the CrunchBase dataset. Business proximity is then measured as the cosine similarity between the topic distributions of firm pairs. The method is shown to outperform a baseline of using common industry membership and provides a validated measure of firms' technological and business relatedness.
OpenAIRE - Implementing Open Science (presentation by Natalia Manola at Food ...OpenAIRE
OpenAIRE provides services to support open science including:
1. A helpdesk that answers questions about open access policies and requirements.
2. Tools to explore and link publications, grants, and other research outputs.
3. Dashboards to help research communities monitor and connect their research outcomes.
Similar to ALLDATA 2015 - RDF Based Linked Data Management as a DaaS Platform (20)
Everything You Need to Know About X-Sign: The eSign Functionality of XfilesPr...XfilesPro
Wondering how X-Sign gained popularity in a quick time span? This eSign functionality of XfilesPro DocuPrime has many advancements to offer for Salesforce users. Explore them now!
14 th Edition of International conference on computer visionShulagnaSarkar2
About the event
14th Edition of International conference on computer vision
Computer conferences organized by ScienceFather group. ScienceFather takes the privilege to invite speakers participants students delegates and exhibitors from across the globe to its International Conference on computer conferences to be held in the Various Beautiful cites of the world. computer conferences are a discussion of common Inventions-related issues and additionally trade information share proof thoughts and insight into advanced developments in the science inventions service system. New technology may create many materials and devices with a vast range of applications such as in Science medicine electronics biomaterials energy production and consumer products.
Nomination are Open!! Don't Miss it
Visit: computer.scifat.com
Award Nomination: https://x-i.me/ishnom
Conference Submission: https://x-i.me/anicon
For Enquiry: Computer@scifat.com
UI5con 2024 - Keynote: Latest News about UI5 and it’s EcosystemPeter Muessig
Learn about the latest innovations in and around OpenUI5/SAPUI5: UI5 Tooling, UI5 linter, UI5 Web Components, Web Components Integration, UI5 2.x, UI5 GenAI.
Recording:
https://www.youtube.com/live/MSdGLG2zLy8?si=INxBHTqkwHhxV5Ta&t=0
Hand Rolled Applicative User ValidationCode KataPhilip Schwarz
Could you use a simple piece of Scala validation code (granted, a very simplistic one too!) that you can rewrite, now and again, to refresh your basic understanding of Applicative operators <*>, <*, *>?
The goal is not to write perfect code showcasing validation, but rather, to provide a small, rough-and ready exercise to reinforce your muscle-memory.
Despite its grandiose-sounding title, this deck consists of just three slides showing the Scala 3 code to be rewritten whenever the details of the operators begin to fade away.
The code is my rough and ready translation of a Haskell user-validation program found in a book called Finding Success (and Failure) in Haskell - Fall in love with applicative functors.
Flutter is a popular open source, cross-platform framework developed by Google. In this webinar we'll explore Flutter and its architecture, delve into the Flutter Embedder and Flutter’s Dart language, discover how to leverage Flutter for embedded device development, learn about Automotive Grade Linux (AGL) and its consortium and understand the rationale behind AGL's choice of Flutter for next-gen IVI systems. Don’t miss this opportunity to discover whether Flutter is right for your project.
E-Invoicing Implementation: A Step-by-Step Guide for Saudi Arabian CompaniesQuickdice ERP
Explore the seamless transition to e-invoicing with this comprehensive guide tailored for Saudi Arabian businesses. Navigate the process effortlessly with step-by-step instructions designed to streamline implementation and enhance efficiency.
UI5con 2024 - Boost Your Development Experience with UI5 Tooling ExtensionsPeter Muessig
The UI5 tooling is the development and build tooling of UI5. It is built in a modular and extensible way so that it can be easily extended by your needs. This session will showcase various tooling extensions which can boost your development experience by far so that you can really work offline, transpile your code in your project to use even newer versions of EcmaScript (than 2022 which is supported right now by the UI5 tooling), consume any npm package of your choice in your project, using different kind of proxies, and even stitching UI5 projects during development together to mimic your target environment.
Top Benefits of Using Salesforce Healthcare CRM for Patient Management.pdfVALiNTRY360
Salesforce Healthcare CRM, implemented by VALiNTRY360, revolutionizes patient management by enhancing patient engagement, streamlining administrative processes, and improving care coordination. Its advanced analytics, robust security, and seamless integration with telehealth services ensure that healthcare providers can deliver personalized, efficient, and secure patient care. By automating routine tasks and providing actionable insights, Salesforce Healthcare CRM enables healthcare providers to focus on delivering high-quality care, leading to better patient outcomes and higher satisfaction. VALiNTRY360's expertise ensures a tailored solution that meets the unique needs of any healthcare practice, from small clinics to large hospital systems.
For more info visit us https://valintry360.com/solutions/health-life-sciences
Consistent toolbox talks are critical for maintaining workplace safety, as they provide regular opportunities to address specific hazards and reinforce safe practices.
These brief, focused sessions ensure that safety is a continual conversation rather than a one-time event, which helps keep safety protocols fresh in employees' minds. Studies have shown that shorter, more frequent training sessions are more effective for retention and behavior change compared to longer, infrequent sessions.
Engaging workers regularly, toolbox talks promote a culture of safety, empower employees to voice concerns, and ultimately reduce the likelihood of accidents and injuries on site.
The traditional method of conducting safety talks with paper documents and lengthy meetings is not only time-consuming but also less effective. Manual tracking of attendance and compliance is prone to errors and inconsistencies, leading to gaps in safety communication and potential non-compliance with OSHA regulations. Switching to a digital solution like Safelyio offers significant advantages.
Safelyio automates the delivery and documentation of safety talks, ensuring consistency and accessibility. The microlearning approach breaks down complex safety protocols into manageable, bite-sized pieces, making it easier for employees to absorb and retain information.
This method minimizes disruptions to work schedules, eliminates the hassle of paperwork, and ensures that all safety communications are tracked and recorded accurately. Ultimately, using a digital platform like Safelyio enhances engagement, compliance, and overall safety performance on site. https://safelyio.com/
UI5con 2024 - Bring Your Own Design SystemPeter Muessig
How do you combine the OpenUI5/SAPUI5 programming model with a design system that makes its controls available as Web Components? Since OpenUI5/SAPUI5 1.120, the framework supports the integration of any Web Components. This makes it possible, for example, to natively embed own Web Components of your design system which are created with Stencil. The integration embeds the Web Components in a way that they can be used naturally in XMLViews, like with standard UI5 controls, and can be bound with data binding. Learn how you can also make use of the Web Components base class in OpenUI5/SAPUI5 to also integrate your Web Components and get inspired by the solution to generate a custom UI5 library providing the Web Components control wrappers for the native ones.
E-commerce Development Services- Hornet DynamicsHornet Dynamics
For any business hoping to succeed in the digital age, having a strong online presence is crucial. We offer Ecommerce Development Services that are customized according to your business requirements and client preferences, enabling you to create a dynamic, safe, and user-friendly online store.
Most important New features of Oracle 23c for DBAs and Developers. You can get more idea from my youtube channel video from https://youtu.be/XvL5WtaC20A
Oracle 23c New Features For DBAs and Developers.pptx
ALLDATA 2015 - RDF Based Linked Data Management as a DaaS Platform
1. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
Seonho Kim
RDF based Linked Data
Management as a DaaS Platform!
LODaaS (Linked Open Data as a Service)
2. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
About Me
• Affiliations
• A Research Engineer of Saltlux Inc., Korea (http://saltlux.com/en)
• Managing Director of ODI Node Seoul (http://theodi.kr)
• A member of OKF Korea
• PhD candidate / Medical Informatics, Kyunghee Univ., Korea
• Interests
• Semantic Web, Linked Open Data, Artificial Intelligence, knowledge
representation, logical reasoning
• Contacts
• Email: ksseono@gmail.com
Facebook:@ksseono / Twitter:@SeonhoKim
2
3. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
• If you search ‘linked open data’ from Google…
3
4. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
• If you search ‘linked open data’ from Google…
3
5. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 294
• If you search ‘linked open data’ from Google…
Linked Open Data World
6. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
5
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
7. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
6
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
8. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
7
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
• But essentially, you need to query the data
9. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
7
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
• But essentially, you need to query the data
10. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
8
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
• But essentially, you need to query the data
• -> You need to access the databases
11. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
8
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
• But essentially, you need to query the data
• -> You need to access the databases
HOW?@#$!&
12. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
9
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
• But essentially, you need to query the data
• -> You need to access the databases, or store all
the data on your databse
13. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Open Data World
9
• If you search ‘linked open data’ from Google…
• Then how you could access those data(resources)?
• But essentially, you need to query the data
• -> You need to access the databases, or store all
the data on your databse
www.fujitsu.com
‘Overview of LOD data
store technology'
14. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Data Platform
It’s only about agreements for
protocols, standard formats and
their definitions, not about
accessing/querying those
‘distributed' data.
10
15. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Data Platform
• We need ontology schema to query (to write
SPARQL queries) on Linked Data.
11
16. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Linked Data Platform
• We need ontology schema to query (to write
SPARQL queries) on Linked Data.
11
17. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
LODaaS (Linked Open Data as a Service)
• We need ontology schema to query (to write
SPARQL queries) on Linked Data for real
service development
12
linked data ontology schema
service databse
(LODaaS)
application
optional mapping
/ additional modeling supporting querying
18. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
LODaaS (Linked Open Data as a Service)
• Two ways to provide endpoints
13
19. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Stakeholders of LODaaS
14
20. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 2915
• Data Owners
• owns original datasets to be opened, e.g., public sectors
• not professional on the technologies
• Data Publishers
• collect datasets needed and transform to store the open data into
service repositories
• normally developers
• Data Consumers
• service developers who want to use endpoints or end users
accessing the data via the services
Stakeholders of LODaaS
21. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
• PLUQI (Personalized and Localized Urban
Quality Index)
• datasets - open datasets, physical sensor data, social
sensor data
• web application representing and visualizing the
level of well-being and sustainability for given cities
• applying individual preferences
• based on DaPaaS platform
16
22. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
17
23. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
18
• Key Components (1/3)
• PLUQI Ontology Schema
• standard model for service developments based on those
datasets
24. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
18
• Key Components (1/3)
• PLUQI Ontology Schema
• standard model for service developments based on those
datasets
25. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
19
• Key Components (1/3)
• PLUQI Ontology Schema provides
• standard model for service developments based on those
datasets
• a principle to integrate datasets regarding the index
model
• mapping information with external linked data resources
26. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
20
• Key Components (2/3)
• ETL (Extract-Transform-Load) framework for the datasets
- Grafter (http://grafter.org)
• to transform various datasets following PLUQI schema
27. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
20
• Key Components (2/3)
• ETL (Extract-Transform-Load) framework for the datasets
- Grafter (http://grafter.org)
• to transform various datasets following PLUQI schema
28. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
21
• Key Components (3/3)
• DaPaaS platform (http://dapaas.eu)
• provides linked data management based on PLUQI schema
• provides data transforming functionality (Grafter)
• allows to deploy PLUQI service apps on
29. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
How does it work?
22
• Advantages
• Data owners don’t have to know technologies to be used end
users as linked data.
• Data publishers just could focus on their services, if they
want to use new datasets then they just add the
transforming template.
• Data consumers (service developers) don’t have to care the
new datasets : although new datasets come up, they just
keep the queries following the schema, then new datasets
will be retrieved.
30. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Exobrain Project
• Korean version of IBM Watson - Artificial
Intelligence QA system (http://exobrain.kr)
23
31. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Exobrain Project
• Korean version of IBM Watson - Artificial
Intelligence QA system (http://exobrain.kr)
23
32. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Exobrain Project
• Korean version of IBM Watson - Artificial
Intelligence QA system (http://exobrain.kr)
24
33. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Exobrain Project
• Korean version of IBM Watson - Artificial
Intelligence QA system (http://exobrain.kr)
• big difference between two:
• IBM Waton - uses search technologies on big data
• Exobrain - uses semantic querying on KB + search
technologies
25
34. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Exobrain Project
26
linked data ontology schema
service databse
(LODaaS)
application
data curation
(by human, machines)
Exobrain
ontology schema
브로솃 DB
Exobrain
Knowledge bases
Exobrain
QA system
35. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Summary
• LODaaS
• considering developments practical(domain specific)
applications using linked data
• service ontology schema needed for separate roles between
stakeholders so that they can focus on theirs
• Use cases
• DaPaaS (http://dapaas.eu)
• Exobrain (http://exobrain.kr)
27
36. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Future works
• experiment to transform various datasets
• modify the PLUQI ontology schema and find
proper ways to design it more flexible and
applicable for applications
28
37. ALLDATA 2015 / April 19-24, 2015 @Barcelona, Spain
/ 29
Thank you for your patience
29