Web open standards for linked data and knowledge graphs as enablers of EU digital sovereignty
ENDORSE Keynote by Fabien GANDON, 19/03/2021
https://op.europa.eu/en/web/endorse
This course is a quick overview of the fundamentals of graph databases and graph queries, with a focus on RDF and SPARQL. It includes both simple and challenging hands-on exercises to practice and test your understanding.
The material for this course can be downloaded form the following link: https://github.com/paolo7/Introduction-to-Graph-Databases
This course is a quick overview of the fundamentals of graph databases and graph queries, with a focus on RDF and SPARQL. It includes both simple and challenging hands-on exercises to practice and test your understanding.
The material for this course can be downloaded form the following link: https://github.com/paolo7/Introduction-to-Graph-Databases
One Web of pages, One Web of peoples, One Web of Services, One Web of Data, O...Fabien Gandon
Keynote Fabien GANDON, at WIM2016: One Web of pages, One Web of peoples, One Web of Services, One Web of Data, One Web of Things…and with the Semantic Web bind them.
JURIX talk on representing and reasoning on the deontic aspects of normative rules relying only on standard Semantic Web languages.
The corresponding paper is at https://hal.inria.fr/hal-01643769v1
Wimmics Research Team 2015 Activity ReportFabien Gandon
Extract of the activity report of the Wimmics joint research team between Inria Sophia Antipolis - Méditerranée and I3S (CNRS and Université Nice Sophia Antipolis). Wimmics stands for web-instrumented man-machine interactions, communities and semantics. The team focuses on bridging social semantics and formal semantics on the web.
on the ontological necessity of the multidisciplinary development of the webFabien Gandon
Talk on the ontological necessity of the multidisciplinary development of the web at the panel CLOSER/WEBIST 2014 on "social, political and economic implications of cloud and web"
How to Build Linked Data Sites with Drupal 7 and RDFascorlosquet
Slides of the tutorial Stéphane Corlosquet, Lin Clark and Alexandre Passant presented at SemTech 2010 in San Francisco http://semtech2010.semanticuniverse.com/sessionPop.cfm?confid=42& proposalid=2889
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
DISIT Lab overview: smart city, big data, semantic computing, cloudPaolo Nesi
Smart City
• Projects: http://www.disit.org/5501
– Sii-Mobility, http://www.sii-mobility.org
– Service Map: http://servicemap.disit.org
– Social Innovation: Coll@bora http://www.disit.org/5479
– Navigation Indoor/outdoor: Mobile Emergency http://www.disit.org/5404
– Mobility and Transport: TRACE-IT, RAISSS, TESYSRAIL
• Tools: http://www.disit.org/5489
– Data gathering, data mining and reconciliation
– Data reasoning, deduction, prediction
– Smart city ontology and reasoning tools
– Service analysis and recommendations
– Autonomous train operator, train signaling
– Risk analysis, decision support systems
– Mobile Applications
Data Analytics - Big data
• Projects: http://www.disit.org/5501
– Linked Open Graph: http://LOG.disit.org
– Sii-Mobility, http://www.sii-mobility.org
– Service on a number of projects
• Tools: http://www.disit.org/5489
– Open data and Linked Open Data
– LOG LOD service and tools
– Data mining and reconciliation
– Data reasoning, deduction, prediction, decision support
– SN Analysis and recommendations
– User behavior monitoring and analysis
Smart Cloud - Computing
• Projects: http://www.disit.org/5501
– ICARO: http://www.disit.org/5482
– Cloud ontology: http://www.disit.org/5604
– Cloud simulator:
– Smart Cloud: http://www.disit.org/6544
• Tools: http://www.disit.org/5489
– Cloud Monitoring
– Smart Cloud Engine and reasoner,
– Service Level Analyzer and control
– Configuration analysis and checker
– Cloud Simulation
Text and Web Mining
• Projects: http://www.disit.org/5501
– OSIM: http://www.disit.org/5482
– SACVAR: http://www.disit.org/5604
– Blog/Twitter Vigilance
• Tools: http://www.disit.org/5489
– Text and web mining, Natural Language Processing
– Service localization
– Web Crawling
– Competence analysis
– Blog Vigiliance, sentiment analysis
Social Media and e-Learning
• Projects: http://www.disit.org/5501
– ECLAP, http://www.eclap.eu
– ApreToscana: http://www.apretoscana.org
– Others: AXMEDIS, VARIAZIONI, SMNET, etc.
– Samsung Smart TV: http://www.disit.org/6534
• Tools: http://www.disit.org/5489
– XLMS, Cross Media Learning System
– IPR and content protection and distribution
– Mobile and SmartTv Applications
– Suggestions and recommendations
– Matchmaking solutions
– Media Tools for cross media content
Mobile Computing
• Projects:
– ECLAP: http://www.eclap.eu
– Mobile Medicine: http://mobmed.axmedis.org
– Mobile Emergency: http://www.disit.org/5500
– Smart City, FODD 2015: http://www.disit.org/6593
– Resolute: Mobiles as sensors
• Tools and support:
– Content distribution: e-learning
– Integrated Indoor/outdoor navigation
– User networking and collaboration
– Service localization
– Smart city and services
– OS: iOS, Android, Windows Phone, etc.
– Tech: IOT, iBeacoms, NFC, QR, ….
One Web of pages, One Web of peoples, One Web of Services, One Web of Data, O...Fabien Gandon
Keynote Fabien GANDON, at WIM2016: One Web of pages, One Web of peoples, One Web of Services, One Web of Data, One Web of Things…and with the Semantic Web bind them.
JURIX talk on representing and reasoning on the deontic aspects of normative rules relying only on standard Semantic Web languages.
The corresponding paper is at https://hal.inria.fr/hal-01643769v1
Wimmics Research Team 2015 Activity ReportFabien Gandon
Extract of the activity report of the Wimmics joint research team between Inria Sophia Antipolis - Méditerranée and I3S (CNRS and Université Nice Sophia Antipolis). Wimmics stands for web-instrumented man-machine interactions, communities and semantics. The team focuses on bridging social semantics and formal semantics on the web.
on the ontological necessity of the multidisciplinary development of the webFabien Gandon
Talk on the ontological necessity of the multidisciplinary development of the web at the panel CLOSER/WEBIST 2014 on "social, political and economic implications of cloud and web"
How to Build Linked Data Sites with Drupal 7 and RDFascorlosquet
Slides of the tutorial Stéphane Corlosquet, Lin Clark and Alexandre Passant presented at SemTech 2010 in San Francisco http://semtech2010.semanticuniverse.com/sessionPop.cfm?confid=42& proposalid=2889
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
DISIT Lab overview: smart city, big data, semantic computing, cloudPaolo Nesi
Smart City
• Projects: http://www.disit.org/5501
– Sii-Mobility, http://www.sii-mobility.org
– Service Map: http://servicemap.disit.org
– Social Innovation: Coll@bora http://www.disit.org/5479
– Navigation Indoor/outdoor: Mobile Emergency http://www.disit.org/5404
– Mobility and Transport: TRACE-IT, RAISSS, TESYSRAIL
• Tools: http://www.disit.org/5489
– Data gathering, data mining and reconciliation
– Data reasoning, deduction, prediction
– Smart city ontology and reasoning tools
– Service analysis and recommendations
– Autonomous train operator, train signaling
– Risk analysis, decision support systems
– Mobile Applications
Data Analytics - Big data
• Projects: http://www.disit.org/5501
– Linked Open Graph: http://LOG.disit.org
– Sii-Mobility, http://www.sii-mobility.org
– Service on a number of projects
• Tools: http://www.disit.org/5489
– Open data and Linked Open Data
– LOG LOD service and tools
– Data mining and reconciliation
– Data reasoning, deduction, prediction, decision support
– SN Analysis and recommendations
– User behavior monitoring and analysis
Smart Cloud - Computing
• Projects: http://www.disit.org/5501
– ICARO: http://www.disit.org/5482
– Cloud ontology: http://www.disit.org/5604
– Cloud simulator:
– Smart Cloud: http://www.disit.org/6544
• Tools: http://www.disit.org/5489
– Cloud Monitoring
– Smart Cloud Engine and reasoner,
– Service Level Analyzer and control
– Configuration analysis and checker
– Cloud Simulation
Text and Web Mining
• Projects: http://www.disit.org/5501
– OSIM: http://www.disit.org/5482
– SACVAR: http://www.disit.org/5604
– Blog/Twitter Vigilance
• Tools: http://www.disit.org/5489
– Text and web mining, Natural Language Processing
– Service localization
– Web Crawling
– Competence analysis
– Blog Vigiliance, sentiment analysis
Social Media and e-Learning
• Projects: http://www.disit.org/5501
– ECLAP, http://www.eclap.eu
– ApreToscana: http://www.apretoscana.org
– Others: AXMEDIS, VARIAZIONI, SMNET, etc.
– Samsung Smart TV: http://www.disit.org/6534
• Tools: http://www.disit.org/5489
– XLMS, Cross Media Learning System
– IPR and content protection and distribution
– Mobile and SmartTv Applications
– Suggestions and recommendations
– Matchmaking solutions
– Media Tools for cross media content
Mobile Computing
• Projects:
– ECLAP: http://www.eclap.eu
– Mobile Medicine: http://mobmed.axmedis.org
– Mobile Emergency: http://www.disit.org/5500
– Smart City, FODD 2015: http://www.disit.org/6593
– Resolute: Mobiles as sensors
• Tools and support:
– Content distribution: e-learning
– Integrated Indoor/outdoor navigation
– User networking and collaboration
– Service localization
– Smart city and services
– OS: iOS, Android, Windows Phone, etc.
– Tech: IOT, iBeacoms, NFC, QR, ….
presented at WORKS 2021
https://works-workshop.org/
16th Workshop on Workflows in Support of Large-Scale Science
November 15, 2021
Held in conjunction with SC21: The International Conference for High Performance Computing, Networking, Storage and Analysis
RDMkit, a Research Data Management Toolkit. Built by the Community for the ...Carole Goble
https://datascience.nih.gov/news/march-data-sharing-and-reuse-seminar 11 March 2022
Starting in 2023, the US National Institutes of Health (NIH) will require institutes and researchers receiving funding to include a Data Management Plan (DMP) in their grant applications, including the making their data publicly available. Similar mandates are already in place in Europe, for example a DMP is mandatory in Horizon Europe projects involving data.
Policy is one thing - practice is quite another. How do we provide the necessary information, guidance and advice for our bioscientists, researchers, data stewards and project managers? There are numerous repositories and standards. Which is best? What are the challenges at each step of the data lifecycle? How should different types of data? What tools are available? Research Data Management advice is often too general to be useful and specific information is fragmented and hard to find.
ELIXIR, the pan-national European Research Infrastructure for Life Science data, aims to enable research projects to operate “FAIR data first”. ELIXIR supports researchers across their whole RDM lifecycle, navigating the complexity of a data ecosystem that bridges from local cyberinfrastructures to pan-national archives and across bio-domains.
The ELIXIR RDMkit (https://rdmkit.elixir-europe.org (link is external)) is a toolkit built by the biosciences community, for the biosciences community to provide the RDM information they need. It is a framework for advice and best practice for RDM and acts as a hub of RDM information, with links to tool registries, training materials, standards, and databases, and to services that offer deeper knowledge for DMP planning and FAIR-ification practices.
Launched in March 2021, over 120 contributors have provided nearly 100 pages of content and links to more than 300 tools. Content covers the data lifecycle and specialized domains in biology, national considerations and examples of “tool assemblies” developed to support RDM. It has been accessed by over 123 countries, and the top of the access list is … the United States.
The RDMkit is already a recommended resource of the European Commission. The platform, editorial, and contributor methods helped build a specialized sister toolkit for infectious diseases as part of the recently launched BY-COVID project. The toolkit’s platform is the simplest we could manage - built on plain GitHub - and the whole development and contribution approach tailored to be as lightweight and sustainable as possible.
In this talk, Carole and Frederik will present the RDMkit; aims and context, content, community management, how folks can contribute, and our future plans and potential prospects for trans-Atlantic cooperation.
Data policy must be partnered with data practice. Our researchers need to be the best informed in order to meet these new data management and data sharing mandates.
COMBINE 2019, EU-STANDS4PM, Heidelberg, Germany 18 July 2019
FAIR: Findable Accessable Interoperable Reusable. The “FAIR Principles” for research data, software, computational workflows, scripts, or any other kind of Research Object one can think of, is now a mantra; a method; a meme; a myth; a mystery. FAIR is about supporting and tracking the flow and availability of data across research organisations and the portability and sustainability of processing methods to enable transparent and reproducible results. All this is within the context of a bottom up society of collaborating (or burdened?) scientists, a top down collective of compliance-focused funders and policy makers and an in-the-middle posse of e-infrastructure providers.
Making the FAIR principles a reality is tricky. They are aspirations not standards. They are multi-dimensional and dependent on context such as the sensitivity and availability of the data and methods. We already see a jungle of projects, initiatives and programmes wrestling with the challenges. FAIR efforts have particularly focused on the “last mile” – “FAIRifying” destination community archive repositories and measuring their “compliance” to FAIR metrics (or less controversially “indicators”). But what about FAIR at the first mile, at source and how do we help Alice and Bob with their (secure) data management? If we tackle the FAIR first and last mile, what about the FAIR middle? What about FAIR beyond just data – like exchanging and reusing pipelines for precision medicine?
Since 2008 the FAIRDOM collaboration [1] has worked on FAIR asset management and the development of a FAIR asset Commons for multi-partner researcher projects [2], initially in the Systems Biology field. Since 2016 we have been working with the BioCompute Object Partnership [3] on standardising computational records of HTS precision medicine pipelines.
So, using our FAIRDOM and BioCompute Object binoculars let’s go on a FAIR safari! Let’s peruse the ecosystem, observe the different herds and reflect what where we are for FAIR personalised medicine.
References
[1] http://www.fair-dom.org
[2] http://www.fairdomhub.org
[3] http://www.biocomputeobject.org
FAIRy stories: the FAIR Data principles in theory and in practiceCarole Goble
https://ucsb.zoom.us/meeting/register/tZYod-ippz4pHtaJ0d3ERPIFy2QIvKqjwpXR
FAIRy stories: the FAIR Data principles in theory and in practice
The ‘FAIR Guiding Principles for scientific data management and stewardship’ [1] launched a global dialogue within research and policy communities and started a journey to wider accessibility and reusability of data and preparedness for automation-readiness (I am one of the army of authors). Over the past 5 years FAIR has become a movement, a mantra and a methodology for scientific research and increasingly in the commercial and public sector. FAIR is now part of NIH, European Commission and OECD policy. But just figuring out what the FAIR principles really mean and how we implement them has proved more challenging than one might have guessed. To quote the novelist Rick Riordan “Fairness does not mean everyone gets the same. Fairness means everyone gets what they need”.
As a data infrastructure wrangler I lead and participate in projects implementing forms of FAIR in pan-national European biomedical Research Infrastructures. We apply web-based industry-lead approaches like Schema.org; work with big pharma on specialised FAIRification pipelines for legacy data; promote FAIR by Design methodologies and platforms into the researcher lab; and expand the principles of FAIR beyond data to computational workflows and digital objects. Many use Linked Data approaches.
In this talk I’ll use some of these projects to shine some light on the FAIR movement. Spoiler alert: although there are technical issues, the greatest challenges are social. FAIR is a team sport. Knowledge Graphs play a role – not just as consumers of FAIR data but as active contributors. To paraphrase another novelist, “It is a truth universally acknowledged that a Knowledge Graph must be in want of FAIR data.”
[1] Wilkinson, M., Dumontier, M., Aalbersberg, I. et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci Data 3, 160018 (2016). https://doi.org/10.1038/sdata.2016.18
Data management plans – EUDAT Best practices and case study | www.eudat.euEUDAT
| www.eudat.eu | Presentation given by Stéphane Coutin during the PRACE 2017 Spring School joint training event with the EU H2020 VI-SEEM project (https://vi-seem.eu/) organised by CaSToRC at The Cyprus Institute. Science and more specifically projects using HPC is facing a digital data explosion. Instruments and simulations are producing more and more volume; data can be shared, mined, cited, preserved… They are a great asset, but they are facing risks: we can miss storage, we can lose them, they can be misused,… To start this session, we will review why it is important to manage research data and how to do this by maintaining a Data Management Plan. This will be based on the best practices from EUDAT H2020 project and European Commission recommendation. During the second part we will interactively draft a DMP for a given use case.
Ontology Building vs Data Harvesting and Cleaning for Smart-city ServicesPaolo Nesi
Presently, a very large number of public and private data sets are available around the local governments. In most cases, they are not semantically interoperable and a huge human effort is needed to create integrated ontologies and knowledge base for smart city. Smart City ontology is not yet standardized, and a lot of research work is needed to identify models that can easily support the data reconciliation, the management of the complexity and reasoning. In this paper, a system for data ingestion and reconciliation of smart cities related aspects as road graph, services available on the roads, traffic sensors etc., is proposed. The system allows managing a big volume of data coming from a variety of sources considering both static and dynamic data. These data are mapped to smart-city ontology and stored into an RDF-Store where they are available for applications via SPARQL queries to provide new services to the users. The paper presents the process adopted to produce the ontology and the knowledge base and the mechanisms adopted for the verification, reconciliation and validation. Some examples about the possible usage of the coherent knowledge base produced are also offered and are accessible from the RDF-Store and related services. The article also presented the work performed about reconciliation algorithms and their comparative assessment and selection. Keywords Smart city, knowledge base construction, reconciliation, validation and verification of knowledge base, smart city ontology, linked open graph.
FAIR Data
Principles
FAIR vs Open Data
Implementing FAIR & FAIRmetrics
FAIRness de ASIO-HERCULES
Research Objects
Definition
Standard RO-CRATE
Usage examples
Make our Scientific Datasets Accessible and Interoperable on the WebFranck Michel
The presentation investigates the challenges that we must face to share scientific datasets on the Web following the Linked Open Data principles. We present the standards of the Semantic Web and investigate how they can help address those challenges. We give tips as to how to choose vocabularies to describe data and metadata, link datasets to other related datasets by making appropriate alignments, translate existing data sources to RDF and publish it on the Web as linked data.
This is one out of a series of presentations which I have given during a recent trip to the United States. I will make them all public, but content does not vary a lot between some of them
Infrastructures Supporting Inter-disciplinary Research - Exemplars from the UK NeISSProject
Infrastructures Supporting Inter-disciplinary Research - Exemplars from the UK . Talk given by Richard Sinnott at Urban Research Infrastructure Network Workshops, Melbourne, Brisbane, Sydney, September 2010.
Similar to Web open standards for linked data and knowledge graphs as enablers of EU digital sovereignty (20)
Walking Our Way to the Web - Fabien Gandon
The Web: Scientific Creativity, Technological Innovation and Society
XXVIII Conference on Contemporary Philosophy and Methodology of Science
9 and 10 March 2023
University of A Coruña
The prospect of Walking our Way to the Web may sound strange to contemporary readers of this article for whom the Web is omnipresent. However, the slogan of the World Wide Web Consortium (W3C) has been, for years, and remains today, to lead “the Web to its full potential” meaning we haven’t reached that potential yet, whatever it is. The first architect of the Web himself, Tim Berners-Lee, said in an interview in 2009: “The Web as I envisaged it, we have not seen it yet. The future is still so much bigger than the past”. And he is still very active, together with the W3C members and Web experts world-wide, in proposing evolutions of the Web architecture to improve its growing usages and applications. In this article we will review the path that led us to the actual Web, the shape it is taking now and the possible evolutions, good and bad, we can identify today. This will lead us to consider the distance that we witness between the initial vision and the reality of the Web today, and to reflect on the possible divergence between the potential we see in the Web and the directions it could take. Our goal in this article is to reflect on how we could walk the delicate path to the full potential of the Web, finding the missing links and avoiding the one too many links.
a shift in our research focus: from knowledge acquisition to knowledge augmen...Fabien Gandon
EKAW 2022 keynote by Fabien GANDON: "a shift in our research focus: from knowledge acquisition to knowledge augmentation"
While EKAW started in 1987 as the European Knowledge Acquisition Workshop, in 2000 it transformed into a conference where we advance knowledge engineering and modelling in general. At the time, this transition also echoed shifts of focus such as moving from the paradigm of expert systems to the more encompassing one of knowledge-based systems. Nowadays, with the current strong interest for knowledge graphs, it is important again to reaffirm that our ultimate goal is not the acquisition of bigger siloed knowledge bases but to support knowledge requisition by and for all kinds of intelligence. Knowledge without intelligence is a highly perishable resource. Intelligence without knowledge is doomed to stagnation. We will defend that intelligence and knowledge, and their evolutions, have to be considered jointly and that the Web is providing a social hypermedia to link them in all their forms. Using examples from several projects, we will suggest that, just like intelligence augmentation and amplification insist on putting humans at the center of the design of artificial intelligence methods, we should think in terms of knowledge augmentation and amplification and we should design a knowledge web to be an enabler of the futures we want.
A Never-Ending Project for Humanity Called “the Web”Fabien Gandon
A Never-Ending Project for Humanity Called "the Web"
Fabien Gandon, Wendy Hall
https://hal.inria.fr/WIMMICS/hal-03633526
In this paper we summarized the main historical steps in making the Web, its foundational principles and its evolution. First we mention some of the influences and streams of thought that interacted to bring the Web about. Then we recall that its birthplace, the CERN, had a need for a global hypertext system and at the same time was the perfect microcosm to provide a cradle for the Web. We stress how this invention required to strike a balance between the integration of and the departure from the existing and emerging paradigms of the day. We then review the pillars of the Web architecture and the features that made the Web so viral compared to competitors. Finally we survey the multiple mutations the Web underwent no sooner it was born, evolving in multiple directions. We conclude on the fact the Web is now an architecture, an artefact, a science object and a research and development object, and of which we haven't seen the full potential yet.
CovidOnTheWeb : covid19 linked data published on the WebFabien Gandon
The Covid-on-the-Web project aims to allow biomedical researchers to access, query and make sense of COVID-19 related literature. To do so, it adapts, combines and extends tools to process, analyze and enrich the "COVID-19 Open Research Dataset" (CORD-19) that gathers 50,000+ full-text scientific articles related to the coronaviruses. We report on the RDF dataset and software resources produced in this project by leveraging skills in knowledge representation, text, data and argument mining, as well as data visualization and exploration. The dataset comprises two main knowledge graphs describing (1) named entities mentioned in the CORD-19 corpus and linked to DBpedia, Wikidata and other BioPortal vocabularies, and (2) arguments extracted using ACTA, a tool automating the extraction and visualization of argumentative graphs, meant to help clinicians analyze clinical trials and make decisions. On top of this dataset, we provide several visualization and exploration tools based on the Corese Semantic Web platform, MGExplorer visualization library, as well as the Jupyter Notebook technology. All along this initiative, we have been engaged in discussions with healthcare and medical research institutes to align our approach with the actual needs of the biomedical community, and we have paid particular attention to comply with the open and reproducible science goals, and the FAIR principles.
from linked data & knowledge graphs to linked intelligence & intelligence graphsFabien Gandon
ISWC Vision track talk "from linked data & knowledge graphs to linked intelligence & intelligence graphs or the potential of the semantic Web to break the walls between semantic networks and computational networks"
Retours sur le MOOC "Web Sémantique et Web de données"Fabien Gandon
Présentation des caractéristiques et résultats de la première session en 2015 du MOOC "Web Sémantique et Web de données" par Inria, Université de Nice, FUN et UNIT.
Nous lisons régulièrement que le Web révolutionne notre monde et provoque des évolutions dans toutes les dimensions de notre société. Mais le Web lui-même, ses usages et la compréhension que nous en avons n’ont pas cessé d’évoluer depuis la proposition à l’origine de sa création en 1989. C’est un espace en perpétuelle recréation qui nous demande sans cesse de nouvelles explorations et reconsidérations. Ce sont certains de ces changements passés, actuels, et à venir du Web que nous allons regarder ensemble en insistant sur la complexité de cet artefact qui en fait un objet de recherches pluridisciplinaires.
On Youtube: https://youtu.be/jNjHdqS-1Ko
Données de la culture et culture des donnéesFabien Gandon
Présentation "Données de la culture et culture des données" ou le web sémantique et les données liées sur le web dans le domaine de la culture à l'occasion de la conférence "Transmettre la culture à l’ère du numérique" dans le programme Automne Numérique du ministère de la Culture et de la Communication.
La vidéo de la conférence est ici:
http://www.dailymotion.com/video/x17i1g6_conference-transmettre-la-culture-a-l-age-du-numerique-fabien-gandon_tech
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Web open standards for linked data and knowledge graphs as enablers of EU digital sovereignty
1. Web open standards for linked data and
knowledge graphs as enablers of EU
digital sovereignty
Fabien Gandon, http://fabien.info
2. PROFILE
Graduated Engineer INSA Applied Math, DEA/Master Image & Vision
PHD & HDR (Habilitation) in computer science
Research Director / Senior researcher, INRIA
Leader Wimmics (UCA, Inria, CNRS, I3S) on Campus Sophia Antipolis
Advisory Committee of W3C
Responsible research convention French Ministry of Culture – Inria
Vice-head of Science for Inria Sophia Antipolis
3. WIMMICS TEAM
DR/Professors:
Fabien GANDON, Inria, AI, KRR, Semantic Web, Social Web, K. Graphs
Nhan LE THANH, UCA, Logics, KR, Emotions, Workflows, K. Graphs
Peter SANDER, UCA, Web, Emotions
Andrea TETTAMANZI, UCA, AI, Logics, Evo, Learning, Agents, K. Graphs
Marco WINCKLER, UCA, Human-Computer Interaction, Web, K. Graphs
CR/Assistant Professors:
Michel BUFFA, UCA, Web, Social Media, Web Audio, K. Graphs
Elena CABRIO, UCA, NLP, KR, Linguistics, Q&A, Text Mining, K. Graphs
Olivier CORBY, Inria, KR, AI, Sem. Web, Programming, K. Graphs
Catherine FARON-ZUCKER, UCA, KR, AI, Semantic Web, K. Graphs
Damien GRAUX, Inria, Linked Data, Sem. Web, Querying, K. Graphs
Serena VILLATA, CNRS, AI, Argumentation, Licenses, Rights, K. Graphs
Research engineer: Franck MICHEL, CNRS, Linked Data, Integration, DB, K. Graphs
External:
Andrei Ciortea (University of St. Gallen) Agents, WoT, Sem. Web, K. Graphs
Nicolas DELAFORGE (Mnemotix) Sem. Web, KM, Integration, K. Graphs
Alain GIBOIN, (Retired CR Inria), Interaction Design, KE, User & Task, K. Graphs
Freddy LECUE (Thales, Montreal) AI, Logics, Mining, Big Data, S. Web , K. Graphs
4. URI, IRI, URL, HTTP URI
STANDARDS FOR DATA & KNOWLEDGE GRAPHS ON THE WEB
JSON
RDF
JSON LD
N-Triple
N-Quad
Turtle/N3
TriG
RDFS
OWL
SPARQL
XML
HTML
RDF XML
HTTP
Linked Data
CSV-LD R2RML
GRDDL
RDFa
SHACL
LDP
6. World Wide Web Consortium
an international community leading the Web to its full potential since 1994
i.e. building an open, interoperable Web that works for everyone,
by developing freely available and open standards for it.
In 2016, Tim Berners-Lee received the
Turing Award for his invention of the Web
7. World Wide Web Consortium
Over 430 Members org. around the world
The not-for-profit organization’s staff of 50
supported by Membership dues
Over 12,000 developers worldwide
38 working groups + 10 interest groups
+ 350 Business Groups and Community Groups
Hundreds of open technologies that power…
browsers, smart phones, ebook readers, set top
boxes, automobiles, search engines, social media,
trillions of dollars of online commerce, and more
than a billion Web sites
=
8. for instance…
examples of former or current members
html http
url
uri
iri atag
uaag
wcag
aria
mwbp
earl
ra cc/pp
assx
css
ddrsa xml eve. exi
geo api
dom xform
grddl inkml its cmwww ruby an.
xhtml rdfa
ets omr m. ok emma
p3p
mathml mf
pics qa rif sec cont. sawsdl
png powder
sml soap
wsdl
svg awww
ttml smile
rdf owl
rdfs
sparql
woff
webcgm
xbl xkms xlink
wscdl wsp
skos
ns canon. x dtxml xproc xfrag
xml xbase
xschema
xml:id xpath xpointer
xquery
xsignat. xbop
xslt
xslfo
examples of standards
…
…
9. (2/8) Web open standards for…
distributed, interoperable hypermedia
11. three components of the Web architecture
1. identification (URI) & address (URL)
ex. http://www.inria.fr
URL
12. three components of the Web architecture
1. identification (URI) & address (URL)
ex. http://www.inria.fr
2. communication / protocol (HTTP)
GET /centre/sophia HTTP/1.1
Host: www.inria.fr
HTTP
URL
address
13. three components of the Web architecture
1. identification (URI) & address (URL)
ex. http://www.inria.fr
2. communication / protocol (HTTP)
GET /centre/sophia HTTP/1.1
Host: www.inria.fr
3. representation language (HTML)
Fabien works at
<a href="http://inria.fr">Inria</a>
HTTP
URL
HTML
reference address
communication
WEB
15. (3/8) Web open standards for…
distributed, interoperable identifiers
16. Universal Resource Locator / Indentifier
HTTP
URL
HTML
reference address
communication
WEB
HTTP
URI
HTML
reference address
communication
WEB
17. identify what
exists on the
web
http://my-site.fr
identify,
on the web, what
exists
http://animals.org/this-zebra
18. URIs for everything
• URI for Paris in DBpedia:
http://dbpedia.org/resource/Paris
• URI for name of Victor Hugo in the Library of Congress:
http://id.loc.gov/authorities/names/n79091479
• The MUC18 protein at UniProt
http://www.uniprot.org/uniprot/P43121
• Xavier Dolan in Wikidata
https://www.wikidata.org/wiki/Special:EntityData/Q551861
• The book with doi:10.1007/3-540-45741-0_18
http://dx.doi.org/10.1007/3-540-45741-0_18
•
28. linked open data(sets) cloud on the Web
0
200
400
600
800
1000
1200
1400
5/1/2007 10/8/2007 11/7/2007 11/10/2007 2/28/2008 3/31/2008 9/18/2008 3/5/2009 3/27/2009 7/14/2009 9/22/2010 9/19/2011 8/30/2014 1/26/2017
number of linked open datasets on the Web
29. Smarter Cities’ knowledge graphs
IBM Dublin [Lécué et al., 2015] (also for private KGs behind firewalls)
30. (5/8) Web open standards for…
distributed interoperable access
32. DBPEDIA.FR
180 000 000 arcs in an
encyclopedic knowledge
graph
number of queries per day
70 000 on average
2.5 millions max
185 377 686 RDF triples extracted and mapped
public dumps, endpoints, interfaces, APIs…
33. COVID LINKED DATA
integrate multiple datasets in heterogeneous formats
perform information extraction, inferences, validation
provide a public end-point and visualization services
[Gandon, Michel, Gazzotti, Mayer, Cabrio, Corby, Menin, Winckler, Villata et al. 2020]
34. (6/8) Web open standards for…
distributed interoperable validation
35. SHACL is a language for
describing and validating pieces (shapes) of
RDF knowledge graphs
eg. every Person must have one and only one name
used for validation, description, interaction,
integration, code generation,…
36. This project has received funding from the European Union's Horizon 2020
research and innovation programme under grant agreement 825619.
ONTOLOGY FOR AI ITSELF
ontology and metadata of AI resources
SHACL to validate AI4EU these RDF graphs
online endpoint http://corese.inria.fr
predefined SPARQL queries, SHACL shapes, display
[Corby et al., 2019]
37. (7/8) Web open standards for…
distributed, interoperable vocabularies
38. RDFS to declare classes of resources and properties, of
your knowledge graph and organize their hierarchy
Document
Report
creator
author
Document Person
39. OWL in one…
algebraic properties
disjoint properties
qualified cardinality
1..1
!
individual prop. neg
chained prop.
enumeration
intersection
union
complement
disjunction
restriction
!
cardinality
1..1
equivalence
[>18]
disjoint union
value restriction
keys …
40. PREDICT HOSPITALIZATION
Predict hospitalization from
Physician’s records classification
[Gazzotti, Faron et al. 2020]
Sexe Date Cause CISP2 ... History Observations
H 25/04/2012 vaccin-antitétanique A44 ... Appendicite EN CP - Bon état général - auscult
pulm libre; bdc rég sans souffle -
tympans ok-
Element Number
Patients
Consultations
Past medical history
Biometric data
Semiotics
Diagnosis
Row of prescribed drugs
Symptoms
Health care procedures
Additional examination
Paramedical prescription
Observations/notes
55 823
364 684
187 290
293 908
250 669
117 442
847 422
23 488
11 850
871 590
17 222
56 143
PRIMEGE
41. PREDICT HOSPITALIZATION
Predict hospitalization from
Physician’s records classification
Augment records data with
Web knowledge graphs
[Gazzotti, Faron et al. 2020]
Sexe Date Cause CISP2 ... History Observations
H 25/04/2012 vaccin-antitétanique A44 ... Appendicite EN CP - Bon état général - auscult
pulm libre; bdc rég sans souffle -
tympans ok-
Element Number
Patients
Consultations
Past medical history
Biometric data
Semiotics
Diagnosis
Row of prescribed drugs
Symptoms
Health care procedures
Additional examination
Paramedical prescription
Observations/notes
55 823
364 684
187 290
293 908
250 669
117 442
847 422
23 488
11 850
871 590
17 222
56 143
(1)
PRIMEGE
42. PREDICT HOSPITALIZATION
Predict hospitalization from
Physician’s records classification
Augment records data with
Web knowledge graphs
Study impact on prediction
[Gazzotti, Faron et al. 2020]
Sexe Date Cause CISP2 ... History Observations
H 25/04/2012 vaccin-antitétanique A44 ... Appendicite EN CP - Bon état général - auscult
pulm libre; bdc rég sans souffle -
tympans ok-
Element Number
Patients
Consultations
Past medical history
Biometric data
Semiotics
Diagnosis
Row of prescribed drugs
Symptoms
Health care procedures
Additional examination
Paramedical prescription
Observations/notes
55 823
364 684
187 290
293 908
250 669
117 442
847 422
23 488
11 850
871 590
17 222
56 143
(1)
(2)
PRIMEGE
44. MonaLIA
reason & query on RDF to build training sets.
350 000 images
of artworks
RDF metadata based
on external thesauri
Joconde database from French museums
(1)
[Bobasheva et al. 2020]
45. MonaLIA
reason & query on RDF to build training sets.
transfer learning & CNN classifiers on targeted
categories (topics, techniques, etc.)
350 000 images
of artworks
RDF metadata based
on external thesauri
Joconde database from French museums
(1)
[Bobasheva et al. 2020]
(2)
46. Image Metadata Score
portrait
50350012455
C:Jocondejoconde0138m503501_d0012455-000_p.jpg
cheval:
0.999
Image Metadata Score
figure (saint Eloi de Noyon, évêque, en pied, bénédiction,
vêtement liturgique, mitre, attribut, cheval, marteau, outil :
ferronnerie)
000SC022652
C:/Joconde/joconde0355/m079806_bsa0030101_p.jpg
cheval:
0.006
MonaLIA
reason & query on RDF to build training sets.
transfer learning & CNN classifiers on targeted
categories (topics, techniques, etc.)
reason & query RDF of results to address
silence, noise and explain
350 000 images
of artworks
RDF metadata based
on external thesauri
Joconde database from French museums
(1)
(3)
[Bobasheva et al. 2020]
(2)
47. Web open standards as enablers of
interoperable platforms e.g.
“Solid (…) is a proposed set of conventions and tools for building decentralized
Web applications based on Linked Data principles. (…)
It relies as much as possible on existing W3C standards and protocols. (…)
RDF 1.1 (…) The WebID 1.0 (…) The FOAF vocabulary (…)
WebID-TLS protocol (…) HTML5 (…) Linked Data Platform (LDP) standard”
https://github.com/solid/solid#standards-used
48. (8/8) Web open standards for…
distributed, interoperable Europe
“I’m right there in the room, but
no one even acknowledges me.”
50. W3C = strategic place to survey and shape Web standards
Personal opinion:
Important to have a neutral place to build open-standards (1 member = 1 vote)
Important to have public and private members at W3C
Important to have a large European participation to W3C
51. Web open standards & world-wide interoperability
are key enablers of EU digital sovereignty
Interoperability is strategic to federate actors/actions. (cf. members)
Web standards are transversal to domains/tasks/… (cf. applications examples)
Importance of knowledge graphs and danger of knowledge silos. (cf. data)
Having established open standards between actors in Europe
(public and private) is a stake for setting up European data spaces.
52. Web open standards & world-wide interoperability
are key enablers of EU digital sovereignty
Interoperability is strategic to federate actors/actions. (cf. members)
Web standards are transversal to domains/tasks/… (cf. applications examples)
Importance of knowledge graphs and danger of knowledge silos. (cf. data)
Having established open standards between actors in Europe
(public and private) is a stake for setting up European data spaces.
• active participation to W3C is a key to build EU digital sovereignty.
53. WIMMICS
Web-Instrumented Man-Machine Interactions, Communities and Semantics
Fabien Gandon - @fabien_gandon - http://fabien.info
he who controls metadata, controls the web
and through the world-wide web many things in our world.
Site: http://wimmics.inria.fr
Overview: http://bit.ly/wimmics-slides
Technical details: http://bit.ly/wimmics-papers
Editor's Notes
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Ces codes sont symptomatiques d’une évolution d’un composant central du web qu’est l’adresse web.
Nous sommes passés d’adresses essentiellement utilisées pour identifier les pages et ressources du web.
A des adresses permettant d’identifier sur le web tout ce qui existe autour de nous et d’en parler sur le web.
This evolution of the use of identifiers on the Web together with the ability to change the languages to exchange representations open a new perspective where we can use the Web to identify and exchange any kind of data about everything around us.
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo