How to be one step ahead of leveraging knowledge technologies for your apps!
When: Dec 8, 2017
Where: Fl. 6, Multimedia Tower, Central Jakarta
Thanks to Ragil for the invitation!
Assessing, Creating and Using Knowledge Graph RestrictionsSven Lieber
The presentation of my public PhD defense on March 10, 2022. The related video is available at https://www.youtube.com/watch?v=NofQSwc3Svk
This doctoral thesis tackles the support of users when assessing, creating and using Knowledge Graph restrictions.
More concretely, in this dissertation the FAIR Montolo statistics are contributed, supporting users in assessing existing Knowledge Graphs based on used restrictions.
The two visual notations ShapeUML and ShapeVOWL are presented and evaluated: they represent all constraint types of the Shapes Constraint Language (SHACL) and thus advance the state of the art.
Finally, the use of restrictions to represent formal meaning and to assess data quality is demonstrated for a social media archiving use case in the BESOCIAL project of the Royal Library of Belgium (KBR).
Statistics about Data Shape Use in RDF DataSven Lieber
The presentation of the poster paper "Statistics about Data Shape Use in RDF Data" presented during the demo/poster session at the International Semantic Web Conference (ISWC) 2020.
Joint work with Ben De Meester, Anastasia Dimou and Ruben Verborgh.
The related video is online available at YouTube: https://www.youtube.com/watch?v=6-OdjYdEpeU
What is Linked Data?
Presented at the Linked Data for Libraries on Thursday, November 6, 2014 at Trinity College Dublin
http://www.dri.ie/linked-data-libraries
BESOCIAL A Knowledge Graph for Social Media ArchivingSven Lieber
The presentation of our paper "BESOCIAL: A Sustainable Knowledge Graph-based Workflow for Social Media Archiving" presented at the SEMANTiCS EU conference 2021 in Amsterdam.
Joint work with Dylan Van Assche, Sally Chambers, Fien Messens, Friedel Geeraert. Julie M. Birkholz and Anastasia Dimou
The relate video is online available at https://youtu.be/oYmzD3e8rBE?t=1912
Assessing, Creating and Using Knowledge Graph RestrictionsSven Lieber
The presentation of my public PhD defense on March 10, 2022. The related video is available at https://www.youtube.com/watch?v=NofQSwc3Svk
This doctoral thesis tackles the support of users when assessing, creating and using Knowledge Graph restrictions.
More concretely, in this dissertation the FAIR Montolo statistics are contributed, supporting users in assessing existing Knowledge Graphs based on used restrictions.
The two visual notations ShapeUML and ShapeVOWL are presented and evaluated: they represent all constraint types of the Shapes Constraint Language (SHACL) and thus advance the state of the art.
Finally, the use of restrictions to represent formal meaning and to assess data quality is demonstrated for a social media archiving use case in the BESOCIAL project of the Royal Library of Belgium (KBR).
Statistics about Data Shape Use in RDF DataSven Lieber
The presentation of the poster paper "Statistics about Data Shape Use in RDF Data" presented during the demo/poster session at the International Semantic Web Conference (ISWC) 2020.
Joint work with Ben De Meester, Anastasia Dimou and Ruben Verborgh.
The related video is online available at YouTube: https://www.youtube.com/watch?v=6-OdjYdEpeU
What is Linked Data?
Presented at the Linked Data for Libraries on Thursday, November 6, 2014 at Trinity College Dublin
http://www.dri.ie/linked-data-libraries
BESOCIAL A Knowledge Graph for Social Media ArchivingSven Lieber
The presentation of our paper "BESOCIAL: A Sustainable Knowledge Graph-based Workflow for Social Media Archiving" presented at the SEMANTiCS EU conference 2021 in Amsterdam.
Joint work with Dylan Van Assche, Sally Chambers, Fien Messens, Friedel Geeraert. Julie M. Birkholz and Anastasia Dimou
The relate video is online available at https://youtu.be/oYmzD3e8rBE?t=1912
Development of Semantic Web based Disaster Management SystemNIT Durgapur
Semantic Web model In the field of disaster management to structurise the data such that any information needed during emergency will be easily available.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
Semantic Web technologies such as RDF and OWL have become World Wide Web Consortium (W3C) standards for knowledge representation and reasoning. RDF triples about triples, or meta triples, form the basis for a contextualized knowledge graph. They represent the contextual information about individual triples such as the source, the occurring time or place, or the certainty.
However, an efficient RDF representation for such meta-knowledge of triples remains a major limitation of the RDF data model. The existing reification approach allows such meta-knowledge of RDF triples to be expressed in RDF by using four triples per reified triple. While reification is simple and intuitive, this approach does not have a formal foundation and is not commonly used in practice as described in the RDF Primer.
This dissertation presents the foundations for representing, querying, reasoning and traversing the contextualized knowledge graphs (CKG) using Semantic Web technologies.
A triple-based compact representation for CKGs. We propose a principled approach and construct RDF triples about triples by extending the current RDF data model with a new concept, called singleton property (SP), as a triple identifier. The SP representation needs two triples to the RDF datasets and can be queried with SPARQL.
A formal model-theoretic semantics for CKGs. We formalize the semantics of the singleton property and its relationships with the triple it represents. We extend the current RDF model-theoretic semantics to capture the semantics of the singleton properties and provide the interpretation at three levels: simple, RDF, and RDFS. It provides a single interpretation of the singleton property semantics across applications and systems.
A sound and complete inference mechanism for CKGs. Based on the semantics we propose, we develop a set of inference rules for validating and inferring new triples based on the SP syntax. We also develop different sets of context-based inference rules for provenance, time, and uncertainty.
A graph-based formalism for CKGs. We propose a formal contextualized graph model for the SP representation. We formalize the RDF triples as a mathematical graph by combining the model theory and the graph theory into a hybrid RDF formal semantics. The unified semantics allows the RDF formal semantics to be leveraged in the graph-based algorithms.
Libraries around the world have a long tradition of maintaining authority files to assure the consistent presentation and indexing of names. As library authority files have become available online, the authority data has become accessible -- and many have been published as Linked Open Data (LOD) -- but names in one library authority file typically had no link to corresponding records for persons and organizations in other library authority files. After a successful experiment in matching the Library of Congress/NACO authority file with the German National Library's authority file, an online system called the Virtual International Authority File was developed to facilitate sharing by ingesting, matching, and displaying the relations between records in multiple authority files.
The Virtual International Authority File (VIAF) has grown from three source files in 2007 to more than two dozen files today. The system harvests authority records, enhances them with bibliographic information and brings them together into clusters when it is confident the records describe the same identity. Although the most visible part of VIAF is a HTML interface, the API beneath it supports a linked data view of VIAF with URIs representing the identities themselves, not just URIs for the clusters. It supports names for person, corporations, geographic entities, works, and expressions. With English, French, German, Spanish interfaces (and a Japanese in process), the system is used around the world, with over a million queries per day.
Speaker
Thomas Hickey is Chief Scientist at OCLC where he helped found OCLC Research. Current interests include metadata creation and editing systems, authority control, parallel systems for bibliographic processing, and information retrieval and display. In addition to implementing VIAF, his group looks into exploring Web access to metadata, identification of FRBR works and expressions in WorldCat, the algorithmic creation of authorities, and the characterization of collections. He has an undergraduate degree in Physics and a Ph.D. in Library and Information Science.
This presentation is an updated version of my Data Management 101 talk, which covers the basics of research data management in the categories of: storage and backup, documentation, organization, and making files usable for the future.
Slides from NCURA's webinar "Part I: Public Access: Practical Ways To Assist Faculty To Comply With Public Access Policies". This is the last section on the webinar on open data.
This tutorial explains the Data Web vision, some preliminary standards and technologies as well as some tools and technological building blocks developed by AKSW research group from Universität Leipzig.
RDF is a general method to decompose knowledge into small pieces, with some rules about the semantics or meaning of those pieces. The point is to have a method so simple that it can express any fact, and yet so structured that computer applications can do useful things with knowledge expressed in RDF.
Lecture at the advanced course on Data Science of the SIKS research school, May 20, 2016, Vught, The Netherlands.
Contents
-Why do we create Linked Open Data? Example questions from the Humanities and Social Sciences
-Introduction into Linked Open Data
-Lessons learned about the creation of Linked Open Data (link discovery, knowledge representation, evaluation).
-Accessing Linked Open Data
Presentation given at Barcamp Chiang Mai 4 on the basics of Semantic Web. A simple introduction with examples, aimed for those with a little Web development experience.
Raises questions about the true identity of Tim Berners-Lee.
Development of Semantic Web based Disaster Management SystemNIT Durgapur
Semantic Web model In the field of disaster management to structurise the data such that any information needed during emergency will be easily available.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
Semantic Web technologies such as RDF and OWL have become World Wide Web Consortium (W3C) standards for knowledge representation and reasoning. RDF triples about triples, or meta triples, form the basis for a contextualized knowledge graph. They represent the contextual information about individual triples such as the source, the occurring time or place, or the certainty.
However, an efficient RDF representation for such meta-knowledge of triples remains a major limitation of the RDF data model. The existing reification approach allows such meta-knowledge of RDF triples to be expressed in RDF by using four triples per reified triple. While reification is simple and intuitive, this approach does not have a formal foundation and is not commonly used in practice as described in the RDF Primer.
This dissertation presents the foundations for representing, querying, reasoning and traversing the contextualized knowledge graphs (CKG) using Semantic Web technologies.
A triple-based compact representation for CKGs. We propose a principled approach and construct RDF triples about triples by extending the current RDF data model with a new concept, called singleton property (SP), as a triple identifier. The SP representation needs two triples to the RDF datasets and can be queried with SPARQL.
A formal model-theoretic semantics for CKGs. We formalize the semantics of the singleton property and its relationships with the triple it represents. We extend the current RDF model-theoretic semantics to capture the semantics of the singleton properties and provide the interpretation at three levels: simple, RDF, and RDFS. It provides a single interpretation of the singleton property semantics across applications and systems.
A sound and complete inference mechanism for CKGs. Based on the semantics we propose, we develop a set of inference rules for validating and inferring new triples based on the SP syntax. We also develop different sets of context-based inference rules for provenance, time, and uncertainty.
A graph-based formalism for CKGs. We propose a formal contextualized graph model for the SP representation. We formalize the RDF triples as a mathematical graph by combining the model theory and the graph theory into a hybrid RDF formal semantics. The unified semantics allows the RDF formal semantics to be leveraged in the graph-based algorithms.
Libraries around the world have a long tradition of maintaining authority files to assure the consistent presentation and indexing of names. As library authority files have become available online, the authority data has become accessible -- and many have been published as Linked Open Data (LOD) -- but names in one library authority file typically had no link to corresponding records for persons and organizations in other library authority files. After a successful experiment in matching the Library of Congress/NACO authority file with the German National Library's authority file, an online system called the Virtual International Authority File was developed to facilitate sharing by ingesting, matching, and displaying the relations between records in multiple authority files.
The Virtual International Authority File (VIAF) has grown from three source files in 2007 to more than two dozen files today. The system harvests authority records, enhances them with bibliographic information and brings them together into clusters when it is confident the records describe the same identity. Although the most visible part of VIAF is a HTML interface, the API beneath it supports a linked data view of VIAF with URIs representing the identities themselves, not just URIs for the clusters. It supports names for person, corporations, geographic entities, works, and expressions. With English, French, German, Spanish interfaces (and a Japanese in process), the system is used around the world, with over a million queries per day.
Speaker
Thomas Hickey is Chief Scientist at OCLC where he helped found OCLC Research. Current interests include metadata creation and editing systems, authority control, parallel systems for bibliographic processing, and information retrieval and display. In addition to implementing VIAF, his group looks into exploring Web access to metadata, identification of FRBR works and expressions in WorldCat, the algorithmic creation of authorities, and the characterization of collections. He has an undergraduate degree in Physics and a Ph.D. in Library and Information Science.
This presentation is an updated version of my Data Management 101 talk, which covers the basics of research data management in the categories of: storage and backup, documentation, organization, and making files usable for the future.
Slides from NCURA's webinar "Part I: Public Access: Practical Ways To Assist Faculty To Comply With Public Access Policies". This is the last section on the webinar on open data.
This tutorial explains the Data Web vision, some preliminary standards and technologies as well as some tools and technological building blocks developed by AKSW research group from Universität Leipzig.
RDF is a general method to decompose knowledge into small pieces, with some rules about the semantics or meaning of those pieces. The point is to have a method so simple that it can express any fact, and yet so structured that computer applications can do useful things with knowledge expressed in RDF.
Lecture at the advanced course on Data Science of the SIKS research school, May 20, 2016, Vught, The Netherlands.
Contents
-Why do we create Linked Open Data? Example questions from the Humanities and Social Sciences
-Introduction into Linked Open Data
-Lessons learned about the creation of Linked Open Data (link discovery, knowledge representation, evaluation).
-Accessing Linked Open Data
Presentation given at Barcamp Chiang Mai 4 on the basics of Semantic Web. A simple introduction with examples, aimed for those with a little Web development experience.
Raises questions about the true identity of Tim Berners-Lee.
Why do they call it Linked Data when they want to say...?Oscar Corcho
The four Linked Data publishing principles established in 2006 seem to be quite clear and well understood by people inside and outside the core Linked Data and Semantic Web community. However, not only when discussing with outsiders about the goodness of Linked Data but also when reviewing papers for the COLD workshop series, I find myself, in many occasions, going back again to the principles in order to see whether some approach for Web data publication and consumption is actually Linked Data or not. In this talk we will review some of the current approaches that we have for publishing data on the Web, and we will reflect on why it is sometimes so difficult to get into an agreement on what we understand by Linked Data. Furthermore, we will take the opportunity to describe yet another approach that we have been working on recently at the Center for Open Middleware, a joint technology center between Banco Santander and Universidad Politécnica de Madrid, in order to facilitate Linked Data consumption.
A lecture/conversation focusing on the first 12 years of Semantic Web - delivered on February 21, 2012.
See http://j.mp/SWIntro for more details. More detailed course material is at http://knoesis.org/courses/web3/
These slides were presented at the "graph databases in life sciences workshop". There is an accompanying Neo4j guide that will walk you through importing data into Neo4j using web services form a number of databases at EMBL-EBI.
https://github.com/simonjupp/importing-lifesci-data-into-neo4j
Beyond document retrieval using semantic annotations Roi Blanco
Traditional information retrieval approaches deal with retrieving full-text document as a response to a user's query. However, applications that go beyond the "ten blue links" and make use of additional information to display and interact with search results are becoming increasingly popular and adopted by all major search engines. In addition, recent advances in text extraction allow for inferring semantic information over particular items present in textual documents. This talks presents how enhancing a document with structures derived from shallow parsing is able to convey a different user experience in search and browsing scenarios, and what challenges we face as a consequence.
Presentation at ELAG 2011, European Library Automation Group Conference, Prague, Czech Republic. 25th May 2011
http://elag2011.techlib.cz/en/815-lifting-the-lid-on-linked-data/
Providing open data is of interest for its societal and commercial value, for transparency, and because more people can do fun things with data. There is a growing number of initiatives to provide open data, from, for example, the UK government and the World Bank. However, much of this data is provided in formats such as Excel files, or even PDF files. This raises the question of
- How best to provide access to data so it can be most easily reused?
- How to enable the discovery of relevant data within the multitude of available data sets?
- How to enable applications to integrate data from large numbers of formerly unknown data sources?
One way to address these issues to to use the design principles of linked data (http://www.w3.org/DesignIssues/LinkedData.html), which suggest best practices for how to publish and connect structured data on the Web. This presentation gives an overview of linked data technologies (such as RDF and SPARQL), examples of how they can be used, as well as some starting points for people who want to provide and use linked data.
The presentation was given on August 8, at the Hacknight event (http://hacknight.se/) of Forskningsavdelningen (http://forskningsavd.se/) (Swedish: “Research Department”) a hackerspace in Malmö.
Given at the annual Open Universiteit Informatics faculty research meeting on March 6, 2012. Video is at http://video.intranet.ou.nl/mediadienst/_website/php/external_video.php?Q=1056|videoID
This 2-hour lecture was held at Amsterdam University of Applied Sciences (HvA) on October 16th, 2013. It represents a basic overview over core technologies used by ICT companies such as Google, Twitter or Facebook. The lecture does not require a strong technical background and stays at conceptual level.
When we created this quiz of Java programming course, we did that with Fasilkom UI students in mind.
Fast forward, we now thought that the quiz could be of greater use if it's shared to anyone, not just Fasilkom UI students.
Yes, our students of our course are everyone, including you!
So please find attached, fresh from the oven, Java programming quiz part 01 (with key answers). More parts are coming whenever they are ready.
#java #programming #universitasindonesia #opencourse #openaccess #openeducation #opentridharma
Featuring pointers for: Single-layer neural networks and multi-layer neural networks, gradient descent, backpropagation. Slides are for introduction, for deep explanation on deep learning, please consult other slides.
Current situation: focus is limited to only implement Tridharma, that is, education, research, and community service, with little concern on openness aspect.
The openness of Tridharma can potentially be a breakthrough in mitigating the quality gap issue: opening Tridharma outputs for public would help to increase the citizen inclusion in accessing the quality content of Tridharma, hence narrowing the quality gap in higher education.
[ISWC 2013] Completeness statements about RDF data sources and their use for ...Fariz Darari
This was presented at ISWC 2013 in Sydney, Australia.
Abstract:
With thousands of RDF data sources available on the Web covering disparate and possibly overlapping knowledge domains, the problem of providing high-level descriptions (in the form of metadata) of their content becomes crucial. In this paper we introduce a theoretical framework for describing data sources in terms of their completeness. We show how existing data sources can be described with completeness statements expressed in RDF. We then focus on the problem of the completeness of query answering over plain and RDFS data sources augmented with completeness statements. Finally, we present an extension of the completeness framework for federated data sources.
Dissertation Defense - Managing and Consuming Completeness Information for RD...Fariz Darari
The ever increasing amount of Semantic Web data gives rise to the question: How complete is the data? Though generally data on the Semantic Web is incomplete, many parts of data are indeed complete, such as the children of Barack Obama and the crew of Apollo 11. This thesis aims to study how to manage and consume completeness information about Semantic Web data. In particular, we first discuss how completeness information can guarantee the completeness of query answering. Next, we propose optimization techniques of completeness reasoning and conduct experimental evaluations to show the feasibility of our approaches. We also provide a technique to check the soundness of queries with negation via reduction to query completeness checking. We further enrich completeness information with timestamps, enabling query answers to be checked up to when they are complete. We then introduce two demonstrators, i.e., CORNER and COOL-WD, to show how our completeness framework can be realized. Finally, we investigate an automated method to generate completeness statements from text on the Web via relation cardinality extraction.
KOI - Knowledge Of Incidents - SemEval 2018Fariz Darari
We present KOI (Knowledge Of Incidents), a system that given news articles as input, builds a knowledge graph (KOI-KG) of incidental events.
KOI-KG can then be used to efficiently answer questions such as "How many killing incidents happened in 2017 that involve Sean?" The required steps in building the KG include:
(i) document preprocessing involving word sense disambiguation, named-entity recognition, temporal expression recognition and normalization, and semantic role labeling;
(ii) incidental event extraction and coreference resolution via document clustering; and (iii) KG construction and population.
Slides made and presented by Paramita.
Italy Agriculture Equipment Market Outlook to 2027harveenkaur52
Agriculture and Animal Care
Ken Research has an expertise in Agriculture and Animal Care sector and offer vast collection of information related to all major aspects such as Agriculture equipment, Crop Protection, Seed, Agriculture Chemical, Fertilizers, Protected Cultivators, Palm Oil, Hybrid Seed, Animal Feed additives and many more.
Our continuous study and findings in agriculture sector provide better insights to companies dealing with related product and services, government and agriculture associations, researchers and students to well understand the present and expected scenario.
Our Animal care category provides solutions on Animal Healthcare and related products and services, including, animal feed additives, vaccination
Instagram has become one of the most popular social media platforms, allowing people to share photos, videos, and stories with their followers. Sometimes, though, you might want to view someone's story without them knowing.
APNIC Foundation, presented by Ellisha Heppner at the PNG DNS Forum 2024APNIC
Ellisha Heppner, Grant Management Lead, presented an update on APNIC Foundation to the PNG DNS Forum held from 6 to 10 May, 2024 in Port Moresby, Papua New Guinea.
Meet up Milano 14 _ Axpo Italia_ Migration from Mule3 (On-prem) to.pdfFlorence Consulting
Quattordicesimo Meetup di Milano, tenutosi a Milano il 23 Maggio 2024 dalle ore 17:00 alle ore 18:30 in presenza e da remoto.
Abbiamo parlato di come Axpo Italia S.p.A. ha ridotto il technical debt migrando le proprie APIs da Mule 3.9 a Mule 4.4 passando anche da on-premises a CloudHub 1.0.
Gen Z and the marketplaces - let's translate their needsLaura Szabó
The product workshop focused on exploring the requirements of Generation Z in relation to marketplace dynamics. We delved into their specific needs, examined the specifics in their shopping preferences, and analyzed their preferred methods for accessing information and making purchases within a marketplace. Through the study of real-life cases , we tried to gain valuable insights into enhancing the marketplace experience for Generation Z.
The workshop was held on the DMA Conference in Vienna June 2024.
2. Fariz Darari
• 1988: Born in Malang
• 2010: BSc in Computer Science at Universitas Indonesia
• 2013: MSc in Computational Logic at University of Bolzano,
Italy and TU Dresden, Germany
Best Thesis Award and Enno-Heidebroek Award
• 2017: PhD in Computational Logic at University of Bolzano,
Italy and TU Dresden, Germany
• 2017: Lecturer at Faculty of CS, Universitas Indonesia
6. • Knowledge Technologies: Motivations
• Semantic Web
• Knowledge Bases These Days (= Zaman Now)
• Wikidata
• DBpedia
• Applications
• Discussion: Challenges & Opportunities
Menu
7. What if the knowledge in your brains,*
can be queried by computers?
*notice the plural form
8. What if the knowledge in your brains,*
can be queried by computers?
*notice the plural form
9. What if the knowledge in your brains,*
can be queried by computers?
*notice the plural form
10. What if the knowledge in your brains,
can be queried by computers?
Can you imagine what kind of advancements
can be made to humanity?
11. What if the knowledge in your brains,
can be queried by computers?
Can you imagine what kind of advancements
can be made to humanity?
Stay tuned, will present you an answer to this question some slides later!
12. ... if properly designed,
the Semantic Web can assist
the evolution of human knowledge
as a whole.
– Tim Berners-Lee
Inventor of the (Semantic) Web
15. What is the Semantic Web?
The set of technologies to put knowledge on the Web,
that is based on the following four principles:
1. Use URIs (Universal Resource Identifiers)* for identifying things
2. Use HTTP** URIs so people can look up those names
3. When someone looks up a URI, provide useful knowledge
using the standards: RDF and SPARQL.
4. Include links to other URIs, so they can discover more things
https://www.w3.org/DesignIssues/LinkedData.html
* URI = just like URL (web address), but you use it to identify things just like barcode for supermarket stuff!
** HTTP = the mechanism you use every time you access the Web!
18. RDF in one slide
the data guy
• Data model, based on S-P-O triple structure (Subject, Predicate, Object)
• Used for describing things, yes, every, single, thing
And anyway, RDF = Resource Description Framework
• Key features:
• RDF data can be exported in JSON and XML
• RDF links things, not just documents
• RDF links are typed
TelkomUniversity somelink Bandung
TelkomUniversity locatedIn Bandung
19. OWL in one slide
• Schema (=Ontology) language, describing vocabularies
• Yes, it is on the meta-level!
• Short for: Web Ontology Language (WOL? No, it is OWL!)
• Key features:
• Reasoning: you can check if your knowledge is consistent/not!
• Reasoning again: you can conclude new things
based on existing facts.
• Very simple example:
owl SubClassOf bird + bird SubClassOf animal + owl EquivalentClass strigifomes
Now, if Bobi is a Strigifomes, do you think Bobi is an animal?
the schema guy
20. OWL in one slide
• Schema (=Ontology) language, describing vocabularies
• Yes, it is on the meta-level!
• Short for: Web Ontology Language (WOL? No, it is OWL!)
• Key features:
• Reasoning: you can check if your knowledge is consistent/not!
• Reasoning again: you can conclude new things
based on existing facts.
• Very simple example:
owl SubClassOf bird + bird SubClassOf animal + owl EquivalentClass strigifomes
Now, if Bobi is a Strigifomes, do you think Bobi is an animal? OWL will say:
the schema guy
21. OWL in one slide
• Schema (=Ontology) language, describing vocabularies
• Yes, it is on the meta-level!
• Short for: Web Ontology Language (WOL? No, it is OWL!)
• Key features:
• Reasoning: you can check if your knowledge is consistent/not!
• Reasoning again: you can conclude new things
based on existing facts.
• Very simple example:
owl SubClassOf bird + bird SubClassOf animal + owl EquivalentClass strigifomes
Now, if Bobi is a Strigifomes, do you think Bobi is an animal? OWL will say: "YES!"
the schema guy
22. SPARQL in one slide
the query guy
• Query language: If RDF captures knowledge,
SPARQL asks questions about knowledge!
• Short for: SPARQL Protocol and RDF Query Language
• Key features: Asking for knowledge, is a KEY feature!
• Very simple example:
TelkomUniversity locatedIn Bandung
Bandung headOfGov RidwanKamil
TelkomUniversity instanceOf University
It is SPARQLing!
SELECT ?university WHERE {
?university instanceOf University .
?university locatedIn ?city .
?city headOfGov RidwanKamil }
Guess what this query is asking for?
HINT: Question mark (?) represents variables to match
with RDF data!
25. Knowledge Bases (KBs) These Days (Zaman Now)
KB NOW
Subject
Predicate
Predicate
Predicate
Object
Object
Object
Reminds you of something?
26. Knowledge Bases (KBs) These Days (Zaman Now)
KB NOW
Subject
Predicate
Predicate
Predicate
Object
Object
Object
Reminds you of something?
the data guy
27. Knowledge Bases (KBs) These Days (Zaman Now)
KB NOW
Subject
Predicate
Predicate
Predicate
Object
Object
Object
Reminds you of something?
the data guy
btw, every subject in Wikidata has its own identifier, the URI is made by: Wikidata domain + identifier
28. Knowledge Bases (KBs) These Days (Zaman Now)
KB NOW
Subject
Predicate
Predicate
Predicate
Object
Object
Object
Reminds you of something?
the data guy
btw, every subject in Wikidata has its own identifier, the URI is made by: Wikidata domain + identifier
= P31
= P571
= Q4830453
= Q10389
29. Knowledge Bases (KBs) These Days (Zaman Now)
the query guy
http://tinyurl.com/yc6jsmhv
30. Knowledge Bases (KBs) These Days (Zaman Now)
the query guy
http://tinyurl.com/y84kyl4d
31. Knowledge Bases (KBs) These Days (Zaman Now)
the schema guy
OwlInWinnieThePooh instanceOf fictionalOwl
fictionalOwl subClassOf fictionalBird
fictionalBird subClassOf fictionalAnimal
32. Knowledge Bases (KBs) These Days (Zaman Now)
Wikidata key features:
• It is like Wikipedia but for data!
• It is under Wikimedia foundation
• It is crowdsourced, anyone can add data
• It is free
• It's got 326 million facts about 40 million
subjects! (Wikipedia only has 5 million subjects!)
• It loves the Semantic Web
34. Knowledge Bases (KBs) These Days (Zaman Now)
DBpedia key features:
• It extracts data from Wikipedia infoboxes
(summary box on top right corner).
• It is free
• It's got 13 BILLION facts about 7 million
subjects!
• It loves the Semantic Web
35. Knowledge Bases (KBs) These Days (Zaman Now)
DBpedia key features:
• It extracts data from Wikipedia infoboxes
(summary box on top right corner).
• It is free
• It's got 13 BILLION facts about 7 million
subjects!
• It loves the Semantic Web
• DBpedia Indonesia is available, hosted by
Faculty of Computer Science, Univ. Indonesia
42. Question: When was Soekarno born?
http://id.dbpedia.org/page/Soekarno
Application: DBpedia-powered Answer Engine
43. Application: DBpedia-powered Answer Engine
Question: When was Soekarno born?
Borrow Techniques from
Natural Language Processing
SELECT ?birthDate
WHERE {
<http://id.dbpedia.org/resource/Soekarno> <http://dbpedia.org/ontology/birthDate> ?birthDate
}
SPARQL Query over DBpedia
http://id.dbpedia.org/sparql
45. Application: Timeline Infographics
Task: Create timeline of Indonesian national heroes based on their birthdates!
Without Wikidata:
- Read by eyes websites about national heroes (there are all 173 heroes!)
- Gather information manually
- Visualize information manually
Total time spent: 24+ hours!
46. Application: Wikidata-powered Timeline Infographics
Task: Create timeline of Indonesian national heroes based on their birthdates!
With Wikidata (and Histropedia):
- Formulate and evaluate the query
- VOILA: Beautiful timeline infographics created!
Total time spent: 10 minutes!
55. Application: Wikidata-powered Virtual Doctor
dr Wikidata: Tell me your symptoms
Patient: I feel like fatigue, headache,
joint pain, and vomitting
dr Wikidata: From what I know,
you most likely get dengue fever!
65. Completeness: Is the data complete
enough? Is it of sufficient breadth and
depth?
Accuracy: How accurate is the data? Is
it reliable and verifiable?
Timeliness: Is the data up-to-date? Is
the latest data included?
Data Quality