Simon Dixon, Cedric Mesnage and Barry Norton (Centre for Digital Music, Queen Mary University of London).
Music Linked Data Workshop, 12 May 2011, JISC, London.
This presentation introduces the semantic web concepts that enable the publication of linked open data. It also introduces LodLive, a linked open data visualization, and discover-me-semantically, an RDF authoring tool.
Linked Open Data Fundamentals for Libraries, Archives and Museumstrevorthornton
This document provides an overview of linked open data concepts for libraries, archives, and museums. It discusses what linked open data is, potential benefits for cultural institutions, and technical concepts like URIs, HTTP, RDF, ontologies, and SPARQL. The document also covers publishing linked open data by establishing URIs for resources and using content negotiation. Trust and attribution of linked data sources are addressed. Open data licensing, including options from Creative Commons, is also summarized.
This chapter discusses metadata and ontologies for digitally documenting cultural heritage. It introduces XML, RDF, Dublin Core, and the Semantic Web as standards for representing metadata. It also discusses OWL for defining ontologies and CIDOC-CRM as an ontology for cultural heritage documentation. The chapter aims to explain how these standards help achieve interoperability when sharing digital cultural heritage information on the internet.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
The document summarizes the Smithsonian Libraries' efforts to build a new open linked library by exposing their digital collections as linked open data using semantic web standards. They analyzed their existing digital content to identify which data elements could be exposed as linked data. They migrated their website to Drupal to natively support RDFa and allow querying between systems. They provided examples of how book metadata and records from their Taxonomic Literature database would be represented as linked data.
The document discusses search and how it works under the hood. It begins with an overview of common search problems and limitations. It then demonstrates how search works by indexing documents into an inverted index of tokens and associated document references. Key steps include analyzing text by splitting, downcasing, and removing stopwords, and then storing the token postings in the index. Search queries can then be executed by looking up token postings in the index. A Ruby example class demonstrates indexing sample documents and searching the generated index.
Presentation at the Big Data interest group at Los Alamos National Laboratory, regarding with with the IIPC to merge the indexes of multiple web archives.
Linked data demystified:Practical efforts to transform CONTENTDM metadata int...Cory Lampert
This document outlines a presentation about transforming metadata from a CONTENTdm digital collection into linked data. It discusses the concepts of linked data, including defining linked data, linked data principles, technologies and standards. It then explains how these concepts can be applied to digital collection records, including anticipated challenges working with CONTENTdm. The document describes a linked data project at UNLV Libraries to transform collection records into linked data and publish it on the linked data cloud. It provides tips for creating metadata that is more suitable for linked data.
This presentation introduces the semantic web concepts that enable the publication of linked open data. It also introduces LodLive, a linked open data visualization, and discover-me-semantically, an RDF authoring tool.
Linked Open Data Fundamentals for Libraries, Archives and Museumstrevorthornton
This document provides an overview of linked open data concepts for libraries, archives, and museums. It discusses what linked open data is, potential benefits for cultural institutions, and technical concepts like URIs, HTTP, RDF, ontologies, and SPARQL. The document also covers publishing linked open data by establishing URIs for resources and using content negotiation. Trust and attribution of linked data sources are addressed. Open data licensing, including options from Creative Commons, is also summarized.
This chapter discusses metadata and ontologies for digitally documenting cultural heritage. It introduces XML, RDF, Dublin Core, and the Semantic Web as standards for representing metadata. It also discusses OWL for defining ontologies and CIDOC-CRM as an ontology for cultural heritage documentation. The chapter aims to explain how these standards help achieve interoperability when sharing digital cultural heritage information on the internet.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
The document summarizes the Smithsonian Libraries' efforts to build a new open linked library by exposing their digital collections as linked open data using semantic web standards. They analyzed their existing digital content to identify which data elements could be exposed as linked data. They migrated their website to Drupal to natively support RDFa and allow querying between systems. They provided examples of how book metadata and records from their Taxonomic Literature database would be represented as linked data.
The document discusses search and how it works under the hood. It begins with an overview of common search problems and limitations. It then demonstrates how search works by indexing documents into an inverted index of tokens and associated document references. Key steps include analyzing text by splitting, downcasing, and removing stopwords, and then storing the token postings in the index. Search queries can then be executed by looking up token postings in the index. A Ruby example class demonstrates indexing sample documents and searching the generated index.
Presentation at the Big Data interest group at Los Alamos National Laboratory, regarding with with the IIPC to merge the indexes of multiple web archives.
Linked data demystified:Practical efforts to transform CONTENTDM metadata int...Cory Lampert
This document outlines a presentation about transforming metadata from a CONTENTdm digital collection into linked data. It discusses the concepts of linked data, including defining linked data, linked data principles, technologies and standards. It then explains how these concepts can be applied to digital collection records, including anticipated challenges working with CONTENTdm. The document describes a linked data project at UNLV Libraries to transform collection records into linked data and publish it on the linked data cloud. It provides tips for creating metadata that is more suitable for linked data.
RDA: Are We There Yet?
This document discusses the progress of Resource Description and Access (RDA) since its publication in 2010. It notes recommendations from libraries that tested RDA, including rewriting instructions in plain English and improving the RDA Toolkit. The implementation date for RDA is March 31, 2013. Differences after implementing RDA include lack of abbreviations, more transcription of elements, new MARC fields, and richer authority records. Fully implementing RDA may involve changes to search options and semantic web/linked data approaches. Tips are provided for libraries on deciding when to implement, talking to vendors, and planning training.
The document discusses using linked open data and linked data principles for libraries. It covers key concepts like URIs, RDF triples, ontologies and vocabularies. It then outlines options for libraries to both consume and publish linked data, such as enriching existing catalog data by linking to external sources, creating new information aggregates, and publishing library holdings and metadata as linked open data. Challenges include a lack of common identifiers, FRBRization of existing data, and the need for content curation and new technical systems to fully realize the benefits of linked open data for libraries.
This document provides an overview of Resource Description and Access (RDA), the new cataloging standard that will replace AACR2. It discusses what RDA is, why it was developed, how it relates to FRBR, and some of the key differences between RDA and AACR2, such as changes to terminology, transcription, and MARC fields. It also explores potential future directions for RDA, such as linked data and semantic web applications. The document concludes by offering suggestions for how libraries can prepare to implement RDA.
Intro to Linked Open Data in Libraries Archives & Museums.Jon Voss
This document discusses a presentation on Linked Open Data in libraries, archives, and museums. The presentation introduces Linked Open Data and how it is being used in cultural heritage institutions. It discusses representing data as graphs using triples and RDF, important vocabularies and ontologies, and following Tim Berners-Lee's principles of Linked Data. The presentation also covers legal and licensing considerations for publishing open cultural data on the web.
Libraries and Linked Data: Looking to the Future (2)ALATechSource
The document discusses options for new bibliographic frameworks after MARC. It describes three scenarios: 1) a relational/object-oriented RDA database, 2) linked bibliographic and authority records, and 3) flat files without links. It then discusses three approaches to implementing a new framework: 1) going native by using URIs for things, elements and values, 2) extracting data from existing MARC records, and 3) serializing data into key-value pairs, XML, or JSON. Advantages and disadvantages of each approach are outlined.
I used these slides in the context of a cultural heritage presentation so the examples are relevant to that community. For example the choice of CIDOC CRM is obvious in that community.
Introduction to linked open data, RDF: the Resource Description Framework, Tools to convert data to RDF, Tools for linking/reconciliation/resolution, Storing and maintaining the data, BBC and Linked Data
SPARQL1.1 Tutorial, given in UChile by Axel Polleres (DERI)net2-project
This document provides an introduction to SPARQL 1.1. It begins by explaining that SPARQL is a query language for the semantic web that allows users to query RDF data stores similarly to how SQL queries relational databases. It then describes SPARQL 1.0, the initial standard version, and the new features being added in SPARQL 1.1, including aggregate functions, subqueries, property paths and federated querying. The document concludes by discussing SPARQL implementations and the status of the 1.1 specification.
The document discusses data standards for describing cultural collections on the web. It advocates for using URIs, HTTP, and semantic web standards like RDF and SPARQL to provide structured data about cultural objects that is interlinked and can be queried. Alternative approaches like CSV, XML, and relational databases are discussed and their limitations explained. The benefits of a linked data approach for sharing cultural information on the web are emphasized.
This document provides an overview of linked data and the SPARQL query language. It defines linked data as a method of publishing structured data on the web so that it can be interlinked and queried. The key aspects covered include linked data principles of using URIs to identify things and including links to other related data. SPARQL is introduced as the query language for retrieving and manipulating linked data.
Semantic Web technologies such as RDF and OWL have become World Wide Web Consortium (W3C) standards for knowledge representation and reasoning. RDF triples about triples, or meta triples, form the basis for a contextualized knowledge graph. They represent the contextual information about individual triples such as the source, the occurring time or place, or the certainty.
However, an efficient RDF representation for such meta-knowledge of triples remains a major limitation of the RDF data model. The existing reification approach allows such meta-knowledge of RDF triples to be expressed in RDF by using four triples per reified triple. While reification is simple and intuitive, this approach does not have a formal foundation and is not commonly used in practice as described in the RDF Primer.
This dissertation presents the foundations for representing, querying, reasoning and traversing the contextualized knowledge graphs (CKG) using Semantic Web technologies.
A triple-based compact representation for CKGs. We propose a principled approach and construct RDF triples about triples by extending the current RDF data model with a new concept, called singleton property (SP), as a triple identifier. The SP representation needs two triples to the RDF datasets and can be queried with SPARQL.
A formal model-theoretic semantics for CKGs. We formalize the semantics of the singleton property and its relationships with the triple it represents. We extend the current RDF model-theoretic semantics to capture the semantics of the singleton properties and provide the interpretation at three levels: simple, RDF, and RDFS. It provides a single interpretation of the singleton property semantics across applications and systems.
A sound and complete inference mechanism for CKGs. Based on the semantics we propose, we develop a set of inference rules for validating and inferring new triples based on the SP syntax. We also develop different sets of context-based inference rules for provenance, time, and uncertainty.
A graph-based formalism for CKGs. We propose a formal contextualized graph model for the SP representation. We formalize the RDF triples as a mathematical graph by combining the model theory and the graph theory into a hybrid RDF formal semantics. The unified semantics allows the RDF formal semantics to be leveraged in the graph-based algorithms.
The document discusses linked data and its potential impact on libraries. It describes linked data as connecting the world's libraries by publishing structured data about 290 million resources using common schemas, embedding RDFa, and linking to controlled vocabularies. While linked data presents challenges like metadata for different types of materials, it offers opportunities to describe resources as part of the web and link catalog data to related concepts through identifiers.
The presentation discusses OCLC's efforts to create a cultural linked data backbone by connecting library data through open linked data standards. It provides examples of existing linked data projects like VIAF, Dewey, and WorldCat and outlines OCLC's work to publish 270 million WorldCat resources as linked open data using the Schema.org vocabulary. The presentation argues that linking library data in this way allows the world's libraries to be more connected on the web.
The presentation discusses linked data and its potential impact on libraries. Linked data uses URIs and HTTP to identify things on the web and link related resources. It enables libraries to publish structured data on the web and connect their resources to other data sources. While challenging, linked data offers a way for libraries to share data on the web of data and transition from cataloging to "catalinking", connecting their resources on the semantic web.
This document discusses the need for named graphs in RDF to represent contextual information like provenance and source of RDF data. It proposes extensions to the RDF/XML syntax to associate RDF descriptions and statements with named graphs. This allows modeling things like different hypotheses, temporal aspects, points of view, and distributed storage in a way that is currently not possible without named graphs in the RDF model.
The document discusses exposing library holdings data on the web using linked data. It notes that OCLC has exposed over 300 million resources using Schema.org, RDFa, and links to controlled vocabularies. The data is available via various formats like RDF/XML, JSON-LD and Turtle. BIBFRAME is presented as the new standard for bibliographic description that allows library data to be shared as part of the web. Libraries are encouraged to make their resources discoverable on the web of data by linking to other institutions and authorities.
TPDL2013 tutorial linked data for digital libraries 2013-10-22jodischneider
Tutorial on Linked Data for Digital Libraries, given by me, Uldis Bojars, and Nuno Lopes in Valletta, Malta at TPDL2013 on 2013-10-22.
http://tpdl2013.upatras.gr/tut-lddl.php
This half-day tutorial is aimed at academics and practitioners interested in creating and using Library Linked Data. Linked Data has been embraced as the way to bring complex information onto the Web, enabling discoverability while maintaining the richness of the original data. This tutorial will offer participants an overview of how digital libraries are already using Linked Data, followed by a more detailed exploration of how to publish, discover and consume Linked Data. The practical part of the tutorial will include hands-on exercises in working with Linked Data and will be based on two main case studies: (1) linked authority data and VIAF; (2) place name information as Linked Data.
For practitioners, this tutorial provides a greater understanding of what Linked Data is, and how to prepare digital library materials for conversion to Linked Data. For researchers, this tutorial updates the state of the art in digital libraries, while remaining accessible to those learning Linked
Data principles for the first time. For library and iSchool instructors, the tutorial provides a valuable introduction to an area of growing interest for information organization curricula. For digital library project managers, this tutorial provides a deeper understanding of the principles of Linked Data, which is needed for bespoke projects that involve data mapping and the reuse of existing metadata models.
RDA implementation is scheduled for March 31, 2013. Testers of RDA recommended improvements like rewriting instructions in plain English and ensuring community involvement. Differences from AACR2 include lack of abbreviations, more transcription of what is seen, and new fields in MARC like 336, 337, 338 for content/media/carrier types. Linked data and semantic web approaches may make relationships between works more explicit over time. Preparing for RDA involves decisions about cataloging workflows and training.
IFLA LIDASIG Open Session 2017: Introduction to Linked DataLars G. Svensson
At the IFLA Linked Data Special Interest Group open session in Wroclaw we briefly introduced the mission of the SIG and then went on to a brief introduction to what linked data is and why that topic is important to libraries.
The presentation was held jointly by Astrid Verheusen (general introduction to the SIG) and Lars G. Svensson (introduction to Linked Data)
RDA: Are We There Yet?
This document discusses the progress of Resource Description and Access (RDA) since its publication in 2010. It notes recommendations from libraries that tested RDA, including rewriting instructions in plain English and improving the RDA Toolkit. The implementation date for RDA is March 31, 2013. Differences after implementing RDA include lack of abbreviations, more transcription of elements, new MARC fields, and richer authority records. Fully implementing RDA may involve changes to search options and semantic web/linked data approaches. Tips are provided for libraries on deciding when to implement, talking to vendors, and planning training.
The document discusses using linked open data and linked data principles for libraries. It covers key concepts like URIs, RDF triples, ontologies and vocabularies. It then outlines options for libraries to both consume and publish linked data, such as enriching existing catalog data by linking to external sources, creating new information aggregates, and publishing library holdings and metadata as linked open data. Challenges include a lack of common identifiers, FRBRization of existing data, and the need for content curation and new technical systems to fully realize the benefits of linked open data for libraries.
This document provides an overview of Resource Description and Access (RDA), the new cataloging standard that will replace AACR2. It discusses what RDA is, why it was developed, how it relates to FRBR, and some of the key differences between RDA and AACR2, such as changes to terminology, transcription, and MARC fields. It also explores potential future directions for RDA, such as linked data and semantic web applications. The document concludes by offering suggestions for how libraries can prepare to implement RDA.
Intro to Linked Open Data in Libraries Archives & Museums.Jon Voss
This document discusses a presentation on Linked Open Data in libraries, archives, and museums. The presentation introduces Linked Open Data and how it is being used in cultural heritage institutions. It discusses representing data as graphs using triples and RDF, important vocabularies and ontologies, and following Tim Berners-Lee's principles of Linked Data. The presentation also covers legal and licensing considerations for publishing open cultural data on the web.
Libraries and Linked Data: Looking to the Future (2)ALATechSource
The document discusses options for new bibliographic frameworks after MARC. It describes three scenarios: 1) a relational/object-oriented RDA database, 2) linked bibliographic and authority records, and 3) flat files without links. It then discusses three approaches to implementing a new framework: 1) going native by using URIs for things, elements and values, 2) extracting data from existing MARC records, and 3) serializing data into key-value pairs, XML, or JSON. Advantages and disadvantages of each approach are outlined.
I used these slides in the context of a cultural heritage presentation so the examples are relevant to that community. For example the choice of CIDOC CRM is obvious in that community.
Introduction to linked open data, RDF: the Resource Description Framework, Tools to convert data to RDF, Tools for linking/reconciliation/resolution, Storing and maintaining the data, BBC and Linked Data
SPARQL1.1 Tutorial, given in UChile by Axel Polleres (DERI)net2-project
This document provides an introduction to SPARQL 1.1. It begins by explaining that SPARQL is a query language for the semantic web that allows users to query RDF data stores similarly to how SQL queries relational databases. It then describes SPARQL 1.0, the initial standard version, and the new features being added in SPARQL 1.1, including aggregate functions, subqueries, property paths and federated querying. The document concludes by discussing SPARQL implementations and the status of the 1.1 specification.
The document discusses data standards for describing cultural collections on the web. It advocates for using URIs, HTTP, and semantic web standards like RDF and SPARQL to provide structured data about cultural objects that is interlinked and can be queried. Alternative approaches like CSV, XML, and relational databases are discussed and their limitations explained. The benefits of a linked data approach for sharing cultural information on the web are emphasized.
This document provides an overview of linked data and the SPARQL query language. It defines linked data as a method of publishing structured data on the web so that it can be interlinked and queried. The key aspects covered include linked data principles of using URIs to identify things and including links to other related data. SPARQL is introduced as the query language for retrieving and manipulating linked data.
Semantic Web technologies such as RDF and OWL have become World Wide Web Consortium (W3C) standards for knowledge representation and reasoning. RDF triples about triples, or meta triples, form the basis for a contextualized knowledge graph. They represent the contextual information about individual triples such as the source, the occurring time or place, or the certainty.
However, an efficient RDF representation for such meta-knowledge of triples remains a major limitation of the RDF data model. The existing reification approach allows such meta-knowledge of RDF triples to be expressed in RDF by using four triples per reified triple. While reification is simple and intuitive, this approach does not have a formal foundation and is not commonly used in practice as described in the RDF Primer.
This dissertation presents the foundations for representing, querying, reasoning and traversing the contextualized knowledge graphs (CKG) using Semantic Web technologies.
A triple-based compact representation for CKGs. We propose a principled approach and construct RDF triples about triples by extending the current RDF data model with a new concept, called singleton property (SP), as a triple identifier. The SP representation needs two triples to the RDF datasets and can be queried with SPARQL.
A formal model-theoretic semantics for CKGs. We formalize the semantics of the singleton property and its relationships with the triple it represents. We extend the current RDF model-theoretic semantics to capture the semantics of the singleton properties and provide the interpretation at three levels: simple, RDF, and RDFS. It provides a single interpretation of the singleton property semantics across applications and systems.
A sound and complete inference mechanism for CKGs. Based on the semantics we propose, we develop a set of inference rules for validating and inferring new triples based on the SP syntax. We also develop different sets of context-based inference rules for provenance, time, and uncertainty.
A graph-based formalism for CKGs. We propose a formal contextualized graph model for the SP representation. We formalize the RDF triples as a mathematical graph by combining the model theory and the graph theory into a hybrid RDF formal semantics. The unified semantics allows the RDF formal semantics to be leveraged in the graph-based algorithms.
The document discusses linked data and its potential impact on libraries. It describes linked data as connecting the world's libraries by publishing structured data about 290 million resources using common schemas, embedding RDFa, and linking to controlled vocabularies. While linked data presents challenges like metadata for different types of materials, it offers opportunities to describe resources as part of the web and link catalog data to related concepts through identifiers.
The presentation discusses OCLC's efforts to create a cultural linked data backbone by connecting library data through open linked data standards. It provides examples of existing linked data projects like VIAF, Dewey, and WorldCat and outlines OCLC's work to publish 270 million WorldCat resources as linked open data using the Schema.org vocabulary. The presentation argues that linking library data in this way allows the world's libraries to be more connected on the web.
The presentation discusses linked data and its potential impact on libraries. Linked data uses URIs and HTTP to identify things on the web and link related resources. It enables libraries to publish structured data on the web and connect their resources to other data sources. While challenging, linked data offers a way for libraries to share data on the web of data and transition from cataloging to "catalinking", connecting their resources on the semantic web.
This document discusses the need for named graphs in RDF to represent contextual information like provenance and source of RDF data. It proposes extensions to the RDF/XML syntax to associate RDF descriptions and statements with named graphs. This allows modeling things like different hypotheses, temporal aspects, points of view, and distributed storage in a way that is currently not possible without named graphs in the RDF model.
The document discusses exposing library holdings data on the web using linked data. It notes that OCLC has exposed over 300 million resources using Schema.org, RDFa, and links to controlled vocabularies. The data is available via various formats like RDF/XML, JSON-LD and Turtle. BIBFRAME is presented as the new standard for bibliographic description that allows library data to be shared as part of the web. Libraries are encouraged to make their resources discoverable on the web of data by linking to other institutions and authorities.
TPDL2013 tutorial linked data for digital libraries 2013-10-22jodischneider
Tutorial on Linked Data for Digital Libraries, given by me, Uldis Bojars, and Nuno Lopes in Valletta, Malta at TPDL2013 on 2013-10-22.
http://tpdl2013.upatras.gr/tut-lddl.php
This half-day tutorial is aimed at academics and practitioners interested in creating and using Library Linked Data. Linked Data has been embraced as the way to bring complex information onto the Web, enabling discoverability while maintaining the richness of the original data. This tutorial will offer participants an overview of how digital libraries are already using Linked Data, followed by a more detailed exploration of how to publish, discover and consume Linked Data. The practical part of the tutorial will include hands-on exercises in working with Linked Data and will be based on two main case studies: (1) linked authority data and VIAF; (2) place name information as Linked Data.
For practitioners, this tutorial provides a greater understanding of what Linked Data is, and how to prepare digital library materials for conversion to Linked Data. For researchers, this tutorial updates the state of the art in digital libraries, while remaining accessible to those learning Linked
Data principles for the first time. For library and iSchool instructors, the tutorial provides a valuable introduction to an area of growing interest for information organization curricula. For digital library project managers, this tutorial provides a deeper understanding of the principles of Linked Data, which is needed for bespoke projects that involve data mapping and the reuse of existing metadata models.
RDA implementation is scheduled for March 31, 2013. Testers of RDA recommended improvements like rewriting instructions in plain English and ensuring community involvement. Differences from AACR2 include lack of abbreviations, more transcription of what is seen, and new fields in MARC like 336, 337, 338 for content/media/carrier types. Linked data and semantic web approaches may make relationships between works more explicit over time. Preparing for RDA involves decisions about cataloging workflows and training.
IFLA LIDASIG Open Session 2017: Introduction to Linked DataLars G. Svensson
At the IFLA Linked Data Special Interest Group open session in Wroclaw we briefly introduced the mission of the SIG and then went on to a brief introduction to what linked data is and why that topic is important to libraries.
The presentation was held jointly by Astrid Verheusen (general introduction to the SIG) and Lars G. Svensson (introduction to Linked Data)
Presentation at ELAG 2011, European Library Automation Group Conference, Prague, Czech Republic. 25th May 2011
http://elag2011.techlib.cz/en/815-lifting-the-lid-on-linked-data/
Talk at 3th Keystone Training School - Keyword Search in Big Linked Data - Institute for Software Technology and Interactive Systems, TU Wien, Austria, 2017
ESWC SS 2013 - Tuesday Tutorial 1 Maribel Acosta and Barry Norton: Providing ...eswcsummerschool
This document discusses providing linked data. It presents the core tasks of creating, interlinking, and publishing linked data based on linked data principles and lifecycles. Creating linked data involves extracting data, representing it using the RDF data model, assigning URIs to name things, and selecting appropriate vocabularies. Interlinking involves creating RDF links between different data sets at the instance and schema levels. Publishing linked data consists of making the data set and associated metadata accessible.
Usage of Linked Data: Introduction and Application ScenariosEUCLID project
This presentation introduces the main principles of Linked Data, the underlying technologies and background standards. It provides basic knowledge for how data can be published over the Web, how it can be queried, and what are the possible use cases and benefits. As an example, we use the development of a music portal (based on the MusicBrainz dataset), which facilitates access to a wide range of information and multimedia resources relating to music.
Providing open data is of interest for its societal and commercial value, for transparency, and because more people can do fun things with data. There is a growing number of initiatives to provide open data, from, for example, the UK government and the World Bank. However, much of this data is provided in formats such as Excel files, or even PDF files. This raises the question of
- How best to provide access to data so it can be most easily reused?
- How to enable the discovery of relevant data within the multitude of available data sets?
- How to enable applications to integrate data from large numbers of formerly unknown data sources?
One way to address these issues to to use the design principles of linked data (http://www.w3.org/DesignIssues/LinkedData.html), which suggest best practices for how to publish and connect structured data on the Web. This presentation gives an overview of linked data technologies (such as RDF and SPARQL), examples of how they can be used, as well as some starting points for people who want to provide and use linked data.
The presentation was given on August 8, at the Hacknight event (http://hacknight.se/) of Forskningsavdelningen (http://forskningsavd.se/) (Swedish: “Research Department”) a hackerspace in Malmö.
In October of 2011, the Library of Congress released a statement outlining its efforts to move away from the MARC 21 format and toward another carrier for library data. According to the statement, "Linked Data principles and mechanisms" will be the focus of this project. You may be wondering, what is Linked Data? What could it mean for our library catalogs? How do we create Linked Data? In this session, Emily Nimsakont, the NLC’s Cataloging Librarian, will answer those questions and more.
NCompass Live - Jan. 11, 2012.
J-P. Fauconnier, J. Roumier. Musonto - A Semantic Search Engine Dedicated to ...MusicNet
Jean-Philippe Fauconnier (Université Catholique de Louvain, Belgium) and Joseph Roumier (CETIC, Belgium).
Music Linked Data Workshop, 12 May 2011, JISC, London.
Radically Open Cultural Heritage Data on the WebJulie Allinson
What happens when tens of thousands of archival photos are shared with open licenses, then mashed up with geolocation data and current photos? Or when app developers can freely utilize information and images from millions of books? On this panel, we'll explore the fundamental elements of Linked Open Data and discover how rapidly growing access to metadata within the world's libraries, archives and museums is opening exciting new possibilities for understanding our past, and may help in predicting our future. Our panelists will look into the technological underpinnings of Linked Open Data, demonstrate use cases and applications, and consider the possibilities of such data for scholarly research, preservation, commercial interests, and the future of cultural heritage data.
This document discusses providing linked data. It covers the core tasks of creating, interlinking, and publishing linked data. For creating data, it describes extracting data, naming things with URIs, and selecting vocabularies. Interlinking involves creating RDF links between datasets using properties like owl:sameAs, rdfs:seeAlso, and SKOS mapping properties. Publishing linked data involves creating metadata to describe the dataset, making the data accessible, exposing it in repositories, and validating it.
This presentation covers the whole spectrum of Linked Data production and exposure. After a grounding in the Linked Data principles and best practices, with special emphasis on the VoID vocabulary, we cover R2RML, operating on relational databases, Open Refine, operating on spreadsheets, and GATECloud, operating on natural language. Finally we describe the means to increase interlinkage between datasets, especially the use of tools like Silk.
Publishing and Using Linked Open Data - Day 1 Richard Urban
This document provides an agenda and schedule for Monday's Linked Open Data class. The day includes introductions, sessions on introducing linked data and exploring use cases, breaks for discussion, and a concluding session on kicking off participant projects. Evening events include an outside lecture and networking social for graduate students.
Smart Interfaces through Domain Knowledge: Facets, Metadata Displays, Analysi...Charleston Conference
The document discusses improving displays of bibliographic data from MARC records. It provides examples of how MARC records are normally displayed versus improved displays that provide additional context and details. It also discusses using XSLT to transform MARCXML into HTML to produce more informative displays, and the benefits of libraries having tools to customize their displays.
Thinking of Linking: A random series of ideas, concepts, Platonic ideals, a y...Martin Kalfatovic
Thinking of Linking: A random series of ideas, concepts, Platonic ideals, a yeoman's miscellany, and nonesuch guide to Linked Data, especially as it relates to libraries, archives, and museums. Martin R. Kalfatovic. American Library Association Annual Meeting. Anaheim, CA. 23 June 2012.
DSNotify - Detecting and Fixing Broken Links in Linked Data SetsBernhard Haslhofer
The document discusses DSNotify, a system for detecting and fixing broken links in linked data sets. DSNotify periodically monitors resources in linked data sources, extracts descriptive features for each resource, and stores this in indices. It uses the features to detect if resources have been removed or moved, and adds relationships between old and new resources if moved. The system aims to maintain link integrity in linked data through automatically detecting and helping to fix broken links.
This document discusses how archives can use semantic web technologies like linked data to improve access to archival descriptions and resources. It provides background on the semantic web and linked data, and examples of how libraries are already using these approaches. While archival description standards like EAD currently focus on human-readable documents rather than linked data, the presenter argues the standards should evolve to represent information in a more computer-friendly and interoperable way, such as the emerging EAC standard. Overall, the presentation promotes the idea that archives can benefit from adopting semantic web best practices to better connect and expose archival information online.
Similar to S. Dixon, C. Mesnage, B. Norton. LinkedBrainz Live (20)
S. Dixon, C. Mesnage, B. Norton. LinkedBrainz Live
1. LinkedBrainz Live!
Live
Simon Dixon, Cedric Mesnage, Barry Norton
Centre for Digital Music, Queen Mary
SPARQL
+ =
RDF
RDFa
HTTP
URIs
2. MusicBrainz
• MusicBrainz is “an open music encyclopedia ...
collects & makes available... music metadata”
• It aims to be:
– The ultimate source of music information
by allowing anyone to contribute and releasing
the data under open licenses.
– The universal lingua franca for music
by providing a reliable and unambiguous form of music
identification, enabling both people & machines to have
meaningful conversations about music.
• It currently describes >9M tracks, soon to be >30M
LinkedBrainz Live, Dixon, Mesnage and Norton
2
Music Linked Data Workshop, 12th May 2011
3. Linked Data
• Defined in a W3C Technical Note including
these core principles:
1. Use URIs as names for things
2. Use HTTP URIs so that people can look up those
names.
3. When someone looks up a URI, provide useful
information, using the standards (RDF*, SPARQL)
4. Include links to other URIs. so that they can
discover more things.
LinkedBrainz Live, Dixon, Mesnage and Norton
3
Music Linked Data Workshop, 12th May 2011
4. Linked Open Data
• The Linking Open Data (LOD) project of the
W3C Semantic Web Outreach and Education
Task Force has
developed a
good deal of
best practice
and exposed
a large number
of interlinked datasets
LinkedBrainz Live, Dixon, Mesnage and Norton
4
Music Linked Data Workshop, 12th May 2011
5. Comparison of MusicBrainz & LOD 1
• MusicBrainz is a relational schema where
(internal) primary keys are externalised as
UUIDs, “Universally Unique” Identifiers
• Disambiguation easier if entity type included:
• artist/
1. Use URIs as names for things b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d
2. Use HTTP URIs so that people can look up (The Beatles)
those names.
3. When someone looks up a URI, provide useful • recording/
information, using the standards (RDF*, SPARQL) efe365d3-a1d8-42dc-a63d-0b4cc6ecfaed
4. Include links to other URIs. so that they can (‘With a Little Help from my Friends’)
discover more things.
LinkedBrainz Live, Dixon, Mesnage and Norton
5
Music Linked Data Workshop, 12th May 2011
6. Comparison of MusicBrainz & LOD 2
http://musicbrainz.org/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d
1. Use URIs as names for things
2. Use HTTP URIs so that people can look up
those names.
3. When someone looks up a URI, provide useful
information, using the standards (RDF*, SPARQL)
4. Include links to other URIs. so that they can
discover more things.
LinkedBrainz Live, Dixon, Mesnage and Norton
6
Music Linked Data Workshop, 12th May 2011
7. Comparison of MusicBrainz & LOD 3
<?xml version="1.0" encoding="UTF-8"?>
<metadata xmlns="http://musicbrainz.org/ns
/mmd-1.0#">
<artist id="b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d" type="Group">
<name>The Beatles</name>
<sort-name>Beatles, The</sort-name>
<life-span begin="1957" end="1970-04-10"/>
</artist>
</metadata>
http://musicbrainz.org/ws/1/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d?type=xml
http://musicbrainz.org/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d
1. Use URIs as names for things
2. Use HTTP URIs so that people can look up
those names.
3. When someone looks up a URI, provide useful
information, using the standards (RDF*, SPARQL)
4. Include links to other URIs. so that they can
discover more things.
LinkedBrainz Live, Dixon, Mesnage and Norton
7
Music Linked Data Workshop, 12th May 2011
8. Comparison of MusicBrainz & LOD 2’
• Non-informational resource:
http://musicbrainz.org/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d#_
• Informational resources:
http://musicbrainz.org/ws/1/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d?type=rdf
http://musicbrainz.org/ws/1/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d?type=xml
http://musicbrainz.org/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d
1. Use URIs as names for things
2. Use HTTP URIs so that people can look up
those names.
3. When someone looks up a URI, provide useful
information, using the standards (RDF*, SPARQL)
4. Include links to other URIs. so that they can
discover more things.
LinkedBrainz Live, Dixon, Mesnage and Norton
8
Music Linked Data Workshop, 12th May 2011
9. Comparison of MusicBrainz & LOD 4
1. Use URIs as names for things
2. Use HTTP URIs so that people can look up
those names.
3. When someone looks up a URI, provide useful
information, using the standards (RDF*, SPARQL)
4. Include links to other URIs. so that they can
discover more things.
LinkedBrainz Live, Dixon, Mesnage and Norton
9
Music Linked Data Workshop, 12th May 2011
10. MB Next Generation Schema
• Artist
As before, but further
attributes
• Artist Credit
Allows joint credit
• Release Group
Generalises
‘release type’ (cf. album) entity, cf.
• Release • Track • Work
• Medium • Track List • Recording
LinkedBrainz Live, Dixon, Mesnage and Norton
10
Music Linked Data Workshop, 12th May 2011
11. Music Ontology
LinkedBrainz Live, Dixon, Mesnage and Norton
11
Music Linked Data Workshop, 12th May 2011
12. Music Ontology
d2rq:ClassMap
d2rq:ClassMap
map:Recording a d2rq:ClassMap;
d2rq:dataStorage map:database;
map:recording_name a d2rq:PropertyBridge;
d2rq:class mo:Signal;
d2rq:belongsToClassMap map:Recording;
d2rq:classDefinitionLabela"Signal"@en;
map:recording_credit d2rq:PropertyBridge;
d2rq:property dct:title, rdfs:label;
... map:DBpediaTrans a d2rq:TranslationTable;
d2rq:uriPattern "recording/@@musicbrainz.recording.gid@@#_".
d2rq:join "musicbrainz.recording.name =>
d2rq:join "musicbrainz.recording.artist_credit =>
d2rq:javaClass "translators.WikipediaTranslator".
musicbrainz.track_name.id";
musicbrainz.artist_credit.id";
d2rq:column "musicbrainz.track_name.name".
d2rq:join "musicbrainz.artist_credit_name.artist_credit =>
map:artist_dbpedia a d2rq:PropertyBridge;
musicbrainz.artist_credit.id";
... d2rq:property owl:sameAs;
d2rq:join "musicbrainz.artist_credit_name.artist =>
... d2rq:condition "musicbrainz.link_type.id=213";
musicbrainz.artist.id". map:DBpediaTrans.
LinkedBrainz Live, Dixon, Mesnage and Norton
... d2rq:translateWith 12
Music Linked Data Workshop, 12th May 2011
13. RDFa Roll-out
<div class="recordingheader"
about="http://ngs.musicbrainz.org/recording/efe365d3-a1d8-42dc-a63d-0b4cc6ecfaed#_"
typeof="mo:Signal">
<h1>“<a rel="foaf:isPrimaryTopicOf"
href="http://ngs.musicbrainz.org/recording/efe365d3-a1d8-42dc-a63d-0b4cc6ecfaed"
property="dct:title rdfs:label" xml:lang="" datatype="">
With a Little Help From My Friends</a>” </h1>
<p class="subheader"> <span class="prefix">~</span>
Recording by
<a title="Beatles, The" rel="foaf:maker“
resource="[mbz:artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d#_]“
LinkedBrainz Live, Dixon, Mesange and Norton
href="http://ngs.musicbrainz.org/artist/b10bbbfc-cf9e-42e0-be17-e2c3e1d2600d"> 13
Music Linked Data Workshop, 12th May 2011
The Beatles</a></p></div>
14. Future Work
• Complete D2R mappings
– Regular dumps to be made available
– Deferencing to RDF from (NI) MB URIs
– SPARQL endpoint
– Completion of mappings (advanced relationships,
interlinkage to further datasets), consistency testing
• Tutorials on using LinkedBrainz
Summer School (23rd May)
Linked Services Tutorial (30th May)
LinkedBrainz Live, Dixon, Mesnage and Norton
14
Music Linked Data Workshop, 12th May 2011