Presented at the seminar Libraries and the Semantic Web: the role of International Standard Bibliographic Description (ISBD), National Library of Scotland, Edinburgh, 25 Feb 2011
This document provides an introduction to the RDF data model. It describes RDF as a data model that represents data as subject-predicate-object triples that can be used to describe resources. These triples form a directed graph. The document provides examples of RDF triples and graphs, and compares the RDF data model to relational and XML data models. It also describes common RDF formats like RDF/XML, Turtle, N-Triples, and how RDF graphs from different sources can be merged.
This document discusses rules and the Semantic Web Rule Language (SWRL). It defines rules as a means of representing knowledge similar to if-then statements. SWRL combines OWL and rule-based languages by allowing users to write rules that can refer to OWL classes, properties, individuals and datatypes. SWRL has an abstract and XML syntax and supports built-in predicates for manipulating data types. Rules provide more expressivity than RDFS and OWL in some cases, such as defining application behaviors, but rule-based reasoning is less performant so they should not be overused when RDFS/OWL suffice.
This document discusses RDFS semantics, inference techniques, and using RDFS inference with Sesame. It covers the core concepts of reasoning engines and RDFS rule-based semantics. It describes implementing RDFS semantics using forward chaining and backward chaining inference techniques. It also provides an overview of how RDFS inference is implemented in Sesame using a forward chaining SAIL that performs inferences on load.
VALA Tech Camp 2017: Intro to Wikidata & SPARQLJane Frazier
A hands-on introduction to interrogation of Wikidata content using SPARQL, the query language used to query data represented in RDF, SKOS, OWL, and other Semantic Web standards.
Presented by myself and Peter Neish, Research Data Specialist @ University of Melbourne.
RDF is a general-purpose language for representing information on the web. It allows for describing resources and the relationships between them using subject-predicate-object expressions called triples. RDF is used as a foundation for the semantic web and allows machines to mechanically process and interpret the logical pieces of meaning in data. While RDF does not define specific properties or vocabularies, it provides mechanisms for describing properties and classes of resources in a human- or machine-readable format.
The document discusses test-driven quality assessment of RDF data. It proposes a methodology called the Test-driven Quality Assessment Methodology (TDQAM) where test cases are generated automatically from the RDF schema to validate data constraints. Test cases are written as SPARQL queries and can check for issues like a person having a birthdate after a deathdate. Pattern-based test generators analyze the schema to instantiate test cases. The methodology provides a unified way to validate RDF data against different schema languages to improve data quality.
Using NLP to Explore Entity Relationships in COVID-19 LiteratureDatabricks
In this talk, we will cover how to extract entities from text using both rule-based and deep learning techniques. We will also cover how to use rule-based entity extraction to bootstrap a named entity recognition model. The other important aspect of this project we will cover is how to infer relationships between entities, and combine them with explicit relationships found in the source data sets. Although this talk is focused on the CORD-19 data set, the techniques covered are applicable to a wide variety of domains. This talk is for those who want to learn how to use NLP to explore relationships in text.
RDF is a general method to decompose knowledge into small pieces, with some rules about the semantics or meaning of those pieces. The point is to have a method so simple that it can express any fact, and yet so structured that computer applications can do useful things with knowledge expressed in RDF.
This document provides an introduction to the RDF data model. It describes RDF as a data model that represents data as subject-predicate-object triples that can be used to describe resources. These triples form a directed graph. The document provides examples of RDF triples and graphs, and compares the RDF data model to relational and XML data models. It also describes common RDF formats like RDF/XML, Turtle, N-Triples, and how RDF graphs from different sources can be merged.
This document discusses rules and the Semantic Web Rule Language (SWRL). It defines rules as a means of representing knowledge similar to if-then statements. SWRL combines OWL and rule-based languages by allowing users to write rules that can refer to OWL classes, properties, individuals and datatypes. SWRL has an abstract and XML syntax and supports built-in predicates for manipulating data types. Rules provide more expressivity than RDFS and OWL in some cases, such as defining application behaviors, but rule-based reasoning is less performant so they should not be overused when RDFS/OWL suffice.
This document discusses RDFS semantics, inference techniques, and using RDFS inference with Sesame. It covers the core concepts of reasoning engines and RDFS rule-based semantics. It describes implementing RDFS semantics using forward chaining and backward chaining inference techniques. It also provides an overview of how RDFS inference is implemented in Sesame using a forward chaining SAIL that performs inferences on load.
VALA Tech Camp 2017: Intro to Wikidata & SPARQLJane Frazier
A hands-on introduction to interrogation of Wikidata content using SPARQL, the query language used to query data represented in RDF, SKOS, OWL, and other Semantic Web standards.
Presented by myself and Peter Neish, Research Data Specialist @ University of Melbourne.
RDF is a general-purpose language for representing information on the web. It allows for describing resources and the relationships between them using subject-predicate-object expressions called triples. RDF is used as a foundation for the semantic web and allows machines to mechanically process and interpret the logical pieces of meaning in data. While RDF does not define specific properties or vocabularies, it provides mechanisms for describing properties and classes of resources in a human- or machine-readable format.
The document discusses test-driven quality assessment of RDF data. It proposes a methodology called the Test-driven Quality Assessment Methodology (TDQAM) where test cases are generated automatically from the RDF schema to validate data constraints. Test cases are written as SPARQL queries and can check for issues like a person having a birthdate after a deathdate. Pattern-based test generators analyze the schema to instantiate test cases. The methodology provides a unified way to validate RDF data against different schema languages to improve data quality.
Using NLP to Explore Entity Relationships in COVID-19 LiteratureDatabricks
In this talk, we will cover how to extract entities from text using both rule-based and deep learning techniques. We will also cover how to use rule-based entity extraction to bootstrap a named entity recognition model. The other important aspect of this project we will cover is how to infer relationships between entities, and combine them with explicit relationships found in the source data sets. Although this talk is focused on the CORD-19 data set, the techniques covered are applicable to a wide variety of domains. This talk is for those who want to learn how to use NLP to explore relationships in text.
RDF is a general method to decompose knowledge into small pieces, with some rules about the semantics or meaning of those pieces. The point is to have a method so simple that it can express any fact, and yet so structured that computer applications can do useful things with knowledge expressed in RDF.
The Semantic Web is the extension of the World Wide Web that enables people to share content beyond the boundaries of applications and websites.
It has been described in rather different ways: as a utopic vision, as a web of data, or merely as a natural paradigm shift in our daily use of the Web.
Most of all, the Semantic Web has inspired and engaged many people to create innovative and intelligent technologies and applications.
In this presentation we describe the underlying principles and key features of the semantic web along with where and how they fit in with server side and client side technologies supported by ColdFusion.
The document discusses the Semantic Web and Resource Description Framework (RDF). It defines the Semantic Web as making web data machine-understandable by describing web resources with metadata. RDF uses triples to describe resources, properties, and relationships. RDF data can be visualized as a graph and serialized in formats like RDF/XML. RDF Schema (RDFS) provides a basic vocabulary for defining classes, properties, and hierarchies to enable reasoning about RDF data.
This document provides an overview of the Resource Description Framework (RDF). It begins with background information on RDF including URIs, URLs, IRIs and QNames. It then describes the RDF data model, noting that RDF is a schema-less data model featuring unambiguous identifiers and named relations between pairs of resources. It also explains that RDF graphs are sets of triples consisting of a subject, predicate and object. The document also covers RDF syntax using Turtle and literals, as well as modeling with RDF. It concludes with a brief overview of common RDF tools including Jena.
Lightening talk for Semantic Web in Libraries (SWIB13) conference at 2013-11-27 about another method of expressing RDF data. See http://gbv.github.io/aREF/ for a preliminary specification.
Efficient Query Answering against Dynamic RDF DatabasesAlexandra Roatiș
The document describes efficient query answering against dynamic RDF databases. It discusses RDF as a graph-based data model and standard, blank nodes, RDF Schema (RDFS) for semantic constraints, the open-world assumption and RDF entailment through implicit triples and saturation. It also covers basic graph pattern (BGP) queries in SPARQL and the need to decouple RDF entailment from query evaluation through data saturation or query reformulation to obtain complete query answers.
The document discusses different methods for embedding semantic information in HTML pages, including microformats which use standard HTML tags and attributes to describe information, and RDFa which uses XML attributes to carry metadata. It provides examples of using microformats and RDFa to add semantic annotations and explores tools for processing and consuming these semantic data formats. The conclusion questions which technology, microformats or RDFa, web developers will embrace more in the future for adding semantics to web pages.
This document discusses different data structures including records, files, and trees. It provides the following information:
1) A record is a collection of fields of different data types that can be used to store related data. Fields are also called members. Records can be created in Python using namedtuple.
2) A file is a sequence of characters that is stored on secondary storage. There are different types of files like text files. Basic file operations in Python include opening, closing, reading and writing to files.
3) A tree is a nonlinear data structure that stores nodes hierarchically. A binary tree stores nodes with at most two children. Trees in Python can be implemented as a Node class with left and right
This document provides an overview of describing web resources using the Resource Description Framework (RDF). It discusses the basic concepts of RDF including resources, properties, statements, and the XML syntax used to represent them. It also covers RDF Schema which adds vocabulary for describing properties and classes of RDF resources, and provides a critical view of some aspects of RDF such as its use of binary predicates and treatment of properties.
RDFS can be used to define taxonomies and ontologies for organizing information. It allows defining relationships like broader/narrower terms which enables enhanced search capabilities and faceted browsing. Large, curated taxonomies in domains like biology have been developed and are widely used to power semantic annotation services and applications that deliver benefits like improved content management, search, and knowledge reuse.
2013 CrossRef Workshops System Update: CrossMark, Fundref and LicenseRef Car...Crossref
This document discusses the proper placement and handling of Crossmark, Fundref, and License Ref elements when depositing and updating XML documents. It notes that:
- These elements can only exist in certain places within a deposit, with Crossmark containing any Fundref or License Ref elements.
- When updating a document, any existing Crossmark, Fundref, or License Ref elements on file must be preserved or merged with new elements as needed.
- Crossmark, Fundref, and License Ref can now be deposited individually, requiring processing of nine possible combinations when handling deposits.
- The document outlines the step-by-step flow for accepting, verifying, processing, and properly placing these elements when
Theory behind Image Compression and Semantic SearchSanti Adavani
Singular Value Decomposition (SVD) is a matrix decomposition technique developed during the 18th century and has been in use ever since. SVD has applications in several areas including image processing, natural language processing (NLP), genomics, and data compression. In NLP context, SVD is called latent semantic indexing (LSI) and used for concept based search and topic modeling. In this talk, we will describe the math and intuition behind eigenvalues, eigenvectors and their relation to SVD. We will also discuss specific applications of SVD in image processing and NLP with examples.
Best practices for generating Bio2RDF linked dataalison.callahan
This document provides best practices for generating linked data, including assigning URIs and identifiers, assigning labels, declaring types, and providing dataset provenance. It discusses assigning a URI to all entities using a base URI pattern and maintaining source identifiers. It also covers assigning labels, identifiers, and types using RDF and OWL predicates according to patterns. Finally, it discusses linking resources to their dataset using void:inDataset.
This document discusses RDFS (Resource Description Framework Schema), which is a standard ontology language for the Semantic Web. RDFS introduces predefined meanings for resources through axioms and allows for basic inferences over RDF data through mechanisms like type propagation between classes and properties. The document provides examples of how RDFS can be used to classify resources in an RDF graph and automatically infer additional types for resources based on their properties and class memberships.
Indexing, searching, and aggregation with redi search and .netStephen Lorello
This document discusses indexing, searching, and aggregation in Redis using RediSearch and .NET. It provides an introduction to Redis data structures and building secondary indices. It then covers using RediSearch to define schemas, query data through full text search and filters, and perform aggregations through grouping, reductions, and applying functions. RediSearch provides an easier way to index and query Redis compared to building secondary indices in vanilla Redis.
A view on data quality in the real estate domain.
Presented at the LDQ workshop, colocated with SEMANTICS 2017 conference.
see https://2017.semantics.cc/satellite-events/linked-data-quality-assessment-and-improvement-academia-industry
for more details
The document discusses scaling web data at low cost. It begins by presenting Javier D. Fernández and providing context about his work in semantic web, open data, big data management, and databases. It then discusses techniques for compressing and querying large RDF datasets at low cost using binary RDF formats like HDT. Examples of applications using these techniques include compressing and sharing datasets, fast SPARQL querying, and embedding systems. It also discusses efforts to enable web-scale querying through projects like LOD-a-lot that integrate billions of triples for federated querying.
First steps towards publishing library data on the semantic webhorvadam
First steps towards publishing library data on the semantic web. Implementing:
CoolUri
RDFDC
SKOS
RDF database and SPARQL interface
Content negotiation
Two graph data models : RDF and Property Graphsandyseaborne
This document provides an overview of two graph data models: RDF and Property Graphs. It describes the key components of each model, including triples for RDF and nodes/edges/properties for Property Graphs. It also discusses Apache projects that work with each model like Apache Jena for RDF and Apache TinkerPop, Spark, Giraph and Flink for Property Graphs. Finally, it notes that while the models have different focuses, they could potentially share technologies like storage and query capabilities.
What's been happening in the world of cataloguing and educational metadata? 2013 saw the introduction of a new cataloguing standard known as Resource Description and Access (RDA) to replace AACR2. This webinar provides an overview of cataloguing changes for school library staff.
Why a new cataloguing standard?
The concepts underpinning RDA
The implication for resource discovery in schools
Impacts on Library Management Systems
The presentation is relevant for all school staff, whether they are trained as cataloguers or not, and will assist library staff understand the catalogue records they receive from SCIS.
This document describes the British Library Data Model, which defines classes, properties and relationships for representing bibliographic metadata and linked data. It shows how concepts like works, expressions, manifestations, and items are modeled, along with authors, subjects, and publication events. Properties are defined to link these concepts and describe their relationships according to standards like FRBR, RDA, and SKOS.
Resource Discovery Infrastructure - what if we were starting from scratch?ostephens
This document discusses the history and evolution of library cataloging and metadata standards over time. It mentions several important standards, systems, and leaders in the field such as Ranganathan's 5 Laws of Library Science, MARC, AACR, RDA, Z39.50, and OCLC. It also acknowledges several important figures and resources that advanced the field such as Melvil Dewey, Charles Cutter, and Sir Tim Berners-Lee through the development of new technologies.
The Semantic Web is the extension of the World Wide Web that enables people to share content beyond the boundaries of applications and websites.
It has been described in rather different ways: as a utopic vision, as a web of data, or merely as a natural paradigm shift in our daily use of the Web.
Most of all, the Semantic Web has inspired and engaged many people to create innovative and intelligent technologies and applications.
In this presentation we describe the underlying principles and key features of the semantic web along with where and how they fit in with server side and client side technologies supported by ColdFusion.
The document discusses the Semantic Web and Resource Description Framework (RDF). It defines the Semantic Web as making web data machine-understandable by describing web resources with metadata. RDF uses triples to describe resources, properties, and relationships. RDF data can be visualized as a graph and serialized in formats like RDF/XML. RDF Schema (RDFS) provides a basic vocabulary for defining classes, properties, and hierarchies to enable reasoning about RDF data.
This document provides an overview of the Resource Description Framework (RDF). It begins with background information on RDF including URIs, URLs, IRIs and QNames. It then describes the RDF data model, noting that RDF is a schema-less data model featuring unambiguous identifiers and named relations between pairs of resources. It also explains that RDF graphs are sets of triples consisting of a subject, predicate and object. The document also covers RDF syntax using Turtle and literals, as well as modeling with RDF. It concludes with a brief overview of common RDF tools including Jena.
Lightening talk for Semantic Web in Libraries (SWIB13) conference at 2013-11-27 about another method of expressing RDF data. See http://gbv.github.io/aREF/ for a preliminary specification.
Efficient Query Answering against Dynamic RDF DatabasesAlexandra Roatiș
The document describes efficient query answering against dynamic RDF databases. It discusses RDF as a graph-based data model and standard, blank nodes, RDF Schema (RDFS) for semantic constraints, the open-world assumption and RDF entailment through implicit triples and saturation. It also covers basic graph pattern (BGP) queries in SPARQL and the need to decouple RDF entailment from query evaluation through data saturation or query reformulation to obtain complete query answers.
The document discusses different methods for embedding semantic information in HTML pages, including microformats which use standard HTML tags and attributes to describe information, and RDFa which uses XML attributes to carry metadata. It provides examples of using microformats and RDFa to add semantic annotations and explores tools for processing and consuming these semantic data formats. The conclusion questions which technology, microformats or RDFa, web developers will embrace more in the future for adding semantics to web pages.
This document discusses different data structures including records, files, and trees. It provides the following information:
1) A record is a collection of fields of different data types that can be used to store related data. Fields are also called members. Records can be created in Python using namedtuple.
2) A file is a sequence of characters that is stored on secondary storage. There are different types of files like text files. Basic file operations in Python include opening, closing, reading and writing to files.
3) A tree is a nonlinear data structure that stores nodes hierarchically. A binary tree stores nodes with at most two children. Trees in Python can be implemented as a Node class with left and right
This document provides an overview of describing web resources using the Resource Description Framework (RDF). It discusses the basic concepts of RDF including resources, properties, statements, and the XML syntax used to represent them. It also covers RDF Schema which adds vocabulary for describing properties and classes of RDF resources, and provides a critical view of some aspects of RDF such as its use of binary predicates and treatment of properties.
RDFS can be used to define taxonomies and ontologies for organizing information. It allows defining relationships like broader/narrower terms which enables enhanced search capabilities and faceted browsing. Large, curated taxonomies in domains like biology have been developed and are widely used to power semantic annotation services and applications that deliver benefits like improved content management, search, and knowledge reuse.
2013 CrossRef Workshops System Update: CrossMark, Fundref and LicenseRef Car...Crossref
This document discusses the proper placement and handling of Crossmark, Fundref, and License Ref elements when depositing and updating XML documents. It notes that:
- These elements can only exist in certain places within a deposit, with Crossmark containing any Fundref or License Ref elements.
- When updating a document, any existing Crossmark, Fundref, or License Ref elements on file must be preserved or merged with new elements as needed.
- Crossmark, Fundref, and License Ref can now be deposited individually, requiring processing of nine possible combinations when handling deposits.
- The document outlines the step-by-step flow for accepting, verifying, processing, and properly placing these elements when
Theory behind Image Compression and Semantic SearchSanti Adavani
Singular Value Decomposition (SVD) is a matrix decomposition technique developed during the 18th century and has been in use ever since. SVD has applications in several areas including image processing, natural language processing (NLP), genomics, and data compression. In NLP context, SVD is called latent semantic indexing (LSI) and used for concept based search and topic modeling. In this talk, we will describe the math and intuition behind eigenvalues, eigenvectors and their relation to SVD. We will also discuss specific applications of SVD in image processing and NLP with examples.
Best practices for generating Bio2RDF linked dataalison.callahan
This document provides best practices for generating linked data, including assigning URIs and identifiers, assigning labels, declaring types, and providing dataset provenance. It discusses assigning a URI to all entities using a base URI pattern and maintaining source identifiers. It also covers assigning labels, identifiers, and types using RDF and OWL predicates according to patterns. Finally, it discusses linking resources to their dataset using void:inDataset.
This document discusses RDFS (Resource Description Framework Schema), which is a standard ontology language for the Semantic Web. RDFS introduces predefined meanings for resources through axioms and allows for basic inferences over RDF data through mechanisms like type propagation between classes and properties. The document provides examples of how RDFS can be used to classify resources in an RDF graph and automatically infer additional types for resources based on their properties and class memberships.
Indexing, searching, and aggregation with redi search and .netStephen Lorello
This document discusses indexing, searching, and aggregation in Redis using RediSearch and .NET. It provides an introduction to Redis data structures and building secondary indices. It then covers using RediSearch to define schemas, query data through full text search and filters, and perform aggregations through grouping, reductions, and applying functions. RediSearch provides an easier way to index and query Redis compared to building secondary indices in vanilla Redis.
A view on data quality in the real estate domain.
Presented at the LDQ workshop, colocated with SEMANTICS 2017 conference.
see https://2017.semantics.cc/satellite-events/linked-data-quality-assessment-and-improvement-academia-industry
for more details
The document discusses scaling web data at low cost. It begins by presenting Javier D. Fernández and providing context about his work in semantic web, open data, big data management, and databases. It then discusses techniques for compressing and querying large RDF datasets at low cost using binary RDF formats like HDT. Examples of applications using these techniques include compressing and sharing datasets, fast SPARQL querying, and embedding systems. It also discusses efforts to enable web-scale querying through projects like LOD-a-lot that integrate billions of triples for federated querying.
First steps towards publishing library data on the semantic webhorvadam
First steps towards publishing library data on the semantic web. Implementing:
CoolUri
RDFDC
SKOS
RDF database and SPARQL interface
Content negotiation
Two graph data models : RDF and Property Graphsandyseaborne
This document provides an overview of two graph data models: RDF and Property Graphs. It describes the key components of each model, including triples for RDF and nodes/edges/properties for Property Graphs. It also discusses Apache projects that work with each model like Apache Jena for RDF and Apache TinkerPop, Spark, Giraph and Flink for Property Graphs. Finally, it notes that while the models have different focuses, they could potentially share technologies like storage and query capabilities.
What's been happening in the world of cataloguing and educational metadata? 2013 saw the introduction of a new cataloguing standard known as Resource Description and Access (RDA) to replace AACR2. This webinar provides an overview of cataloguing changes for school library staff.
Why a new cataloguing standard?
The concepts underpinning RDA
The implication for resource discovery in schools
Impacts on Library Management Systems
The presentation is relevant for all school staff, whether they are trained as cataloguers or not, and will assist library staff understand the catalogue records they receive from SCIS.
This document describes the British Library Data Model, which defines classes, properties and relationships for representing bibliographic metadata and linked data. It shows how concepts like works, expressions, manifestations, and items are modeled, along with authors, subjects, and publication events. Properties are defined to link these concepts and describe their relationships according to standards like FRBR, RDA, and SKOS.
Resource Discovery Infrastructure - what if we were starting from scratch?ostephens
This document discusses the history and evolution of library cataloging and metadata standards over time. It mentions several important standards, systems, and leaders in the field such as Ranganathan's 5 Laws of Library Science, MARC, AACR, RDA, Z39.50, and OCLC. It also acknowledges several important figures and resources that advanced the field such as Melvil Dewey, Charles Cutter, and Sir Tim Berners-Lee through the development of new technologies.
The document discusses trends in digital media and mobile technology and their implications for libraries. It notes that e-readers like Kindle are growing in popularity, more people have mobile phones, and mobile networks are expanding access in developing areas. While print is at risk, this also presents opportunities for libraries to rethink their services and focus on engaging experiences rather than just collections. The future of libraries depends on adapting to changing formats and finding new ways to contribute to knowledge production and consumption in their communities.
Dokumen tersebut membahas tentang standar pengatalogan bahan perpustakaan, yaitu AACR2 dan RDA. Ia menjelaskan struktur, perbandingan, dan contoh penerapan kedua standar tersebut dalam format MARC21.
This document introduces the fourth edition of The Concise AACR2, which provides a simplified version of the Anglo-American Cataloguing Rules (AACR2) intended for cataloguing students, librarians in small libraries, and others who do not need the full complexity and detail of AACR2. The Concise AACR2 presents the commonly applicable descriptive cataloging rules from AACR2 in a rewritten, simplified form with new examples, while referring users to AACR2 for problems not covered and for fuller explanations. It aims to convey the essence and basic principles of AACR2 cataloging practice in a more accessible way.
The document provides instructions for cataloging different types of library materials according to established standards, including books, audiovisual materials, maps, computer software, sound recordings, kits, filmstrips, and filing rules for manual and dictionary card catalogs. Key details covered include sources of information, main entry rules, and physical description areas for different material types.
Keynote address (Feb, 2016) to the educators in the Fort Nelson school district. We all know that we cannot teach a child without a concection... without a relationship. In the hustle and bustle of our jobs as educators, we often forget our why, the reason we got into education, of trying to make a difference with kids. In this talk, 6 Keys to Connecting are shared and discussed with the challenge of creating a more positive climate and better connections with kids in our classrooms, schools, and organizations.
With the explosion of the maker movement, schools are beginning to embrace creativity. However, what does this mean for assessment? Should we assess the creative process? Should we assess the finished product? Does assessing creativity actually make kids more risk-averse? In this workshop we explore what it means to assess both the creative process and the creative product without leading to risk aversion.
Artificial intelligence (AI) is everywhere, promising self-driving cars, medical breakthroughs, and new ways of working. But how do you separate hype from reality? How can your company apply AI to solve real business problems?
Here’s what AI learnings your business should keep in mind for 2017.
The International Federation of Library Associations and Institutions (IFLA) is responsible for the development and maintenance of International Standard Bibliographic Description (ISBD), UNIMARC, and the "Functional Requirements" family for bibliographic records (FRBR), authority data (FRAD), and subject authority data (FRSAD). ISBD underpins the MARC family of formats used by libraries world-wide for many millions of catalog records, while FRBR is a relatively new model optimized for users and the digital environment. These metadata models, schemas, and content rules are now being expressed in the Resource Description Framework language for use in the Semantic Web.
This webinar provides a general update on the work being undertaken. It describes the development of an Application Profile for ISBD to specify the sequence, repeatability, and mandatory status of its elements. It discusses issues involved in deriving linked data from legacy catalogue records based on monolithic and multi-part schemas following ISBD and FRBR, such as the duplication which arises from copy cataloging and FRBRization. The webinar provides practical examples of deriving high-quality linked data from the vast numbers of records created by libraries, and demonstrates how a shift of focus from records to linked-data triples can provide more efficient and effective user-centered resource discovery services.
Presentation at ELAG 2011, European Library Automation Group Conference, Prague, Czech Republic. 25th May 2011
http://elag2011.techlib.cz/en/815-lifting-the-lid-on-linked-data/
These slides were presented as part of a W3C tutorial at the CSHALS 2010 conference (http://www.iscb.org/cshals2010). The slides are adapted from a longer introduction to the Semantic Web available at http://www.slideshare.net/LeeFeigenbaum/semantic-web-landscape-2009 .
A PDF version of the slides is available at http://thefigtrees.net/lee/sw/cshals/cshals-w3c-semantic-web-tutorial.pdf .
This document discusses how ISBD (International Standard Bibliographic Description) metadata can be expressed using semantic web standards like RDF, URIs, and namespaces to make the metadata machine-readable. Key points include: metadata is expressed as subject-predicate-object triples using URIs; namespaces provide bases for URIs and can be abbreviated; common vocabularies define predicates and properties for labels, scopes, etc.; and an application profile specifies how a useful "record" can be constructed by defining which triples from which namespaces are involved, their sequence, repeatability, and mandatory status.
Lotus: Linked Open Text UnleaShed - ISWC COLD '15Filip Ilievski
The document describes LOTUS, a system for finding Linked Open Data (LOD) resources based on natural text queries. LOTUS indexes over 5 billion text literals from the LOD Laundromat. It supports various query modes like phrase matching and term matching. Evaluation shows LOTUS retrieves more resources compared to SPARQL and previous approaches. Current work involves adding language tags and improving ranking. LOTUS provides a single access point for exploring the LOD cloud via natural text queries.
RDA (Resource Description and Access) is a new standard for describing library resources, designed to replace AACR2. Library staff, including public services, systems personnel, and catalogers, may have heard mention of RDA but not know much about it or how it will change their daily work. You may have many questions. What is RDA? We'll give a very little bit of history and theoretical background. What is this going to mean for catalogers, ILS managers, and users in the near term? What are the future implications, or, why are we doing this? What are the juicy bits of controversy in cataloger-land? And finally, Do we HAVE to? We'll talk for a while, have some activities that get you thinking, and find out your thoughts on RDA.
Presented at "Captains & Crew Collaborating," the 8th annual paraprofessional conference at J.Y. Joyner Library, East Carolina University.
The document discusses lessons learned in transforming metadata from XML formats to RDF. It describes how libraries and cultural heritage institutions are working to express existing metadata standards like MODS and PBCore in RDF to take advantage of capabilities like linked data. Challenges include mapping XML schemas to RDF ontologies and ensuring RDF can meet identified use cases. Examples are provided of institutions that have transformed metadata to RDF to share across systems or publish as linked open data.
SPARQL1.1 Tutorial, given in UChile by Axel Polleres (DERI)net2-project
This document provides an introduction to SPARQL 1.1. It begins by explaining that SPARQL is a query language for the semantic web that allows users to query RDF data stores similarly to how SQL queries relational databases. It then describes SPARQL 1.0, the initial standard version, and the new features being added in SPARQL 1.1, including aggregate functions, subqueries, property paths and federated querying. The document concludes by discussing SPARQL implementations and the status of the 1.1 specification.
The document discusses the concepts and implementation of linked data and the semantic web. It describes Cambridge University Library's COMET project which converted bibliographic records from MARC21 format to RDF triples and published them as linked open data with HTTP URIs. The project aimed to release data for open use and gain experience working with semantic web technologies like RDF, SPARQL and triplestores. Key challenges included dealing with IPR issues in MARC21 records and developing tools to transform and link the data.
The document discusses plans by the Japan Link Center (JaLC) to expand its Digital Object Identifier (DOI) registration services. Currently JaLC registers DOIs for journal articles and will soon add additional content types like books, theses, reports, and research data. To gain experience with registering DOIs for research data, JaLC will conduct an experimental project involving participant organizations. The project aims to establish workflows for stable research data DOI registration and integration with DataCite standards. Testing is scheduled to begin in the fall. The addition of data and other content will help JaLC further its goal of supporting all researcher activities through persistent identification with DOIs.
... or how to query an RDF graph with 28 billion triples in a standard laptop
These slides correspond to my talk at the Stanford Center for Biomedical Informatics, on 25th April 2018
Enriching the semantic web tutorial session 1Tobias Wunner
The document discusses challenges and opportunities in natural language processing for the multilingual semantic web. It provides examples of how content on the web and semantic web exhibits linguistic variations within and across languages. It also summarizes several NLP applications like information extraction and natural language generation that utilize ontologies, and notes that these applications require domain and multilingual adaptation of lexicons and extraction rules. The document argues that efficient adaptation and sharing of linguistic resources between ontology-based NLP applications is needed.
The document discusses the Semantic Web and Linked Data. It provides an overview of RDF syntaxes, storage and querying technologies for the Semantic Web. It also discusses issues around scalability and reasoning over large amounts of semantic data. Examples are provided to illustrate SPARQL querying of RDF data, including graph patterns, conjunctions, optional patterns and value testing.
Full version of http://www.slideshare.net/valexiev1/gvp-lodcidocshort. Same is available on http://vladimiralexiev.github.io/pres/20140905-CIDOC-GVP/index.html
CIDOC Congress, Dresden, Germany
2014-09-05: International Terminology Working Group: full version.
2014-09-09: Getty special session: short version
The document discusses the semantic web and case-based reasoning. It provides an overview of key concepts like ontology languages, RDF, OWL, and describes how case-based reasoning works and how it can be applied to the semantic web through a conversational case-based reasoning approach and prototype. The document also includes references for further information.
The document discusses the semantic web and case-based reasoning. It provides an overview of key concepts like ontology languages, RDF, OWL, and how case-based reasoning can utilize semantic web technologies. It also describes a prototype conversational case-based reasoning application for retrieving earthquake science codes.
The document discusses semantic web, ontology languages, and case-based reasoning. It provides an overview of semantic web and its motivations. It describes ontology languages like RDF, RDF Schema, OWL, and others. It then discusses how case-based reasoning can utilize semantic web technologies by applying an AI technique to retrieve metadata related to codes for earthquake science.
Lecture at the advanced course on Data Science of the SIKS research school, May 20, 2016, Vught, The Netherlands.
Contents
-Why do we create Linked Open Data? Example questions from the Humanities and Social Sciences
-Introduction into Linked Open Data
-Lessons learned about the creation of Linked Open Data (link discovery, knowledge representation, evaluation).
-Accessing Linked Open Data
Semantic Web and Linked Data for cultural heritage materials - Approaches in ...Antoine Isaac
The document discusses using semantic web technologies like linked data and the Europeana Data Model (EDM) to improve access to cultural heritage materials by enabling semantic search and exploiting relationships between concepts, objects, and vocabularies. EDM aims to preserve original metadata while allowing for interoperability by using standards like Dublin Core, SKOS, and OAI ORE. Linked data approaches can ease getting and publishing data across cultural heritage datasets by direct access to RDF descriptions via URIs.
Similar to Getting triples from records: the role of ISBD (20)
A presentation by Gill Hamilton, Digital Access Manager at the National Library of Scotland (NLS).
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
A presentation by Richard Wallis, Technology Evangelist at OCLC.
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
A presentation by Daniel Lewis of the Open Knowledge Foundation.
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
A presentation by Muriel Mewissen, Project Manager of the Shakespeare Registry Project.
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
A presentation by Susanne Thorbord, Bibliographic Consultant at the Danish Bibliographic Centre (DBC).
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
A presentation by Gordon Dunsire.
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
A presentation by Daniel Vila Suero of the Ontology Engineering Group at the Universidad Politecnica de Madrid.
Delivered at the Cataloguing and Indexing Group Scotland (CIGS) Linked Open Data (LOD) Conference which took place Fri 21 September 2012 at the Edinburgh Centre for Carbon Innovation.
Event: Metadata and Web 2.0 seminar
Organised by: Cataloguing & Indexing Group in Scotland
Held on Friday 2nd March at the National Library of Scotland, Edinburgh
Presented by: Martin Morrey
Event: Metadata and Web 2.0 seminar
Organised by: Cataloguing & Indexing Group in Scotland
Held on Friday 2nd March at the National Library of Scotland, Edinburgh
Presented by: Bryan Christie, National Library of Scotland
Event: Metadata and Web 2.0 seminar
Organised by: Cataloguing & Indexing Group in Scotland
Held on Friday 2nd March at the National Library of Scotland, Edinburgh
Presented by: Nora McGregor, British Library
Event: Metadata and Web 2.0 seminar
Organised by: Cataloguing & Indexing Group in Scotland
Held on Friday 2nd March at the National Library of Scotland, Edinburgh
Presented by: Vicki Cormie
The University of Glasgow Library developed a strategy to deliver mobile services over multiple phases.
Phase 1 involved developing a mobile catalogue, implementing QR codes, testing instant messaging and e-books, and conducting user surveys. Based on survey findings that most use mobile internet for email, social media and browsing, the top requested services were searching articles, checking out books and reading e-books.
Phase 2 will focus on improving the mobile website, expanding access to e-resources beyond e-books, using QR codes in physical locations, teaching digital skills, and utilizing mobile devices to improve workflows. The strategy is being implemented and promoted in phases based on user research.
The document summarizes several ongoing reviews that will impact post-16 education in Scotland. Key points include:
1) Reviews cover college governance, ICT infrastructure, and the role of colleges. Recommendations include regionalization and shared services.
2) Budget cuts of 7-23% are expected over the next few years for colleges. Capital funding is also being cut significantly.
3) The "Putting Learners at the Centre" review proposes a more interconnected, sustainable, and learner-centered post-16 system focused on employment.
4) Colleges will receive regional funding to enhance regional provision and outcomes. Courses must focus on recognized qualifications and job skills. Funding models will
The document summarizes the work of SLIC (Scottish Library and Information Council) in supporting further education (FE) libraries in Scotland. It discusses SLIC's annual FE library survey which benchmarks resources, technology, and staffing. The survey found that most colleges cite JISC ebooks as their most used online resource and are making good use of technology like QR codes. It also notes future challenges for FE libraries like reduced budgets, restructured services, and changing user expectations. SLIC helps FE libraries respond by promoting collaboration, knowledge sharing, workforce development, and digital initiatives to support quality library services across Scotland.
The document summarizes the SCURL Walk-In Access to E-resources Project which investigated technical and operational solutions to enable higher education institutions to provide access to electronic resources for external and walk-in users where licenses permit. The project aimed to deliver a toolkit identifying good practices and built on outcomes of previous UK projects. It discusses various approaches taken by member institutions to provide access within the terms of licenses, including the use of Shibboleth, EZProxy, dedicated computers and IP ranges.
The British Library is facing significant budget cuts while still aiming to fulfill its mission of enabling access and supporting research. This has required developing new digitization and funding models, moving from selective "boutique" digitization to larger-scale "mass digitization" projects pursued through public/private partnerships. Examples discussed include newspaper digitization through a partnership with Brightsolid and book digitization through agreements with Microsoft and Google. While innovative approaches are needed, traditional efficiency improvements also remain important to make progress despite budget challenges.
The document summarizes research from two surveys on students' use of mobile technologies and attitudes toward using library services on their phones. It finds that most students own smartphones and browse the web daily, with Apple and Samsung being the most popular brands. While over 90% of students would like to access library services on their phones, about a third said data limits restrict their mobile web use. The document concludes that libraries need to design accessible services for a variety of mobile devices and operating systems while keeping pace with the rapid changes in mobile technology.
The document discusses the University of Glasgow's experiments with demand-driven acquisition (DDA) models for ebooks. It implemented a small patron-driven acquisition trial through an ebook aggregator, spending a limited amount on about 1500 ebook titles over 4 months. Usage was high, with 10% of titles receiving 10 or more access and the average cost per used title being £55.96. Only 10% of titles saw no repeat use after initial purchase. The university is considering further DDA trials and evaluating different models to help build its ebook collection in a responsible way while supporting user needs.
The document outlines three deliverables for a project: 1) Install a prototype digital research space, 2) Design and build user interfaces and workstation arrays demonstrating a vision for multimedia digital research, 3) Evaluate user experiences in the technology-rich environment. It then summarizes findings from evaluating the prototype space, including that researchers want advanced search functionality and better organization of research papers, and that Wi-Fi access and using their own devices were important to users. The recommendation is to focus on greater remote access to resources, advanced search and reference tools, and training to facilitate their use while acknowledging a role in the full digital research process beyond content provision.
More from Scottish Library & Information Council (SLIC), CILIP in Scotland (CILIPS) (20)
Nora Dale - Growing Knowledge: The evolution of research
Getting triples from records: the role of ISBD
1. Getting triples from records: the role of ISBD Gordon Dunsire Presented at the seminar Libraries and the Semantic Web: the role of International Standard Bibliographic Description (ISBD), National Library of Scotland, Edinburgh, 25 Feb 2011
2. Linking Open Data cloud (LOD) Diagram by Richard Cyganiak and AnjaJentzsch. http://lod-cloud.net/
4. Why get involved? To share our data We work for “society” To share our expertise and experience 150 + years To promote the power of libraries (and archives and museums) To survive
5. From record to triples (in 9 stages) Very large numbers of records Catalogue records, finding aids, etc. 300 million; 1 billion? High quality metadata In comparison with other communities Each record may generate many triples 200 “raw” triples (no inferences) per MARC record? Very, very large numbers of triples Billions? Trillions?
8. 3. Create URI for record Must be unique, so 54321 no good on its own http URIs are a good thing (W3C) So add record ID to a unique http domain E.g. http://MyLibraryX.com (unique to the library) + 54321 http://MyLibraryX.com/54321 (or http://MyLibraryX.com#54321) This is not a URL!
9. 4. Replace record ID with URI “mlx” = qname (xmlns) = shorthand for “http://MyLibraryX.com/”
10. 5. Find URIs for attributes Attributes are modelled as RDF properties (predicates) in “element set” namespaces E.g. Dublin Core terms (dct); ISBD (isbd); FRBR (frbrer); RDA (rdaxxx); Bibliographic Ontology (bibo); etc. Choose a namespace, find property with same (or closest) “meaning” (e.g. definition) as attribute Nearest property minimises loss of information Get URI for property If no suitable property, choose another namespace Properties do not have to come from single namespace Match and mix!
11. 5 (cont). Find URI for title http://purl.org/dc/terms/title (dct:title) http://iflastandards.info/ns/isbd/elements/P1014 (isbd:P1014) hasTitleProper http://RDVocab.info/Elements/titleProper (rdaGR1:titleProper)
12. 5 (cont). Find URI for author dct:creator rdarole:author (isbd does not cover “headings”)
13. 5 (cont). Find URI for date dct:date isbd:P1018 hasDateOfPublicationProductionDistribution rdaGr1:dateOfPublication
14. 5 (cont). Find URI for LCSH LCSH is a subject vocabulary Controlled terms So attribute is really “subject” And the term itself is the value dct:subject
15. 5 (cont). Find URI for media type Assuming record uses new ISBD Area 0 ... isbd:P1003 hasMediaType
16. 5 (cont). Find URI for content form Assuming record uses new ISBD Area 0 ... isbd: P1001 hasContentForm
18. 7. Find URIs for values If object of a triple is a URI, it can link to the subject of another triple with the same URI Linked data! Values from controlled vocabularies may have URIs Possible vocabularies: author, subject, ISBD Area 0 NOT: title, date For author: Virtual International Authority File (VIAF) For LCSH: Library of Congress Authorities & Vocabularies For ISBD Area 0: Open Metadata Registry
19. 7 (cont). Find URI for author Author: Wythe, Deborah VIAF: http://www.viaf.org/ viaf:31899419/#Wythe,+Deborah
20. 7 (cont). Find URI for subject (LCSH) LCSH: Museum archives LoC: http://id.loc.gov/authorities/ lcsh:/sh85088707#concept
21. 7 (cont). Find URIs for ISBD Area 0 Media type: Electronic ISBD media type isbdmt:T1002 Content form: Text ISBD Content form isbdcf:T1009