The explosion in growth of the Web of Linked Data has provided, for the first time, a plethora of information in disparate locations, yet bound together by machine-readable, semantically typed relations. Utilisation of the Web of Data has been, until now, restricted to the members of the community, eating their own dogfood, so to speak. To the regular web user browsing Facebook and watching YouTube, this utility is yet to be realised. The primary factor inhibiting uptake is the usability of the Web of Data, where users are required to have prior knowledge of elements from the Semantic Web technology stack. Our solution to this problem is to hide the stack, allowing end users to browse the Web of Data, explore the information it contains, discover knowledge, and use Linked Data. We propose a template-based visualisation approach where information attributed to a given resource is rendered according to the rdf:type of the instance.
From the Feb 19 2014 NISO Virtual Conference: The Semantic Web Coming of Age: Technologies and Implementations
The Web of Data - Ralph Swick, Domain Lead of the Information and Knowledge Domain at W3C
This presentation was given by Tim Thompson of Princeton University during the NISO Virtual Conference, BIBFRAME & Real World Applications for Linked Bibliographic Data, held on June 15, 2016.
Open data is a crucial prerequisite for inventing and disseminating the innovative practices needed for agricultural development. To be usable, data must not just be open in principle—i.e., covered by licenses that allow re-use. Data must also be published in a technical form that allows it to be integrated into a wide range of applications. The webinar will be of interest to any institution seeking ways to publish and curate data in the Linked Data cloud.
This webinar describes the technical solutions adopted by a widely diverse global network of agricultural research institutes for publishing research results. The talk focuses on AGRIS, a central and widely-used resource linking agricultural datasets for easy consumption, and AgriDrupal, an adaptation of the popular, open-source content management system Drupal optimized for producing and consuming linked datasets.
Agricultural research institutes in developing countries share many of the constraints faced by libraries and other documentation centers, and not just in developing countries: institutions are expected to expose their information on the Web in a re-usable form with shoestring budgets and with technical staff working in local languages and continually lured by higher-paying work in the private sector. Technical solutions must be easy to adopt and freely available.
Feb 19, 2014: NISO Virtual Conference: The Semantic Web Coming of Age: Technologies and Implementations
Deck includes presentations from:
Ramanathan V. Guha, Google Fellow; Founder of Schema.org; Pierre-Paul Lemyre, Director of Business Development, Lexum; Bob Du Charme, Director of Digital Media Solutions, TopQuadrant
From the Feb 19 2014 NISO Virtual Conference: The Semantic Web Coming of Age: Technologies and Implementations
The Web of Data - Ralph Swick, Domain Lead of the Information and Knowledge Domain at W3C
This presentation was given by Tim Thompson of Princeton University during the NISO Virtual Conference, BIBFRAME & Real World Applications for Linked Bibliographic Data, held on June 15, 2016.
Open data is a crucial prerequisite for inventing and disseminating the innovative practices needed for agricultural development. To be usable, data must not just be open in principle—i.e., covered by licenses that allow re-use. Data must also be published in a technical form that allows it to be integrated into a wide range of applications. The webinar will be of interest to any institution seeking ways to publish and curate data in the Linked Data cloud.
This webinar describes the technical solutions adopted by a widely diverse global network of agricultural research institutes for publishing research results. The talk focuses on AGRIS, a central and widely-used resource linking agricultural datasets for easy consumption, and AgriDrupal, an adaptation of the popular, open-source content management system Drupal optimized for producing and consuming linked datasets.
Agricultural research institutes in developing countries share many of the constraints faced by libraries and other documentation centers, and not just in developing countries: institutions are expected to expose their information on the Web in a re-usable form with shoestring budgets and with technical staff working in local languages and continually lured by higher-paying work in the private sector. Technical solutions must be easy to adopt and freely available.
Feb 19, 2014: NISO Virtual Conference: The Semantic Web Coming of Age: Technologies and Implementations
Deck includes presentations from:
Ramanathan V. Guha, Google Fellow; Founder of Schema.org; Pierre-Paul Lemyre, Director of Business Development, Lexum; Bob Du Charme, Director of Digital Media Solutions, TopQuadrant
As described in the April NISO/DCMI webinar by Dan Brickley, schema.org is a search-engine initiative aimed at helping webmasters use structured data markup to improve the discovery and display of search results. Drupal 7 makes it easy to markup HTML pages with schema.org terms, allowing users to quickly build websites with structured data that can be understood by Google and displayed as Rich Snippets.
Improved search results are only part of the story, however. Data-bearing documents become machine-processable once you find them. The subject matter, important facts, calendar events, authorship, licensing, and whatever else you might like to share become there for the taking. Sales reports, RSS feeds, industry analysis, maps, diagrams and process artifacts can now connect back to other data sets to provide linkage to context and related content. The key to this is the adoption standards for both the data model (RDF) and the means of weaving it into documents (RDFa). Drupal 7 has become the leading content platform to adopt these standards.
This webinar will describe how RDFa and Drupal 7 can improve how organizations publish information and data on the Web for both internal and external consumption. It will discuss what is required to use these features and how they impact publication workflow. The talk will focus on high-level and accessible demonstrations of what is possible. Technical people should learn how to proceed while non-technical people will learn what is possible.
NISO Webinar: 21st Century Resource Sharing: Which Inter-Library Loan Standard Should I Use?
October 15, 2014
1:00 p.m. - 2:30 p.m. (Eastern Time)
Has “Rethinking Resource Sharing” Succeeded? – A Survey of Resource Sharing Protocols Ten Years Later
Ted Koppel, Product Manager, VERSO® ILS – Auto-Graphics, Inc.
Invisible Alphabet Soup: How Libraries Use a Variety of ILL Standards Everyday and Don't Necessarily Know It
Margaret Ellingson, Head of Interlibrary Loan and Course Reserves, Robert W. Woodruff Library, Emory University
Occams Reader and the Interlibrary Loan of E-books
Kenny Ketner, Software Development Manager, Texas Tech University Libraries
Ryan Litsey, Document Delivery/Interlibrary Loan Assistant Librarian, Texas Tech University Library
Libraries around the world have a long tradition of maintaining authority files to assure the consistent presentation and indexing of names. As library authority files have become available online, the authority data has become accessible -- and many have been published as Linked Open Data (LOD) -- but names in one library authority file typically had no link to corresponding records for persons and organizations in other library authority files. After a successful experiment in matching the Library of Congress/NACO authority file with the German National Library's authority file, an online system called the Virtual International Authority File was developed to facilitate sharing by ingesting, matching, and displaying the relations between records in multiple authority files.
The Virtual International Authority File (VIAF) has grown from three source files in 2007 to more than two dozen files today. The system harvests authority records, enhances them with bibliographic information and brings them together into clusters when it is confident the records describe the same identity. Although the most visible part of VIAF is a HTML interface, the API beneath it supports a linked data view of VIAF with URIs representing the identities themselves, not just URIs for the clusters. It supports names for person, corporations, geographic entities, works, and expressions. With English, French, German, Spanish interfaces (and a Japanese in process), the system is used around the world, with over a million queries per day.
Speaker
Thomas Hickey is Chief Scientist at OCLC where he helped found OCLC Research. Current interests include metadata creation and editing systems, authority control, parallel systems for bibliographic processing, and information retrieval and display. In addition to implementing VIAF, his group looks into exploring Web access to metadata, identification of FRBR works and expressions in WorldCat, the algorithmic creation of authorities, and the characterization of collections. He has an undergraduate degree in Physics and a Ph.D. in Library and Information Science.
This presentation was delivered by Gloria Gonzalez of Zepheira during the NISO Virtual Conference, BIBFRAME & Real World Applications of Linked Bibliographic Data, held on June 15, 2016.
4.2.15 Slides, “Hydra: many heads, many connections. Enriching Fedora Reposit...DuraSpace
Hot Topics: The DuraSpace Community Webinar Series
Series 11: Integrating ORCID Persistent Identifiers with DSpace, Fedora and VIVO
Webinar 2: “Hydra: many heads, many connections. Enriching Fedora Repositories with ORCID.”
Thursday, April 2, 2015
Curated by Josh Brown, ORCID
Presented by: Laura Paglione, Technical Director, ORCID and Rick Johnson, Head of Digital Library Services, University of Notre Dame
Should We Expect a Bang or a Whimper? Will Linked Data Revolutionize Scholar Authoring and Workflow Tools?
Jeff Baer, Senior Director of Product Management, Research Development Services, Proquest
This presentation was delivered by Carolyn Hansen of the University of Cincinnati during the NISO VIrtual Conference, BIBFRAME & Real World Applications of Linked Bibliographic Data, held on June 15, 2016
This presentation was provided by Diana Brooking of the University of Washington during the 11th Annual NISO-BISG Forum, Delivering the Integrated Information Experience, on June 23, 2017 and held at the ALA Annual Conference.
This presentation was provided by Carolyn Hansen of the University of Cincinnati during the NISO Training Thursday event, Metadata and the IR, held on Thursday, February 23, 2017.
Analysing & Improving Learning Resources Markup on the WebStefan Dietze
Talk at WWW2017 on LRMI adoption, quality and usage. Full paper here: http://papers.www2017.com.au.s3-website-ap-southeast-2.amazonaws.com/companion/p283.pdf.
About the Webinar
In May 2012, the Library of Congress announced a new modeling initiative focused on reflecting the MARC 21 library standard as a Linked Data model for the Web, with an initial model to be proposed by the consulting company Zepheira. The goal of the initiative is to translate the MARC 21 format to a Linked Data model while retaining the richness and benefits of existing data in the historical format.
In this webinar, Eric Miller of Zepheira will report on progress towards this important goal, starting with an analysis of the translation problem and concluding with potential migration scenarios for a broad-based transition from MARC to a new bibliographic framework.
The Progress of BIBFRAME, by Angela KroegerAngela Kroeger
Presentation given at the OLAC-MOUG 2014 conference. Abstract: BIBFRAME is the Library of Congress's current effort to develop a linked data replacement for MARC. BIBFRAME is a work in progress, not yet ready for implementation. In this two-hour session, we will examine how BIBFRAME works, what it is intended to accomplish, and the progress that has been made toward that goal. We'll take a look at the BIBFRAME tools that are under development, including the prototype editor for creating new records. And we'll share a glimpse of what the future holds for library catalogs and cataloging. NOTE: SlideShare seems to have garbled the formatting of some of my slides. To receive a clean copy via email, contact me at angelajkroeger [at] gmail [dot] com.
Open science can contribute to AI trustworthiness. This talk is a categorization of scientific data platforms, and a framing of AI trustworthiness with pointers to open science contributions.
Sensing Presence (PreSense) Ontology - User Modelling in the Semantic Sensor Webaba-sah
Increasingly, people's digital identities are attached to, and expressed through, their mobile devices. At the same time digital sensors pervade smart environments in which people are immersed.
This paper explores different perspectives in which users' modelling features can be expressed through the information obtained by their attached personal sensors. We introduce the PreSense Ontology, which is designed to assign meaning to sensors' observations in terms of user modelling features. We believe that the Sensing Presence PreSense Ontology is a first step toward the integration of user modelling and "smart environments". In order to motivate our work we present a scenario and demonstrate how the ontology could be applied in order to enable context-sensitive services.
As described in the April NISO/DCMI webinar by Dan Brickley, schema.org is a search-engine initiative aimed at helping webmasters use structured data markup to improve the discovery and display of search results. Drupal 7 makes it easy to markup HTML pages with schema.org terms, allowing users to quickly build websites with structured data that can be understood by Google and displayed as Rich Snippets.
Improved search results are only part of the story, however. Data-bearing documents become machine-processable once you find them. The subject matter, important facts, calendar events, authorship, licensing, and whatever else you might like to share become there for the taking. Sales reports, RSS feeds, industry analysis, maps, diagrams and process artifacts can now connect back to other data sets to provide linkage to context and related content. The key to this is the adoption standards for both the data model (RDF) and the means of weaving it into documents (RDFa). Drupal 7 has become the leading content platform to adopt these standards.
This webinar will describe how RDFa and Drupal 7 can improve how organizations publish information and data on the Web for both internal and external consumption. It will discuss what is required to use these features and how they impact publication workflow. The talk will focus on high-level and accessible demonstrations of what is possible. Technical people should learn how to proceed while non-technical people will learn what is possible.
NISO Webinar: 21st Century Resource Sharing: Which Inter-Library Loan Standard Should I Use?
October 15, 2014
1:00 p.m. - 2:30 p.m. (Eastern Time)
Has “Rethinking Resource Sharing” Succeeded? – A Survey of Resource Sharing Protocols Ten Years Later
Ted Koppel, Product Manager, VERSO® ILS – Auto-Graphics, Inc.
Invisible Alphabet Soup: How Libraries Use a Variety of ILL Standards Everyday and Don't Necessarily Know It
Margaret Ellingson, Head of Interlibrary Loan and Course Reserves, Robert W. Woodruff Library, Emory University
Occams Reader and the Interlibrary Loan of E-books
Kenny Ketner, Software Development Manager, Texas Tech University Libraries
Ryan Litsey, Document Delivery/Interlibrary Loan Assistant Librarian, Texas Tech University Library
Libraries around the world have a long tradition of maintaining authority files to assure the consistent presentation and indexing of names. As library authority files have become available online, the authority data has become accessible -- and many have been published as Linked Open Data (LOD) -- but names in one library authority file typically had no link to corresponding records for persons and organizations in other library authority files. After a successful experiment in matching the Library of Congress/NACO authority file with the German National Library's authority file, an online system called the Virtual International Authority File was developed to facilitate sharing by ingesting, matching, and displaying the relations between records in multiple authority files.
The Virtual International Authority File (VIAF) has grown from three source files in 2007 to more than two dozen files today. The system harvests authority records, enhances them with bibliographic information and brings them together into clusters when it is confident the records describe the same identity. Although the most visible part of VIAF is a HTML interface, the API beneath it supports a linked data view of VIAF with URIs representing the identities themselves, not just URIs for the clusters. It supports names for person, corporations, geographic entities, works, and expressions. With English, French, German, Spanish interfaces (and a Japanese in process), the system is used around the world, with over a million queries per day.
Speaker
Thomas Hickey is Chief Scientist at OCLC where he helped found OCLC Research. Current interests include metadata creation and editing systems, authority control, parallel systems for bibliographic processing, and information retrieval and display. In addition to implementing VIAF, his group looks into exploring Web access to metadata, identification of FRBR works and expressions in WorldCat, the algorithmic creation of authorities, and the characterization of collections. He has an undergraduate degree in Physics and a Ph.D. in Library and Information Science.
This presentation was delivered by Gloria Gonzalez of Zepheira during the NISO Virtual Conference, BIBFRAME & Real World Applications of Linked Bibliographic Data, held on June 15, 2016.
4.2.15 Slides, “Hydra: many heads, many connections. Enriching Fedora Reposit...DuraSpace
Hot Topics: The DuraSpace Community Webinar Series
Series 11: Integrating ORCID Persistent Identifiers with DSpace, Fedora and VIVO
Webinar 2: “Hydra: many heads, many connections. Enriching Fedora Repositories with ORCID.”
Thursday, April 2, 2015
Curated by Josh Brown, ORCID
Presented by: Laura Paglione, Technical Director, ORCID and Rick Johnson, Head of Digital Library Services, University of Notre Dame
Should We Expect a Bang or a Whimper? Will Linked Data Revolutionize Scholar Authoring and Workflow Tools?
Jeff Baer, Senior Director of Product Management, Research Development Services, Proquest
This presentation was delivered by Carolyn Hansen of the University of Cincinnati during the NISO VIrtual Conference, BIBFRAME & Real World Applications of Linked Bibliographic Data, held on June 15, 2016
This presentation was provided by Diana Brooking of the University of Washington during the 11th Annual NISO-BISG Forum, Delivering the Integrated Information Experience, on June 23, 2017 and held at the ALA Annual Conference.
This presentation was provided by Carolyn Hansen of the University of Cincinnati during the NISO Training Thursday event, Metadata and the IR, held on Thursday, February 23, 2017.
Analysing & Improving Learning Resources Markup on the WebStefan Dietze
Talk at WWW2017 on LRMI adoption, quality and usage. Full paper here: http://papers.www2017.com.au.s3-website-ap-southeast-2.amazonaws.com/companion/p283.pdf.
About the Webinar
In May 2012, the Library of Congress announced a new modeling initiative focused on reflecting the MARC 21 library standard as a Linked Data model for the Web, with an initial model to be proposed by the consulting company Zepheira. The goal of the initiative is to translate the MARC 21 format to a Linked Data model while retaining the richness and benefits of existing data in the historical format.
In this webinar, Eric Miller of Zepheira will report on progress towards this important goal, starting with an analysis of the translation problem and concluding with potential migration scenarios for a broad-based transition from MARC to a new bibliographic framework.
The Progress of BIBFRAME, by Angela KroegerAngela Kroeger
Presentation given at the OLAC-MOUG 2014 conference. Abstract: BIBFRAME is the Library of Congress's current effort to develop a linked data replacement for MARC. BIBFRAME is a work in progress, not yet ready for implementation. In this two-hour session, we will examine how BIBFRAME works, what it is intended to accomplish, and the progress that has been made toward that goal. We'll take a look at the BIBFRAME tools that are under development, including the prototype editor for creating new records. And we'll share a glimpse of what the future holds for library catalogs and cataloging. NOTE: SlideShare seems to have garbled the formatting of some of my slides. To receive a clean copy via email, contact me at angelajkroeger [at] gmail [dot] com.
Open science can contribute to AI trustworthiness. This talk is a categorization of scientific data platforms, and a framing of AI trustworthiness with pointers to open science contributions.
Sensing Presence (PreSense) Ontology - User Modelling in the Semantic Sensor Webaba-sah
Increasingly, people's digital identities are attached to, and expressed through, their mobile devices. At the same time digital sensors pervade smart environments in which people are immersed.
This paper explores different perspectives in which users' modelling features can be expressed through the information obtained by their attached personal sensors. We introduce the PreSense Ontology, which is designed to assign meaning to sensors' observations in terms of user modelling features. We believe that the Sensing Presence PreSense Ontology is a first step toward the integration of user modelling and "smart environments". In order to motivate our work we present a scenario and demonstrate how the ontology could be applied in order to enable context-sensitive services.
Does Size Matter? When Small is Good Enoughaba-sah
This paper reports the observation of the influence of the size of documents on the accuracy of a defined text processing task.
Our hypothesis is that based on a specific task (in this case, topic classification), results obtained using longer texts may be approximated by short texts, of micropost size, i.e., maximum length 140 characters.
Using an email dataset as the main corpus, we generate several fixed-size corpora, consisting of truncated emails, from micropost size (140 characters), and successive multiples thereof, to the full size of each email. Our methodology consists of two steps: (1) corpus-driven topic extraction and (2) document topic classification. We build the topic representation model using the main corpus, through k-means clustering, with each \emph{k}-derived topic represented as a weighted number of terms. We then perform document classification according to the k topics: first over the main corpus, then over each truncated corpus, and observe the variance in classification accuracy with document size.
The results obtained show that the accuracy of topic classification for micropost-size texts is a suitable approximation of classification performed on longer texts.
The impact of innovation on travel and tourism industries (World Travel Marke...Brian Solis
From the impact of Pokemon Go on Silicon Valley to artificial intelligence, futurist Brian Solis talks to Mathew Parsons of World Travel Market about the future of travel, tourism and hospitality.
We’re all trying to find that idea or spark that will turn a good project into a great project. Creativity plays a huge role in the outcome of our work. Harnessing the power of collaboration and open source, we can make great strides towards excellence. Not just for designers, this talk can be applicable to many different roles – even development. In this talk, Seasoned Creative Director Sara Cannon is going to share some secrets about creative methodology, collaboration, and the strong role that open source can play in our work.
The Six Highest Performing B2B Blog Post FormatsBarry Feldman
If your B2B blogging goals include earning social media shares and backlinks to boost your search rankings, this infographic lists the size best approaches.
Each technological age has been marked by a shift in how the industrial platform enables companies to rethink their business processes and create wealth. In the talk I argue that we are limiting our view of what this next industrial/digital age can offer because of how we read, measure and through that perceive the world (how we cherry pick data). Companies are locked in metrics and quantitative measures, data that can fit into a spreadsheet. And by that they see the digital transformation merely as an efficiency tool to the fossil fuel age. But we need to stretch further…
10-15-13 “Metadata and Repository Services for Research Data Curation” Presen...DuraSpace
“Hot Topics: The DuraSpace Community Webinar Series," Series Six: Research Data in Repositories” Curated by David Minor, Research Data Curation Program, UC San Diego Library. Webinar 2: “Metadata and Repository Services for Research Data Curation”
Presented by Declan Fleming, Chief Technology Strategist, Arwen Hutt, Metadata Librarian & Matt Critchlow, Manager of Development and Web ServicesUC, San Diego Library.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
About the Webinar
The library and cultural institution communities have generally accepted the vision of moving to a Linked Data environment that will align and integrate their resources with those of the greater Semantic Web. But moving from vision to implementation is not easy or well-understood. A number of institutions have begun the needed infrastructure and tools development with pilot projects to provide structured data in support of discovery and navigation services for their collections and resources.
Join NISO for this webinar where speakers will highlight actual Linked Data projects within their institutions—from envisioning the model to implementation and lessons learned—and present their thoughts on how linked data benefits research, scholarly communications, and publishing.
Speakers:
Jon Voss - Strategic Partnerships Director, We Are What We Do
LODLAM + Historypin: A Collaborative Global Community
Matt Miller - Front End Developer, NYPL Labs at the New York Public Library
The Linked Jazz Project: Revealing the Relationships of the Jazz Community
Cory Lampert - Head, Digital Collections , UNLV University Libraries
Silvia Southwick - Digital Collections Metadata Librarian, UNLV University Libraries
Linked Data Demystified: The UNLV Linked Data Project
Duraspace Hot Topics Series 6: Metadata and Repository ServicesMatthew Critchlow
Presented by Declan Fleming, Arwen Hutt, and Matt Critchlow. The second in a three part Webinar series on Research Data Curation at UC San Diego, as part of the larger Research Cyberinfrastructure initiative.
Engaging Information Professionals in the Process of Authoritative Interlinki...Lucy McKenna
Through the use of Linked Data (LD), Libraries, Archives and Museums (LAMs) have the potential to expose their collections to a larger audience and to allow for more efficient user searches. Despite this, relatively few LAMs have invested in LD projects and the majority of these display limited interlinking across datasets and institutions. A survey was conducted to understand Information Professionals' (IPs') position with regards to LD, with a particular focus on the interlinking problem. The survey was completed by 185 librarians, archivists, metadata cataloguers and researchers. Results indicated that, when interlinking, IPs find the process of ontology and property selection to be particularly challenging, and LD tooling to be technologically complex and unsuitable for their needs.
Our research is focused on developing an authoritative interlinking framework for LAMs with a view to increasing IP engagement in the linking process. Our framework will provide a set of standards to facilitate IPs in the selection of link types, specifically when linking local resources to authorities. The framework will include guidelines for authority, ontology and property selection, and for adding provenance data. A user-interface will be developed which will direct IPs through the resource interlinking process as per our framework. Although there are existing tools in this domain, our framework differs in that it will be designed with the needs and expertise of IPs in mind. This will be achieved by involving IPs in the design and evaluation of the framework. A mock-up of the interface has already been tested and adjustments have been made based on results. We are currently working on developing a minimal viable product so as to allow for further testing of the framework. We will present our updated framework, interface, and proposed interlinking solutions.
Sands Fish - Knowing in the Age of Networked Knowledgesandsfish
Knowledge representation has become extremely complex since the advent of the internet, online education, and commons-based peer production. This talk discusses the thresholds we've crossed and what it means to know something when knowledge is massively interlinked.
Opening up and linking data is becoming a priority for many data producers because of institutional requirements, or to consume data in newer applications, or simply to keep pace with current development. Since 2014, this priority has gaining momentum with the Global Open Data in Agriculture and Nutrition initiative (GODAN). However, typical small and medium-size institutions have to deal with constrained resources, which often hamper their possibilities for making their data publicly available. This webinar will be of interest to any institution seeking ways to publish and curate data in the Linked Data World.
BIBFLOW and the Libhub Initiative: Leveraging our past to define our future
Eric Miller, President, Zepheira
Jeff Penka, Director of Channel and Product Development, Zepheira
Semantic Similarity and Selection of Resources Published According to Linked ...Riccardo Albertoni
The position paper aims at discussing the potential of exploiting linked data best practice to provide metadata documenting domain specific resources created through verbose acquisition-processing pipelines. It argues that resource selection, namely the process engaged to choose a set of resources suitable for a given analysis/design purpose, must be supported by a deep comparison of their metadata. The semantic similarity proposed in our previous works is discussed for this purpose and the main issues to make it scale up to the web of data are introduced. Discussed issues contribute beyond the re-engineering of our similarity since they largely apply to every tool which is going to exploit information made available as linked data. A research plan and an exploratory phase facing the presented issues are described remarking the lessons we have learnt so far.
Why do they call it Linked Data when they want to say...?Oscar Corcho
The four Linked Data publishing principles established in 2006 seem to be quite clear and well understood by people inside and outside the core Linked Data and Semantic Web community. However, not only when discussing with outsiders about the goodness of Linked Data but also when reviewing papers for the COLD workshop series, I find myself, in many occasions, going back again to the principles in order to see whether some approach for Web data publication and consumption is actually Linked Data or not. In this talk we will review some of the current approaches that we have for publishing data on the Web, and we will reflect on why it is sometimes so difficult to get into an agreement on what we understand by Linked Data. Furthermore, we will take the opportunity to describe yet another approach that we have been working on recently at the Center for Open Middleware, a joint technology center between Banco Santander and Universidad Politécnica de Madrid, in order to facilitate Linked Data consumption.
Facets and Pivoting for Flexible and Usable Linked Data ExplorationRoberto García
The success of Open Data initiatives has increased the amount of data available on the Web. Unfortunately, most of this data is only available in raw tabular form, what makes analysis and reuse quite difficult for non-experts. Linked Data principles allow for a more sophisticated approach by making explicit both the structure and semantics of the data. However, from the end-user viewpoint, they continue to be monolithic files completely opaque or difficult to explore by making tedious semantic queries. Our objective is to facilitate the user to grasp what kind of entities are in the dataset, how they are interrelated, which are their main properties and values, etc. Rhizomer is a tool for data publishing whose interface provides a set of components borrowed from Information Architecture (IA) that facilitate awareness of the dataset at hand. It automatically generates navigation menus and facets based on the kinds of things in the dataset and how they are described through metadata properties and values. Moreover, motivated by recent tests with end-users, it also provides the possibility to pivot among the faceted views created for each class of resources in the dataset.
Presentación del Dr. Getaneh Alemu (Solent University, Reino Unido), en el II Congreso de Información, Comunicación e Investigación (CICI 2018) “Metadatos y Organización de la Información”. Facultad de Filosofía y Letras de la Universidad Autónoma de Chihuahua, México. Evento organizado por el Cuerpo Académico 'Estudios de la Información' y el Grupo Disciplinar ‘Información, Lenguaje, Comunicación y Desarrollo Sostenible’. 29 de octubre de 2018.
Similar to Hide the Stack:Toward Usable Linked Data (20)
1. Hide the Stack:
Toward Usable Linked Data
A.‐S. Dadzie1, M. Rowe2 & D. Petrelli3
1. The OAK Group, Dept. of Computer Science, The University of Sheffield
2. The Knowledge Media InsPtute, The Open University
3. Art & Design Research Centre, Sheffield Hallam University
2. Key Message
• Linked Data
– connecPons between disparate, independent (albeit related) data
– rendering public interest data accessible
• allowing hidden informaPon to be discovered more easily
• enabling quesPons to be answered more fully
• PotenPal widely recognised, but
– very large‐scale, wide‐coverage, highly inter‐linked data repositories
– under‐uPlised outside SemanPc Web community
• Aims of the research
– explore new methods for presenPng Linked Data to wider audience
– support more intuiPve exploraPon and knowledge retrieval
– encourage wider use by web‐savvy but non‐technical users
5. Challenges in Linked Data ConsumpPon
1. CombaPng informaPon overload
2. ExploraPon starPng point
3. Returning something useful
4. Enabling interacPon
How can we make Linked Data usable to real, end users?
• where end users broadly classified into one of:
– SemanPc Web experts
– web‐savvy but non‐technical
8. Scenario
• InformaPon‐seeking scenario
– end user: a primary school teacher
– task: looking for research in local university on ‘Web Technology’
– tools typically used for informaPon seeking acPviPes:
• web search/browse
• library
• consider the university department’s web site built on top of
Data.dcs
10. Tools for consuming Linked Data
• SemanPc Web user
– well catered for
– typical tasks
• browsing RDF
• validaPng data and models
• extracPng data using formal query syntax
• mainstream web user
– lower tool support
– typical tasks ‐ exploratory informaPon seeking
• search and query (using less formal methods, e.g., forms)
• browsing to discover informaPon
• sharing of informaPon discovered, results of any analysis
11. Tools for consuming Linked Data
Tool Type Examples of Tools
formaded text display of RDF (e.g., Sig.ma, Marbles, URI Burner, Haystack, Tabulator
using HTML tables, templates)
RDF graph model W3C RDF Validator, Sindice Inspector
other graph visualisaPon IsaViz, RDFGravity, Cytoscape, RelFinder
other domain‐specific visualisaPon DBPedia Mobile, Talis Research Funding Explorer
Expected Skill Set Examples of Tools
understanding of SW technology Sig.ma, Marbles, URI Burner, W3C RDF Validator,
stack RelFinder, Tabulator
formal querying, e.g., SPARQL LESS
basic to advanced knowledge DBPedia Mobile, RelFinder, Talis RFE, IsaViz
seeking, exploratory navigaPon
web browsing (desktop, mobile) LESS, DBPedia Mobile
See also:
• Dadzie, A.-S. & Rowe, M. (In press). Approaches to Visualising Linked Data: A Survey, the Semantic Web Journal —
Special Call for Survey articles on Semantic Web topics.
• Katifori, A., Halatsis, C., Lepouras, G., Vassilakis, C. & Giannopoulou, E. (2007). Ontology visualization methods — a
survey, ACM Computing Surveys.
13. A Template‐based SoluPon
• taking advantage of self‐describing RDF data
– look up class of a given resource
– load template based on class – i.e., rdf:type of the instance
• focus on perPnent informaPon in a dataset
• highlight relaPonships within data
• allow end users to retrieve detail in ROIs (regions of interest)
• combine templates with informaPon/knowledge visualisaPon
– hide the complexity of the underlying data
– remove the need for specialist SW knowledge or skill
14. Why this approach?
• Templates
– value seen in the wide use of Fresnel lenses and other template
development tools and languages, e.g., IsaViz, LENA, LESS
– simplicity, reusability, extensibility, flexibility
• VisualisaPon
– overview to support detecPon of data structure
• exploratory informaPon seeking
• idenPfying/highlighPng relaPonships
• recognising anomalies, errors
– reducPon in cogniPve load – through advanced human percepPon
• especially useful for analysis of large, complex data
15. Template Design
• idenPfy key concepts & relevant metadata to define templates
– match to standard ontologies, e.g., FOAF, PRV, SWRC, BIB
– SPARQL queries – built based on Fresnel lens SPARQL selectors
• presentaPon methods
– visual overview ‐ node‐link graph
• collapse informaPon related to key concepts into compound nodes
• filter out less immediately relevant data – provide more room for ROIs
– visual encoding
• colour coding based on RDF type (nodes and links)
• icons based on RDF type (nodes)
• size to encode node properPes, e.g., no. of outlinks
– detail view
• text and thumbnails/icons
16. Data.dcs templates
• informaPon of main interest – key concepts
– organisaPonal structure ‐ research groups
– people
– relaPonships within structure
• main concepts
– Organisa.on [<http://xmlns.com/foaf/0.1/Group>]
– Person [<http://xmlns.com/foaf/0.1/Person>]
– Publica.on [<http://zeitkunst.org/bibtex/0.1/bibtex.owl#Entry>]
21. Challenge 1: SoluPon Proposed
• Comba.ng informa.on overload
– very large amounts of distributed, heterogeneous data
– high inter‐linking
• our soluPon:
– visual overview + filters
– highlighPng key relaPonships
– detail view
• focus on graph ROI within context of surrounding data
+ text detail & thumbnails or representaPve icons
22. Challenge 2:
Explora.on star.ng point
• SW users
– may have a specific URI to explore
• mainstream users
– may or may not have a specific starPng point
– onen start with a vague idea and browse to find if there is anything
interesPng
24. Challenge 2: SoluPon proposed
• Explora.on star.ng point
– SW users
• may have a specific URI to explore
– mainstream users
• may or may not have a specific starPng point
• onen start with a vague idea and browse to find what they want
• our soluPon:
– support both
• where input, specific URI as focus at start
• extract list of potenPal start points
– selected from key RDF types in a dataset
– randomly chosen focus – influenced by context
• centre graph on focus
25. Challenge 3:
Returning something useful
• what is the end user looking for?
• how can we present the data so they are able to find it?
29. Under the Hood‐ Detail View
• e.g., SPARQL query template for the full publicaPon view
PREFIX foaf: <http://xmlns.com/foaf/0.1/>
PREFIX bib: <http://zeitkunst.org/bibtex/0.1/bibtex.owl#>
SELECT DISTINCT ?publicationTitle ?year ?bookTitle ?personUri ?author ?
imageUri
WHERE {
<data.dcs:publicationUri> bib:title ?publicationTitle ;
bib:hasYear ?year ;
bib:hasBookTitle ?bookTitle ;
foaf:maker ?personUri .
?personUri foaf:name ?author ;
foaf:img ?imageUri
} ORDER BY DESC(?year) ?publicationTitle
30. Challenge 3: SoluPon
• Returning something useful
– what is the end user looking for?
– how can we present the data so they are able to find it?
• Our soluPon
– graph overview + detail template view
– reuse familiar web browser look and feel (detail)
– interacPve graph
• to support exploratory navigaPon
• retain context of surrounding informaPon
• colour coding to highlight key resource types and relaPonships
31. Challenge 4:
Enabling interac.on
• Our soluPon
– graph overview + detail template view
– reuse familiar web browsing interacPon
– click to navigate through data (in both views)
– pan + zoom for graph view
– filters to remove less relevant informaPon
33. FormaPve EvaluaPon
• at ESWC 2010 ‘EssenPal HCI for the SemanPc Web’ tutorial
– focus group of (14) “expert reviewers”
– assessing usability for both mainstream and expert users
– training task and an informaPon exploraPon exercise
• graphs found to be expressive
• graph view effecPve in giving a sense of data distribuPon
• detail view effecPvely displayed key resources in a neat and
concise way
• prototype seen to have potenPal for exploring and debugging LD
• some difficulty for users not familiar with interacPve graph layout
– eventually you got a big picture of the data
– I liked the direct manipulaPon but the graph should stay put
[when I click]
35. Conclusions
• explored human‐centred soluPon for consuming Linked Data
– exploiPng templates (via SW technology)
– combined with visualisaPon
• evaluaPon highlighted challenges sPll remaining ‐ among others:
– scale
– complexity
• however – promising start....
• Next Steps
– more user control ‐ (intuiPve) support for defining templates, filters
– dynamic update with new Linked Data
– formal usability evaluaPon with wider range of users
36. Outline
• Challenges
• IllustraPve Scenario
• ExisPng Work
• Approach
• IniPal EvaluaPon
• Conclusions & Next Steps
• Acknowledgements
– parPcipants of ESWC 2010 ‘EssenPal HCI for the SemanPc Web’ tutorial
– Funding
• A‐S Dadzie – SmartProducts & WeKnowIt (EU FP7), X‐Media (EU FP6)
• M Rowe – WeGov (EU FP7)
• D Petrelli – X‐Media (EU FP6)