LANL Research Library
March 12, 2009
Martin Klein & Michael L. Nelson
Department of Computer Science
Old Dominion University
Norfolk VA
www.cs.odu.edu/~{mklein,mln}
Invited talk at USEWOD2014 (http://people.cs.kuleuven.be/~bettina.berendt/USEWOD2014/)
A tremendous amount of machine-interpretable information is available in the Linked Open Data Cloud. Unfortunately, much of this data remains underused as machine clients struggle to use the Web. I believe this can be solved by giving machines interfaces similar to those we offer humans, instead of separate interfaces such as SPARQL endpoints. In this talk, I'll discuss the Linked Data Fragments vision on machine access to the Web of Data, and indicate how this impacts usage analysis of the LOD Cloud. We all can learn a lot from how humans access the Web, and those strategies can be applied to querying and analysis. In particular, we have to focus first on solving those use cases that humans can do easily, and only then consider tackling others.
Invited talk at USEWOD2014 (http://people.cs.kuleuven.be/~bettina.berendt/USEWOD2014/)
A tremendous amount of machine-interpretable information is available in the Linked Open Data Cloud. Unfortunately, much of this data remains underused as machine clients struggle to use the Web. I believe this can be solved by giving machines interfaces similar to those we offer humans, instead of separate interfaces such as SPARQL endpoints. In this talk, I'll discuss the Linked Data Fragments vision on machine access to the Web of Data, and indicate how this impacts usage analysis of the LOD Cloud. We all can learn a lot from how humans access the Web, and those strategies can be applied to querying and analysis. In particular, we have to focus first on solving those use cases that humans can do easily, and only then consider tackling others.
Guest lecture at the Syracuse University School of Information Studies eScience Librarianship Lecture Series (08 Dec 2011).
Description: It’s your government, is it your data? New approaches to building interlinked catalogs of government-produced data. Dr. John S. Erickson, Director of Web Science Operations for the Tetherless World Constellation at Rensselaer Polytechnic Institute will present technical methods being developed to manage the delivery of large-scale open government data projects based on semantic web and linked data best practices.
This presentation was provided by Robert Sebek of Virginia Tech during the NISO webinar, Discovery: Where Researchers Start, held on Wednesday, August 8, 2018.
Poster - Completeness Statements about RDF Data Sources and Their Use for Qu...Fariz Darari
Thousands of RDF data sources are today available on the Web.
Machine-readable qualitative descriptions of their content are crucial.
We focus on data completeness, an important aspect of data quality.
How to formalize and express in a machine-readable way completeness information about RDF data sources?
How to leveragesuch completeness information?
Formal framework for expressing completeness information.
Study of query completeness from completeness information in various settings.
Interlinking Standardized OpenStreetMap Data and Citizen Science Data in the ...Werner Leyh
Abstract. The aim of this work is to explore the opportunities offered by
semantic standardization to interlink primary “spatial data” (GI) from “Open-
StreetMap” (OSM) with repositories of the “Linked Open Data Cloud” (LOD).
Research in natural sciences can generate vast amounts of spatial data, where
Wikidata could be considered as the central hub between more detailed natural
science hubs on the spatial semantic web. Wikidata is a world readable and
writable community-driven knowledge base. It offers the opportunity to collaboratively
construct an open access knowledge graph that spans biology,
medicine, and all other domains of knowledge. In this study, we discuss
the opportunities and challenges provided by exploring Wikidata as a central
integration facility by interlink it with OSM, a popular, community driven
collection of free geographic data. This is empowered by the reuse of terms
and properties from commonly understood controlled vocabularies that
represent their respective well-identified knowledge domains.
URL: https://www.springerprofessional.de/en/interlinking-standardized-openstreetmap-data-and-citizen-science/13302088
DOI: https://doi.org/10.1007/978-3-319-60366-7_9
Werner Leyh, Homero Fonseca Filho
University of São Paulo (USP), São Paulo, Brazil
WernerLeyh@yahoo.com
Guest lecture at the Syracuse University School of Information Studies eScience Librarianship Lecture Series (08 Dec 2011).
Description: It’s your government, is it your data? New approaches to building interlinked catalogs of government-produced data. Dr. John S. Erickson, Director of Web Science Operations for the Tetherless World Constellation at Rensselaer Polytechnic Institute will present technical methods being developed to manage the delivery of large-scale open government data projects based on semantic web and linked data best practices.
This presentation was provided by Robert Sebek of Virginia Tech during the NISO webinar, Discovery: Where Researchers Start, held on Wednesday, August 8, 2018.
Poster - Completeness Statements about RDF Data Sources and Their Use for Qu...Fariz Darari
Thousands of RDF data sources are today available on the Web.
Machine-readable qualitative descriptions of their content are crucial.
We focus on data completeness, an important aspect of data quality.
How to formalize and express in a machine-readable way completeness information about RDF data sources?
How to leveragesuch completeness information?
Formal framework for expressing completeness information.
Study of query completeness from completeness information in various settings.
Interlinking Standardized OpenStreetMap Data and Citizen Science Data in the ...Werner Leyh
Abstract. The aim of this work is to explore the opportunities offered by
semantic standardization to interlink primary “spatial data” (GI) from “Open-
StreetMap” (OSM) with repositories of the “Linked Open Data Cloud” (LOD).
Research in natural sciences can generate vast amounts of spatial data, where
Wikidata could be considered as the central hub between more detailed natural
science hubs on the spatial semantic web. Wikidata is a world readable and
writable community-driven knowledge base. It offers the opportunity to collaboratively
construct an open access knowledge graph that spans biology,
medicine, and all other domains of knowledge. In this study, we discuss
the opportunities and challenges provided by exploring Wikidata as a central
integration facility by interlink it with OSM, a popular, community driven
collection of free geographic data. This is empowered by the reuse of terms
and properties from commonly understood controlled vocabularies that
represent their respective well-identified knowledge domains.
URL: https://www.springerprofessional.de/en/interlinking-standardized-openstreetmap-data-and-citizen-science/13302088
DOI: https://doi.org/10.1007/978-3-319-60366-7_9
Werner Leyh, Homero Fonseca Filho
University of São Paulo (USP), São Paulo, Brazil
WernerLeyh@yahoo.com
Santa Fe Complex
March 13, 2009
Martin Klein, Frank McCown,
Joan Smith, Michael L. Nelson
Department of Computer Science
Old Dominion University
Norfolk VA
Using timed-release cryptography to mitigate the preservation risk of embargo...Michael Nelson
Slides for:
Rabia Haq, Michael L. Nelson: Using timed-release cryptography to mitigate the preservation risk of embargo periods. 2009 ACM/IEEE Joint Conference on Digital Libraries (JCDL), pp. 183-192.
Mathematics & Computer Science Seminar
Emory University
October 2, 2009
Martin Klein & Michael L. Nelson
Department of Computer Science
Old Dominion University
Norfolk VA
A set of slides we've used in various presentations to show that replaying an experience via archived web pages is more compelling than reading a summary of the event.
Finding knowledge, data and answers on the Semantic Webebiquity
Web search engines like Google have made us all smarter by providing ready access to the world's knowledge whenever we need to look up a fact, learn about a topic or evaluate opinions. The W3C's Semantic Web effort aims to make such knowledge more accessible to computer programs by publishing it in machine understandable form.
<p>
As the volume of Semantic Web data grows software agents will need their own search engines to help them find the relevant and trustworthy knowledge they need to perform their tasks. We will discuss the general issues underlying the indexing and retrieval of RDF based information and describe Swoogle, a crawler based search engine whose index contains information on over a million RDF documents.
<p>
We will illustrate its use in several Semantic Web related research projects at UMBC including a distributed platform for constructing end-to-end use cases that demonstrate the semantic web’s utility for integrating scientific data. We describe ELVIS (the Ecosystem Location Visualization and Information System), a suite of tools for constructing food webs for a given location, and Triple Shop, a SPARQL query interface which searches the Semantic Web for data relevant to a given query ELVIS functionality is exposed as a collection of web services, and all input and output data is expressed in OWL, thereby enabling its integration with Triple Shop and other semantic web resources.
Aqua Browser Implementation at Oklahoma State Universityyouthelectronix
On Wednesday November 7th Dr. Anne Prestamo discussed "AquaBrowser Implementation at Oklahoma
State Univerity Library" as part of a program on Next Generation Catalogs held at at the University of Massachusetts at Amherst and co-sponsored by the Five Colleges' Librarians
Council and Simmons College Graduate School of Library and Information Science (GSLIS).
Information Extraction and Linked Data CloudDhaval Thakker
In the media industry there is a great emphasis on providing descriptive metadata as part of the media assets to the consumers. Information extraction (IE) is considered an important tool for metadata generation process and its performance largely depend on the knowledge base it utilizes. The advances in the “Linked Data Cloud” research provide a great opportunity for generating such knowledge base that benefit from the participation of wider community. In this talk, I will discuss our experiences of utilizing Linked Data Cloud in conjunction with a GATE-based IE system.
Presentation given at Barcamp Chiang Mai 4 on the basics of Semantic Web. A simple introduction with examples, aimed for those with a little Web development experience.
Raises questions about the true identity of Tim Berners-Lee.
An introduction deck for the Web of Data to my team, including basic semantic web, Linked Open Data, primer, and then DBpedia, Linked Data Integration Framework (LDIF), Common Crawl Database, Web Data Commons.
The common use by archaeologists of ubiquitous technologies such as computers and digital cameras means that archaeological research projects now produce huge amounts of diverse, digital documentation. However, while the technology is available to collect this documentation, we still largely lack community accepted dissemination channels appropriate for such torrents of data. Open Context (http://www.opencontext.org) aims to help fill this gap by providing open access data publication services for archaeology. Open Context has a flexible and generalized technical architecture that can accommodate most archaeological datasets, despite the lack of common recording systems or other documentation standards. Open Context includes a variety of tools to make data dissemination easier and more worthwhile. Authorship is clearly identified through citation tools, a web-based publication systems enables individuals upload their own data for review, and collaboration is facilitated through easy download and other features. While we have demonstrated a potentially valuable approach for data sharing, we face significant challenges in scaling Open Context up for serving large quantities of data from multiple projects.
HIBERLINK: Reference Rot and Linked Data: Threat and RemedyPRELIDA Project
Peter Burnhill (EDINA, University of Edinburgh), presented at the 3rd PRELIDA Consolidation and Dissemination Workshop, Riva, Italy, October, 17, 2014. More information about the workshop at: prelida.eu
Delivered by Peter Burnhill, Director of EDINA, at the PRELIDA Consolidation and Dissemination workshop on 17/18 October 2014 (http://prelida.eu/consolidation-workshop).
Summary: The web changes over time, and significant reference rot inevitably occurs. Web archiving delivers only a 50% chance of success. So in addition to the original URI, the link should be augmented with temporal context to increase robustness.
Web data management has been a topic of interest for many years during which a number of different modelling approaches have been tried. The latest in this approaches is to use RDF (Resource Description Framework), which seems to provide real opportunity for querying at least some of the web data systematically. RDF has been proposed by the World Wide Web Consortium (W3C) for modeling Web objects as part of developing the “semantic web”. W3C has also proposed SPARQL as the query language for accessing RDF data repositories. The publication of Linked Open Data (LOD) on the Web has gained tremendous momentum over the last number of years, and this provides a new opportunity to accomplish web data integration. A number of approaches have been proposed for running SPARQL queries over RDFencoded Web data: data warehousing, SPARQL federation, and live linked query execution. In this talk, I will review these approaches with particular emphasis on some of our research within the context of gStore project (joint project with Prof. Lei Zou of Peking University and Prof. Lei Chen of Hong Kong University of Science and Technology), chameleondb project (joint work with Günes Aluç, Dr. Olaf Hartig, and Prof. Khuzaima Daudjee of University of Waterloo), and live linked query execution (joint work with Dr. Olaf Hartig).
A lecture/conversation focusing on the first 12 years of Semantic Web - delivered on February 21, 2012.
See http://j.mp/SWIntro for more details. More detailed course material is at http://knoesis.org/courses/web3/
Web Archiving in the Year eaee1902f186819154789ee22ca30035Michael Nelson
(Web Archiving in the Year 2025)
My Vision for Trustworthy
Web Archiving in 2025
Michael L. Nelson
@phonedude_mln
with: Scott Ainsworth, Sawood Alam, Mohamed Aturban, John Berlin, Justin Brunelle, Kritika Garg, Hussam Hallak, Himarsha Jayanetti, Mat Kelly, Michele C. Weigle
@WebSciDL
Trust in Web Archives Panel, 2021 Web Archiving Conference
2021-06-16
Uncertainty in replaying archived Twitter pagesMichael Nelson
Michael L. Nelson
@phonedude_mln
with: Sawood Alam, Kritika Garg, Himarsha Jayanetti,
Shawn M. Jones, Nauman Siddique, Michele C. Weigle
@WebSciDL
Ethics and Archiving the Web: How to ethically collect and use web archives
2021-03-30
Web Archives at the Nexus of Good Fakes and Flawed OriginalsMichael Nelson
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group @WebSciDL, @phonedude_mln
Drexel CCI IS Department Distinguished Speaker Series, 2020-03-09
Web Archives at the Nexus of Good Fakes and Flawed OriginalsMichael Nelson
Web Archives at the Nexus of Good Fakes and Flawed Originals
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
@WebSciDL, @phonedude_mln
With:
ODU: Michele C. Weigle, John Berlin, Mohamed Aturban, Justin Whitlock
LANL: Martin Klein, DANS: Herbert Van de Sompel
CNI Spring 2019 Membership Meeting, 2019-04-09,
@phonedude_mln, @WebSciDL
Blockchain Can Not Be Used To Verify Replayed Archived Web PagesMichael Nelson
Blockchain Can Not Be Used To Verify Replayed Archived Web Pages
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
@WebSciDL, @phonedude_mln
With:
ODU: Michele C. Weigle, Mohamed Aturban
Los Alamos National Laboratory: Herbert Van de Sompel, Martin Klein
CNI Fall 2018 Membership Meeting, 2018-12-11,
@phonedude_mln, @WebSciDL
Blockchain Can Not Be Used To Verify Replayed Archived Web PagesMichael Nelson
Blockchain Can Not Be Used To Verify Replayed Archived Web Pages
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
@WebSciDL, @phonedude_mln
With:
ODU: Michele C. Weigle, Mohamed Aturban
Los Alamos National Laboratory: Herbert Van de Sompel, Martin Klein
Weaponized Web Archives: Provenance Laundering of Short Order Evidence Michael Nelson
Weaponized Web Archives: Provenance Laundering of Short Order Evidence
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
@WebSciDL, @phonedude_mln
With:
ODU: Michele C. Weigle, Mohamed Aturban, John Berlin, Sawood Alam, Plinio Vargas
Los Alamos National Laboratory: Herbert Van de Sompel, Martin Klein
Weaponized Web Archives: Provenance Laundering of Short Order Evidence Michael Nelson
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
@WebSciDL, @phonedude_mln
With:
ODU: Michele C. Weigle, Mohamed Aturban, John Berlin, Sawood Alam, Plinio Vargas
Los Alamos National Laboratory: Herbert Van de Sompel, Martin Klein
ODU Computer Science Colloquium 2018-04-06
based on a 2018-03-23 presentation at the National Forum on Ethics and Archiving the Web
Weaponized Web Archives: Provenance Laundering of Short Order Evidence Michael Nelson
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
@WebSciDL, @phonedude_mln
With:
ODU: Michele C. Weigle, Mohamed Aturban, John Berlin, Sawood Alam, Plinio Vargas
Los Alamos National Laboratory: Herbert Van de Sompel, Martin Klein
National Forum on Ethics and Archiving the Web
2018-03-23, #eaw18, @phonedude_mln
Web Archiving Activities of ODU’s Web Science and Digital Library Research G...Michael Nelson
Michael L. Nelson
@phonedude_mln
Michele C. Weigle
@weiglemc
National Symposium on Web Archiving Interoperability
2017-02-21
Many projects joint with LANL
Funding from NSF, IMLS, NEH, and AMF
Summarizing archival collections using storytelling techniquesMichael Nelson
Summarizing archival collections using storytelling techniques
Yasmin AlNoamany
Michele C. Weigle
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
www.cs.odu.edu/~mln/
@phonedude_mln
Research Funded by IMLS LG-71-15-0077-15
Dodging the Memory Hole
Los Angeles, CA, 2016-10-14
The Memento Protocol and Research Issues With Web ArchivingMichael Nelson
Michael L. Nelson
Old Dominion University
Web Science & Digital Libraries Research Group
www.cs.odu.edu/~mln/
University of Virginia Colloquium
2016-09-12
Combining Heritrix and PhantomJS for Better Crawling of Pages with JavascriptMichael Nelson
Justin F. Brunelle
Michele C. Weigle
Michael L. Nelson
Web Science and Digital Libraries Research Group
Old Dominion University
@WebSciDL
IIPC 2016
Reykjavik, Iceland, April 11, 2016
Storytelling for Summarizing Collections in Web ArchivesMichael Nelson
Yasmin AlNoamany
Michele C. Weigle
Michael L. Nelson
Old Dominion University
Web Science and Digital Libraries Group
@WebSciDL
This work is supported in part by IMLS LG-71-15-0077
CNI Spring 2016
2016-04-05
Yasmin AlNoamany
Michele C. Weigle
Michael L. Nelson
Old Dominion University
Web Science and Digital Libraries Group
ws-dl.cs.odu.edu
@WebSciDL
This work is supported in part by IMLS LG-71-15-0077
Old Dominion University ECE Department Colloquium
2015-11-13
@WebSciDL PhD Student Project Reviews August 5&6, 2015Michael Nelson
Herbert Van de Sompel (LANL) visisted the Web Science & Digital Libraries Group @ ODU on August 5--7, 2015. The seven PhD students who were in town at that time reviewed their current status for him.
Evaluating the Temporal Coherence of Archived PagesMichael Nelson
Evaluating the Temporal Coherence of Archived Pages
Scott G. Ainsworth
Michael L. Nelson
Herbert Van De Sompel
IIPC 2015
April 27–May 1, 2015
Stanford University