A closing talk I gave at the JISC/DPC 'Missing Links' conference on web archiving in July 2009. The talks were on the DPC site but ironically the link is now broken.
GLAMorous LOD and ResearchSpace introductionBarry Norton
This document discusses the development of ResearchSpace (RS), a platform that allows researchers to make claims by adding to and modifying data from cultural heritage institutions in a way that preserves canonical data. RS components include search, data annotation, image annotation, a "data basket" for collecting items, a dashboard, and conjunctive search. It also discusses fundamental relationships that can be represented in linked cultural data.
This document summarizes past and recent web archive activities conducted by the Oxford Internet Institute (OII). It describes projects from 2008-2015 funded by JISC, AHRC, and IIPC that involved analyzing large web archive datasets capturing the UK domain. These projects extracted metadata, links, and temporal data from over 30TB of archived web pages to study topics like the growth of UK universities online and how government and media presence has changed over time. Current work includes additional case studies and making the processed UK domain data openly available to support research on the history and evolution of the British web space.
What is a digital library, repository or platform?Jisc
A digital library, repository or platform is a place to discover, create and share knowledge and ideas through open access to digital information and connections between people. It acts as both a repository for preserving collections of knowledge and a facilitator for developing and sharing new ideas. An important role is providing a trust framework to ensure the provenance or origins of information are verifiable. The goals are to apply the technologies and culture of the internet to respond to people's expectations of access while serving the traditional library functions of being a place of record and discovery in a digital context.
This document discusses the development of linked data and the semantic web over the past 13 years. It outlines how initially the goal was to build the semantic web as a precursor to use, but that approach changed to focus on publishing data so that people could start building applications using that data incrementally. Two examples are given of published linked data sets from the British Museum and LinkedBrainz. The document argues that linked data is now about enabling systems integration across different applications and domains. It also addresses concerns about publishing linked data leading to untrue claims, and introduces ResearchSpace, a platform for researchers to make annotated claims and arguments about GLAM data using linked data techniques.
An introduction to the British Library's digital collections, resources and partnerships. Presented at the 'Discovering Collections, Discovering Communities' 2015 conference (Salford, 13 October 2015)
Peter webster interrogating the archived uk webDigital History
This document summarizes a project that analyzes the JISC UK Web Domain Dataset from 1996-2013 to understand the development of UK web space over time. The project aims to establish frameworks for analyzing web archives and explore ethical implications. It will produce tools to support analysis, case studies across disciplines, and training materials. The dataset contains around 300 million resources from the UK web captured by the Internet Archive, but lacks metadata about subjects and dates. The project highlights the value of web archives as historical sources.
Open Access Week 2013: Events at DartmouthShirley Zhao
This document announces events for Open Access Week from October 21-25, 2013 at Dartmouth College. There will be information tables set up around campus on different days during the week to provide information about open access. Two panel discussions are scheduled - one on Monday with the World Bank on open access and impact, and another on Wednesday about publishing essentials. On Thursday there will be a workshop about copyright laws and exceptions. The document also provides background on open access, digital initiatives at Dartmouth Library, and open access publishing services available through the library.
GLAMorous LOD and ResearchSpace introductionBarry Norton
This document discusses the development of ResearchSpace (RS), a platform that allows researchers to make claims by adding to and modifying data from cultural heritage institutions in a way that preserves canonical data. RS components include search, data annotation, image annotation, a "data basket" for collecting items, a dashboard, and conjunctive search. It also discusses fundamental relationships that can be represented in linked cultural data.
This document summarizes past and recent web archive activities conducted by the Oxford Internet Institute (OII). It describes projects from 2008-2015 funded by JISC, AHRC, and IIPC that involved analyzing large web archive datasets capturing the UK domain. These projects extracted metadata, links, and temporal data from over 30TB of archived web pages to study topics like the growth of UK universities online and how government and media presence has changed over time. Current work includes additional case studies and making the processed UK domain data openly available to support research on the history and evolution of the British web space.
What is a digital library, repository or platform?Jisc
A digital library, repository or platform is a place to discover, create and share knowledge and ideas through open access to digital information and connections between people. It acts as both a repository for preserving collections of knowledge and a facilitator for developing and sharing new ideas. An important role is providing a trust framework to ensure the provenance or origins of information are verifiable. The goals are to apply the technologies and culture of the internet to respond to people's expectations of access while serving the traditional library functions of being a place of record and discovery in a digital context.
This document discusses the development of linked data and the semantic web over the past 13 years. It outlines how initially the goal was to build the semantic web as a precursor to use, but that approach changed to focus on publishing data so that people could start building applications using that data incrementally. Two examples are given of published linked data sets from the British Museum and LinkedBrainz. The document argues that linked data is now about enabling systems integration across different applications and domains. It also addresses concerns about publishing linked data leading to untrue claims, and introduces ResearchSpace, a platform for researchers to make annotated claims and arguments about GLAM data using linked data techniques.
An introduction to the British Library's digital collections, resources and partnerships. Presented at the 'Discovering Collections, Discovering Communities' 2015 conference (Salford, 13 October 2015)
Peter webster interrogating the archived uk webDigital History
This document summarizes a project that analyzes the JISC UK Web Domain Dataset from 1996-2013 to understand the development of UK web space over time. The project aims to establish frameworks for analyzing web archives and explore ethical implications. It will produce tools to support analysis, case studies across disciplines, and training materials. The dataset contains around 300 million resources from the UK web captured by the Internet Archive, but lacks metadata about subjects and dates. The project highlights the value of web archives as historical sources.
Open Access Week 2013: Events at DartmouthShirley Zhao
This document announces events for Open Access Week from October 21-25, 2013 at Dartmouth College. There will be information tables set up around campus on different days during the week to provide information about open access. Two panel discussions are scheduled - one on Monday with the World Bank on open access and impact, and another on Wednesday about publishing essentials. On Thursday there will be a workshop about copyright laws and exceptions. The document also provides background on open access, digital initiatives at Dartmouth Library, and open access publishing services available through the library.
Blogs are online diaries that can be updated from anywhere and include multimedia content and links. They allow for commenting and sharing of opinions. Libraries use blogs to efficiently disseminate current information to large audiences and allow communication between staff and patrons. Examples provided are a graphic novel blog from a state library targeting that collection's demographic, and a readers' advisory blog about Web 2.0 tools for librarians.
Digital Social Science Lab: Connecting academia with data literacy Christian Lauersen
Talk at Linköping University Library 15/12 - 217 about the creation of Digital Social Science Lab at Faculty Library of Social Sciences, Copenhagen University Library and how to support data literacy within academia and higher education
The document discusses the benefits of open access research over traditional publishing models. Open access research is available online for free without restrictions. It has a wider audience which can increase citations and impact. However, traditional publishers charge high subscription costs that are difficult for universities to sustain. The document advocates sharing research outputs through open access platforms to maximize exposure and dissemination of scholarly work.
This is a short run through the activities of the Office of Scholarly Communication at the University of Cambridge presented to the Cambridge University Press Library Board meeting on 28 November 2016.
How to cope when you open your digital doorsPaul Rowe
Museums are striving to share information about their collections online. What impact does this have on their organisations and how can they cope?
This presentation is the result of a survey of 16 different online collections where the organisations have chosen to publish the majority of their records online.
WMS presentation for OCLC conference July 2013Nicola Perry
The document discusses Bishop Grosseteste University's transition from their previous library system, OLIB, to the cloud-based WorldShare Management Services (WMS) system provided by OCLC. It describes the university, which has 2,000 students and a library with 135,000 items. The migration to WMS took one year to complete and involved loading the library's catalog data, patron and circulation records, and configuring the new system. Staff found WMS to be web-based, provide regular updates, and allow them to take advantage of the resources of being part of the larger OCLC network.
Maphub und Pelagios: Anwendung von Linked Data in den Digitalen Geisteswissen...Bernhard Haslhofer
In recent years, scientists at the Austrian Institute of Technology have been involved in numerous projects in the digital humanities area. In this talk, Dr. Bernhard Haslhofer will present two of them, both having a strong focus on applying the Linked Open Data method on datasets produced throughout the project. The first is Maphub (http://maphub.github.io/), an open source Web application which allows users to create annotations on historical maps, link these annotations with other Web sources (e.g., Wikipedia), and share annotations as Linked Open Data following the Open Annotation model. The second is Pelagios (http://pelagios-project.blogspot.co.at/), a community initiative that aims to facilitate better linking between online resources documenting the past, based on the places they refer to. To date, Pelagios interconnects 900.000+ heterogeneous digital objects - literature, archaeology, epigraphy, cartography - from 40+ international partners. The current focus of the project is to annotate Early Geospatial Documents - documents that use written or visual representation to describe geographic space prior to the European discovery of the Americas in 1492, and make the annotations available as (Linked) Open Data.
An overview of Wikipedia and its potential for libraries, also covering cataloguing issues. Part of the Cataloguing and Indexing Group in Scotland (CIGS) seminar "Toto, I've got a feeling we're not in Kansas anymore": metadata issues and Web2.0 services.
For repositories to succeed they have to end. Reflections on (not just) the U...Torsten Reimer
Presentation given at the Open Repositories 2018 conference in Bozeman, Montana, 6th June 2018. Starting with an assessment of the UK open access repository environment, this presentation asks broader questions about the state of the open repository landscape globally. In response to a report to the UK government on open access, Universities UK have set up a repositories working group to identify issues where common benefit can be delivered and actions that can be taken. In this talk I will combine my own assessment of the repository landscape with a summary of the work of the working group and its recommendations. The presentation will also introduce work underway at the British Library to address some of the issues the working group has identified, including an assessment of a national OA preservation solution and a shared-services repository infrastructure. I will make the case that to realise the benefits of open repositories we need to move away from the model of locally hosted repositories.
Linda Spurdle Pre Raphaelite Online ResourceMartin Bazley
Talk presented as part of Creating Online Exhibitions on 2 Nov 09 at the British Museum, run by the E-Learning Group for Museums, Libraries and Archives
1. The document discusses the history and evolution of books and publishing, from the printing press to modern digital formats and standards.
2. It traces milestones like the development of hypertext in the 1960s, the invention of the World Wide Web in the 1980s/90s, and modern standards for ebooks and web publications.
3. The document argues that while books have long provided structure and organization of content, digital formats now integrate more tightly with the web and allow for new capabilities like text mining and semantic augmentation while still providing standardization and citation mechanisms.
The presentation will provide an overview of DPLA, including the current partnership model and future plans for growth. The talk will also describe DPLA’s infrastructure and technologies, metadata model, open access and rights policies, as well as DPLA outreach and engagement programs.
The data journal: incentivizing open scholarship or 'a convenient fiction'?Brian Hole
The document discusses the role of data journals and data papers in incentivizing open scholarship. It notes that scientific practice involves validation, dissemination and further development of research, but that currently publishers, researchers and repositories are not fulfilling these obligations with respect to data. Data journals and data papers can help by providing incentives for researchers to publish and cite data, similarly to how conventional papers and citations currently work. This would help formalize data sharing as part of the normal scholarly process.
The document proposes the creation of a South Eastern European Knowledge Network to build a knowledge repository for the region and facilitate exchange of good practices. It would use open source tools to efficiently share experiences, materials, and artifacts in a digital library and open access journal. The network would initially build a directory of practitioners and institutions in the region to establish cooperation.
As an educator, the instructor would accept research from wikis as they provide up-to-date and extensive information, though alternatives would be needed for students without internet access. As a student, the individual would rely on wikis for assignments as they allow gathering a large amount of information from diverse sources and authors to expand one's knowledge. Wikis are developing into a more prominent internet tool as technology advances society and knowledge on wikis increases, enhancing how information is applied across different fields to promote a more knowledgeable society.
What can the DCC do for you? Sheffield RoadshowKevin Ashley
A description of the ways in which the Digital Curation can work with institutions to improve research data management at institutional level. Delivered at the 2nd DCC roadshow, Sheffield, 2011-03-01
Blogs are online diaries that can be updated from anywhere and include multimedia content and links. They allow for commenting and sharing of opinions. Libraries use blogs to efficiently disseminate current information to large audiences and allow communication between staff and patrons. Examples provided are a graphic novel blog from a state library targeting that collection's demographic, and a readers' advisory blog about Web 2.0 tools for librarians.
Digital Social Science Lab: Connecting academia with data literacy Christian Lauersen
Talk at Linköping University Library 15/12 - 217 about the creation of Digital Social Science Lab at Faculty Library of Social Sciences, Copenhagen University Library and how to support data literacy within academia and higher education
The document discusses the benefits of open access research over traditional publishing models. Open access research is available online for free without restrictions. It has a wider audience which can increase citations and impact. However, traditional publishers charge high subscription costs that are difficult for universities to sustain. The document advocates sharing research outputs through open access platforms to maximize exposure and dissemination of scholarly work.
This is a short run through the activities of the Office of Scholarly Communication at the University of Cambridge presented to the Cambridge University Press Library Board meeting on 28 November 2016.
How to cope when you open your digital doorsPaul Rowe
Museums are striving to share information about their collections online. What impact does this have on their organisations and how can they cope?
This presentation is the result of a survey of 16 different online collections where the organisations have chosen to publish the majority of their records online.
WMS presentation for OCLC conference July 2013Nicola Perry
The document discusses Bishop Grosseteste University's transition from their previous library system, OLIB, to the cloud-based WorldShare Management Services (WMS) system provided by OCLC. It describes the university, which has 2,000 students and a library with 135,000 items. The migration to WMS took one year to complete and involved loading the library's catalog data, patron and circulation records, and configuring the new system. Staff found WMS to be web-based, provide regular updates, and allow them to take advantage of the resources of being part of the larger OCLC network.
Maphub und Pelagios: Anwendung von Linked Data in den Digitalen Geisteswissen...Bernhard Haslhofer
In recent years, scientists at the Austrian Institute of Technology have been involved in numerous projects in the digital humanities area. In this talk, Dr. Bernhard Haslhofer will present two of them, both having a strong focus on applying the Linked Open Data method on datasets produced throughout the project. The first is Maphub (http://maphub.github.io/), an open source Web application which allows users to create annotations on historical maps, link these annotations with other Web sources (e.g., Wikipedia), and share annotations as Linked Open Data following the Open Annotation model. The second is Pelagios (http://pelagios-project.blogspot.co.at/), a community initiative that aims to facilitate better linking between online resources documenting the past, based on the places they refer to. To date, Pelagios interconnects 900.000+ heterogeneous digital objects - literature, archaeology, epigraphy, cartography - from 40+ international partners. The current focus of the project is to annotate Early Geospatial Documents - documents that use written or visual representation to describe geographic space prior to the European discovery of the Americas in 1492, and make the annotations available as (Linked) Open Data.
An overview of Wikipedia and its potential for libraries, also covering cataloguing issues. Part of the Cataloguing and Indexing Group in Scotland (CIGS) seminar "Toto, I've got a feeling we're not in Kansas anymore": metadata issues and Web2.0 services.
For repositories to succeed they have to end. Reflections on (not just) the U...Torsten Reimer
Presentation given at the Open Repositories 2018 conference in Bozeman, Montana, 6th June 2018. Starting with an assessment of the UK open access repository environment, this presentation asks broader questions about the state of the open repository landscape globally. In response to a report to the UK government on open access, Universities UK have set up a repositories working group to identify issues where common benefit can be delivered and actions that can be taken. In this talk I will combine my own assessment of the repository landscape with a summary of the work of the working group and its recommendations. The presentation will also introduce work underway at the British Library to address some of the issues the working group has identified, including an assessment of a national OA preservation solution and a shared-services repository infrastructure. I will make the case that to realise the benefits of open repositories we need to move away from the model of locally hosted repositories.
Linda Spurdle Pre Raphaelite Online ResourceMartin Bazley
Talk presented as part of Creating Online Exhibitions on 2 Nov 09 at the British Museum, run by the E-Learning Group for Museums, Libraries and Archives
1. The document discusses the history and evolution of books and publishing, from the printing press to modern digital formats and standards.
2. It traces milestones like the development of hypertext in the 1960s, the invention of the World Wide Web in the 1980s/90s, and modern standards for ebooks and web publications.
3. The document argues that while books have long provided structure and organization of content, digital formats now integrate more tightly with the web and allow for new capabilities like text mining and semantic augmentation while still providing standardization and citation mechanisms.
The presentation will provide an overview of DPLA, including the current partnership model and future plans for growth. The talk will also describe DPLA’s infrastructure and technologies, metadata model, open access and rights policies, as well as DPLA outreach and engagement programs.
The data journal: incentivizing open scholarship or 'a convenient fiction'?Brian Hole
The document discusses the role of data journals and data papers in incentivizing open scholarship. It notes that scientific practice involves validation, dissemination and further development of research, but that currently publishers, researchers and repositories are not fulfilling these obligations with respect to data. Data journals and data papers can help by providing incentives for researchers to publish and cite data, similarly to how conventional papers and citations currently work. This would help formalize data sharing as part of the normal scholarly process.
The document proposes the creation of a South Eastern European Knowledge Network to build a knowledge repository for the region and facilitate exchange of good practices. It would use open source tools to efficiently share experiences, materials, and artifacts in a digital library and open access journal. The network would initially build a directory of practitioners and institutions in the region to establish cooperation.
As an educator, the instructor would accept research from wikis as they provide up-to-date and extensive information, though alternatives would be needed for students without internet access. As a student, the individual would rely on wikis for assignments as they allow gathering a large amount of information from diverse sources and authors to expand one's knowledge. Wikis are developing into a more prominent internet tool as technology advances society and knowledge on wikis increases, enhancing how information is applied across different fields to promote a more knowledgeable society.
What can the DCC do for you? Sheffield RoadshowKevin Ashley
A description of the ways in which the Digital Curation can work with institutions to improve research data management at institutional level. Delivered at the 2nd DCC roadshow, Sheffield, 2011-03-01
Audit and outsourcing: their role in creating interoperable repository infras...Kevin Ashley
A brief presentation for the REPRISE workshop before IDCC09 (2009-12-02) in London. I look at the role that audit and outsourcing play in helping deliver interoperable preservation in repositories.
JISC repositories and preservation programme: Plenary presentation 2009Kevin Ashley
The document summarizes the Repositories and Preservation Programme that was conducted by JISC, looking back at what was asked of participants and what was accomplished, and looking forward to the future direction. Specifically:
1) JISC asked participants to create more repositories, enhance existing ones, and provide services to help and exploit repository content through specific targeted projects.
2) Participants established more repositories, built on existing successes, and created services to help with discovery, deposit, and application profiles.
3) Looking ahead, the document suggests moving away from individual projects and toward more joined-up international activities, exposing and sharing content across repositories to better support research, teaching, and learning.
Research data for repository managers Kevin Ashley
A presentation given at ULCC's Institutional Repository Manager's workshop 2012 on 2012-06-15. Aimed at getting traditional repository managers to think about their role in research data management.
Inverting the data pyramid: maximising the value of data reuse (IMCW2014/ICKM...Kevin Ashley
This document summarizes a presentation on research data management and reuse. It discusses:
1. The Digital Curation Centre's (DCC) mission to increase research data services capabilities in UK institutions and how this is an international issue.
2. How data reuse is already occurring but could be expanded, providing benefits for research quality, speed, and costs. Proper data management can also help ensure research integrity.
3. Barriers to increased data reuse including lack of infrastructure and services in some domains, and variability in data management practices between fields. Overcoming these issues requires attention from senior researchers, librarians, and policymakers.
Supporting open research - how to help your researchers - Vitae15Kevin Ashley
A talk given at a Vitae event in Leeds, 2015-12-01, on how universities and other research organisations can help their researchers practice open research, with a special focus on the training resources provided by FOSTER.
The document introduces library resources for students at Lingnan University, including digital library resources, online databases, and tips for searching the web for information. It provides the library homepage address and describes tools for finding news stories and online lectures. It also gives guidance on evaluating web information and recommends subject directories and search engines for broad or specific topics. Tips are included for searching tools like Google.
Simple Knowledge Organization System (SKOS) in the Context of Semantic Web De...gardensofmeaning
The document discusses the development and use of SKOS (Simple Knowledge Organization System) for representing knowledge organization systems like thesauri and classification schemes as structured data on the semantic web. It describes how LCSH (Library of Congress Subject Headings) has been converted to RDF using SKOS and published as linked open data. It suggests further steps like linking LCSH to other metadata and developing RDF representations of additional bibliographic schemas.
This is a very basic workshop to introduce novice users to Omeka with an eye towards providing hands-on experience to decide whether it can serve their own research needs.
Talk given at Society of Southwest Archivists 2013 annual meeting, discussing web archiving in the context of university archives. Explores why web archiving is important for university archives, some methods for web archiving, and technological and ethical challenges.
Presentation given by Marieke Guy on "Preservation for the Next Generation" at the Internet Librarian International 2008 conference held at the Novotel London West, London on 16th October 2008.
The common use by archaeologists of ubiquitous technologies such as computers and digital cameras means that archaeological research projects now produce huge amounts of diverse, digital documentation. However, while the technology is available to collect this documentation, we still largely lack community accepted dissemination channels appropriate for such torrents of data. Open Context (http://www.opencontext.org) aims to help fill this gap by providing open access data publication services for archaeology. Open Context has a flexible and generalized technical architecture that can accommodate most archaeological datasets, despite the lack of common recording systems or other documentation standards. Open Context includes a variety of tools to make data dissemination easier and more worthwhile. Authorship is clearly identified through citation tools, a web-based publication systems enables individuals upload their own data for review, and collaboration is facilitated through easy download and other features. While we have demonstrated a potentially valuable approach for data sharing, we face significant challenges in scaling Open Context up for serving large quantities of data from multiple projects.
SENESCHAL: Semantic ENrichment Enabling Sustainability of arCHAeological Link...CIGScotland
Presented at Linked Open Data: current practice in libraries and archives (Cataloguing & Indexing Group in Scotland 3rd Linked Open Data Conference), Edinburgh, 18 Nov 2013
The document discusses the relevance of classification and indexing for organizing internet resources. It argues that while the internet has grown without formal organization tools like classification schemes and vocabularies, bringing such library techniques to the web could help address the problem of information being difficult to find, access, and retrieve from the vast, unorganized data available online. Applying concepts from knowledge organization, like subject descriptors and relationships between resources, is an approach taken by the semantic web to help machines better understand and process information on the web.
The document discusses using wikis in education. It defines wikis and their basics like easy collaborative editing. It provides many examples of using wikis in the classroom for activities like class pages, student projects, and exam preparation. Wikis can also be used outside the classroom for projects, training, and outreach. The document discusses considerations for wiki software and hosting options. It demos the WetPaint wiki platform and provides additional resources.
This document discusses how libraries are using the open source content management system Drupal. Drupal provides a flexible architecture that allows libraries to support different types of content on their websites and create custom modules. It also enables libraries to build intranets with modules for events, forms, and taxonomies. Additionally, Drupal functions as a social networking tool with features like groups, feeds, and chat. Libraries have implemented Drupal for digital collections, merging catalogs with websites, and other innovative projects. Drupal provides a platform for libraries to adapt technologies to their needs.
Web 2.0 session for library staff - 2008 versionlearning20
This document discusses the concept of Web 2.0 and its potential impact for information professionals. It defines Web 2.0 as the second generation of web-based communities and services that aim to facilitate creativity, collaboration, and sharing among users. Some key Web 2.0 technologies discussed include blogs, wikis, RSS feeds, social bookmarking, social networking, podcasts/webcasts, and instant messaging. The document explores how libraries can use these technologies both to interact with and better serve their users.
This document discusses the concepts of Web 2.0 and Library 2.0. It provides examples of how libraries are adopting Web 2.0 technologies like blogs, wikis, RSS feeds, and engaging users in more participatory ways. Examples mentioned include libraries that have implemented blogs, podcasts, tagging features for catalogues, and virtual libraries in Second Life. The document also discusses who the main users are of these new technologies, namely younger "millennial" generations who have high usage of social software and expect to access services anytime on any device.
The document discusses the evolution of the World Wide Web from a global communication medium to an open programmable platform for data processing services. It introduces some key Web 2.0 services like social networks, maps, news, and photo sharing. It then shows how these services can be combined into "mashups" with basic programming knowledge to create new applications.
This document discusses how Web 2.0 technologies like RSS, ATOM, and XML can be applied to library research tools and services. It provides examples of how these formats can be used to share and aggregate research-related content like tags on Delicious, keywords in WorldCat, and updates on Wikipedia. The document encourages setting up alerts and widgets using these feeds to integrate library materials into the research workflow.
Hypertext2007 Wendy Hall - "Whatever Happened to Hypertext?"hypertext2007
The document discusses the history of hypertext conferences from 1987 to 2007 and the transition to web and semantic web conferences. It notes that early hypertext conferences in the late 1980s saw significant participation from technical and non-technical authors. Conference attendance declined in the late 1990s as the web conference series grew rapidly in popularity. The semantic web represents a rebirth of hypertext ideas but does not attract the original hypertext community.
In 2015, I created a web archiving fundamentals course for the Society of American Archivists (SAA) Digital Archives Specialist (DAS) program. This is a portion of the slide deck I used for that course.
The document provides an overview of the history and principles of the Semantic Web and linked open data. It discusses how the Semantic Web aims to add meaning to information on the web by identifying things with URIs and linking related resources. The core components are RDF triples that describe resources and ontologies that define classes and properties. Linked open data publishes structured data on the web in ways that allow them to be interlinked and accessed via SPARQL queries. Examples demonstrate how semantic search and applications are using these techniques.
Similar to Missing links closing talk - with notes (20)
RISE - the DCC's Research Infrastructure Self-Evaluation FrameworkKevin Ashley
The document introduces RISE, a research infrastructure self-evaluation framework developed by the Digital Curation Centre. RISE aims to help research institutions assess the maturity of their research data services, identify gaps, prioritize improvements, and benchmark against peers. It comprises 22 capabilities across different levels of achievement. The framework was created based on the DCC's experience and incorporates standards from existing models. It is freely available online and several institutions have conducted self-assessments using RISE.
An analysis of open data and open science policies in Europe - a SPARCEurope ...Kevin Ashley
A short presentation given at the SPARCEurope members meeting on July 5th in Patras, Greece. It summarises the findings of a recent joint report by the DCC and SPARCEurope on European national open data and open science policy.
This document summarizes a presentation on the benefits of research data management. It discusses how data management can benefit researchers through increased citations and compliance with funder requirements. It also benefits society by enabling data sharing, reuse and discovery. However, many researchers do not practice good data management due to a lack of skills, resources or incentives. The presentation provides information on data management best practices and their importance for research excellence.
My data, your data, our data - increasing data value through reuse (Eurocris2...Kevin Ashley
My keynote talk for Eurocris2014, Rome. I make the case for reuse of research data, discuss the barriers and look at ways we are trying to overcome them.
Use and reuse: research data locally & globally #esipfedKevin Ashley
The document discusses the importance of research data reuse and the growing demands by funders for data management and sharing. It notes that properly managing and sharing research data can improve research quality, speed, and cost effectiveness. However, many researchers remain reluctant to share data due to various excuses. The document advocates for national research data infrastructure and services to support universities in meeting funder requirements and overcoming barriers to data sharing.
Data Quality and Data Curation - a personal viewKevin Ashley
- The document discusses data quality and curation from the perspective of Kevin Ashley, director of the Digital Curation Centre.
- It notes that different stakeholders have varying definitions of data quality, as some aspects of quality, like accuracy, may conflict with others like timeliness or completeness.
- It suggests that current curation practices often only cater to single consumer groups and domains, and that taking a more generic approach could increase data mobility and reuse across different domains.
The document discusses the importance of good research data management. It notes that good data is needed for good research and outlines funder requirements for data management plans and long-term data preservation. The Digital Curation Centre (DCC) provides tools, services, and support to help research institutions develop their research data management capabilities and policies.
Research Data Management: the UK national change programme (Nordbib)Kevin Ashley
The UK National Change Program aims to realize the maximum value of research data through a 5-year program of national services and support coordinated by the Digital Curation Centre. The program focuses on building capacity and capability for research data management within research institutions. Proper research data management is important because data is expensive to create, facilitates reuse and reproducibility, and is increasingly subject to legal and regulatory requirements from research funders.
This document reviews challenges in digital preservation research by examining past reports that identified key research areas. It discusses work that has been done, is currently being done, and remains to be done. Some areas explored include format migration, repository models, metadata standards, and preserving newer digital formats and software. The document emphasizes the need for both pragmatic and theoretical research that can inform practice and help define problems more specifically to guide future work.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
I’m here to talk with you today about some use cases for web archives, and more widely about what people might want to know about the history of the web and what’s on the web. And the reason for thinking about these things is that they can and should influence how we collect stuff, what we collect, but more importantly how we provide access to it. But first, to my title. It is somewhat contrived. And I can reveal that the reason I chose it is simply because it can be shortened, so. This enables me to lay claim to be the first person to mention web 8.0. And that’s pretty pointless, isn’t it. The one thing we can be sure about is that whenever we need to refer to another shift in the usage model of the web similar to that which inspired the name ‘web 2.0’, we won’t be calling it ‘web 3.0’ (although some marketers already are.) We’ll need some other entirely new metaphor to describe an entirely new shift. But if we did stick to using numbers, web 8.0 wouldn’t be with us for another 60 to 90 years. Most of us will be dead. None of us can imagine what it will be like. And some people will be looking back at the web and web 2.0 and marvelling at how primitive, quaint and amusing we all were. And how frustrating - because we can guarantee that we won’t have captured something of importance to the future.
Perhaps we can get some perspective on this by considering our present-day view of a similar technological revolution of the past, the telegraph. This popular history by Tom Standage explicitly tries to draw parallels between the change which the telegraph brought to its day, and that which the internet has brought us now. The comparison is a bit strained, but amongst the points of interest to me are the aspects of the history of the telegraph that make up the book. The contains of messages and their authorship is certainly a part of the history. But just as important are contemporary reactions to it and descriptions of how the technology affected other aspects of life. One image that has stayed with me is of the receptacles which appeared on the back of horse-drawn trams in Paris into which people could drop telegraph messages on pre-paid stationery, from whence they would be delivered to a telegraph office en route for speedy onward delivery. Historians of communication technology are interested in more than simply what was communicated, and by who.
Let’s look at another communication technology, which also allows individuals to send messages to each other. If we have a small number of these messages we might well be interested in analysing their content. If we have a few tens of millions of them, though, other things become interesting, as this visualisation demonstrates. It shows SMS traffic in Amsterdam in the days leading up to and following new year’s eve. There’s a lot of information in this visualisation and a lot more that could be done with the data behind it. Yet it was done without knowing the contents of any of the messages, nor who they were from nor who they were to. We simply know where they were sent and when. When you have enough data of a given type, the useless becomes useful.
So my concern about web archives comes partly from a feeling that some of them are too document-centred, imaging that the only user is one sitting in front of a screen selecting past pages to view one by one. That is one possible use of a web archive, but by no means the only one. We can be interested in content in aggregate; we can be interested in the properties of content, the web of data and as data, and material about the web as well as from it - as Standage’s history makes clear.
So I’m not suggesting thhat no one wants to view single pages. Sometimes we’re interested in their content more than their presentation, as we heard from the archivepress folk today. And sometimes the presentation is key, and sometimes it’s both. We might be interested in timeslices - a bunch of pages from a given point in time - or in investigating how one page or set of pages change over time. Brian Kelly’s presentation of the history of the University of Bath homeppage is a useful illustration, painstakingly constructed by him using the Internet Archive. Couldn’t we make access like this a little easier for people ?
But when we have content in aggregate new possibilities arise. Textual analysis can tell us many things. We can look at the contrasting use of language between different types of sites, we can track the spread of neoligisms, concepts and rumours, or we can do something as simple as constructing word clouds from pages over time. Neil Grindley is currently examining the text of the entire JISC website to look for overuse of particular forms of language, or to constrast mentions of teaching and learning as opposed to research. And none of these involve a human looking at a web page.
But properties of the content can be just as interesting. To take one simple example, we might want to see how the takeup of a particular image format such as PNG took place. Was it geographically uniform, or uniform over types of site? What did it replace - all image formats, or GIF, BMP or JPG preferentially ? And looking back further, whatever happened to XPM, the original icon format used by early graphical web browsers ?
We expect archives to offer some sort of search. But how should it operate ? This was the result of carrying out a search using a well known search engine yesterday for ‘web archive’. What would the results have been if I carried out that search 10 years ago ? Is any archive’s search interface capable of telling me that, or even getting close to it ?
And on a larger scale, the web itself is data, connected data, connected metadata, with connections that shift and break and change meaning over time. There are a host of ways of visualising this, some more suited to individual sites, some to cluster of sites on a topic or over time and some to the entire web, or large parts of it. I want to be able to do these things with archived web content.
And increasingly there is talk of the web of data - the use of the web to link together data and construct something greater than the sum of the parts doing it. This well-known diagram is a visualisation of that state of the web of data a few months ago and it’s since been joined by much more government and research data from a variety of sources. Now one can argue that dealing with this isn’t the job of a web archive - the data is not of the web, but simply on the web, and it’s the job of database archives to preserve it. But some of the data really does make sense only in a web context, and there’s no doubt that the web is being used to do stuff with it that wou;dn’t happen otherwise. I think we should be able to look at the web of data as it was as well as as it is and will be. I confess I don’t know how we might do that, but the possibility should not be ruled out.
APIs are key to allowing much of this to happen. Those which are web-friendly encourage the developmentof innovative ways of accessing content. They allow archives to concentrate on collecting material, protecting it and providing permanent references for their content, whilst permitting a variety of viewing and access methods to emerge. Moreover, those that permit bulk access enable intelligent agents to work for us or to work alongside us in exploring archived content - agents that don’t need to be developed by the archives themselves, but which can be exploited by them. If the archive space fragments, APIs are particlarly important to allow people to spread research over a variety of archives. Let’s look at one reason that should persuade you why APIs are a good thing. Taggalaxy allows the exploration of flickr content in a way that is completely different from flickr, yet depends only on the metadata that flickr exposes via its API. Imagine exploring a web archive this way.
And let’s not forget that there’s more stuff than web pages that might be of future interest. Some of it is already being preserved, such as traditional media about the web - the flurry of consumer magazines that emerged in the 1990s, for instance, although not the premium-rate telephone line that told you the ‘cool site of the day.’ Usage logs, server configs, etc are all part of the history of the web in some form, as is the software that makes it possible. I’m not suggesting that web archivists deal with this stuff - but someone should.
But finally a few more of those content visualisations, all taken from Martin Dodge’s cyber-geography pages - a discipline that’s been going long enough that his pages are no longer maintained and are out of date.