Keynote presentation delivered at ELAG 2013 in Gent, Belgium, on May 29 2013. Discusses Research Objects and the relationship to work my team has been involved in during the past couple of years: OAI-ORE, Open Annotation, Memento.
The presentation explores the trend towards a scholarly communication system that is friendly to machines. It presents 3 exhibits illustrating the trend and 1 exhibit illustrating inertia in the system. It makes the point that machine-actionability can be much easier achieved if content and metadata are available in Open Access and under a permissive Creative Commons license. It also observes that even with content and metadata openly available, new costs related to advanced tools to explore the scholarly record will emerge. Finally, it points at significant challenges regarding the persistence of the scholarly record in light of increasingly interconnected and actionable content and advanced tools to interact with it.
The slides were used for a plenary presentation at the LIBER 2011 Conference in Barcelona, Spain, on June 30 2011.
Keynote presentation delivered at ELAG 2013 in Gent, Belgium, on May 29 2013. Discusses Research Objects and the relationship to work my team has been involved in during the past couple of years: OAI-ORE, Open Annotation, Memento.
The presentation explores the trend towards a scholarly communication system that is friendly to machines. It presents 3 exhibits illustrating the trend and 1 exhibit illustrating inertia in the system. It makes the point that machine-actionability can be much easier achieved if content and metadata are available in Open Access and under a permissive Creative Commons license. It also observes that even with content and metadata openly available, new costs related to advanced tools to explore the scholarly record will emerge. Finally, it points at significant challenges regarding the persistence of the scholarly record in light of increasingly interconnected and actionable content and advanced tools to interact with it.
The slides were used for a plenary presentation at the LIBER 2011 Conference in Barcelona, Spain, on June 30 2011.
As the scholarly communication system evolves to become natively web-based and starts supporting the communication of a wide variety of objects, the manner in which its essential functions – registration, certification, awareness, archiving - are fulfilled co-evolves. This presentation focuses on the nature of the archival function based on a perspective of the future scholarly communication infrastructure. This presentation, prepared for a meeting in June 2014, is based on and updates a previous one that was prepared for a January 2014 meeting. The latter is available at http://www.slideshare.net/atreloar/scholarly-archiveofthefuture
The European Open Science Cloud: just what is it?Carole Goble
Presented at Jisc and CNI leaders conference 2018, 2 July 2018, Oxford, UK (https://www.jisc.ac.uk/events/jisc-and-cni-leaders-conference-02-jul-2018). The European Open Science Cloud. What exactly is it? In principle it is conceived as a virtual environment with open and seamless services for storage, management, analysis and re-use of research data, across borders and scientific disciplines. How? By federating existing scientific data infrastructures, currently dispersed across disciplines and Member States. In practice, what it is depends on the stakeholder. To European Research Infrastructures it’s a coordinated mission to organise and exchange their data, metadata, software and services to be FAIR – Findable, Accessible, Interoperable, Reusable – and to use e-Infrastructures, either EU or commercial. To EU e-Infrastructures offering data storage and cloud services, it’s a funding mission to integrate their services, policies and organisational structures, and to be used by the Research Infrastructures. To agencies it’s a means to promote Open Science, standardisation, cross-disciplinary research and coordinated investment with a dream of a “one stop shop” for researchers. And for Libraries?
Specimen-level mining: bringing knowledge back 'home' to the Natural History ...Ross Mounce
A talk given at the Geological Society of London, UK on 2016/03/09 as part of the Lyell meeting on Palaeoinformatics. http://www.geolsoc.org.uk/lyell16 #lyell16
Open Science: how to serve the needs of the researcher? Carole Goble
Open science Jisc CNI roundtable 2018
Lightning talk
What should the future look like?
What are the essential characteristics we desire in a relatively near future system to support scholarly communication across the full research life cycle?
What are the key areas requiring attention, action, or investment today to reach the future that we want to reach?
What are the best opportunities to build upon existing practices, investments and infrastructure, both
open and commercially provided?
Where must alternatives be developed?
What areas are already on good trajectories and can be left to evolve without additional intervention
Presentation given to Pubmet 2015, Zadar, Croatia.
For the live presentation having the rich media content, please access: http://kosson.ro/webpedia/presentationsnicolaiec/Croatia2015/#/
Short talk on Research Object and their use for reproducibility and publishing in the Systems Biology Commons Platform FAIRDOMHub, and the underlying software SEEK.
Reproducible and citable data and models: an introduction.FAIRDOM
Prepared and presented by Carole Goble (University of Manchester), Wolfgang Mueller (HITS), Dagmar Waltermath (University of Rostock), at the Reproducible and Citable Data and Models Workshop, Warnemünde, Germany. September 14th - 16th 2015.
This presentation introduces the Memento solution to allow time travel on the Web. Slides used at the first presentation about Memento at the Library of Congress, November 16 2009. Please consult the February 2010 slides (http://www.slideshare.net/hvdsomp/memento-updated-technical-details-february-2010) for up-to-date technical details. More info at http://www.mementoweb.org
As the scholarly communication system evolves to become natively web-based and starts supporting the communication of a wide variety of objects, the manner in which its essential functions – registration, certification, awareness, archiving - are fulfilled co-evolves. This presentation focuses on the nature of the archival function based on a perspective of the future scholarly communication infrastructure. This presentation, prepared for a meeting in June 2014, is based on and updates a previous one that was prepared for a January 2014 meeting. The latter is available at http://www.slideshare.net/atreloar/scholarly-archiveofthefuture
The European Open Science Cloud: just what is it?Carole Goble
Presented at Jisc and CNI leaders conference 2018, 2 July 2018, Oxford, UK (https://www.jisc.ac.uk/events/jisc-and-cni-leaders-conference-02-jul-2018). The European Open Science Cloud. What exactly is it? In principle it is conceived as a virtual environment with open and seamless services for storage, management, analysis and re-use of research data, across borders and scientific disciplines. How? By federating existing scientific data infrastructures, currently dispersed across disciplines and Member States. In practice, what it is depends on the stakeholder. To European Research Infrastructures it’s a coordinated mission to organise and exchange their data, metadata, software and services to be FAIR – Findable, Accessible, Interoperable, Reusable – and to use e-Infrastructures, either EU or commercial. To EU e-Infrastructures offering data storage and cloud services, it’s a funding mission to integrate their services, policies and organisational structures, and to be used by the Research Infrastructures. To agencies it’s a means to promote Open Science, standardisation, cross-disciplinary research and coordinated investment with a dream of a “one stop shop” for researchers. And for Libraries?
Specimen-level mining: bringing knowledge back 'home' to the Natural History ...Ross Mounce
A talk given at the Geological Society of London, UK on 2016/03/09 as part of the Lyell meeting on Palaeoinformatics. http://www.geolsoc.org.uk/lyell16 #lyell16
Open Science: how to serve the needs of the researcher? Carole Goble
Open science Jisc CNI roundtable 2018
Lightning talk
What should the future look like?
What are the essential characteristics we desire in a relatively near future system to support scholarly communication across the full research life cycle?
What are the key areas requiring attention, action, or investment today to reach the future that we want to reach?
What are the best opportunities to build upon existing practices, investments and infrastructure, both
open and commercially provided?
Where must alternatives be developed?
What areas are already on good trajectories and can be left to evolve without additional intervention
Presentation given to Pubmet 2015, Zadar, Croatia.
For the live presentation having the rich media content, please access: http://kosson.ro/webpedia/presentationsnicolaiec/Croatia2015/#/
Short talk on Research Object and their use for reproducibility and publishing in the Systems Biology Commons Platform FAIRDOMHub, and the underlying software SEEK.
Reproducible and citable data and models: an introduction.FAIRDOM
Prepared and presented by Carole Goble (University of Manchester), Wolfgang Mueller (HITS), Dagmar Waltermath (University of Rostock), at the Reproducible and Citable Data and Models Workshop, Warnemünde, Germany. September 14th - 16th 2015.
This presentation introduces the Memento solution to allow time travel on the Web. Slides used at the first presentation about Memento at the Library of Congress, November 16 2009. Please consult the February 2010 slides (http://www.slideshare.net/hvdsomp/memento-updated-technical-details-february-2010) for up-to-date technical details. More info at http://www.mementoweb.org
These slides accompany the LDOW2010 paper "An HTTP-Based Versioning Mechanism for Linked Data". The paper is available at http://arxiv.org/abs/1003.3661. It describes how the combination of the Memento (Time Travel for the Web) framework, and a resource versioning approach that is aligned both with the Cool URI notion and with Tim Berners-Lee concept of Time-Generic and Time-Specific, yields the ability to collect current and prior versions of resource merely using "follow your nose" HTTP navigation. The proposed combination further extends the value of a URI, and allows the emergence of a novel realm of temporal Web applications.
Slides used for a presentation at the CNI 2013 Fall meeting. Discusses the problem domain of the Hiberlink project, a collaboration between the Los Alamos National Laboratory and the University of Edinburgh, funded by the Andrew W. Mellon Foundation. Hiberlink investigates reference rot in web-based scholarly communication.
Presentation for PIDapalooza 2016. PIDs need to be used to achieve their intended persistence. Our research (reported at WWW2016, see http://arxiv.org/1602.09102) found that a disturbing percentage of references to papers that have DOIs actually use the landing page HTTP URI instead of the DOI HTTP URI. The problem is likely related to tools used for collecting references such as bookmarks and reference managers. These select the landing page URI instead of the DOI URI because the former is what's available in the address bar. It can safely be assumed that the same problem exists for other types of PIDs. The net result is that the true potential of PIDs is not realized. In order to ameliorate this problem we propose a Signposting pattern for PIDs (http://signposting.org/identifier/). It consists of adding a Link header to HTTP HEAD/GET responses for all resources identified by a DOI, including the landing page and content resources such as "the PDF" and "the dataset". The Link header contains a link, which points with the "identifier" relation type to the DOI HTTP URI. When such a link is available, tools can automatically discover and use the DOI URI instead of the other URIs (landing page, PDF, dataset) associated with the DOI-identified object.
Memento: Big Leaps Towards Seamless Navigation of the Web of the PastHerbert Van de Sompel
These slides provide an explanation of the Memento Framework (time travel for the Web) from the perspective of resource versioning. It also details progress that has been made with deploying the framework since it was first introduced in November 2009, including standardization, development of tools, and advocacy. In addition, it touches upon new challenges (discovery, branding) and announces plans to make transactional Web archiving software available in the course of 2011.
DBpedia Archive using Memento, Triple Pattern Fragments, and HDTHerbert Van de Sompel
DBpedia is the Linked Data version of Wikipedia. Starting in 2007, several DBpedia dumps have been made available for download. In 2010, the Research Library at the Los Alamos National Laboratory used these dumps to deploy a Memento-compliant DBpedia Archive, in order to demonstrate the applicability and appeal of accessing temporal versions of Linked Data sets using the Memento “Time Travel for the Web” protocol. The archive supported datetime negotiation to access various temporal versions of RDF descriptions of DBpedia subject URIs.
In a recent collaboration with the iMinds Group of Ghent University, the DBpedia Archive received a major overhaul. The initial MongoDB storage approach, which was unable to handle increasingly large DBpedia dumps, was replaced by HDT, the Binary RDF Representation for Publication and Exchange. And, in addition to the existing subject URI access point, Triple Pattern Fragments access, as proposed by the Linked Data Fragments project, was added. This allows datetime negotiation for URIs that identify RDF triples that match subject/predicate/object patterns. To add this powerful capability, native Memento support was added to the Linked Data Fragments Server of Ghent University.
In this talk, we will include a brief refresher of Memento, and will cover Linked Data Fragments, Triple Pattern Fragments, and HDT in more detail. We will share lessons learned from this effort and demo the new DBpedia Archive, which, at this point, holds over 5 billion RDF triples.
Smart Mobility Policies with Evolutionary Algorithms: The Adapting Info Panel...Daniel H. Stolfi
In this article we propose the Yellow Swarm architecture for reducing travel times, greenhouse gas emissions and fuel consumption of road traffic by using several LED panels to suggest changes in the direction of vehicles (detours) for different time slots. These time intervals are calculated using an evolutionary algorithm, specifically designed for our proposal, which evaluates many working scenarios based on real cities, imported from OpenStreetMap into the SUMO traffic simulator. Our results show an improvement in average travel times, emissions, and fuel consumption even when only a small percentage of drivers follow the indications provided by our panels.
http://doi.acm.org/10.1145/2739480.2754742
Keynote talk to LEARN (LERU/H2020 project) for research data management. Emphasizes that problems are cultural not technical. Promotes modern approaches such as Git / continuousIntegration, announces DAT. Asserts that the Right to Read in the Right to Mine. Calls for widespread development of contentmining (TDM)
Interlinking Data and Knowledge in Enterprises, Research and Society with Lin...Christoph Lange
The Linked Data paradigm has emerged as a powerful enabler for data and knowledge interlinking and exchange using standardised Web technologies.
In this article, we discuss our vision how the Linked Data paradigm can be employed to evolve the intranets of large organisations -- be it enterprises, research organisations or governmental and public administrations -- into networks of internal data and knowledge.
In particular for large enterprises data integration is still a key challenge. The Linked Data paradigm seems a promising approach for integrating enterprise data. Like the Web of Data, which now complements the original document-centred Web, data intranets may help to enhance and flexibilise the intranets and service-oriented architectures that exist in large organisations. Furthermore, using Linked Data gives enterprises access to 50+ billion facts from the growing Linked Open Data (LOD) cloud. As a result, a data intranet can help to bridge the gap between structured data management (in ERP, CRM or SCM systems) and semi-structured or unstructured information in documents, wikis or web portals, and make all of these sources searchable in a coherent way.
Keynote at Baltic DB&IS 2014, 9 June 2014, Tallinn, Estonia
ONTOLOGY SERVICE CENTER: A DATAHUB FOR ONTOLOGY APPLICATIONIJwest
With the growth of data-oriented research in humanities, a large number of research datasets have been
created and published through web services. However, how to discover, integrate and reuse these distributed
heterogeneous research datasets is a challenging task. Ontology is the soul between series digital humanities
resources, which provides a good way for people to discover and understand these datasets. With the release
of more and more linked open data and knowledge bases, a large number of ontologies have been produced
at the same time. These ontologies have different publishing formats, consumption patterns, and interactions
ways, which are not conductive to the user’s understanding of the datasets and the reuse of the ontologies.
The Ontology Service Center platform consists of Ontology Query Center and Ontology Validation Center,
mainly using linked data and ontology-based technologies. The Ontology Query Center realizes the functions
of ontology publishing, querying, data interaction and online browsing, while the Ontology Validation
Center can verify the status of using certain ontologies in the linked datasets. The empirical part of the paper
uses the Confucius portrait as an example of how OSC can be used in the semantic annotation of images. In
a word, the purpose of this paper is to construct the applied ecology of ontology to promote the development
of knowledge graphs and the spread of ontology.
ONTOLOGY SERVICE CENTER: A DATAHUB FOR ONTOLOGY APPLICATION dannyijwest
With the growth of data-oriented research in humanities, a large number of research datasets have been
created and published through web services. However, how to discover, integrate and reuse these distributed
heterogeneous research datasets is a challenging task. Ontology is the soul between series digital humanities
resources, which provides a good way for people to discover and understand these datasets. With the release
of more and more linked open data and knowledge bases, a large number of ontologies have been produced
at the same time
Published on Jan 29, 2016 by PMR
Keynote talk to LEARN (LERU/H2020 project) for research data management. Emphasizes that problems are cultural not technical. Promotes modern approaches such as Git / continuous Integration, announces DAT. Asserts that the Right to Read in the Right to Mine. Calls for widespread development of content mining (TDM)
The Culture of Research Data, by Peter Murray-RustLEARN Project
1st LEARN Workshop. Embedding Research Data as part of the research cycle. 29 Jan 2016. Presentation by Peter Murray-Rust, ContentMine.org and University of Cambridge
The presentation I held at #ocg12, based on the paper "The case for an open science in technology enhanced learning" by P. Kraker, D. Leony, W. Reinhardt, and G. Beham
“Open Research Data: Implications for Science and Society”, Warsaw, Poland, May 28–29, 2015, conference organized by the Open Science Platform — an initiative of the Interdisciplinary Centre for Mathematical and Computational Modelling at the University of Warsaw. pon.edu.pl @OpenSciPlatform #ORD2015
Open Data in a Big Data World: easy to say, but hard to do?LEARN Project
Presentation at 3rd LEARN workshop on Research Data Management, “Make research data management policies work”
Helsinki, 28 June 2016, by Sarah Callaghan, STFC Rutherford Appleton Laboratory
folksonomy, social tagging, tag clouds, automatic folksonomy construction, word clouds, wordle,context-preserving word cloud visualisation, CPEWCV, seam carving, inflate and push, star forest, cycle cover, quantitative metrics, realized adjacencies, distortion, area utilization, compactness, aspect ratio, running time, semantics in language technology
Presentation about reference rot given at the Complexity Science Hub in Vienna, November 2021.
Links to web resources frequently break (link rot), and linked content can change at unpredictable rates (content drift). These dynamics of the Web are detrimental when references to web resources provide evidence or supporting information.
This presentation will report on research that assessed the extent of these problems for links to web resources in scholarly literature, by using three vast corpora of publications and a range of public web archives. It will also describe the Robust Link approach that offers a proactive, uniform, and machine-actionable way to combat link rot and content drift. Finally, it will introduce the Robustify web service and API that was devised to generate links that remain functional over time, paying special attention to challenges related to deploying infrastructure that is required to be long lasting.
Researcher Pod: Scholarly Communication Using the Decentralized WebHerbert Van de Sompel
The presentation provides an overview of the motivation and direction of the Mellon-funded Researcher Pod project that investigates technical aspects of scholarly communication in a decentralized web setting.
Presentation for a workshop about persistent identifiers organized by the Royal Library of The Netherlands and DANS. Highlights the non-trivial commitments required of all parties involved in persistent identifier systems to actually keep links based on persistent identifiers ... err ... persistent.
Various FAIR criteria pertaining to machine interaction with scholarly artifacts can commonly be addressed by means of repository-wide affordances that are uniformly provided for all hosted artifacts rather than through artifact-specific interventions. If various repository platforms provide such affordances in an interoperable manner, devising tools - for both human and machine use - that leverage them becomes easier.
My involvement, over the years, in a range of interoperability efforts has brought the insight that two factors strongly influence adoption: addressing a burning issue and delivering a KISS solution to tackle it. Undoubtedly, FAIR and FAIR DOs are burning issues. FAIR Signposting <https://signposting.org/FAIR/> is an ad-hoc repository interoperability effort that squarely fits in this problem space and that purposely specifies a KISS solution, hoping to inspire wide adoption.
Registration / Certification Interoperability Architecture (overlay peer-review)Herbert Van de Sompel
Presentation for the COAR meeting on Overlay Peer-Review held at INRIA, Paris, France. It provides overall context regarding a scholarly communication system in which the core functions of scholarly communication (registration, certification, awareness, archiving) are implemented in a decoupled manner and whereby each function can simultaneously be fulfilled by different parties, potentially in different ways. It shows how notifications can be used to achieve loosely coupled, point-to-point interoperability in such an environment, zooming in on interoperability between registration and certification aka interoperability between repositories and overlay peer-review services.
Slides used for a keynote presentation at the VIVO 2019 Conference in Podgorica, Montenegro.
Abstract: The invitation to present a keynote at the VIVO Conference and the goal of the VIVO platform, as stated on the DuraSpace site, to create an integrated record of the scholarly work of an organisation reminded me of various efforts that I have been involved in over the past years that had similar goals. EgoSystem (2014) attempted to gather information about postdocs that had left the organisation, leaving little or no contact details behind. Autoload (2017), an operational service, discovers papers by organisational researchers in order to upload them in the institutional repository. myresearch.institute (2018), an experiment that is still in progress, discovers artefacts that researchers deposit in web productivity portals and subsequently archives them. More recently, I have been involved in thinking about the future of NARCIS, a portal that provides an overview of research productivity in The Netherlands. The approach taken in all these efforts share a characteristic motivated by a desire to devise scalable and sustainable solutions: let machines rather than humans do the work. In this talk, I will provide an overview of these efforts, their motivations, the challenges involved, and the nature of success (if any).
Presentation for PIDapalooza 2019, Dublin, Ireland.
The Scholarly Orphans project, funded by the Andrew W. Mellon Foundation, explores technical approaches aimed at capturing and archiving scholarly artifacts that researchers deposit in web productivity portals as a means to collaborate and communicate with their peers. These artifacts are not collected by other frameworks aimed at archiving the scholarly record (e.g., LOCKSS, Portico, Institutional Repositories) and are only incidentally captured by web archives. The project explores an institution-driven approach inspired by web archiving. To demonstrate the ongoing thinking, the project has devised an experimental automated pipeline that continuously discovers, captures, and archives artifacts. These are created by actual researchers who, for the purpose of the experiment, were virtually enlisted in a fictive research institution. A portal at myresearch.institute provides an overview of the artifacts that were discovered and provides access to archived versions stored in both an institutional and a cross-institutional archive. The set-up leverages a range of technologies that share a flavor of persistence: Memento, Memento Tracer, Robust Links, Signposting.
As a memento of my last week of working at LANL, I put together a slide deck that provides an overview of major efforts conducted during the time I was there.
Presentation given at EuropeanaTech 2018 in Rotterdam, The Netherlands. Provides a summary of insights gained from working for about a decade on challenges related to temporal aspects of the web, persistence.
"Scholarly Communication: Deconstruct and Decentralize" was presented at the Fall 2017 Meeting of the Coalition for Networked Information. It explores working towards a Scholarly Commons by applying decentralized web ideas to scholarly communication.
Looks at hyperlinks from the perspective of a managed collection of resources for which link persistence/integrity is considered a quality of service concern. Distinguishes between links into other managed collections and to the web at large. Considers link rot and content drift.
This slide deck provides an overview of proposals to use HTTP Links as a means to address some long standing problems related to scholarly resources on the web.
This slide deck provides an overview of proposals to use HTTP Links as a means to address some long standing problems related to scholarly resources on the web.
These slides go with the paper "Reminiscing About 15 Years of Interoperability Efforts" which is available at http://dx.doi.org/10.1045/november2015-vandesompel
Slides were used for a presentation at the Fall 2015 Membership Meeting of the Coalition for Networked Information.
This presentation looks back at several efforts, conducted in the past fifteen years, aimed at establishing interoperability for web-based scholarly communication. It tries to characterize the perspectives/approaches taken by these efforts and, based upon that, proposes an HATEOS-based approach to interlink scholarly nodes on the web. This was first presented at the Research Data Alliance meeting in Paris, France, September 22 2015.
Extended version of slides presented at the "404/File Not Found" symposium held at Georgetown University on October 24 2014, see http://www.law.georgetown.edu/library/404/ . The presentation provides a brief overview of the link/reference rot problem and then discusses three complimentary strategies to combat it: Pro-actively capturing web resources that are linked from a seed collection; Referencing the captures by means of annotated links; Accessing the captures using Memento infrastructure.
This presentation introduces ResourceSync, a specification aimed to enable web-based synchronization of resources. The specification is the result of a collaboration between NISO and the Open Archives Initiative funded by the Sloan Foundation and JISC. The proposed resource synchronization approach is based on several existing specifications (e.g. Sitemaps, PubSubHubbub, well-known URI) and is aligned with common architectural principles (e.g. REST, follow your nose).
A 15 minute video version of these slides is available at https://www.youtube.com/watch?v=ASQ4jMYytsA
This presentation provides an overview of the Memento "Time Travel for the Web" framework that is aligned with the stable version of the Memento protocol, specified in RFC 7089.
The slides were used to accompany an overview of the outcomes of the ResourceSync project at the 2014 Spring Membership Meeting of the Coalition for Networked Information (CNI).
The launch of ResourceSync, a joint project of the National Information Standards Organization (NISO) and the Open Archives Initiative (OAI) funded by the Alfred P. Sloan Foundation, was motivated by the ubiquitous need to synchronize resources for applications in the realm of cultural heritage and research communication. After an initial problem definition and scoping phase, the project has designed, specified, and tested a framework for web-based synchronization that is based on SiteMaps, a protocol widely used by web servers to advertise the resources they make available to search engines for indexing. This choice allows repositories to address both search engine optimization and resource synchronization needs using the same technology.
The ResourceSync framework specifies various modular capabilities that a repository can support in order to allow third party systems to remain synchronized with its evolving resources. For example, a Resource List provides an inventory of resources whereas a Change List details resources that were created, deleted or updated during a given temporal interval. Support for capabilities can be combined in order to meet local or community requirements. The framework specifies capabilities that require a third party to recurrently poll for up-to-date information about a repositories’ resources but also publish/subscribe capabilities that keep third parties informed about changes through notifications, thereby significantly reducing synchronization latency.
Persistent Identifiers and the Web: The Need for an Unambiguous MappingHerbert Van de Sompel
Presentation given at the International Digital Curation Conference in San Francisco, February 26 2014. Highlights the lack of machine-actionability of persistent identifiers assigned to scholarly communication assets. Proposes an approach to address the issue that meets requirements that take into account the changing nature of web based research communication. A draft paper provides more details: http://public.lanl.gov/herbertv/papers/Papers/2014/IDCC2014_vandesompel.pdf
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
The OAI-ORE Interoperability Framework in the Context of the Current Scholarly Communication System
1. http://www.slideshare.net/hvdsomp/the-oaiore-interoperability-framework
The OAI-ORE Interoperability Framework
in the current
Scholarly Communication Context
<http://www.openarchives.org/ore/toc>
Herbert Van de Sompel
Los Alamos National Laboratory, USA
OAI-ORE was funded
by the Andrew W. Mellon Foundation, the National
Science Foundation, JISC, and Microsoft
The ORE Editors are: Carl Lagoze (Cornell U.), Herbert Van de Sompel (LANL), Pete Johnston (Eduserv Found.),
Michael Nelson (Old Dominion University), Robert Sanderson (U. of Liverpool), Simeon Warner (Cornell U.)
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
2. In This Talk
Some thoughts about scholarly communication based
on the forthcoming paper:
Herber Van de Sompe & Car Lagoze (2009) Al
t l l l
Aboard Towards a Machi
: ne-Fr l Scho ly
iend y lar
Co m munication System.
Written as a contribution to a book compiled in honor
of Jim Gray.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
3. Herbert’s slogan used for effect in many talks
The current scholarly communication system is
nothing but a scanned copy of the paper-based
system.
Systemic issues
Economic strains
Technical aspects
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
4. Working Towards Change
About 10 years ago, the OAI & OA started out of
dissatisfaction with the existing scholarly
communication system.
And many efforts since then were launched to try and
introduce real change.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
5. Working Towards Change is Starting to Pay Off
Rapid changes are occurring along various
dimensions in scholarly communication.
Technical trends:
1. Augmentation of the scholarly record with a
machine-actionable substrate.
2. Integration of datasets into the scholarly record.
3. Exposure of process and its integration into the
scholarly record.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
6. Working Towards Change is Starting to Pay Off
Rapid changes are occurring along various
dimensions in scholarly communication.
Technical trends:
1. Augmentation of the scholarly record with a
machine-actionable substrate.
2. Integration of datasets into the scholarly record.
3. Exposure of process and its integration into the
scholarly record.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
7. Augmentation with a machine-actionable substrate
• There is just too much literature for researchers to be able to keep up [1].
• There is a need to absorb literature across disciplines, to connect the dots, to
combine existing disparate findings to arrive at new insights.
• E.g. Life sciences: many interconnected disciplines & the lack of uniformly
structured data across disciplines is a barrier for moving discoveries in basic
research to clinical applications [2].
• There is a need for machine-agents that filter, read and reason on behalf of
researchers.
[1] Palmer, C.L., Cragin, M.H. and Hogan, T.P. (2006) Weak information work in scientific discovery.
Information Processing & Management, 43 (3). 808-820. doi:10.1016/j.ipm.2006.06.003
[2] Ruttenberg, A., Clark, T., Bug, W., et al. (2007) Advancing translational research with the Semantic Web.
BMC Bioinformatics, 8 Suppl 3. S2. doi:10.1186/1471-2105-8-S3-S2
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
8. Augmentation with a machine-actionable substrate
• There is a strong push towards a machine-actionable representation of knowledge
that is embedded in scientific literature.
• Post publication: Entity/Relation extraction via text mining.
• At publication time: Semantic publications [1].
• Shared ontologies across disciplines for uniform knowledge representation
[2].
• The Elsevier Grand Challenge [3] and the Concept Web Alliance [4] are strong
indications that both the information industry and the research community are
taking this seriously.
• Knowlets as snippets of semantically expressed knowledge that can be
shared and … traded. The emergence of A&I databases for machine
consumption …
[1] Shotton, D. (2009) Semantic Publishing: the coming revolution in scientific journal publishing. Learned
Publishing, 22 (2), pp 85-94. doi:10.1087/2009202
[2] The Open Biomedical Ontologies. http://www.obofoundry.org/
[3] Elsevier Grand Challenge. http://www.elseviergrandchallenge.com
[4] The Concept Web Alliance. http://conceptweblog.wordpress.com/declaration/
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
9. Working Towards Change is Starting to Pay Off
Rapid changes are occurring along various
dimensions in scholarly communication.
Technical trends:
1. Augmentation of the scholarly record with a
machine-actionable substrate.
2. Integration of datasets into the scholarly record.
3. Exposure of process and its integration into the
scholarly record.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
10. Integration of datasets in the scholarly record
• Data has always been important to scholarship.
• But until recently data was not treated as a first-class scholarly object.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
11. Integration of datasets in the scholarly record
• Now all functions of scholarly communication – registration, certification,
awareness, archiving, rewarding – are being discussed/implemented for datasets:
• Data Pyramid [1] indicates how trust – certification – and preservation –
archiving – are crucial as the applicability of data reaches beyond the
personal realm.
[1] Berman, F. (2008) Got data?: a guide to data preservation in the information age. Communications of the
ACM, 51 (12). 50-56. doi:10.1145/1409360.1409376
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
12. Integration of datasets in the scholarly record
• Now all functions of scholarly communication – registration, certification,
awareness, archiving, rewarding – are being discussed/implemented for datasets:
• Data Pyramid [1] indicates how trust – certification – and preservation –
archiving – are crucial as the applicability of data reaches beyond the
personal realm.
• General recognition of the necessity of capabilities to discover datasets –
awareness [2].
• Efforts aimed at identifying and referencing datasets in a standard way [3]
take their position as primary artifacts for granted (and are partly motivated by
allowing researchers to gain credit – rewarding – for sharing data).
[1] Berman, F. (2008) Got data?: a guide to data preservation in the information age. Communications of the
ACM, 51 (12). 50-56. doi:10.1145/1409360.1409376
[2] Ruusalepp, R. (2008) Infrastructure Planning And Data Curation: A Comparative Study Of International
Approaches To Enabling The Sharing Of Research
Datahttp://www.dcc.ac.uk/docs/publications/reports/Data_Sharing_Report.pdf.
[3] Altman, M. and King, G. (2007) A Proposed Standard for the Scholarly Citation of Quantitative Data. D-Lib
Magazine, 13 (3/4). doi:10.1045/march2007-altman
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
13. Working Towards Change is Starting to Pay Off
Rapid changes are occurring along various
dimensions in scholarly communication.
Technical trends:
1. Augmentation of the scholarly record with a
machine-actionable substrate.
2. Integration of datasets into the scholarly record.
3. Exposure of process and its integration into the
scholarly record.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
14. Exposure of process and its integration into the scholarly record
• The process of using of prior knowledge to generate new insights has been
exposed for a long time via citations in the literature.
• The citation graph reveals aspects of scholarly dynamics and is heavily studied for
that reason.
• But the citation graph is problematic (imperfect citation extraction; challenging
author disambiguation; partial coverage; proprietary nature).
• A more meaningful graph can emerge:
• Web-oriented citation approach, e.g. CLADDIER [1].
• Author identifiers [2].
• Citation semantics, e.g. Citation Typing Ontology [3;4].
[1] The CLADDIER Project. http://claddier.badc.ac.uk/trac
[2] Enserink, M. (2009) Are you ready to become a number? Science, 323(5922) pp 1662-1664.
doi:10.1126/science.323.5922.1662
[3] Citation Typing Ontology. http://purl.org/net/cito/
[4] Shotton, D. (2009) Semantic publishing: the coming revolution in scientific journal publishing. Learned
Publishing, 22(2) pp 85-94(10). doi:10.1087/2009202
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
15. Exposure of process and its integration into the scholarly record
• New: Exposing the process of consuming or paying attention to scholarly
information (by sharing usage data) as revealed by COUNTER [1], MESUR [2], bX
[3] projects.
• Applications in:
• Collection management.
• Novel metrics of scholarly impact [4].
• Study of scholarly dynamics/trends [5].
• Discovery - Recommender systems [6].
[1] Project COUNTER. http://www.projectcounter.org/
[2] MESUR. http://www.mesur.org/
[3] Ex Libris bX. http://www.exlibrisgroup.com/category/bXOverview
[4] Bollen, J., Van de Sompel, H., Hagberg A., Chute, R. (2009) A principal component analysis of 39
scientific impact measures. http://arxiv.org/abs/0902.2183 . Accepted by PLoS ONE
[5] Bollen J., Van de Sompel, H., Hagberg, A. et al. (2009) Clickstream Data Yields High-Resolution Maps of
Science. PLoS ONE 4(3): e4803. doi:10.1371/journal.pone.0004803
[6] Bollen, J., and Van de Sompel, H. (2006) An architecture for the aggregation and analysis of scholarly
usage data. Proceedings of the 6th ACM/IEEE-CS Joint Conference on Digital Libraries, pp. 298-307.
doi:10.1145/1141753.1141821
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
16. Exposure of process and its integration into the scholarly record.
• New: Exposing scientific workflows (Taverna, Kepler, …) , experiments, bundles of
resources used in workflows as revealed by myExperiment [1;2].
• Applications in:
• Discovery and reuse of data and workflows.
• Novel metrics of scholarly impact.
[1] myExperiment. http://www.myexperiment.org/
[2] De Roure, D.C. and Goble, C. (2007) myExperiment – A Web 2.0 Virtual Research Environment, in
International Workshop on Virtual Research Environments and Collaborative Work Environments,
(Edinburgh, UK, 2007). Available at http://eprints.ecs.soton.ac.uk/13961/.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
17. Exposure of process and its integration into the scholarly record.
• New: Exposing scholarly value chains, and scientific life cycles [1]
[1] Pepe, A., Mayernik, M., Borgman, C., Van de Sompel, H. (2009) Technology to Represent Scientific
Practice: Data, Life Cycles, and Value Chains. arXiv preprint arXiv:0906.2549v1 ; submitted to JASIST.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
18. Exposure of process and its integration into the scholarly record.
• New: Exposing scholarly value chains, and scientific life cycles [1;2]
[1] Pepe, A., Mayernik, M., Borgman, C., Van de Sompel, H. (2009) Technology to Represent Scientific
Practice: Data, Life Cycles, and Value Chains. arXiv preprint arXiv:0906.2549v1; submitted to JASIST.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
19. Working Towards Change is Starting to Pay Off
These trends have a Web-orientation in the
development of scholarly interoperability in
common.
URIs, HTTP URIs to identify scholarly artifacts,
concepts, researchers, institutions, etc.
Use of Web Architecture and associated standards
(XML, RDF, RDFS, OWL, RSS, Atom, … ) to
support the representation and communication of
scholarly information and knowledge.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
20. Working Towards Change is Starting to Pay Off
These trends have a Web-orientation in the
development of scholarly interoperability in
common.
The OAI-ORE interoperability framework fits right into
this wave: based on Web Architecture, and Linked
Data (the Semantic Web made real) guidelines.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
21. OAI Object Reuse and Exchange
Subject: How to handle an Aggregation
of Web resources?
Approach: Publish Resource Maps to the Web that
Instantiate, Describe, and provide an Identity
for the Aggregation
Aggregation: a non-information resource
Resource Map: an information resource
that describes an Aggregation
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
22. The Web
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
23. An Aggregation and the Web
• Resources of an
aggregation are
distinct URI-identified
Web resources
• To handle
aggregations, missing
are:
o The boundary
that delineates
the aggregation in
the Web
o An identity (URI)
for the
aggregation
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
24. Introduce a Resource that stands for the Aggregation
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
25. Publish a Resource Map that describes the Aggregation
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
26. Discover the Resource Map via the Aggregation
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
27. All the OAI-ORE details at …
ORE Specs – http://www-openarchives.org/ore/toc
ORE Tutorial - http://www.slideshare.net/hvdsomp/an-
overview-of-the-oai-object-reuse-and-exchange-
interoperability-framework
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland
28. And now …
Robert Sanderson. Visualizing JSTOR: Exploring OAI-
ORE for Information Topology Navigation.
Maarten Hoogerwerf. A demonstrator of enhanced
publications using OAI-ORE.
Tim DiLauro. Using OAI-ORE to Simplify Data
Publishing Workflow.
OAI-ORE in the Current Scholarly Communication Context
Herbert Van de Sompel
OAI6, June 17 2009, Geneva, Switzerland