Scholarly citations from one publication to another, expressed as reference lists within academic articles, are core elements of scholarly communication. Unfortunately, they usually can be accessed en masse only by paying significant subscription fees to commercial organizations, while those few services that do made them available for free impose strict limitations on their reuse. In this paper we provide an overview of the OpenCitations Project (http://opencitations.net) undertaken to remedy this situation, and of its main product, the OpenCitations Corpus, which is an open repository of accurate bibliographic citation data harvested from the scholarly literature, made available in RDF under a Creative Commons public domain dedication.
Paper at: https://w3id.org/oc/paper/occ-lisc2016.html
Making Use of the Linked Open Data Services for OpenAIRE (DI4R 2016 tutorial ...OpenAIRE
Presentation of the tutorial session at DI4R conference in Krakov (Sept. 2016), by Sahar Vahdati & Giorgos Alexiou. Title: Making Use of the Linked Open Data Services for OpenAIRE: Querying Data about Research Results, Persons, Projects and Organisations
2013 DataCite Summer Meeting - Making Research better
DataCite. Co-sponsored by CODATA.
Thursday, 19 September 2013 at 13:00 - Friday, 20 September 2013 at 12:30
Washington, DC. National Academy of Sciences
http://datacite.eventbrite.co.uk/
Making Use of the Linked Open Data Services for OpenAIRE (DI4R 2016 tutorial ...OpenAIRE
Presentation of the tutorial session at DI4R conference in Krakov (Sept. 2016), by Sahar Vahdati & Giorgos Alexiou. Title: Making Use of the Linked Open Data Services for OpenAIRE: Querying Data about Research Results, Persons, Projects and Organisations
2013 DataCite Summer Meeting - Making Research better
DataCite. Co-sponsored by CODATA.
Thursday, 19 September 2013 at 13:00 - Friday, 20 September 2013 at 12:30
Washington, DC. National Academy of Sciences
http://datacite.eventbrite.co.uk/
Linked Data Implementations—Who, What and Why?OCLC
Presented at the CNI Spring Membership Meeting in San Antonio, Texas 4 April 2016. OCLC Research conducted an International Linked Data Survey for Implementers in 2014 and 2015, receiving responses from a total of 90 institutions in 20 countries. In the 2015 survey, 112 projects or services that consumed or published linked data were described (compared to 76 in 2014). This presentation summarizes the 2015 survey results: 1) which institutions have implemented or are implementing linked data; 2) what linked data sources institutions are consuming, and why; 3) what institutions are publishing, and why; 4) barriers and advice from the implementers.
OpenAIRE guidelines and broker service for repository managers - OpenAIRE #OA...OpenAIRE
Presentation by Pedro Principe and Paolo Manghi at the OpenAIRE Open Access week webinar. Friday October 28, 2016. Webinar on Openaire compatibility guidelines and the dashboard for Repository Managers, with Pedro Principe (University of Minho) and Paolo Manghi (CNR/ISTI).
Analysing Structured Scholarly Data Embedded in Web PagesUjwal Gadiraju
Web pages increasingly embed structured data in the form of
microdata, microformats and RDFa. Through efforts such as schema.org, such embedded markup have become prevalent, with current studies estimating an adoption by about 26% of all web pages. Similar to the early adoption of Linked Data principles by publishers, libraries and other providers of bibliographic data, such organisations have been among the
early adopters, providing an unprecedented source of structured data about scholarly works. Such data, however, is fundamentally different
from traditional Linked Data, by being very sparsely linked and consisting of a large amount of coreferences and redundant statements. So far, the scale and nature of embedded scholarly data on the Web has not been investigated. In this work, we provide a study on embedded scholarly data to answer research questions about the depth, syntactic and semantic characteristics and distribution of extracted data, thereby investigating challenges and opportunities for using embedded data as a structured knowledge graph of scholarly information.
The Protein Data Bank (PDB) is a database for the three-dimensional structural data of large biological molecules, such as proteins and nucleic acids. This presentation deals with what, why, how, where and who of PDB. In this presentation we have also included briefing about various file formats available in PDB with emphasis on PDB file format
Publishing the British National Bibliography as Linked Open Data / Corine Del...CIGScotland
Presented at Linked Open Data: current practice in libraries and archives (Cataloguing & Indexing Group in Scotlland 3rd Linked Open Data Conference), Edinburgh, 18 Nov 2013
NISO Webinar:
Experimenting with BIBFRAME: Reports from Early Adopters
About the Webinar
In May 2011, the Library of Congress officially launched a new modeling initiative, Bibliographic Framework Initiative, as a linked data alternative to MARC. The Library then announced in November 2012 the proposed model, called BIBFRAME. Since then, the library world is moving from mainly theorizing about the BIBFRAME model to attempts to implement practical experimentation and testing. This experimentation is iterative, and continues to shape the model so that it’s stable enough and broadly acceptable enough for adoption.
In this webinar, several institutions will share their progress in experimenting with BIBFRAME within their library system. They will discuss the existing, developing, and planned projects happening at their institutions. Challenges and opportunities in exploring and implementing BIBFRAME in their institutions will be discussed as well.
Agenda
Introduction
Todd Carpenter, Executive Director, NISO
Experimental Mode: The National Library of Medicine and experiences with BIBFRAME
Nancy Fallgren, Metadata Specialist Librarian, National Library of Medicine, National Institutes of Health, US Department of Health and Human Services (DHHS)
Exploring BIBFRAME at a Small Academic Library
Jeremy Nelson, Metadata and Systems Librarian, Colorado College
Working with BIBFRAME for discovery and production: Linked data for Libraries/Linked Data for Production
Nancy Lorimer, Head, Metadata Dept, Stanford University Libraries
These slides were presented at the "graph databases in life sciences workshop". There is an accompanying Neo4j guide that will walk you through importing data into Neo4j using web services form a number of databases at EMBL-EBI.
https://github.com/simonjupp/importing-lifesci-data-into-neo4j
Data integration is intrinsic to how modern research is undertaken in areas such as genomics, drug development and personalised medicine. To better enable this integration a large number of biomedical ontologies have been developed to provide standard semantics for describing metadata. There are now several hundred biomedical ontologies in widespread use that describe concepts such as genes, molecules, drugs and diseases. This amounts to millions of terms that are interconnected via relationships that naturally form a graph of biomedical terminology.
The Ontology Lookup Service (OLS) (http://www.ebi.ac.uk/ols) integrates over 160 ontologies and provide a central point for the biomedical community to query and visualise ontologies. OLS also provide a RESTful API over the ontologies that is used in high-throughput data annotation pipelines. OLS is built on top of a Neo4j database that provides efficient indexes for extracting ontological relationships. We have developed generic tools for loading RDF/OWL ontologies into Neo4j where the indexes are optimised for serving common ontology queries. We are now moving to adopt graph database more widely in applications relating to ontology mapping prediction and recommendation systems for data annotation.
Written and presented by Carole Goble (University of Manchester) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
Citing data in research articles: principles, implementation, challenges - an...FAIRDOM
Prepared and presented by Jo McEntyre (EMBL_EBI) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
One year ago we started ingesting citation data from the Open Access literature into the OpenCitations Corpus (OCC), creating an RDF dataset of scholarly citation data that is open to all. In this presentation we introduce the OCC and we discuss its outcomes and uses after the first year of life.
Keynote presentation delivered at ELAG 2013 in Gent, Belgium, on May 29 2013. Discusses Research Objects and the relationship to work my team has been involved in during the past couple of years: OAI-ORE, Open Annotation, Memento.
Linked Data Implementations—Who, What and Why?OCLC
Presented at the CNI Spring Membership Meeting in San Antonio, Texas 4 April 2016. OCLC Research conducted an International Linked Data Survey for Implementers in 2014 and 2015, receiving responses from a total of 90 institutions in 20 countries. In the 2015 survey, 112 projects or services that consumed or published linked data were described (compared to 76 in 2014). This presentation summarizes the 2015 survey results: 1) which institutions have implemented or are implementing linked data; 2) what linked data sources institutions are consuming, and why; 3) what institutions are publishing, and why; 4) barriers and advice from the implementers.
OpenAIRE guidelines and broker service for repository managers - OpenAIRE #OA...OpenAIRE
Presentation by Pedro Principe and Paolo Manghi at the OpenAIRE Open Access week webinar. Friday October 28, 2016. Webinar on Openaire compatibility guidelines and the dashboard for Repository Managers, with Pedro Principe (University of Minho) and Paolo Manghi (CNR/ISTI).
Analysing Structured Scholarly Data Embedded in Web PagesUjwal Gadiraju
Web pages increasingly embed structured data in the form of
microdata, microformats and RDFa. Through efforts such as schema.org, such embedded markup have become prevalent, with current studies estimating an adoption by about 26% of all web pages. Similar to the early adoption of Linked Data principles by publishers, libraries and other providers of bibliographic data, such organisations have been among the
early adopters, providing an unprecedented source of structured data about scholarly works. Such data, however, is fundamentally different
from traditional Linked Data, by being very sparsely linked and consisting of a large amount of coreferences and redundant statements. So far, the scale and nature of embedded scholarly data on the Web has not been investigated. In this work, we provide a study on embedded scholarly data to answer research questions about the depth, syntactic and semantic characteristics and distribution of extracted data, thereby investigating challenges and opportunities for using embedded data as a structured knowledge graph of scholarly information.
The Protein Data Bank (PDB) is a database for the three-dimensional structural data of large biological molecules, such as proteins and nucleic acids. This presentation deals with what, why, how, where and who of PDB. In this presentation we have also included briefing about various file formats available in PDB with emphasis on PDB file format
Publishing the British National Bibliography as Linked Open Data / Corine Del...CIGScotland
Presented at Linked Open Data: current practice in libraries and archives (Cataloguing & Indexing Group in Scotlland 3rd Linked Open Data Conference), Edinburgh, 18 Nov 2013
NISO Webinar:
Experimenting with BIBFRAME: Reports from Early Adopters
About the Webinar
In May 2011, the Library of Congress officially launched a new modeling initiative, Bibliographic Framework Initiative, as a linked data alternative to MARC. The Library then announced in November 2012 the proposed model, called BIBFRAME. Since then, the library world is moving from mainly theorizing about the BIBFRAME model to attempts to implement practical experimentation and testing. This experimentation is iterative, and continues to shape the model so that it’s stable enough and broadly acceptable enough for adoption.
In this webinar, several institutions will share their progress in experimenting with BIBFRAME within their library system. They will discuss the existing, developing, and planned projects happening at their institutions. Challenges and opportunities in exploring and implementing BIBFRAME in their institutions will be discussed as well.
Agenda
Introduction
Todd Carpenter, Executive Director, NISO
Experimental Mode: The National Library of Medicine and experiences with BIBFRAME
Nancy Fallgren, Metadata Specialist Librarian, National Library of Medicine, National Institutes of Health, US Department of Health and Human Services (DHHS)
Exploring BIBFRAME at a Small Academic Library
Jeremy Nelson, Metadata and Systems Librarian, Colorado College
Working with BIBFRAME for discovery and production: Linked data for Libraries/Linked Data for Production
Nancy Lorimer, Head, Metadata Dept, Stanford University Libraries
These slides were presented at the "graph databases in life sciences workshop". There is an accompanying Neo4j guide that will walk you through importing data into Neo4j using web services form a number of databases at EMBL-EBI.
https://github.com/simonjupp/importing-lifesci-data-into-neo4j
Data integration is intrinsic to how modern research is undertaken in areas such as genomics, drug development and personalised medicine. To better enable this integration a large number of biomedical ontologies have been developed to provide standard semantics for describing metadata. There are now several hundred biomedical ontologies in widespread use that describe concepts such as genes, molecules, drugs and diseases. This amounts to millions of terms that are interconnected via relationships that naturally form a graph of biomedical terminology.
The Ontology Lookup Service (OLS) (http://www.ebi.ac.uk/ols) integrates over 160 ontologies and provide a central point for the biomedical community to query and visualise ontologies. OLS also provide a RESTful API over the ontologies that is used in high-throughput data annotation pipelines. OLS is built on top of a Neo4j database that provides efficient indexes for extracting ontological relationships. We have developed generic tools for loading RDF/OWL ontologies into Neo4j where the indexes are optimised for serving common ontology queries. We are now moving to adopt graph database more widely in applications relating to ontology mapping prediction and recommendation systems for data annotation.
Written and presented by Carole Goble (University of Manchester) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
Citing data in research articles: principles, implementation, challenges - an...FAIRDOM
Prepared and presented by Jo McEntyre (EMBL_EBI) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
One year ago we started ingesting citation data from the Open Access literature into the OpenCitations Corpus (OCC), creating an RDF dataset of scholarly citation data that is open to all. In this presentation we introduce the OCC and we discuss its outcomes and uses after the first year of life.
Keynote presentation delivered at ELAG 2013 in Gent, Belgium, on May 29 2013. Discusses Research Objects and the relationship to work my team has been involved in during the past couple of years: OAI-ORE, Open Annotation, Memento.
This review demonstrates that using these websites can provide researchers with valuable sources of data and research, facilitating access to current literature and specialized scientific content. For optimal results, diversifying sources of research and using multiple search engines based on need and specialization is recommended
Slides of David Shotton's presentation at OASPA 2017 - 20 September 2017, Lisbon, Portugal. These slides describe the Initiative for Open Citations – which is a collaboration between scholarly publishers, researchers, and other interested parties to promote the unrestricted availability of scholarly citation – and OpenCitations – i.e. a small infrastructure organization which hosts and develops the OpenCitations Corpus (OCC), a Linked Open Data repository of scholarly bibliographic citation data.
Closing the scientific literature access gap with CORE - how to gain free acc...Nancy Pontika
Presented during the International Open Access Week 2020 for the Kerala Library Association, October 21, 2020.
The presentation is about CORE, a global harvester of open access scientific content and the CORE services on content discovery, managing content and access to raw data.
OpenAIRE Content Providers Community Call, July 1st, 2020
This call was focused on Data Repositories namely the OpenAIRE Research Graph and Data Repositories, the OpenAIRE Content Acquisition Policy, and the Guidelines for Data Archive Managers.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Follow the Community activities at https://www.openaire.eu/provide-community-calls
Keynote: SemSci 2017: Enabling Open Semantic Science
1st International Workshop co-located with ISWC 2017, October 2017, Vienna, Austria,
https://semsci.github.io/semSci2017/
Abstract
We have all grown up with the research article and article collections (let’s call them libraries) as the prime means of scientific discourse. But research output is more than just the rhetorical narrative. The experimental methods, computational codes, data, algorithms, workflows, Standard Operating Procedures, samples and so on are the objects of research that enable reuse and reproduction of scientific experiments, and they too need to be examined and exchanged as research knowledge.
We can think of “Research Objects” as different types and as packages all the components of an investigation. If we stop thinking of publishing papers and start thinking of releasing Research Objects (software), then scholar exchange is a new game: ROs and their content evolve; they are multi-authored and their authorship evolves; they are a mix of virtual and embedded, and so on.
But first, some baby steps before we get carried away with a new vision of scholarly communication. Many journals (e.g. eLife, F1000, Elsevier) are just figuring out how to package together the supplementary materials of a paper. Data catalogues are figuring out how to virtually package multiple datasets scattered across many repositories to keep the integrated experimental context.
Research Objects [1] (http://researchobject.org/) is a framework by which the many, nested and contributed components of research can be packaged together in a systematic way, and their context, provenance and relationships richly described. The brave new world of containerisation provides the containers and Linked Data provides the metadata framework for the container manifest construction and profiles. It’s not just theory, but also in practice with examples in Systems Biology modelling, Bioinformatics computational workflows, and Health Informatics data exchange. I’ll talk about why and how we got here, the framework and examples, and what we need to do.
[1] Sean Bechhofer, Iain Buchan, David De Roure, Paolo Missier, John Ainsworth, Jiten Bhagat, Philip Couch, Don Cruickshank, Mark Delderfield, Ian Dunlop, Matthew Gamble, Danius Michaelides, Stuart Owen, David Newman, Shoaib Sufi, Carole Goble, Why linked data is not enough for scientists, In Future Generation Computer Systems, Volume 29, Issue 2, 2013, Pages 599-611, ISSN 0167-739X, https://doi.org/10.1016/j.future.2011.08.004
Enabling better science - Results and vision of the OpenAIRE infrastructure a...Paolo Manghi
Enabling better science: presentation on the results and vision of the OpenAIRE infrastructure and RDA Publishing Data Services Working Group in this direction.
David Shotton - OpenCon Oxford, 1st Dec 2017Crossref
David Shotton, Senior Researcher, Oxford eResearch Centre: http://oerc.ox.ac.uk/people/DavidShotton
Director, OpenCitations http://opencitations.net
david.shotton@opencitations.net
Tony Epstein (Sir Michael Anthony Epstein, discoverer of the Epstein-Barr virus) once said to me something I’ve never forgotten:
“Research that is not published is wasted research.”
We live today in an era of open scholarship and open data, in which the Web is the primary means of communication. For many people, information that is not freely published on the Web might as well not exist. It is thus “wasted research”.
However, within academia, we also have to live in the legacy world of subscription-access journals and subscription-access citation indexes such as Web of Science and Scopus – freely available only to members of rich scholarly institutions like Oxford University that pays hundreds of thousands of dollars annually to obtain access for their members, not to the rest of the world including scholars in developing nations.
Today I will briefly discuss the five factors desirable for scholarly publications – the Five Stars of Online Journal Articles: peer review, open access, enriched content, available datasets and machine-readable metadata [1].
I will then discuss how bibliographic citations, which permit an author to give credit to another person's endeavours and integrate our independent acts of scholarship into a global knowledge network, are being freed from commercial restrictions by publication in the OpenCitations Corpus (http://opencitations.net), an open repository of scholarly citation data that others may build upon, enhance and reuse for any purpose [2, 3].
[1] Shotton D (2012). The Five Stars of Online Journal Articles — a Framework for Article Evaluation. D-Lib Magazine 18 (1/2) (January/February 2012 issue). http://dx.doi.org/10.1045/january2012-shotton
[2] David Shotton (2013). Open citations. Nature, 502 (7471): 295-297. http://dx.doi.org/10.1038/502295a
[3] Silvio Peroni, David Shotton, Fabio Vitali (2016). Freedom for bibliographic references: OpenCitations arise. Proceedings of 2016 International Workshop on Linked Data for Information Extraction (LD4IE 2016): 32-43.
https://w3id.org/oc/paper/occ-lisc2016.html
A document-inspired way for tracking changes of RDF data - The case of the Op...University of Bologna
There are several distinct ways to represent data drift in the Linked Open Data world. In this presentation we introduce an approach for tracking data changes that has been used in the context of the OpenCitations Project. Such approach has been inspired by existing works on change tracking mechanisms in documents created through word-processors such as Microsoft Word and OpenOffice Writer.
Towards OpenURL Quality Metrics: Initial Findingsalc28
Presentation on creating a method for benchmarking metadata consistency in OpenURL links. See also: <http: />. Delivered at the July 2009 American Library Association conference in Chicago.
In this presentation we introduce SAMOD, a.k.a. Simplified Agile Methodology for Ontology Development, a novel agile methodology for the development of ontologies by means of small steps of an iterative workflow that focuses on creating well-developed and documented models starting from exemplar domain descriptions.
These slides describe the outcome of an e-government project named FOOD, FOod in Open Data, which was carried out in the context of a collaboration between the Institute of Cognitive Sciences and Technologies of the Italian National Research Council, the Italian Ministry of Agriculture (MIPAAF) and the Italian Digital Agency (AgID). In particular, we implemented several ontologies for describing protected names of products (wine, pasta, fish, oil, etc.). In addition, we present the process carried out for producing and publishing a LOD dataset containing data extracted from existing Italian policy documents on such products and compliant with the aforementioned ontologies.
Paper: http://w3id.org/people/essepuntato/papers/food-iswc2016.html
Ontologies+Data: http://dx.doi.org/10.6084/m9.figshare.3187903
In this paper we introduce the Publishing Workflow Ontology (PWO), i.e., an OWL 2 DL ontology for the description of generic workflows that is particularly suitable for formalising typical publishing processes such as the publication of articles in journals. We support the presentation with a discussion of all the ontology design patterns that have been reused for modelling the main characteristics of workflows.
Semantic lenses to bring digital and semantic publishing togetherUniversity of Bologna
Modern scholarly publishers are making steps towards semantic publishing, i.e. the use of Web and Semantic Web technologies to represent formally the meaning of a published document by specifying information about it as metadata and to publish them as Open Linked Data. In this paper we introduced a way to use a particular semantic publishing model, called semantic lenses, to semantically enhance a published journal article. In addition, we present the main features of TAL, a prototypical application that enables the navigation and understanding of a scholarly document through these semantic lenses, and we describe the outcomes of a user testing session that demonstrates the efficacy of TAL when addressing tasks requiring deeper understanding and fact-finding on the content of the document.
Zeri e LODE : Extracting the Zeri photo archive to Linked Open Data: formaliz...University of Bologna
This paper presents the first steps of a project to convert the notable Italian "Zeri photo archive" to a linked and open dataset. The full project entails the analysis of the records’ description model (Scheda F) in order to define a suitable ontology by exploring existing data models, the creation of the RDF triple store, the creation of links to the cloud, and the definition of the user interface for browsing the linked open dataset. This paper presents and discusses the conceptual modeling of the data stored in the Zeri archival database.
This work presents some experiments in letting humans annotate citations according to CiTO, an OWL ontology for describing the function of citations. We introduce a comparison of the performance of different users, and show strengths and difficulties that emerged when using that particular model to characterise citations of scholarly articles.
Tracking Changes through EARMARK: a Theoretical Perspective and an Implementa...University of Bologna
The Extremely Annotational RDF Markup, a.k.a. EARMARK, is an OWL 2 DL ontology that defines document meta-markup. It is an ontologically precise definition of markup that instantiates the structure of a text document as an independent OWL document outside of the text string it annotates, and through appropriate OWL and SWRL characterizations it can define organizations such as trees or graphs and can be used to generate validity constraints. In this paper we present an extension of EARMARK that allows us to describe how markup documents evolve in time, which complies with concepts expressed in the Functional Requirements for Bibliographic Records (FRBR).
Towards the automatic identification of the nature of citationsUniversity of Bologna
The reasons why an author cites other publications are varied: an author can cite previous works to gain assistance of some sort in the form of background information, ideas, methods, or to review, critique or refute previous works. The problem is that the best possible way to retrieve the nature of citations is very time consuming: one should read article by article to assign a particular characterisation to each citation. In this work we propose an algorithm, called CiTalO, to infer automatically the function of citations by means of Semantic Web technologies and NLP techniques. We also present some preliminary experiments and discuss some strengths and limitations of this approach.
Embedding semantic annotations within texts: the FRETTA approachUniversity of Bologna
In order to make semantic assertions about the text content of a document we need a mechanism to identify and organize the text structures of the document itself. Such mechanism would closely resemble a document-oriented markup language and would be free of the classical constraints of an embedded markup language, having no limitations given by sequentiality, containment, or contiguity of text fragments. In the past years we developed EARMARK, our OWL proposal for expressing arbitrary semantic annota- tions about the structure and the text content of a document. In this paper we describe FRETTA, our mechanism for rendering arbitrary EARMARK annotations (including non-sequential, non-hierarchical and non-contiguous ones) in XML, bringing into a unifying framework a half dozen of syntactic tricks used in literature to handle overlapping structures in a strictly hierarchical language.
A lot of applications handle XML documents where multi- ple overlapping hierarchies are necessary and make use of a number of workarounds to force overlaps into the single hierarchy of an XML for- mat. Although these workarounds are transparent to the users, they are very difficult to handle by applications reading into these formats. This paper proposes an approach to document markup based on Semantic Web technologies. Our model allows the same expressiveness as XML and any other hierarchical meta-markup language, and, rather than re- quiring complex workarounds, allows the explicit expression of overlap- ping structures in such a way that search and manipulation of these structures does not require any specific tool or language. By simply us- ing mainstream technologies such as OWL and SPARQL, our model – called EARMARK (Extremely Annotational RDF Markup) – can per- form rather sophisticated tasks with no special tricks.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.Sérgio Sacani
The return of a sample of near-surface atmosphere from Mars would facilitate answers to several first-order science questions surrounding the formation and evolution of the planet. One of the important aspects of terrestrial planet formation in general is the role that primary atmospheres played in influencing the chemistry and structure of the planets and their antecedents. Studies of the martian atmosphere can be used to investigate the role of a primary atmosphere in its history. Atmosphere samples would also inform our understanding of the near-surface chemistry of the planet, and ultimately the prospects for life. High-precision isotopic analyses of constituent gases are needed to address these questions, requiring that the analyses are made on returned samples rather than in situ.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
A brief information about the SCOP protein database used in bioinformatics.
The Structural Classification of Proteins (SCOP) database is a comprehensive and authoritative resource for the structural and evolutionary relationships of proteins. It provides a detailed and curated classification of protein structures, grouping them into families, superfamilies, and folds based on their structural and sequence similarities.
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
What is greenhouse gasses and how many gasses are there to affect the Earth.
Freedom for bibliographic references: OpenCitations arise
1. Freedom for bibliographic references:
OpenCitations arise
Silvio Peroni, David Shotton, Fabio Vitali
4th International Workshop on
Linked Data for Information Extraction (LD4IE 2016)
Kobe, Japan, October 18, 2016
https://w3id.org/oc/paper/occ-lisc2016.html
2. The Venice analogy
• Island =
scholarly publication
• Bridge = citation
• Current situation:
– local travel to
the next island
is permitted
– unrestricted
travel over the
entire network
of bridges
requires an
expensive
season ticket
– general
populace is
excluded
https://w3id.org/oc/paper/the-venice-analogy.html
3. Opening the bridges
• What – Citation data are one of the main tools used by
researchers to gain knowledge about particular topics, and
they also serve institutional goals, for example in research
assessment
• Problem – The most authoritative databases of citation data,
Scopus and Web of Science, can only be accessed by paying
significant annual access fees
– The University of Bologna pays about 6,000,000 euros per year for
accessing to digital bibliographic resources
• Solution – To create a citation database that freely and legally
makes available citation data in an open repository to assist
scholars with their academic studies and serve knowledge to
the wider public
4. OpenCitations
• The OpenCitations Project aims at creating an open repository of
scholarly citation data – the OpenCitations Corpus (OCC) – made
available under a Creative Commons public domain dedication to
provide in RDF accurate citation information (bibliographic
references) harvested from the scholarly literature
– All scripts are released with Open Source ISC Licence and available on
GitHub at http://github.com/essepuntato/opencitations
• Currently processing papers available in the PubMedCentral Open
Access subset (which contains paper related to the medical,
biological, life science domains) by means of the Europe
PubMedCentral API
• As of October 17, 2016 the OCC contains
– 1,311,196 citing/cited bibliographic resources
– 1,584,945 citation links
http://opencitations.net
5. OpenCitations Ontology
• The OpenCitations Ontology
(OCO) groups existing
complementary ontological
entities from several other
ontologies for the purpose of
providing descriptive metadata
for the OCC
• SPAR Ontologies reused:
– FRBR-aligned Bibliographic
Ontology (FaBiO) http://
purl.org/spar/fabio)
– Publishing Roles Ontology
(PRO, http://purl.org/
spar/pro)
– Bibliographic Reference
Ontology (BiRO, http://
purl.org/spar/biro)
– Citation Counting and Context
Characterization Ontology
(C4O, http://purl.org/
spar/c4o)
– DataCite Ontology (http://
purl.org/spar/datacite)
6. OpenCitations Corpus
• Six distinct kinds of bibliographic entities
– bibliographic resources (citing/cited articles, journals, books, proceedings, etc.)
– resource embodiments (format information about bibliographic resources)
– bibliographic entries (literal textual entries occurring in the reference lists)
– responsible agents (agents having certain roles with respect to the bibliographic
resources)
– agent roles (author, editor, publisher);
– identifiers (DOI, ORCID, PubMedID, URL, etc.)
• Provenance for each entity handled by means of PROV-O – as described in the
Drift-a-LOD 2016 (a workshop held in Bologna next month during EKAW 2016)
paper available at
https://w3id.org/oc/paper/occ-driftalod2016.html
• Access the OCC via
– HTTP (content negotiation, formats: JSON-LD, RDF/XML, Trig, HTML),
e.g. https://w3id.org/oc/corpus/br/1
– SPARQL endpoint, available at https://w3id.org/oc/sparql
– dumps, downloadable at https://opencitations.net/download
7. Ingestion workflow
BEE
EuropeanPubMedCentralProcessor
Parsing the
XML source of
PubMed Central
Open Access
articles.
1
SPACIN
Producing
JSON with DOI
and bib entries.
{
"doi": "10.1590/1414-431x20154655",
"localid": "MED-26577845",
"curator": "BEE EuropeanPubMedCentralProcessor",
"source": "http://www.ebi.ac.uk/europepmc/webservices/rest/PMC4678653/
fullTextXML",
"source_provider": "Europe PubMed Central",
"pmid": "26577845",
"pmcid": “PMC4678653",
"references": [
{
"bibentry": "Wenger, NK. Coronary heart disease: an older woman's major
health risk, BMJ, 1997, 315, 1085, 1090, DOI: 10.1136/bmj.315.7115.1085, PMID:
9366743",
"pmid": "9366743",
"doi": "10.1136/bmj.315.7115.1085",
"pmcid": "PMC2127693",
"process_entry": "True"
} …
]
}
2
For each citing/cited resource,
if an ID (DOI, PMID, PMCID) is
specified check if the resource
exists already. If it does go to 5.
store
ResourceFinder
3
GraphSet
ProvSet
DatasetHandler
Storer
Load all the statements onthe triplestore and storethem in the file system for
easy recovering.
OCC
6
If the resource doesn’t exist,
extract possible IDs from the entry
and query CrossRef and ORCID.
CrossRefProcessor
ORCIDProcessor
4
GraphEntity
New metadata resources are created.
If CrossRef/ORCID returned something, all
the related metadata will be used,
otherwise only basic metadata (IDs and
entries) will be added.
5
8. Test
• Hardware: MacBook Pro, with 2 GHz Intel Core i7 processor, 8 GB
DDR3 1600 MHz, OS X 10.11.3
• BEE: running for 30 minutes (querying Europe PubMedCentral API),
produced 185 JSON files (~6 new JSON files per minute)
• SPACIN
– 45 minutes to process all BEE JSON files related to the 67 papers in the
ISWC 2015 Proceedings (sources kindly made available by Springer-Nature)
– 210 minutes to process BEE JSON files related to 67 papers from Europe
PubMed Central (OA subset)
All these data are available on Figshare – their URLs is included in the article.
9. ISWC2015: most cited papers
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX fabio: <http://purl.org/spar/fabio/>
PREFIX cito: <http://purl.org/spar/cito/>
SELECT ?cited ?title ?tot {
{ SELECT ?cited (count(?citing) as ?tot) { ?cited a fabio:Expression ; ^cito:cites ?citing }
GROUP BY ?cited }
OPTIONAL { ?cited dcterms:title ?title } } ORDER BY DESC(?tot) LIMIT 15
no title?
10. No Crossref metadata
PREFIX biro: <http://purl.org/spar/biro/>
PREFIX c4o: <http://purl.org/spar/c4o/>
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX frbr: <http://purl.org/vocab/frbr/core#>
SELECT ?citing ?entry {
<http://localhost:8000/corpus/br/1302> ^biro:references ?ref .
?ref c4o:hasContent ?entry ; ^frbr:part ?citing
}
How the “no title” paper has
been referenced in the 4
papers citing it
SPACIN used the URL in the textual entries
(i.e. “http://www.w3.org/DesignIssues/LinkedData.html”)
to associate them to the same bibliographic resource:
<http://localhost:8000/corpus/br/1302>
11. Conclusions
• We have introduced the OpenCitations Project, which has created an open
repository of accurate bibliographic references harvested from the scholarly
literature, i.e. the OpenCitations Corpus (OCC)
• The number of citation links is growing day by day (about 25,000 new citation
links per day) as the continuous workflow adds new data dynamically from
Europe PubMedCentral (and other authoritative sources, i.e. Crossref and
ORCID)
• First adopter: Wikidata (via WikiCite)
– The Wikidata community has created a property for associating the OCC bibliographic
resource identifier to the metadata about scholarly papers in Wikidata
– Several links from Wikidata to the OCC have been already added
• Future plans: developing tools for linking the resources within the OCC with those
included in other datasets, e.g. Wikidata, Scholarly Data, Springer LOD
• Don’t hesitate to poke me during the poster and demo session on Wednesday
(panel P30) for additional details about OpenCitations – and don’t forgot to vote
for it, of course :-)
12. Thanks for your attention
Silvio Peroni, David Shotton, Fabio Vitali
4th International Workshop on
Linked Data for Information Extraction (LD4IE 2016)
Kobe, Japan, October 18, 2016