Presentation of the CORE APIv3 which provides seamless programmable access to the metadata and content from across the global repositories network delivered at Open Repositories 2022.
Access the world’s research outputs through the CORE API Matteo Cancellieri
Slides for the webinar: Access the world’s research outputs through the CORE API, 13th January 2022.
Link to the webinar video: https://youtu.be/acRLJNpq4W4
In this webinar, we present our new CORE APIv3.
Presenters Petr Knoth and Matteo Cancellieri walk you through the new features.
At a glance the new APIv3 offers:
- An extended model of the CORE resources to link different versions of a paper.
- Support for medium-size datasets collection.
- Improved analytical tools.
- User management made easier.
- Better documentation.
- A gallery to kick start your journey with the API.
The webinar contains also a quick demo showing the API features and tries to reply to the question "Did research stop during COVID?"
OpenAIRE Content Providers Community Call, July 1st, 2020
This call was focused on Data Repositories namely the OpenAIRE Research Graph and Data Repositories, the OpenAIRE Content Acquisition Policy, and the Guidelines for Data Archive Managers.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Follow the Community activities at https://www.openaire.eu/provide-community-calls
OSFair2017 training | Machine accessibility of Open Access scientific publica...Open Science Fair
Petr Knoth talks about machine accessibility of Open Access scientific publications from publisher systems via ResourceSync
Training title:TDM unlocking a goldmine of information
Training overview:
Text and Data Mining (TDM) is a natural ‘next step’ in open science. It can lead to new and unexpected discoveries and increase the impact of publications and repositories. This workshop showcases examples of successful TDM and infrastructural solutions for researchers. We will also discuss what is needed to make most of infrastructures and how publishers and repositories can open up their content.
DAY 2 - PARALLEL SESSION 4 & 5
The field of Text and Data Mining (TDM) is growing in importance with an increasing number of researchers interested in mining scholarly content. CrossRef Text and Data Mining Services launched in May 2014 and focuses on providing one common way to retrieve the full text of articles for the purposes of TDM for interested parties. This session will provide an introduction to and update on this service, and a short demonstration of it in action.
Access the world’s research outputs through the CORE API Matteo Cancellieri
Slides for the webinar: Access the world’s research outputs through the CORE API, 13th January 2022.
Link to the webinar video: https://youtu.be/acRLJNpq4W4
In this webinar, we present our new CORE APIv3.
Presenters Petr Knoth and Matteo Cancellieri walk you through the new features.
At a glance the new APIv3 offers:
- An extended model of the CORE resources to link different versions of a paper.
- Support for medium-size datasets collection.
- Improved analytical tools.
- User management made easier.
- Better documentation.
- A gallery to kick start your journey with the API.
The webinar contains also a quick demo showing the API features and tries to reply to the question "Did research stop during COVID?"
OpenAIRE Content Providers Community Call, July 1st, 2020
This call was focused on Data Repositories namely the OpenAIRE Research Graph and Data Repositories, the OpenAIRE Content Acquisition Policy, and the Guidelines for Data Archive Managers.
Was also an opportunity to share the most recent updates and novelties in the OpenAIRE Content Provider Dashboard, and to get feedback from community.
Follow the Community activities at https://www.openaire.eu/provide-community-calls
OSFair2017 training | Machine accessibility of Open Access scientific publica...Open Science Fair
Petr Knoth talks about machine accessibility of Open Access scientific publications from publisher systems via ResourceSync
Training title:TDM unlocking a goldmine of information
Training overview:
Text and Data Mining (TDM) is a natural ‘next step’ in open science. It can lead to new and unexpected discoveries and increase the impact of publications and repositories. This workshop showcases examples of successful TDM and infrastructural solutions for researchers. We will also discuss what is needed to make most of infrastructures and how publishers and repositories can open up their content.
DAY 2 - PARALLEL SESSION 4 & 5
The field of Text and Data Mining (TDM) is growing in importance with an increasing number of researchers interested in mining scholarly content. CrossRef Text and Data Mining Services launched in May 2014 and focuses on providing one common way to retrieve the full text of articles for the purposes of TDM for interested parties. This session will provide an introduction to and update on this service, and a short demonstration of it in action.
A user journey in OpenAIRE services through the lens of repository managers -...OpenAIRE
A user journey in OpenAIRE services through the lens of repository managers (II – OpenAIRE dashboard for content providers, usage statistics and the catch-all broker service). OpenAIRE-connect & OpenAIRE Advance workshop at the Open Repositories Conference, June 10, 2019, Hamburg.
Presentation given at the Open Educational Resources International Symposium in the session discovering OERs.
The presentation gives an account of the technical decisions made by the core-materials and open engineering resources pilot projects. Demonstrations of three methods of aggregating OERs is given including Custom Google Search Engine, Yahoo Pipes and searching Jorum Open, Flickr, Scribe, SlideShare, YouTube and Vimeo through their RSS Feeds and APIs.
Conference Opening Science to Meet Future Challenges, Warsaw, March 11, 2014, organized by Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw.
OpenAIRE guidelines and broker service for repository managers - OpenAIRE #OA...OpenAIRE
Presentation by Pedro Principe and Paolo Manghi at the OpenAIRE Open Access week webinar. Friday October 28, 2016. Webinar on Openaire compatibility guidelines and the dashboard for Repository Managers, with Pedro Principe (University of Minho) and Paolo Manghi (CNR/ISTI).
Presented on Tuesday, August 7, at the 2018 LRCN (Librarians' Registration Council of Nigeria) National Workshop on Electronic Resource Management Systems in Libraries, held at the University of Nigeria, Nsukka, Enugu State, Nigeria
A user journey in OpenAIRE services through the lens of repository managers -...OpenAIRE
A user journey in OpenAIRE services through the lens of repository managers (I – OpenAIRE interoperability guidelines, the content acquisition policy and the graph expansion)
Qui Bono? Cumulative advantage in open access publishingpetrknoth
We identify whether barriers related to accessing research literature, such as being located at an institution with limited access to non-OA literature, affect the citation behaviour of scholars. Question 1: Do scholars located in less developed or at less prestigious institutions rely (consume) more OA because their access to subscription literature is limited?
Question 2: Do those who benefit from OA also produce more OA or are production and consumption independent?
OAI Identifiers: Decentralised PIDs for Research Outputs in Repositoriespetrknoth
An OAI (Open Archives Initiative) identifier is a unique identifier of a metadata record. OAI identifiers are used in the context of repositories using the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH), however, the process by which they are assigned can be, in principle, used more broadly elsewhere.
In comparison to DOIs, OAI identifiers are registered in a distributed rather than centralised manner and there is therefore no cost for minting them. OAI identifiers are persistent identifiers in repositories that declare their level of support for deleted documents in the deletedRecord element of the Identify response as persistent. CORE recommends repositories to provide this persistent level of support.
OAI Identifiers are viable PIDs for repositories that can be, as opposed to DOIs, minted in a distributed fashion and cost-free, and which can be resolvable directly to the repository rather than to the publisher.
This approach has the potential to increase the importance of repositories in the process of disseminating knowledge. CORE provides a global OAI Resolver built on top of the CORE research outputs aggregation system.
A user journey in OpenAIRE services through the lens of repository managers -...OpenAIRE
A user journey in OpenAIRE services through the lens of repository managers (II – OpenAIRE dashboard for content providers, usage statistics and the catch-all broker service). OpenAIRE-connect & OpenAIRE Advance workshop at the Open Repositories Conference, June 10, 2019, Hamburg.
Presentation given at the Open Educational Resources International Symposium in the session discovering OERs.
The presentation gives an account of the technical decisions made by the core-materials and open engineering resources pilot projects. Demonstrations of three methods of aggregating OERs is given including Custom Google Search Engine, Yahoo Pipes and searching Jorum Open, Flickr, Scribe, SlideShare, YouTube and Vimeo through their RSS Feeds and APIs.
Conference Opening Science to Meet Future Challenges, Warsaw, March 11, 2014, organized by Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw.
OpenAIRE guidelines and broker service for repository managers - OpenAIRE #OA...OpenAIRE
Presentation by Pedro Principe and Paolo Manghi at the OpenAIRE Open Access week webinar. Friday October 28, 2016. Webinar on Openaire compatibility guidelines and the dashboard for Repository Managers, with Pedro Principe (University of Minho) and Paolo Manghi (CNR/ISTI).
Presented on Tuesday, August 7, at the 2018 LRCN (Librarians' Registration Council of Nigeria) National Workshop on Electronic Resource Management Systems in Libraries, held at the University of Nigeria, Nsukka, Enugu State, Nigeria
A user journey in OpenAIRE services through the lens of repository managers -...OpenAIRE
A user journey in OpenAIRE services through the lens of repository managers (I – OpenAIRE interoperability guidelines, the content acquisition policy and the graph expansion)
Qui Bono? Cumulative advantage in open access publishingpetrknoth
We identify whether barriers related to accessing research literature, such as being located at an institution with limited access to non-OA literature, affect the citation behaviour of scholars. Question 1: Do scholars located in less developed or at less prestigious institutions rely (consume) more OA because their access to subscription literature is limited?
Question 2: Do those who benefit from OA also produce more OA or are production and consumption independent?
OAI Identifiers: Decentralised PIDs for Research Outputs in Repositoriespetrknoth
An OAI (Open Archives Initiative) identifier is a unique identifier of a metadata record. OAI identifiers are used in the context of repositories using the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH), however, the process by which they are assigned can be, in principle, used more broadly elsewhere.
In comparison to DOIs, OAI identifiers are registered in a distributed rather than centralised manner and there is therefore no cost for minting them. OAI identifiers are persistent identifiers in repositories that declare their level of support for deleted documents in the deletedRecord element of the Identify response as persistent. CORE recommends repositories to provide this persistent level of support.
OAI Identifiers are viable PIDs for repositories that can be, as opposed to DOIs, minted in a distributed fashion and cost-free, and which can be resolvable directly to the repository rather than to the publisher.
This approach has the potential to increase the importance of repositories in the process of disseminating knowledge. CORE provides a global OAI Resolver built on top of the CORE research outputs aggregation system.
Tracking compliance of the REF2021 policy with the CORE Repository Dashboardpetrknoth
CORE and REF2021 audit. Where is CORE mentioned in the REF 2021 OA Policy. How CORE collects data. CORE Repository Dashboard demonstration.
Getting access to the Dashboard
Better together: building services for public good on top of content from the...petrknoth
CORE hosts the world’s largest collection of open access full texts, offering seamless, unrestricted access to research for citizens, researchers, libraries, software developers, funders and others. CORE’s aggregated content comes from thousands of institutional and subject repositories as well as journals and covers all research disciplines. In January 2019, CORE has hit the mark of 10 million monthly active users (10.41 million users). In September 2019, core.ac.uk has made it to the top 5k websites globally by user engagement as measured by the independent Alexa Rank, making it clearly one of the world’s most widely used Open Access services.
In this talk, Petr and Nancy will explain the role of CORE in the open science ecosystem. They will introduce the solutions CORE offers for improving the delivery of research literature, including tools for discovering freely available copies of papers that might be behind publishers’ paywalls as well as a recommender system for open access literature. The use of CORE data to monitor compliance with open access policies has also recently received attention. The presenters will then reflect on the challenges in the sector and share their experience of building value-added services for the society on top of open content offered by libraries and their affiliated institutional repositories and open access journals.
Despite being controversial, research metrics are becoming a key component of research evaluation processes globally. Nevertheless, accessing research metrics to support these processes in a timely manner is not a straightforward task, as it requires either having access to expensive commercial solutions such as Elsevier SciVal or Clarivate Analytics' InCites, or having substantial knowledge of existing APIs and data sources as well as the ability and skills needed to analyse large amounts of raw scholarly data in-house. This is especially the case on a department or institutional level where large amounts of data have to be aggregated prior to analysis. To alleviate this problem we have designed and prototyped CORE Analytics Dashboard – a tool for analytical evaluation of research outputs of universities. The aim of the CORE Analytics Dashboard is to help universities analyse their performance using a variety of metrics captured from openly available data sources, including citation counts and social media metrics, and to help them compare their performance with other institutions. This paper presents the motivation behind developing this dashboard and its main features.
Analysing the performance of open access papers discovery toolspetrknoth
Open Access discovery tools aim to locate freely available copies of research papers which might be behind the paywall on a publisher’s website. Our study provides a large scale quantitative performance comparison of several OA discovery tools on a randomly selected sample of 100k DOIs from CrossRef. We use the acquired knowledge from this analysis to build a new discovery tool - CORE Discovery.
Assessing Compliance with the UK REF 2021 Open Access Policypetrknoth
The recent increase in Open Access (OA) policies has brought forth important questions concerning the effect these policies have on the practice of publishing Open Access. In particular, is there evidence to support that mandating OA increases the proportion of OA outputs (in other words, do authors comply with relevant policies)? Furthermore, does mandating OA reduce the time from acceptance to the public availability of research outputs, and can compliance with OA mandates be effectively tracked? This work studies compliance with the UK REF 2021 Open Access policy. We use data from CrossRef and from CORE to create a dataset containing 1.6 million publications. We show that after the introduction of the UK OA policy, the proportion of OA research outputs in the UK has increased significantly, and the time lag between the acceptance of a publication and its Open Access availability has decreased, although there are significant differences in compliance between different repositories. We have developed a tool that can be used to assess publications' compliance with the policy based on a list of DOIs.
Integrating research indicators for use in the repositories infrastructure petrknoth
The current repository infrastructure, which consists of thousands of repositories, does not make effective use of research indicators largely exploited by commercial players in the area. Research indicators, including citation counts and Mendeley reader counts, enable the development and improvement of functionality researchers use on a daily basis. For example, they make it possible to increase the performance in information retrieval and recommendation tasks and serve as an enabler for the development of research analytics & metrics functionality, such as the analysis of research trends or collaboration networks. We believe that there is a strong case for making a better use of these indicators within the repositories infrastructure to improve the functionality of services users rely on.
Towards effective research recommender systems for repositoriespetrknoth
In this paper, we argue why and how the integration of recommender systems for research can enhance the functionality and user experience in repositories. We present the latest technical innovations in the CORE Recommender, which provides research article recommendations across the global network of repositories and journals. The CORE Recommender has been recently redeveloped and released into production in the CORE system and has also been deployed in several third-party repositories. We explain the design choices of this unique system and the evaluation processes we have in place to continue raising the quality of the provided recommendations. By drawing on our experience, we discuss the main challenges in offering a state-of-the-art recommender solution for repositories. We highlight two of the key limitations of the current repository infrastructure with respect to developing research recommender systems: 1) the lack of a standardised protocol and capabilities for exposing anonymised user-interaction logs, which represent critically important input data for recommender systems based on collaborative filtering and 2) the lack of a voluntary global sign-on capability in repositories, which would enable the creation of personalised recommendation and notification solutions based on past user interactions.
COAR Next Generation Repositories WG - Text mining and Recommender system sto...petrknoth
One of the key aims of the COAR NGR group is to help us to overcome the challenges that still make it difficult to move beyond repositories as document silos. The group wants to see a globally interoperable network of repositories and global services built on top of repositories fulfilling the expectations of users in the 21st century. During this talk, I will address two use cases the COAR NGR working group aims to enable: text and data mining and recommender systems.
Semantometrics: Towards Fulltext-based Research Evaluationpetrknoth
Over the recent years, there has been a growing interest in developing new scientometric measures that could go beyond the traditional citation-based bibliometric measures. This interest is motivated on one side by the wider availability or even emergence of new information evidencing research performance, such as article downloads, views and twitter mentions, and on the other side by the continued frustrations and problems surrounding the application of citation-based metrics to evaluate research performance in practice.
Semantometrics are a new class of research evaluation metrics which build on the premise that full text is needed to assess the value of a publication. This talk will present the results of an investigation into the properties of the semantometric contribution measure (Knoth & Herrmannova, 2014). We will provide a comparative evaluation of the contribution measure with traditional bibliometric measures based on citation counting.
My repository is being aggregated: a blessing or a curse?petrknoth
Usage statistics are frequently used by repositories to justify their value to the management who
decide about the funding to support the repository infrastructure. Another reason for collecting usage statistics at
repositories is the increased use of webometrics in the process of assessing the impact of publications and
researchers. Consequently, one of the worries repositories sometimes have about their content being aggregated
is that they feel aggregations have a detrimental effect on the accuracy of statistics they collect. They believe
that this potential decrease in reported usage can negatively influence the funding provided by their own
institutions. This raises the fundamental question of whether repositories should allow aggregators to harvest
their metadata and content. In this paper, we discuss the benefits of allowing content aggregations harvest
repository content and investigate how to overcome the drawbacks.
ESR spectroscopy in liquid food and beverages.pptxPRIYANKA PATEL
With increasing population, people need to rely on packaged food stuffs. Packaging of food materials requires the preservation of food. There are various methods for the treatment of food to preserve them and irradiation treatment of food is one of them. It is the most common and the most harmless method for the food preservation as it does not alter the necessary micronutrients of food materials. Although irradiated food doesn’t cause any harm to the human health but still the quality assessment of food is required to provide consumers with necessary information about the food. ESR spectroscopy is the most sophisticated way to investigate the quality of the food and the free radicals induced during the processing of the food. ESR spin trapping technique is useful for the detection of highly unstable radicals in the food. The antioxidant capability of liquid food and beverages in mainly performed by spin trapping technique.
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxMAGOTI ERNEST
Although Artemia has been known to man for centuries, its use as a food for the culture of larval organisms apparently began only in the 1930s, when several investigators found that it made an excellent food for newly hatched fish larvae (Litvinenko et al., 2023). As aquaculture developed in the 1960s and ‘70s, the use of Artemia also became more widespread, due both to its convenience and to its nutritional value for larval organisms (Arenas-Pardo et al., 2024). The fact that Artemia dormant cysts can be stored for long periods in cans, and then used as an off-the-shelf food requiring only 24 h of incubation makes them the most convenient, least labor-intensive, live food available for aquaculture (Sorgeloos & Roubach, 2021). The nutritional value of Artemia, especially for marine organisms, is not constant, but varies both geographically and temporally. During the last decade, however, both the causes of Artemia nutritional variability and methods to improve poorquality Artemia have been identified (Loufi et al., 2024).
Brine shrimp (Artemia spp.) are used in marine aquaculture worldwide. Annually, more than 2,000 metric tons of dry cysts are used for cultivation of fish, crustacean, and shellfish larva. Brine shrimp are important to aquaculture because newly hatched brine shrimp nauplii (larvae) provide a food source for many fish fry (Mozanzadeh et al., 2021). Culture and harvesting of brine shrimp eggs represents another aspect of the aquaculture industry. Nauplii and metanauplii of Artemia, commonly known as brine shrimp, play a crucial role in aquaculture due to their nutritional value and suitability as live feed for many aquatic species, particularly in larval stages (Sorgeloos & Roubach, 2021).
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...University of Maribor
Slides from talk:
Aleš Zamuda: Remote Sensing and Computational, Evolutionary, Supercomputing, and Intelligent Systems.
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Inter-Society Networking Panel GRSS/MTT-S/CIS Panel Session: Promoting Connection and Cooperation
https://www.etran.rs/2024/en/home-english/
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptxRASHMI M G
Abnormal or anomalous secondary growth in plants. It defines secondary growth as an increase in plant girth due to vascular cambium or cork cambium. Anomalous secondary growth does not follow the normal pattern of a single vascular cambium producing xylem internally and phloem externally.
Nucleophilic Addition of carbonyl compounds.pptxSSR02
Nucleophilic addition is the most important reaction of carbonyls. Not just aldehydes and ketones, but also carboxylic acid derivatives in general.
Carbonyls undergo addition reactions with a large range of nucleophiles.
Comparing the relative basicity of the nucleophile and the product is extremely helpful in determining how reversible the addition reaction is. Reactions with Grignards and hydrides are irreversible. Reactions with weak bases like halides and carboxylates generally don’t happen.
Electronic effects (inductive effects, electron donation) have a large impact on reactivity.
Large groups adjacent to the carbonyl will slow the rate of reaction.
Neutral nucleophiles can also add to carbonyls, although their additions are generally slower and more reversible. Acid catalysis is sometimes employed to increase the rate of addition.
BREEDING METHODS FOR DISEASE RESISTANCE.pptxRASHMI M G
Plant breeding for disease resistance is a strategy to reduce crop losses caused by disease. Plants have an innate immune system that allows them to recognize pathogens and provide resistance. However, breeding for long-lasting resistance often involves combining multiple resistance genes
The ability to recreate computational results with minimal effort and actionable metrics provides a solid foundation for scientific research and software development. When people can replicate an analysis at the touch of a button using open-source software, open data, and methods to assess and compare proposals, it significantly eases verification of results, engagement with a diverse range of contributors, and progress. However, we have yet to fully achieve this; there are still many sociotechnical frictions.
Inspired by David Donoho's vision, this talk aims to revisit the three crucial pillars of frictionless reproducibility (data sharing, code sharing, and competitive challenges) with the perspective of deep software variability.
Our observation is that multiple layers — hardware, operating systems, third-party libraries, software versions, input data, compile-time options, and parameters — are subject to variability that exacerbates frictions but is also essential for achieving robust, generalizable results and fostering innovation. I will first review the literature, providing evidence of how the complex variability interactions across these layers affect qualitative and quantitative software properties, thereby complicating the reproduction and replication of scientific studies in various fields.
I will then present some software engineering and AI techniques that can support the strategic exploration of variability spaces. These include the use of abstractions and models (e.g., feature models), sampling strategies (e.g., uniform, random), cost-effective measurements (e.g., incremental build of software configurations), and dimensionality reduction methods (e.g., transfer learning, feature selection, software debloating).
I will finally argue that deep variability is both the problem and solution of frictionless reproducibility, calling the software science community to develop new methods and tools to manage variability and foster reproducibility in software systems.
Exposé invité Journées Nationales du GDR GPL 2024
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
1. CORE APIv3: Seamless machine access to open
access metadata and full texts from across the
global repositories and publishers network
Matteo Cancellieri, Valerii Budko, Samuel Pearce, Petr Knoth
Knowledge Media institute, The Open University
CORE - https://core.ac.uk
API - https://core.ac.uk/services/api
Twitter - @oacore
2. • What can you do with the CORE API?
• Lessons learned from v2 and new features in v3
Outline
3. In doing so, we:
● enrich scholarly data using state-of-the-art text and data
mining technologies to aid discoverability,
● enable others to develop new tools and use cases on top
of the CORE platform,
● support the network of open access repositories and
journals with innovative technical, solutions and,
● facilitate a scalable, cost-effective route for the delivery
of open scholarship.
CORE’s mission
CORE’s mission is to aggregate all open access research worldwide and deliver
unrestricted access for all.
4. CORE services
Content discovery Raw data services Managing content
Discovery
Recommender
API
Dataset
FastSync
Repository Dashboard
Repository Edition
Search
5. Free to read links to full
text papers
~97 million
Data providers
10,372
Full texts hosted
directly by CORE
28,468,748
Languages
> 90
Countries
190
Metadata records
218,808,331
> 30M
monthly
active
users
6. Recent success stories using the CORE API
- Plagiarism detection
- Open Access papers discovery
- Fact checking
- New approaches to research evaluation
- Innovation engineering
- Content translation
- Trends detection
- Rapid systematic reviews
- Open Access Compliance Monitoring
Find more success stories at:
https://core.ac.uk/about/endorsements AND https://blog.core.ac.uk
7. ● New model abstraction to represent the
scholarly world
● More coherent search queries
● Easier to access large datasets
● Improved analytical tools
● User management made easy
● Better documentation
● A gallery to kick start your journey with the
API
What's new
8. 🖋 Documentation in Swagger
🖋 PHP + Symfony implementation
🚀 Elasticsearch
API clients
•Java https://github.com/oacore/oacore4j
•Python https://github.com/oacore/pyoacore
•R https://github.com/ropensci/rcoreoa
CORE API: where are we?
> 3,000 registered users > 300 active API users
(in the last two months )
9. Works
A deduplicated and polished item, it is made with the best metadata we can use from multiple articles
from different sources, it includes enrichments.
Article (old name) /Output (new name)
It is data coming directly from the data providers. It mostly comes from OAI-PMH but there also other
different data providers. The data is uniform so all the different data providers lead to a single metadata
format.
Data provider
It contains repositories (institutional
and disciplinary), preprint servers, journals and
publishers.
Journal
This dataset contains all journal titles included
in the CORE collection.
How CORE sees the world
1...n versions
contains contains
11. Large dataset access
▪ The API now support querying
for medium size datasets
(1,000-100,000 records) through
the scroll parameter.
▪ For large datasets (>100,000),
consider the CORE dataset
19. Feedback
If you are using the API for research, please cite one of our
research outputs https://core.ac.uk/about/research-
outputs
Show us how you are using the API
Get in touch if you have questions.
Questions? https://bit.ly/core-apiv3