The FAIR data principles were drafted in 2015 to improve the findability, accessibility, interoperability, and reusability of digital assets. They consist of 15 guidelines across the four areas. Findable guidelines ensure data has a unique identifier and is searchable. Accessible guidelines specify how metadata and data can be retrieved. Interoperable guidelines promote standard formats and vocabularies. Reusable guidelines address attribution, licensing, and community standards. The principles aim to make data more easily discovered and used across technology and disciplines.
LIBER Webinar: Are the FAIR Data Principles really fair?LIBER Europe
The FAIR Data Principles are a hot topic in research data managment. Their adoption within the H2020 funding programme means researchers now have to pay much more attention to how their share, publish and archive their data.
In this light, how can libraries help their research communities implement the FAIR principles? And write better data management plans?
This questions were addressed in a LIBER webinar containing some guidance and reflections on the principles themselves. Presented by Alastair Dunning, Head Research Data Services at the TU Delft (hosts of the 4TU.Centre for Research Data), it is based on a study of 37 data repositories (from subject specific repositories, to generic data archives, to national infrastructures), seeing how far they comply with each of the individual facets of the Data principles.
An introduction to the FAIR principles and a discussion of key issues that must be addressed to ensure data is findable, accessible, interoperable and re-usable. The session explored the role of the CDISC and DDI standards for addressing these issues.
Presented by Gareth Knight at the ADMIT Network conference, organised by the Association for Data Management in the Tropics, in Antwerp, Belgium on December 1st 2015.
This document provides an overview of FAIR data principles and the FAIR data ecosystem. It discusses what FAIR data is, including that FAIR data aims to support communities in publishing and utilizing scientific data and knowledge in a findable, accessible, interoperable, and reusable manner. It then describes the different levels of the FAIR data ecosystem, including normative principles, standards in the FAIR data protocol, FAIR data resources that comply with these standards, and systems/tools that use FAIR data. It provides examples of converting raw data into FAIR data resources and the potential applications of a FAIR data ecosystem.
FAIR Data Management and FAIR Data SharingMerce Crosas
Presentation at the Critical Perspective on the Practice of Digiral Archeology symposium: http://archaeology.harvard.edu/critical-perspectives-practice-digital-archaeology
The document describes the metadata provided by a FAIR Data Point. It includes metadata for the FAIR Data Point itself, its catalogs, datasets within catalogs, distributions of datasets, and data records. Samples of metadata are provided for each type following the appropriate prefix, such as <http://dev-vm.fair-dtls.surf-hosted.nl:8082/fdp> for the FAIR Data Point and <http://dev-vm.fair-dtls.surf-hosted.nl:8082/fdp/biobank/77350-collection1> for a dataset. The metadata follows FAIR principles and provides details such as titles, descriptions, identifiers, versions, and links between related
Presentation at Digital Humanities in the Nordics 2020 conference in panel: Towards deterioration, disappearance or destruction? Discussing the critical issue of long-term sustainability of digital humanities projects
Dataset Catalogs as a Foundation for FAIR* DataTom Plasterer
BioPharma and the broader research community is faced with the challenge of simply finding the appropriate internal and external datasets for downstream analytics, knowledge-generation and collaboration. With datasets as the core asset, we wanted to promote both human and machine exploitability, using web-centric data cataloguing principles as described in the W3C Data on the Web Best Practices. To do so, we adopted DCAT (Data CATalog Vocabulary) and VoID (Vocabulary of Interlinked Datasets) for both RDF and non-RDF datasets at summary, version and distribution levels. Further, we’ve described datasets using a limited set of well-vetted public vocabularies, focused on cross-omics analytes and clinical features of the catalogued datasets.
The FAIR data principles were drafted in 2015 to improve the findability, accessibility, interoperability, and reusability of digital assets. They consist of 15 guidelines across the four areas. Findable guidelines ensure data has a unique identifier and is searchable. Accessible guidelines specify how metadata and data can be retrieved. Interoperable guidelines promote standard formats and vocabularies. Reusable guidelines address attribution, licensing, and community standards. The principles aim to make data more easily discovered and used across technology and disciplines.
LIBER Webinar: Are the FAIR Data Principles really fair?LIBER Europe
The FAIR Data Principles are a hot topic in research data managment. Their adoption within the H2020 funding programme means researchers now have to pay much more attention to how their share, publish and archive their data.
In this light, how can libraries help their research communities implement the FAIR principles? And write better data management plans?
This questions were addressed in a LIBER webinar containing some guidance and reflections on the principles themselves. Presented by Alastair Dunning, Head Research Data Services at the TU Delft (hosts of the 4TU.Centre for Research Data), it is based on a study of 37 data repositories (from subject specific repositories, to generic data archives, to national infrastructures), seeing how far they comply with each of the individual facets of the Data principles.
An introduction to the FAIR principles and a discussion of key issues that must be addressed to ensure data is findable, accessible, interoperable and re-usable. The session explored the role of the CDISC and DDI standards for addressing these issues.
Presented by Gareth Knight at the ADMIT Network conference, organised by the Association for Data Management in the Tropics, in Antwerp, Belgium on December 1st 2015.
This document provides an overview of FAIR data principles and the FAIR data ecosystem. It discusses what FAIR data is, including that FAIR data aims to support communities in publishing and utilizing scientific data and knowledge in a findable, accessible, interoperable, and reusable manner. It then describes the different levels of the FAIR data ecosystem, including normative principles, standards in the FAIR data protocol, FAIR data resources that comply with these standards, and systems/tools that use FAIR data. It provides examples of converting raw data into FAIR data resources and the potential applications of a FAIR data ecosystem.
FAIR Data Management and FAIR Data SharingMerce Crosas
Presentation at the Critical Perspective on the Practice of Digiral Archeology symposium: http://archaeology.harvard.edu/critical-perspectives-practice-digital-archaeology
The document describes the metadata provided by a FAIR Data Point. It includes metadata for the FAIR Data Point itself, its catalogs, datasets within catalogs, distributions of datasets, and data records. Samples of metadata are provided for each type following the appropriate prefix, such as <http://dev-vm.fair-dtls.surf-hosted.nl:8082/fdp> for the FAIR Data Point and <http://dev-vm.fair-dtls.surf-hosted.nl:8082/fdp/biobank/77350-collection1> for a dataset. The metadata follows FAIR principles and provides details such as titles, descriptions, identifiers, versions, and links between related
Presentation at Digital Humanities in the Nordics 2020 conference in panel: Towards deterioration, disappearance or destruction? Discussing the critical issue of long-term sustainability of digital humanities projects
Dataset Catalogs as a Foundation for FAIR* DataTom Plasterer
BioPharma and the broader research community is faced with the challenge of simply finding the appropriate internal and external datasets for downstream analytics, knowledge-generation and collaboration. With datasets as the core asset, we wanted to promote both human and machine exploitability, using web-centric data cataloguing principles as described in the W3C Data on the Web Best Practices. To do so, we adopted DCAT (Data CATalog Vocabulary) and VoID (Vocabulary of Interlinked Datasets) for both RDF and non-RDF datasets at summary, version and distribution levels. Further, we’ve described datasets using a limited set of well-vetted public vocabularies, focused on cross-omics analytes and clinical features of the catalogued datasets.
https://www.youtube.com/watch?v=5YqAH3f9LiU
Digital Transformation is a key goal of many large and small companies, as well as of most research institutes today. However, a key prerequisite and enabler of digital transformation is computational accessibility and interoperability of data, as laid out in the FAIR Data principles. The Hyve has been involved in the FAIR Data movement since the start, and for this webinar, our CEO Kees van Bochove will be talking to a very special guest, Ruben Kok, director of DTL. DTL and its predecessor NBIC, as well as ‘spinoff’ GO-FAIR have spent an enormous amount of effort in the past years on outreach, training, tools and community building around the FAIR Data Principles. Where do we stand today? What can we expect to see in the coming years for FAIR and FAIR biomedical data (e.g. Personal Health Train) in particular?
OSFair2017 Training | FAIR metrics - Starring your data setsOpen Science Fair
Peter Doorn, Marjan Grootveld & Elly Dijk talk about FAIR data principles and present the assessment tool that DANS is developing for data repositories | OSFair2017 Workshop
Workshop title: FAIR metrics - Starring your data sets
Workshop overview:
Do you want to join our effort to put the FAIR data principles into practice? Come and explore the assessment tool that DANS, Data Archiving and Networked Services in the Netherlands, is developing for data repositories.
The aim of our work is to implement the FAIR principles into a data assessment tool so that every dataset which is deposited or reused from any digital repository can be assessed in terms of a score on the principles Findable, Accessible, Interoperable, and Reusable, using a ‘FAIRness’ scale from 1 to 5 stars. In this interactive session participants can explore the pilot version of FAIRdat: the FAIR data assessment tool. The organisers would like to inform you about the project, and look forward to all feedback to improve the tool, or to improve the metrics that are used.
DAY 3 - PARALLEL SESSION 7
FAIR data has flown up the hype curve without a clear sense of return from the required data stewardship investment. The killer use case for FAIR data is a science knowledge graph. It enables you to richly address novel questions of your and the world’s data. We started with data catalogues (findability) which exploited linked/referenced data using a few focused vocabularies (interoperability), for credentialed users (accessibility), with provenance and attribution (reusability) to make this happen.
This talk was presented at The Molecular Medicine Tri-Conference/Bio-IT West on March 11, 2019.
FAIR Data Knowledge Graphs–from Theory to PracticeTom Plasterer
This document discusses building FAIR data knowledge graphs from theory to practice. It begins by outlining what R&D researchers want to do with data, such as understanding disease mechanisms and using patient data, but that currently data is fragmented across systems. It then introduces the FAIR data principles and describes building a knowledge graph that incorporates data from multiple sources using standards like the Data Catalog vocabulary. The key challenges discussed are determining canonical representations for entities and linking data to public vocabularies through an enrichment process.
pro-iBiosphere 2013-05 Linked Open Data (Gregor Hagedorn)Gregor Hagedorn
This document discusses using linked open data and semantic web technologies to improve scientific knowledge management systems. It outlines how linking distributed data sources using common web standards allows data to be queried and analyzed across datasets in new ways. This enables researchers to discover relationships in the data that individual datasets did not anticipate. It advocates assigning stable, dereferenceable web identifiers to entities to facilitate linking data both within and across sources.
As BioPharma adapts to incorporate nimble networks of suppliers, collaborators, and regulators the ability to link data is critical for dynamic interoperability. Adoption of linked data paradigm allows BioPharma to focus on core business: delivering valuable therapeutics in a timely manner.
Metadata stores systems in use 20180322Keith Russell
Presentation to Macquarie University research data committee on systems in use in Australia for tracking and exposing research data (including metadata catalogs and metadata stores) 22 March 2018
BioPharma and FAIR Data, a Collaborative AdvantageTom Plasterer
The concept of FAIR (Findable, Accessible, Interoperable and Reusable) data is becoming a reality as stakeholders from industry, academia, funding agencies and publishers are embracing this approach. For BioPharma being able to effectively share and reuse data is a tremendous competitive advantage, within a company, with peer organizations, key opinion leaders and regulatory agencies. A few key drivers, success stories and preliminary results of an industry data stewardship survey are presented.
Creating impact with accessible data in agriculture and nutrition: sharing da...godanSec
Richard Finkers (Wageningen UR) presented at the 2nd International Workshop: Creating Impact with Open Data in Agriculture and Nutrition in The Hague, 11 September 2015.
D4Science Data infrastructure: a facilitator for a FAIR data managementResearch Data Alliance
D4Science is a hybrid data infrastructure that integrates technologies to provide elastic access and usage of data and data management capabilities. It hosts over 50 virtual research environments for over 2500 scientists across 44 countries. D4Science aims to facilitate FAIR (Findable, Accessible, Interoperable, Re-usable) data management by assigning unique identifiers and rich metadata to resources, publishing catalogs to enable discovery, making resources available through standards, adding metadata in multiple formats, and requiring licenses and provenance to promote reuse.
In an expert webinar on April 15th 2020 we discussed (in Finnish) how the FAIR data principles affect service development in RDM services. I presented some relevant outputs from the FAIRsFAIR project. These are the slides (in English). The webinar will be published on the fairdata.fi service site https://www.fairdata.fi/koulutus/koulutuksen-tallenteet/
The document discusses the importance of findability as the first step of the FAIR principles. It emphasizes that assigning globally unique and persistent identifiers (PIDs) to metadata and data is crucial to ensure findability. Secondly, it states that metadata should describe the data in detail so it can be found without its identifier. Finally, it notes that metadata and data need to be searchable through registration in searchable resources to meet the findability principle. The document stresses that PIDs are an important first step to achieving findability and the other FAIR principles.
The document discusses Dublin Core and other metadata schemas for exposing information on the web. It outlines four rules: use URIs as names for things, use HTTP URIs so people can look them up, provide useful information when URIs are looked up, and include links to other URIs. The document also discusses how metadata from different domains and vocabularies can be combined using common frameworks like ontologies. It provides an example of converting XML metadata to a graph using RDF.
Keynote on software sustainability given at the 2nd Annual Netherlands eScience Symposium, November 2014.
Based on the article
Carole Goble ,
Better Software, Better Research
Issue No.05 - Sept.-Oct. (2014 vol.18)
pp: 4-8
IEEE Computer Society
http://www.computer.org/csdl/mags/ic/2014/05/mic2014050004.pdf
http://doi.ieeecomputersociety.org/10.1109/MIC.2014.88
http://www.software.ac.uk/resources/publications/better-software-better-research
Digital Object Identifiers (DOIs) in the context of the International TreatyFAO
http://tiny.cc/faowgsworkshop
FAO's activities relevant to genome sequencing- International Treaty on Plant Genetic Resources for Food and Agriculture (ITPGRFA). Presentation from the FAO expert workshop on practical applications of Whole Genome Sequencing (WGS) for food safety management - 7-8 December 2015, Rome, Italy.
Harnessing Edge Informatics to Accelerate Collaboration in BioPharma (Bio-IT ...Tom Plasterer
As scientists in the life sciences we are trained to pursue singular goals around a publication or a validated target or a drug submission. Our failure rates are exceedingly high especially as we move closer to patients in the attempt to collect sufficient clinical evidence to demonstrate the value of novel therapeutics. This wastes resources as well as time for patients depending upon us for the next breakthrough.
Edge Informatics is an approach to ameliorate these failures. Using both technical and social solutions together knowledge can be shared and leveraged across the drug development process. This is accomplished by making data assets discoverable, accessible, self-described, reusable and annotatable. The Open PHACTS project pioneered this approach and has provided a number of the technical and social solutions to enable Edge Informatics. A number of pre-competitive consortia and some content providers have also embraced this approach, facilitating networks of collaborators within and outside a given organization. When taken together more accurate, timely and inclusive decision-making is fostered.
AFAIR in Astronomy Research - Slides. In this webinar ARDC is partnering with the ADACS project to explore the FAIR data principles in the context of Astronomy research and the ASVO and IVOA as a community exemplars of the implementation of the FAIR data principles.
These slides from: Keith Russell (ARDC): Looking at FAIR
In this talk Keith will provide an overview of the FAIR principles and how it was used in astronomy before it became official. He will conclude the talk by discussing what other disciplines can learn from their approach.
https://www.youtube.com/watch?v=5YqAH3f9LiU
Digital Transformation is a key goal of many large and small companies, as well as of most research institutes today. However, a key prerequisite and enabler of digital transformation is computational accessibility and interoperability of data, as laid out in the FAIR Data principles. The Hyve has been involved in the FAIR Data movement since the start, and for this webinar, our CEO Kees van Bochove will be talking to a very special guest, Ruben Kok, director of DTL. DTL and its predecessor NBIC, as well as ‘spinoff’ GO-FAIR have spent an enormous amount of effort in the past years on outreach, training, tools and community building around the FAIR Data Principles. Where do we stand today? What can we expect to see in the coming years for FAIR and FAIR biomedical data (e.g. Personal Health Train) in particular?
OSFair2017 Training | FAIR metrics - Starring your data setsOpen Science Fair
Peter Doorn, Marjan Grootveld & Elly Dijk talk about FAIR data principles and present the assessment tool that DANS is developing for data repositories | OSFair2017 Workshop
Workshop title: FAIR metrics - Starring your data sets
Workshop overview:
Do you want to join our effort to put the FAIR data principles into practice? Come and explore the assessment tool that DANS, Data Archiving and Networked Services in the Netherlands, is developing for data repositories.
The aim of our work is to implement the FAIR principles into a data assessment tool so that every dataset which is deposited or reused from any digital repository can be assessed in terms of a score on the principles Findable, Accessible, Interoperable, and Reusable, using a ‘FAIRness’ scale from 1 to 5 stars. In this interactive session participants can explore the pilot version of FAIRdat: the FAIR data assessment tool. The organisers would like to inform you about the project, and look forward to all feedback to improve the tool, or to improve the metrics that are used.
DAY 3 - PARALLEL SESSION 7
FAIR data has flown up the hype curve without a clear sense of return from the required data stewardship investment. The killer use case for FAIR data is a science knowledge graph. It enables you to richly address novel questions of your and the world’s data. We started with data catalogues (findability) which exploited linked/referenced data using a few focused vocabularies (interoperability), for credentialed users (accessibility), with provenance and attribution (reusability) to make this happen.
This talk was presented at The Molecular Medicine Tri-Conference/Bio-IT West on March 11, 2019.
FAIR Data Knowledge Graphs–from Theory to PracticeTom Plasterer
This document discusses building FAIR data knowledge graphs from theory to practice. It begins by outlining what R&D researchers want to do with data, such as understanding disease mechanisms and using patient data, but that currently data is fragmented across systems. It then introduces the FAIR data principles and describes building a knowledge graph that incorporates data from multiple sources using standards like the Data Catalog vocabulary. The key challenges discussed are determining canonical representations for entities and linking data to public vocabularies through an enrichment process.
pro-iBiosphere 2013-05 Linked Open Data (Gregor Hagedorn)Gregor Hagedorn
This document discusses using linked open data and semantic web technologies to improve scientific knowledge management systems. It outlines how linking distributed data sources using common web standards allows data to be queried and analyzed across datasets in new ways. This enables researchers to discover relationships in the data that individual datasets did not anticipate. It advocates assigning stable, dereferenceable web identifiers to entities to facilitate linking data both within and across sources.
As BioPharma adapts to incorporate nimble networks of suppliers, collaborators, and regulators the ability to link data is critical for dynamic interoperability. Adoption of linked data paradigm allows BioPharma to focus on core business: delivering valuable therapeutics in a timely manner.
Metadata stores systems in use 20180322Keith Russell
Presentation to Macquarie University research data committee on systems in use in Australia for tracking and exposing research data (including metadata catalogs and metadata stores) 22 March 2018
BioPharma and FAIR Data, a Collaborative AdvantageTom Plasterer
The concept of FAIR (Findable, Accessible, Interoperable and Reusable) data is becoming a reality as stakeholders from industry, academia, funding agencies and publishers are embracing this approach. For BioPharma being able to effectively share and reuse data is a tremendous competitive advantage, within a company, with peer organizations, key opinion leaders and regulatory agencies. A few key drivers, success stories and preliminary results of an industry data stewardship survey are presented.
Creating impact with accessible data in agriculture and nutrition: sharing da...godanSec
Richard Finkers (Wageningen UR) presented at the 2nd International Workshop: Creating Impact with Open Data in Agriculture and Nutrition in The Hague, 11 September 2015.
D4Science Data infrastructure: a facilitator for a FAIR data managementResearch Data Alliance
D4Science is a hybrid data infrastructure that integrates technologies to provide elastic access and usage of data and data management capabilities. It hosts over 50 virtual research environments for over 2500 scientists across 44 countries. D4Science aims to facilitate FAIR (Findable, Accessible, Interoperable, Re-usable) data management by assigning unique identifiers and rich metadata to resources, publishing catalogs to enable discovery, making resources available through standards, adding metadata in multiple formats, and requiring licenses and provenance to promote reuse.
In an expert webinar on April 15th 2020 we discussed (in Finnish) how the FAIR data principles affect service development in RDM services. I presented some relevant outputs from the FAIRsFAIR project. These are the slides (in English). The webinar will be published on the fairdata.fi service site https://www.fairdata.fi/koulutus/koulutuksen-tallenteet/
The document discusses the importance of findability as the first step of the FAIR principles. It emphasizes that assigning globally unique and persistent identifiers (PIDs) to metadata and data is crucial to ensure findability. Secondly, it states that metadata should describe the data in detail so it can be found without its identifier. Finally, it notes that metadata and data need to be searchable through registration in searchable resources to meet the findability principle. The document stresses that PIDs are an important first step to achieving findability and the other FAIR principles.
The document discusses Dublin Core and other metadata schemas for exposing information on the web. It outlines four rules: use URIs as names for things, use HTTP URIs so people can look them up, provide useful information when URIs are looked up, and include links to other URIs. The document also discusses how metadata from different domains and vocabularies can be combined using common frameworks like ontologies. It provides an example of converting XML metadata to a graph using RDF.
Keynote on software sustainability given at the 2nd Annual Netherlands eScience Symposium, November 2014.
Based on the article
Carole Goble ,
Better Software, Better Research
Issue No.05 - Sept.-Oct. (2014 vol.18)
pp: 4-8
IEEE Computer Society
http://www.computer.org/csdl/mags/ic/2014/05/mic2014050004.pdf
http://doi.ieeecomputersociety.org/10.1109/MIC.2014.88
http://www.software.ac.uk/resources/publications/better-software-better-research
Digital Object Identifiers (DOIs) in the context of the International TreatyFAO
http://tiny.cc/faowgsworkshop
FAO's activities relevant to genome sequencing- International Treaty on Plant Genetic Resources for Food and Agriculture (ITPGRFA). Presentation from the FAO expert workshop on practical applications of Whole Genome Sequencing (WGS) for food safety management - 7-8 December 2015, Rome, Italy.
Harnessing Edge Informatics to Accelerate Collaboration in BioPharma (Bio-IT ...Tom Plasterer
As scientists in the life sciences we are trained to pursue singular goals around a publication or a validated target or a drug submission. Our failure rates are exceedingly high especially as we move closer to patients in the attempt to collect sufficient clinical evidence to demonstrate the value of novel therapeutics. This wastes resources as well as time for patients depending upon us for the next breakthrough.
Edge Informatics is an approach to ameliorate these failures. Using both technical and social solutions together knowledge can be shared and leveraged across the drug development process. This is accomplished by making data assets discoverable, accessible, self-described, reusable and annotatable. The Open PHACTS project pioneered this approach and has provided a number of the technical and social solutions to enable Edge Informatics. A number of pre-competitive consortia and some content providers have also embraced this approach, facilitating networks of collaborators within and outside a given organization. When taken together more accurate, timely and inclusive decision-making is fostered.
AFAIR in Astronomy Research - Slides. In this webinar ARDC is partnering with the ADACS project to explore the FAIR data principles in the context of Astronomy research and the ASVO and IVOA as a community exemplars of the implementation of the FAIR data principles.
These slides from: Keith Russell (ARDC): Looking at FAIR
In this talk Keith will provide an overview of the FAIR principles and how it was used in astronomy before it became official. He will conclude the talk by discussing what other disciplines can learn from their approach.
FAIR Ddata in trustworthy repositories: the basicsOpenAIRE
This video illustrates how certified digital repositories contribute to making and keeping research data findable, accessible, interoperable and reusable (FAIR). Trustworthy repositories support Open Access to data, as well as Restricted Access when necessary, and they offer support for metadata, sustainable and interoperable file formats, and persistent identifiers for future citation. Presented by Marjan Grootveld (DANS, OpenAIRE).
Main references
• Core Trust Seal for trustworthy digital repositories: https://www.coretrustseal.org/
• EUDAT FAIR checklist: https://doi.org/10.5281/zenodo.1065991
• European Commission’s Guidelines on FAIR data management: http://ec.europa.eu/research/participants/data/ref/h2020/grants_manual/hi/oa_pilot/h2020-hi-oa-data-mgt_en.pdf
• FAIR data principles: www.force11.org/group/fairgroup/fairprinciples
• Overview of metadata standards and tools: https://rdamsc.dcc.ac.uk/
OSFair2017 workshop | Monitoring the FAIRness of data sets - Introducing the ...Open Science Fair
Elly Dijk & Peter Doorn present the DANS approach to FAIR metrics
Workshop title: Open Science Monitor
Workshop overview:
Which are the measurable components of Open Science? How do we build a trustworthy, global open science monitor? This workshop will discuss a potential framework to measure Open Science, including the path from the publishing of an open policy (registries of policies and how these are represented or machine read), to the use of open methodologies, and the opening up of research results, their recording and measurement.
DAY 2 - PARALLEL SESSION 5
Essentials 4 Data Support: a fine course in FAIR Data SupportEllen Verbakel
The document summarizes the Essentials 4 Data Support (E4DS) course, which teaches people how to support researchers in storing, managing, archiving, and sharing research data according to FAIR principles. The course covers topics like data documentation, identifiers, formats, metadata, and licensing. It is offered online or in a blended format over 6 weeks. The goal is to educate data supporters so that researchers can find, access, interoperate with, and reuse each other's data in a fair manner.
FAIR Data in Trustworthy Data Repositories Webinar - 12-13 December 2016| www...EUDAT
| www.eudat.eu | This webinar was co-organised by DANS, EUDAT and OpenAIRE and was held on 12th and 13th December 2016.
Everybody wants to play FAIR, but how do we put the principles into practice?
There is a growing demand for quality criteria for research datasets. In this webinar we will argue that the DSA (Data Seal of Approval for data repositories) and FAIR principles get as close as possible to giving quality criteria for research data. They do not do this by trying to make value judgements about the content of datasets, but rather by qualifying the fitness for data reuse in an impartial and measurable way. By bringing the ideas of the DSA and FAIR together, we will be able to offer an operationalization that can be implemented in any certified Trustworthy Digital Repository.
In 2014 the FAIR Guiding Principles (Findable, Accessible, Interoperable and Reusable) were formulated. The well-chosen FAIR acronym is highly attractive: it is one of these ideas that almost automatically get stuck in your mind once you have heard it. In a relatively short term, the FAIR data principles have been adopted by many stakeholder groups, including research funders.
The FAIR principles are remarkably similar to the underlying principles of DSA (2005): the data can be found on the Internet, are accessible (clear rights and licenses), in a usable format, reliable and are identified in a unique and persistent way so that they can be referred to. Essentially, the DSA presents quality criteria for digital repositories, whereas the FAIR principles target individual datasets.
In this webinar the two sets of principles will be discussed and compared and a tangible operationalization will be presented.
DataCite and its Members: Connecting Research and Identifying KnowledgeETH-Bibliothek
PIDs and their metadata support scholarly research and its increasing amounts and
variety of scholarly output. DataCite provides services which enable the research community to identify, connect, cite and track these outputs, making content FAIR. New
services include data level metrics and the use of identifiers for organizations and new
types of content, e.g. software, repositories and instruments. As an open, collaborative
and community driven membership organization we rely on our members for their
input and experience to build services that are beneficial for the research community
as a whole. DataCite services as well as current and future initiatives will be described
and it will be shown how members can contribute and benefit. Over the course of the
years, our membership has grown and diversified and we are therefore refreshing and
clarifying our member model. The new member model will be presented and described.
The document discusses the Enabling FAIR Data project, which aims to improve data sharing practices in earth and environmental sciences. It outlines the FAIR data principles, key stakeholders in the project including publishers and repositories, and outputs including a commitment statement, repository finder tool, and shared authoring guidelines. The next steps are to encourage more organizations to sign and implement the commitment statement and guidelines to promote open and interoperable data.
The agenda outlines an introductory meeting to discuss FAIR technology and tools. It includes:
- Welcome and goal setting from 13:00-13:10
- Short introductions from 13:10-13:45
- A presentation on FAIR technology and tools from 13:45-14:30
- A question and answer session from 14:30-15:00
- A wrap up from 15:00
The meeting aims to introduce various organizations to FAIR principles and related technologies through a presentation and discussion.
A presentation of the Dutch Techcentre for Life Sciences FAIR Data ecosystem given at the BlueBridge workshop, a pre-event of the Research Data Alliance's 9th Plenary
Access to biomedical data is increasingly important to enable data driven science in the research community.
The Linked Open Data (LOD) principles (by Tim Berner-Lee) have been suggested to judge the quality of data by its accessibility (open data access), by its format and structures, and by its interoperability with other data sources.
The objective is to use interoperable data sources across the Web with ease.
The FAIR (findable, accessible, interoperable, reusable) data principles have been introduced for similar reasons with a stronger emphasis on achieving reusability.
In this presentation we assess the FAIR principles against the LOD principles to determine, to which degree, the FAIR principles reuse LOD principles, and to which degree they extend the LOD principles.
This assessment helps to clarify the relationship between both schemes and gives a better understanding, what extension FAIR represents in comparison to LOD.
We conclude, that LOD gives a clear mandate to the openness of data, whereas FAIR asks for a stated license for access and thus includes the concept of reusability under consideration of the license agreement.
Furthermore, FAIR makes strong reference to the contextual information required to improve reuse of the data, e.g., provenance information.
According to the LOD principles, such meta-data would be considered interoperable data as well, however, the requirement of extending of data with meta-data does indicate that FAIR is an extension of the LOD (in contrast to the inverse).
A Generic Scientific Data Model and Ontology for Representation of Chemical DataStuart Chalk
The current movement toward openness and sharing of data is likely to have a profound effect on the speed of scientific research and the complexity of questions we can answer. However, a fundamental problem with currently available datasets (and their metadata) is heterogeneity in terms of implementation, organization, and representation.
To address this issue we have developed a generic scientific data model (SDM) to organize and annotate raw and processed data, and the associated metadata. This paper will present the current status of the SDM, implementation of the SDM in JSON-LD, and the associated scientific data model ontology (SDMO). Example usage of the SDM to store data from a variety of sources with be discussed along with future plans for the work.
#1 FINDABLE covers: -- an overview of the FAIR principles: their origins, Australian FAIR initiatives, what FAIR is (and what it is not) -- the 4 FINDABLE principles which underpin the discoverability of data -- resources to support institutional awareness and uptake of Findable principles to make your institutional data globally discoverable
Speakers
1) Keith Russell, ANDS, will introduce FAIR
2) Nick Thieberger, Director of Paradisec, will present how Paradisec has made their data findable via rich metadata, identifiers through Research Data Australia and disciplinary discovery portals.
YouTube : https://youtu.be/vn2pr2dGzCs
Transcript: https://www.slideshare.net/AustralianNationalDataService/transcript-1-fair-intro-into-fair-and-f-for-findable
This document discusses the FAIR data principles and increasing adoption of FAIR. It begins by explaining the 15 FAIR principles for findable, accessible, interoperable and reusable data. It then discusses how adoption is increasing through funder requirements, the role of FAIR within EOSC, and related projects. However, it notes that most data is still not managed or shared according to FAIR principles due to barriers like time and effort required as well as lack of incentives and rewards. The document argues that both cultural and technical aspects must be addressed to fully implement FAIR.
The document discusses the use of linked open data for academia. It introduces linked data and its key principles of using URIs to identify objects and including links between data from different sources. This allows data to be interconnected in a web of data rather than separate silos. Examples are given of applying these principles through projects like Bio2RDF that link life sciences data and LODAC that link academic data about species, museums and locations. Benefits include decentralized data sharing and integration across domains. Requirements for research data are that it be accessible, reusable and sustainable.
The document outlines plans for the VODAN Africa FAIR data project. It discusses the FAIR principles of findability, accessibility, interoperability, and reusability and how they will guide the project. The architecture will include tools like CEDAR for machine-readable data production and a triple store for exposing metadata. An initial minimal viable product will integrate clinical data from DHIS2 to validate the approach before full deployment.
Making Data FAIR (Findable, Accessible, Interoperable, Reusable)Tom Plasterer
What to do About FAIR…
In the experience of most pharma professionals, FAIR remains fairly abstract, bordering on inconclusive. This session will outline specific case studies – real problems with real data, and address opportunities and real concerns.
·
Why making data Findable, Actionable, Interoperable and Reusable is important.
Talk presented at the Data Driven Drug Development (D4) conference on March 20th, 2019.
The document outlines a road map for PID Forum Finland with 3 key steps: 1) Creating engagement around PIDs by raising awareness and building skills and trust. 2) Organizing management and funding by describing use cases, creating proofs of concept, and defining requirements. 3) Creating infrastructure by ensuring interoperability, building a resolver, and organizing support services. The overall goal is to make information traceable across different channels now and in the future.
Esitys kirjastoverkkopäivillä lokakuussa 2021. Puhuin tutkimusaineistoista kuvailun kohteena, pysyvistä tunnisteista ja joistakin muista asioista liittyen tutkimusdatan erityispiirteisiin.
Esitys Kansalliskirjaston Kulttuuriperintöaineistot ja tutkimusdata --yhteistyön rajapintoja verkkotapahtumassa 4.3.2021. In this presentation I discussed research data management and how the Fairdata services enables implementing the FAIR data principles in research data publication.
1) The document summarizes a report on requirements for FAIR (Findable, Accessible, Interoperable, Reusable) data persistence and interoperability.
2) It describes a 36-month, 10 million euro project involving 22 partners from 8 EU member states working on practical implementations of semantic interoperability across research infrastructures.
3) The report analyzes the current landscape of FAIR technologies, semantic artifacts, and infrastructure initiatives; identifies challenges around scope, terminology, and rapid development; and concludes that solutions must be user-friendly, context-sensitive, and transparent while promoting adoption of standards and registries.
Collections meet the researcher. Digitalization, disintegration and disillusi...Jessica Parland-von Essen
Presentation at the LAM3 seminar in Uppsala, 9th of October 2019. On digitalization, researchers and data in the context of cultural heritage collections. The slides mostly contain headings, but the two last slides include a list of relevant reading on the subject.
Presentation on how research data can be divided into categories and how this can help data management for both service providers and researchers. Paper will be published in the journal Informaatiotutkimus in December 2018.
This document discusses best practices for organizing, managing, and publishing research data. It recommends using standardized file naming and folder structures, documenting data through code books and metadata, selecting open formats, and considering issues like data security, versions, and citations. FAIR principles of findable, accessible, interoperable and reusable data are presented. Options in Finland for publishing and archiving research data include repositories like FSD Tietoarkisto and Zenodo. Adopting these practices helps ensure well-organized, documented data that can enable reproducibility and reuse.
This document discusses making data Findable, Accessible, Interoperable and Reusable (FAIR). It provides principles for each component and examples of metadata standards and repositories that help achieve FAIR data. Resources referenced include guidelines for assigning persistent identifiers to data and metadata, describing data with rich metadata using shared vocabularies, and indexing metadata in searchable resources to enable discovery and access.
The document discusses open science and how it has changed research practices. It defines open science as making research data, notes, and processes openly available for collaboration and reuse. It outlines benefits like increasing quality, impact and innovation. Barriers like publishing costs are mentioned. The document recommends openly licensing data and publications, using open peer review and platforms, and sharing materials like code and presentations. Proper data management is important for openness, reproducibility and ensuring research integrity.
This document discusses data management practices in research. It defines research data and emphasizes the importance of good data management for ensuring integrity, reproducibility and excellence in science. Key aspects of data management include planning, documentation, metadata, sustainability, and publication. Funders increasingly require and support open access to publications and research data. The document provides guidance and considerations for implementing responsible data management and open science practices.
Candidate young stellar objects in the S-cluster: Kinematic analysis of a sub...Sérgio Sacani
Context. The observation of several L-band emission sources in the S cluster has led to a rich discussion of their nature. However, a definitive answer to the classification of the dusty objects requires an explanation for the detection of compact Doppler-shifted Brγ emission. The ionized hydrogen in combination with the observation of mid-infrared L-band continuum emission suggests that most of these sources are embedded in a dusty envelope. These embedded sources are part of the S-cluster, and their relationship to the S-stars is still under debate. To date, the question of the origin of these two populations has been vague, although all explanations favor migration processes for the individual cluster members. Aims. This work revisits the S-cluster and its dusty members orbiting the supermassive black hole SgrA* on bound Keplerian orbits from a kinematic perspective. The aim is to explore the Keplerian parameters for patterns that might imply a nonrandom distribution of the sample. Additionally, various analytical aspects are considered to address the nature of the dusty sources. Methods. Based on the photometric analysis, we estimated the individual H−K and K−L colors for the source sample and compared the results to known cluster members. The classification revealed a noticeable contrast between the S-stars and the dusty sources. To fit the flux-density distribution, we utilized the radiative transfer code HYPERION and implemented a young stellar object Class I model. We obtained the position angle from the Keplerian fit results; additionally, we analyzed the distribution of the inclinations and the longitudes of the ascending node. Results. The colors of the dusty sources suggest a stellar nature consistent with the spectral energy distribution in the near and midinfrared domains. Furthermore, the evaporation timescales of dusty and gaseous clumps in the vicinity of SgrA* are much shorter ( 2yr) than the epochs covered by the observations (≈15yr). In addition to the strong evidence for the stellar classification of the D-sources, we also find a clear disk-like pattern following the arrangements of S-stars proposed in the literature. Furthermore, we find a global intrinsic inclination for all dusty sources of 60 ± 20◦, implying a common formation process. Conclusions. The pattern of the dusty sources manifested in the distribution of the position angles, inclinations, and longitudes of the ascending node strongly suggests two different scenarios: the main-sequence stars and the dusty stellar S-cluster sources share a common formation history or migrated with a similar formation channel in the vicinity of SgrA*. Alternatively, the gravitational influence of SgrA* in combination with a massive perturber, such as a putative intermediate mass black hole in the IRS 13 cluster, forces the dusty objects and S-stars to follow a particular orbital arrangement. Key words. stars: black holes– stars: formation– Galaxy: center– galaxies: star formation
Embracing Deep Variability For Reproducibility and Replicability
Abstract: Reproducibility (aka determinism in some cases) constitutes a fundamental aspect in various fields of computer science, such as floating-point computations in numerical analysis and simulation, concurrency models in parallelism, reproducible builds for third parties integration and packaging, and containerization for execution environments. These concepts, while pervasive across diverse concerns, often exhibit intricate inter-dependencies, making it challenging to achieve a comprehensive understanding. In this short and vision paper we delve into the application of software engineering techniques, specifically variability management, to systematically identify and explicit points of variability that may give rise to reproducibility issues (eg language, libraries, compiler, virtual machine, OS, environment variables, etc). The primary objectives are: i) gaining insights into the variability layers and their possible interactions, ii) capturing and documenting configurations for the sake of reproducibility, and iii) exploring diverse configurations to replicate, and hence validate and ensure the robustness of results. By adopting these methodologies, we aim to address the complexities associated with reproducibility and replicability in modern software systems and environments, facilitating a more comprehensive and nuanced perspective on these critical aspects.
https://hal.science/hal-04582287
The cost of acquiring information by natural selectionCarl Bergstrom
This is a short talk that I gave at the Banff International Research Station workshop on Modeling and Theory in Population Biology. The idea is to try to understand how the burden of natural selection relates to the amount of information that selection puts into the genome.
It's based on the first part of this research paper:
The cost of information acquisition by natural selection
Ryan Seamus McGee, Olivia Kosterlitz, Artem Kaznatcheev, Benjamin Kerr, Carl T. Bergstrom
bioRxiv 2022.07.02.498577; doi: https://doi.org/10.1101/2022.07.02.498577
Mechanisms and Applications of Antiviral Neutralizing Antibodies - Creative B...Creative-Biolabs
Neutralizing antibodies, pivotal in immune defense, specifically bind and inhibit viral pathogens, thereby playing a crucial role in protecting against and mitigating infectious diseases. In this slide, we will introduce what antibodies and neutralizing antibodies are, the production and regulation of neutralizing antibodies, their mechanisms of action, classification and applications, as well as the challenges they face.
Sexuality - Issues, Attitude and Behaviour - Applied Social Psychology - Psyc...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
This presentation offers a general idea of the structure of seed, seed production, management of seeds and its allied technologies. It also offers the concept of gene erosion and the practices used to control it. Nursery and gardening have been widely explored along with their importance in the related domain.
Signatures of wave erosion in Titan’s coastsSérgio Sacani
The shorelines of Titan’s hydrocarbon seas trace flooded erosional landforms such as river valleys; however, it isunclear whether coastal erosion has subsequently altered these shorelines. Spacecraft observations and theo-retical models suggest that wind may cause waves to form on Titan’s seas, potentially driving coastal erosion,but the observational evidence of waves is indirect, and the processes affecting shoreline evolution on Titanremain unknown. No widely accepted framework exists for using shoreline morphology to quantitatively dis-cern coastal erosion mechanisms, even on Earth, where the dominant mechanisms are known. We combinelandscape evolution models with measurements of shoreline shape on Earth to characterize how differentcoastal erosion mechanisms affect shoreline morphology. Applying this framework to Titan, we find that theshorelines of Titan’s seas are most consistent with flooded landscapes that subsequently have been eroded bywaves, rather than a uniform erosional process or no coastal erosion, particularly if wave growth saturates atfetch lengths of tens of kilometers.
1. Open Science goes FAIR - what does it
mean?
Jessica Parland-von Essen 4.12.2020
2. OPEN DATA
Open means anyone can freely access,
use, modify, and share for any
purpose (subject, at most, to requirements
that preserve provenance and openness).
https://opendefinition.org/od/2.1/en/
2
7. FAIR principles for data – requirements for research data
Findable
F1. (Meta)data are assigned a globally unique and persistent
identifier
F2. Data are described with rich metadata (defined by R1 below)
F3. Metadata clearly and explicitly include the identifier of the data
they describe
F4. (Meta)data are registered or indexed in a searchable resource
Accessible
A1. (Meta)data are retrievable by their identifier using a
standardised communications protocol
A2. Metadata are accessible, even when the data are no longer
available
Interoperable
I1. (Meta)data use a formal, accessible, shared, and
broadly applicable language for knowledge
representation
I2. (Meta)data use vocabularies that follow FAIR
principles
I3. (Meta)data include qualified references to other
(meta)data
Reusable
R1. Meta(data) are richly described with a plurality of
accurate and relevant attributes
8. 8
F
A
I
R
FINDABLE
• Described in relevant catalog with enough detail
• Landing page with globally unique persistent identifier
ACCESSIBLE
• Can be retrieved over the internet
• Versioning and lifecycle are documented
• Tombstone page if data is deleted
INTEROPERABLE
• Common, documented, and open formats
RE-USABLE
• Well documented and intelligible
• Rights clearly stated
https://doi.org/10.5281/zenodo.4045402
9. Persistent identifiers are unique and unambiguous
Catalog
Resol
ver
Datafile
Contract
Configration
Read me
PID
11. Shallow FAIR and Deep FAIR
11
Necessary
research
information, PIDs,
machine readable
license
All data
elements are
machine
accessible
Research
Information
Research
Data
12. ACTIVE DATA
Raw, continuously
updated
DYNAMIC
RESEARCH DATA
Version
controlled,
possible to cite
RESEARCH
DATASET
PUBLICATION
Immutable
Documentation, validation
Research
Research Data Types
https://doi.org/10.23978/inf.77419
13. FAIR Ecosystem Components and FAIR Digital Objects
13 http://doi.org/10.5281/zenodo.3565428 https://doi.org/doi:10.2777/1524
14. 05/12/202014
• Don’t try to do this alone, contact
your data stewardsResearchers
• Talk to researchers about
reproducibility and semantic artefactsData stewards
• Co-develop and co-create,
support FAIR and document thoroughly
Service
providers
18. • Requirement from many funders to publish data
• GDPR and rights management
• Ensure good reproducibility
• More citations and visibility
• Good documentation supports your own future work
Why data management planning
18
19. Co-creation &
co-development
05/12/202019
Always design a thing by considering it in its next
larger context – a chair in a room, a room in a
house, a house in an environment, an
environment in a city plan.
Eliel Saarinen, Finnish architect (1873--1950)
LA2 / CC BY-S. Wikimedia
(https://creativecommons.org/licenses/by-sa/4.0)
Editor's Notes
F = Findable, kun aineistolla on pysyvä tunniste esim doi, linkki aineistoon toimii aina vaikka säilytyspaikka muuttuisi
A = Accessible, tutkimusaineiston tunniste toimii hyperlinkkinä jonka avulla dataan ja sen kuvailutietoihin pääsee käsiksi verkkoselaimella
I = Interoperable yhteentoimivuuden periaate edellyttää avoimia tiedostomuotoja ja yhteisiä standardeja, ei enää tiedostoja jotka eivät aukea
R = Re-usable (datan kuvailu tukee tätä), dataa voidaan käyttää kun sillä on metatietoja ja käyttöehdoista kertova lisenssi
The first use case is the visibility of your work and outputs. When reporting on your work, to funders, and publishing outputs, a basic level of FAIRness and PID use is sufficient to enable findability, simple citation and output registration with core descriptive metadata. This is the context of what is usually called research information (sometimes referred to as current research information). The most common and useful PIDs for this are the research output DOI and the ORCID for the creator(s)/contributor(s). There are also other systems available to identify other kinds of entities to help further linking of information, such as organisations or protocols. Funders and employers might for instance require linking to some other contextual reference data like lists of grants, funders and affiliated organisations. This kind of information is becoming more important, but the actual data quality is depending on the functionalities each service provides. If the services used for dataset publication or reporting don’t require PIDs or don’t offer reference (meta)datasets or integration with PIDs for these kinds of things, it is difficult for the researcher to provide this information in an unambiguous way.
The other use case for PIDs is the management of the research data itself. Here the PIDs can have different functions: (a) creating deep FAIR research datasets as research outputs, where all individual data elements are machine accessible, see panel F in Figure 1, or (b) when managing and documenting the actual workflow and data and related information during research to ensure reproducibility of research results.
The archive or generic repository usually operates with research dataset publications, that are are a sort of publication, albeit complex, but immutable, archived as output and evidence for research. This case is quite easy, pid wise. But in real life there are many steps and varieties of data before this- This should be taken into account when citing, for instance. How can we support sufficient reproduciblity without overflowing all systems with PID – that should be kept and maintained forever!?
Figure 8 lähde: TFiR https://doi.org/doi:10.2777/1524
Diagram 2 lähde : http://doi.org/10.5281/zenodo.3565428
Researchers: don’t try to do this alone, contact your data-stewards.
Data-stewards: talk to researchers about reproducibility and semantic artefacts.
Service providers: co-develop and co-create, support FAIR and document thoroughly.