Presentation about the IGSN and ongoing initiatives for the Internet of Samples at the EGU 2015 short course "Open Science Goes Geo: Beyond Data and Software".
IGSN: The International Geo Sample Number (DFG Roundtable)Kerstin Lehnert
This presentation provides an overview of the rationale for the IGSN, of the organizational structure and architecture of the IGSN e.V. , and the System for Earth Sample Registration.
Advancing Reproducible Science from Physical Samples: The IGSN and the iSampl...Kerstin Lehnert
Presentation at the Geological Society of America (GSA) meeting 2016 in the session on FOSSIL SPECIMENS 0'S AND 1'S: DATABASES, STANDARDS, & MOBILIZATION
Research Data Infrastructure for Geochemistry (DFG Roundtable)Kerstin Lehnert
This presentation provides an overview of different aspects of data management for geochemistry and resources available at the EarthChem@IEDA data facility.
Digital Representation of Physical Samples in Scientific PublicationsKerstin Lehnert
Presentation about the digital representation of physical samples in scientific publications, given at the European Geoscience Union meeting 2015 in the Splinter Meeting 1.36 "Digital Representation of Physical Samples in Scientific Publications".
USING E-INFRASTRUCTURES FOR BIODIVERSITY CONSERVATION - Module 3Gianpaolo Coro
An e-Infrastructure is a distributed network of service nodes, residing on multiple sites and managed by one or more organizations. e-Infrastructures allow scientists residing at distant places to collaborate. They offer a multiplicity of facilities as-a-service, supporting data sharing and usage at different levels of abstraction, e.g. data transfer, data harmonization, data processing workflows etc. e-Infrastructures are gaining an important place in the field of biodiversity conservation. Their computational capabilities help scientists to reuse models, obtain results in shorter time and share these results with other colleagues. They are also used to access several and heterogeneous biodiversity catalogues.
In this course, the D4Science e-Infrastructure will be used to conduct experiments in the field of biodiversity conservation. D4Science hosts models and contributions by several international organizations involved in the biodiversity conservation field. The course will give students an overview of the models, the practices and the methods that large international organizations like FAO and UNESCO apply by means of D4Science. At the same time, the course will introduce students to the basic concepts under e-Infrastructures, Virtual Research Environments, data sharing and experiments reproducibility.
IGSN: The International Geo Sample Number (DFG Roundtable)Kerstin Lehnert
This presentation provides an overview of the rationale for the IGSN, of the organizational structure and architecture of the IGSN e.V. , and the System for Earth Sample Registration.
Advancing Reproducible Science from Physical Samples: The IGSN and the iSampl...Kerstin Lehnert
Presentation at the Geological Society of America (GSA) meeting 2016 in the session on FOSSIL SPECIMENS 0'S AND 1'S: DATABASES, STANDARDS, & MOBILIZATION
Research Data Infrastructure for Geochemistry (DFG Roundtable)Kerstin Lehnert
This presentation provides an overview of different aspects of data management for geochemistry and resources available at the EarthChem@IEDA data facility.
Digital Representation of Physical Samples in Scientific PublicationsKerstin Lehnert
Presentation about the digital representation of physical samples in scientific publications, given at the European Geoscience Union meeting 2015 in the Splinter Meeting 1.36 "Digital Representation of Physical Samples in Scientific Publications".
USING E-INFRASTRUCTURES FOR BIODIVERSITY CONSERVATION - Module 3Gianpaolo Coro
An e-Infrastructure is a distributed network of service nodes, residing on multiple sites and managed by one or more organizations. e-Infrastructures allow scientists residing at distant places to collaborate. They offer a multiplicity of facilities as-a-service, supporting data sharing and usage at different levels of abstraction, e.g. data transfer, data harmonization, data processing workflows etc. e-Infrastructures are gaining an important place in the field of biodiversity conservation. Their computational capabilities help scientists to reuse models, obtain results in shorter time and share these results with other colleagues. They are also used to access several and heterogeneous biodiversity catalogues.
In this course, the D4Science e-Infrastructure will be used to conduct experiments in the field of biodiversity conservation. D4Science hosts models and contributions by several international organizations involved in the biodiversity conservation field. The course will give students an overview of the models, the practices and the methods that large international organizations like FAO and UNESCO apply by means of D4Science. At the same time, the course will introduce students to the basic concepts under e-Infrastructures, Virtual Research Environments, data sharing and experiments reproducibility.
No specimen left behind: Collections digitisation at the NHM, London*Vince Smith
Presentation on the Natural History Museum, London Digitisation Programme, given at the "Collections for the 21st Century" meeting in Gainesville, Florida, 5-6 May 2014
AusPlots field data collection with AusScribeTERN Australia
The presentation provides an overview of AuScribe, an Android-based ecology field survey App based on AusPLots Rangelands Survey Protocols Manual. The presentation is part of the Workshop on Approaches to Terrestrial Ecosystem Data Management : from collection to synthesis and beyond
USING E-INFRASTRUCTURES FOR BIODIVERSITY CONSERVATION - Module 5Gianpaolo Coro
An e-Infrastructure is a distributed network of service nodes, residing on multiple sites and managed by one or more organizations. e-Infrastructures allow scientists residing at distant places to collaborate. They offer a multiplicity of facilities as-a-service, supporting data sharing and usage at different levels of abstraction, e.g. data transfer, data harmonization, data processing workflows etc. e-Infrastructures are gaining an important place in the field of biodiversity conservation. Their computational capabilities help scientists to reuse models, obtain results in shorter time and share these results with other colleagues. They are also used to access several and heterogeneous biodiversity catalogues.
In this course, the D4Science e-Infrastructure will be used to conduct experiments in the field of biodiversity conservation. D4Science hosts models and contributions by several international organizations involved in the biodiversity conservation field. The course will give students an overview of the models, the practices and the methods that large international organizations like FAO and UNESCO apply by means of D4Science. At the same time, the course will introduce students to the basic concepts under e-Infrastructures, Virtual Research Environments, data sharing and experiments reproducibility.
GBIF BIFA mentoring, Day 4b Event core, July 2016Dag Endresen
GBIF BIFA mentoring in Los Banos, Philippines for the South-East Asian ASEAN Biodiversity Heritage Parks. With Dr. Yu-Huang Wang, Dr. Po-Jen Chiang, and Guan-Shuo Mai from TaiBIF the GBIF node of Taiwan (Chinese Tapei); and the Biodiversity Informatics team at ASEAN Centre For Biodiversity. http://www.gbif.no/events/2016/gbif-bifa-mentoring.html
Data Facilities Workshop - Panel on Current Concepts in Data Sharing & Intero...EarthCube
This series of presentations was given at the EarthCube Data Facilities End-User Workshop held January 15-17, 2014 in Washington, DC. This workshop provided a forum to discuss the unique requirements and challenges associated with developing the communication, collaboration, interoperability, and governance structures that will be required to build EarthCube in conjunction with existing and emerging NSF/GEO facilities.
This panel and discussion, specifically, outlined and explained several current concepts in data sharing and interoperability, featuring presentations by:
Paul Morin (UMN): Polar Cyberinfrastructure
Don Middleton (UCAR): Atmospheric/Climate
Kerstin Lehnert (LDEO): Domain Repositories & Physical Samples
David Schindel (CBOL, GRBio): Biological Perspective & Collections
Hank Leoscher (NEON): Observation Networks
Daniel Fuka (Virginia Tech) and Ruth Duerr (NSIDC): Brokering
Ilya Zaslavsky (UCSD): Cross-Domain Interoperability
CoESRA: Platform for collaborative researchTERN Australia
The presentation provides an overview of CoESRA, a cloud-based platform to build, execute and share complex analysis and synthesis experiments. The infrastructure is offered as "Platform as a Service in virtual desktop. The presentation is part of the Workshop on Approaches to Terrestrial Ecosystem Data Management : from collection to synthesis and beyond
Eco-informatics: Data services for bringing together and publishing the full ...TERN Australia
The presentation provides an overview of Advanced Ecological Knowledge and Observation System and SHaRED services by the TERN Eco-informatics to publish plot-based ecological data. The presentation was part of the Workshop on Approaches to Terrestrial Ecosystem Data Management : from collection to synthesis and beyond which was held on 9th of March 2016 in University of Queensland.
Digital Services Division & The Biodiversity Heritage LibraryMartin Kalfatovic
Digital Services Division & The Biodiversity Heritage Library. Martin R. Kalfatovic. Smithsonian Science Executive Committee. Washington, DC. 12 January 2015
The Biodiversity Heritage Library: Workflow OverviewMartin Kalfatovic
The Biodiversity Heritage Library: Workflow Overview. Martin R. Kalfatovic and Suzanne C. Pilsk. BHL Australian Node Meeting: Melbourne Museum. 2 June 2010. Melbourne, Australia.
Digitizing Entomology: The Biodiversity Heritage Library @ the SmithsonianMartin Kalfatovic
Digitizing Entomology: The Biodiversity Heritage Library @ the Smithsonian. Martin R. Kalfatovic. National Museum of Natural History, Department of Entomology Staff Meeting. Martin R. Kalfatovic. November 26, 2007. Washington, DC.
Data You May Like: A Recommender System for Research Data DiscoveryAnusuriya Devaraju
Various data portals been developed to facilitate access to research datasets from different sources. For example, the Data Publisher for Earth & Environmental Science (PANGAEA), the Registry of Research Data Repositories (re3data.org), and the National Geoscience Data Centre (NGDC). Due to data quantity and heterogeneity, finding relevant datasets on these portals may be difficult and tedious. Keyword searches based on specific metadata elements or multi-key indexes may return irrelevant results. Faceted searches may be unsatisfactory and time consuming, especially when facet values are exhaustive. We need a much more intelligent way to complement existing searching mechanisms in order to enhance user experiences of the data portals.
We developed a recommender system that helps users to find the most relevant research datasets on the CSIRO’s Data Access Portal (DAP). The system is based on content-based filtering. We computed the similarity of datasets based on data attributes (e.g., descriptions, fields of research, location, contributors, and provenance) and inference from transaction logs (e.g., the relations among datasets and between queries and datasets). We improved the recommendation quality by assigning weights to data similarities. The weight values are drawn from a survey involving data users. The recommender results for a given dataset are accessible programmatically via a web service. Taking both data attributes and user actions into account, the recommender system will make it easier for researchers to find and reuse data offered through the data portal.
No specimen left behind: Collections digitisation at the NHM, London*Vince Smith
Presentation on the Natural History Museum, London Digitisation Programme, given at the "Collections for the 21st Century" meeting in Gainesville, Florida, 5-6 May 2014
AusPlots field data collection with AusScribeTERN Australia
The presentation provides an overview of AuScribe, an Android-based ecology field survey App based on AusPLots Rangelands Survey Protocols Manual. The presentation is part of the Workshop on Approaches to Terrestrial Ecosystem Data Management : from collection to synthesis and beyond
USING E-INFRASTRUCTURES FOR BIODIVERSITY CONSERVATION - Module 5Gianpaolo Coro
An e-Infrastructure is a distributed network of service nodes, residing on multiple sites and managed by one or more organizations. e-Infrastructures allow scientists residing at distant places to collaborate. They offer a multiplicity of facilities as-a-service, supporting data sharing and usage at different levels of abstraction, e.g. data transfer, data harmonization, data processing workflows etc. e-Infrastructures are gaining an important place in the field of biodiversity conservation. Their computational capabilities help scientists to reuse models, obtain results in shorter time and share these results with other colleagues. They are also used to access several and heterogeneous biodiversity catalogues.
In this course, the D4Science e-Infrastructure will be used to conduct experiments in the field of biodiversity conservation. D4Science hosts models and contributions by several international organizations involved in the biodiversity conservation field. The course will give students an overview of the models, the practices and the methods that large international organizations like FAO and UNESCO apply by means of D4Science. At the same time, the course will introduce students to the basic concepts under e-Infrastructures, Virtual Research Environments, data sharing and experiments reproducibility.
GBIF BIFA mentoring, Day 4b Event core, July 2016Dag Endresen
GBIF BIFA mentoring in Los Banos, Philippines for the South-East Asian ASEAN Biodiversity Heritage Parks. With Dr. Yu-Huang Wang, Dr. Po-Jen Chiang, and Guan-Shuo Mai from TaiBIF the GBIF node of Taiwan (Chinese Tapei); and the Biodiversity Informatics team at ASEAN Centre For Biodiversity. http://www.gbif.no/events/2016/gbif-bifa-mentoring.html
Data Facilities Workshop - Panel on Current Concepts in Data Sharing & Intero...EarthCube
This series of presentations was given at the EarthCube Data Facilities End-User Workshop held January 15-17, 2014 in Washington, DC. This workshop provided a forum to discuss the unique requirements and challenges associated with developing the communication, collaboration, interoperability, and governance structures that will be required to build EarthCube in conjunction with existing and emerging NSF/GEO facilities.
This panel and discussion, specifically, outlined and explained several current concepts in data sharing and interoperability, featuring presentations by:
Paul Morin (UMN): Polar Cyberinfrastructure
Don Middleton (UCAR): Atmospheric/Climate
Kerstin Lehnert (LDEO): Domain Repositories & Physical Samples
David Schindel (CBOL, GRBio): Biological Perspective & Collections
Hank Leoscher (NEON): Observation Networks
Daniel Fuka (Virginia Tech) and Ruth Duerr (NSIDC): Brokering
Ilya Zaslavsky (UCSD): Cross-Domain Interoperability
CoESRA: Platform for collaborative researchTERN Australia
The presentation provides an overview of CoESRA, a cloud-based platform to build, execute and share complex analysis and synthesis experiments. The infrastructure is offered as "Platform as a Service in virtual desktop. The presentation is part of the Workshop on Approaches to Terrestrial Ecosystem Data Management : from collection to synthesis and beyond
Eco-informatics: Data services for bringing together and publishing the full ...TERN Australia
The presentation provides an overview of Advanced Ecological Knowledge and Observation System and SHaRED services by the TERN Eco-informatics to publish plot-based ecological data. The presentation was part of the Workshop on Approaches to Terrestrial Ecosystem Data Management : from collection to synthesis and beyond which was held on 9th of March 2016 in University of Queensland.
Digital Services Division & The Biodiversity Heritage LibraryMartin Kalfatovic
Digital Services Division & The Biodiversity Heritage Library. Martin R. Kalfatovic. Smithsonian Science Executive Committee. Washington, DC. 12 January 2015
The Biodiversity Heritage Library: Workflow OverviewMartin Kalfatovic
The Biodiversity Heritage Library: Workflow Overview. Martin R. Kalfatovic and Suzanne C. Pilsk. BHL Australian Node Meeting: Melbourne Museum. 2 June 2010. Melbourne, Australia.
Digitizing Entomology: The Biodiversity Heritage Library @ the SmithsonianMartin Kalfatovic
Digitizing Entomology: The Biodiversity Heritage Library @ the Smithsonian. Martin R. Kalfatovic. National Museum of Natural History, Department of Entomology Staff Meeting. Martin R. Kalfatovic. November 26, 2007. Washington, DC.
Data You May Like: A Recommender System for Research Data DiscoveryAnusuriya Devaraju
Various data portals been developed to facilitate access to research datasets from different sources. For example, the Data Publisher for Earth & Environmental Science (PANGAEA), the Registry of Research Data Repositories (re3data.org), and the National Geoscience Data Centre (NGDC). Due to data quantity and heterogeneity, finding relevant datasets on these portals may be difficult and tedious. Keyword searches based on specific metadata elements or multi-key indexes may return irrelevant results. Faceted searches may be unsatisfactory and time consuming, especially when facet values are exhaustive. We need a much more intelligent way to complement existing searching mechanisms in order to enhance user experiences of the data portals.
We developed a recommender system that helps users to find the most relevant research datasets on the CSIRO’s Data Access Portal (DAP). The system is based on content-based filtering. We computed the similarity of datasets based on data attributes (e.g., descriptions, fields of research, location, contributors, and provenance) and inference from transaction logs (e.g., the relations among datasets and between queries and datasets). We improved the recommendation quality by assigning weights to data similarities. The weight values are drawn from a survey involving data users. The recommender results for a given dataset are accessible programmatically via a web service. Taking both data attributes and user actions into account, the recommender system will make it easier for researchers to find and reuse data offered through the data portal.
Irina Bastrakova, GV. Sample management and International geo-sample number (ISGN) implementation at geoscience Australia.
2 Nov 2016, Canberra. International Geo-Sample Number (IGSN) Symposium.
John Morrisseey - CSIRO. CSIRO National research collections Australia (NRCA). Specimen identifiers - possible futures.
2 Nov 2016, Canberra. International Geo-Sampling Number (IGSN) Symposium
Virtual Research Environments supporting tailor-made data management service...Blue BRIDGE
Presented by Pasquale Pagano of CNR at the BlueBRIDGE Workshop at SeaTech Week 2016 in Brest, France. http://www.bluebridge-vres.eu/events/join-bluebridge-10th-biennial-sea-tech-week-brest-france
Identifying and Linking Physical Samples with Data: Using IGSNARDC
7 June 2017
This webinar is the second in a series examining persistent identifiers and their use in research. This webinar:
It introduced the IGSN, outlining its structure, use, application and availability for Australian researchers and research institutions
discussed the international symposium Linking Environmental Data and Samples.
Watch full webinar: https://www.youtube.com/watch?v=mOJRaLwOaCs
Being FAIR: FAIR data and model management SSBSS 2017 Summer SchoolCarole Goble
Lecture 1:
Being FAIR: FAIR data and model management
In recent years we have seen a change in expectations for the management of all the outcomes of research – that is the “assets” of data, models, codes, SOPs, workflows. The “FAIR” (Findable, Accessible, Interoperable, Reusable) Guiding Principles for scientific data management and stewardship [1] have proved to be an effective rallying-cry. Funding agencies expect data (and increasingly software) management retention and access plans. Journals are raising their expectations of the availability of data and codes for pre- and post- publication. The multi-component, multi-disciplinary nature of Systems and Synthetic Biology demands the interlinking and exchange of assets and the systematic recording of metadata for their interpretation.
Our FAIRDOM project (http://www.fair-dom.org) supports Systems Biology research projects with their research data, methods and model management, with an emphasis on standards smuggled in by stealth and sensitivity to asset sharing and credit anxiety. The FAIRDOM Platform has been installed by over 30 labs or projects. Our public, centrally hosted Asset Commons, the FAIRDOMHub.org, supports the outcomes of 50+ projects.
Now established as a grassroots association, FAIRDOM has over 8 years of experience of practical asset sharing and data infrastructure at the researcher coal-face ranging across European programmes (SysMO and ERASysAPP ERANets), national initiatives (Germany's de.NBI and Systems Medicine of the Liver; Norway's Digital Life) and European Research Infrastructures (ISBE) as well as in PI's labs and Centres such as the SynBioChem Centre at Manchester.
In this talk I will show explore how FAIRDOM has been designed to support Systems Biology projects and show examples of its configuration and use. I will also explore the technical and social challenges we face.
I will also refer to European efforts to support public archives for the life sciences. ELIXIR (http:// http://www.elixir-europe.org/) the European Research Infrastructure of 21 national nodes and a hub funded by national agreements to coordinate and sustain key data repositories and archives for the Life Science community, improve access to them and related tools, support training and create a platform for dataset interoperability. As the Head of the ELIXIR-UK Node and co-lead of the ELIXIR Interoperability Platform I will show how this work relates to your projects.
[1] Wilkinson et al, The FAIR Guiding Principles for scientific data management and stewardship Scientific Data 3, doi:10.1038/sdata.2016.18
The BlueBRIDGE approach to collaborative researchBlue BRIDGE
Gianpaolo Coro, ISTI-CNR, at BlueBRIDGE workshop on "Data Management services to support stock assessement", held during the Annual ICES Science conference 2016
FAIR Data, Operations and Model management for Systems Biology and Systems Me...Carole Goble
FAIR Data, Operations and Model management for Systems Biology and Systems Medicine Projects given at 1st Conference of the European Association of Systems Medicine, 26-28 October 2016, Berlin. the FAIRDOM project is described.
Towards an e-infrastructure in agriculture?Blue BRIDGE
Donatella Castelli, CNR-ISTI & BlueBRIDGE Coordinator, gave an introductive talk in the "Towards an e-infrastructure in agriculture?" session at the Euragri workship in Inra, Paris discussing leading an e-infrastructure project in marine research e-Infrastructure and how it refers to a combination of digital technologies (hardware and software), resources (data, services, digital libraries), communications (protocols, access rights and networks), and the people and organisational structures needed to manage them.
Towards A Web-Enabled Geo-Sample Web: An Open Source Resource Registration an...Anusuriya Devaraju
Within the earth sciences the curation and sharing of geo-samples is crucial
to supporting reproducible research, in addition to extending the use of the samples in new
research, and saving costs by avoiding sample loss and duplicating sampling activities. In the
Commonwealth Scientic and Industrial Research Organisation (CSIRO), researchers gather
various geo-samples as part of their eld studies and collaborative projects. The diversity of the
samples and their unsystematic management led ambiguous sample numbers, incomplete sample
descriptions, and diculties in nding the samples and their related data. These problems are
also found in universities, research institutes and government agencies, which usually curate and
manage diverse samples. To address this problem, we developed an open source registration
and management system to identify geo-samples unambiguously and to manage their metadata
and data systematically. The system supports the linking of samples and sample collections to
the real world features from where they were collected, as well as to their data and reports on
the Web. This paper describes the implementation of the system including its underlying design
considerations, and its applications. The system was built upon the International Geo Sample
Number persistent identier system with Semantic Web technologies. It has been implemented
and tested with individual users and three sample repositories in the organization.
PERICLES Workflow for the automated updating of Digital Ecosystem Models with...PERICLES_FP7
This presentation was delivered by Anna Eggers (State and University Library Göttingen) and Fabio Corubolo (University of Liverpool) at PERICLES final project conference 'Acting on Change: New Approaches and Future Practices in LTDP' (Wellcome Collection Conference Centre, London, 30 Nov -1 Dec 2016).
This 'PERICLES in practice' session aimed at demonstrating how the PET and the EcoBuilder tools can be used to generate information to populate Digital Ecosystem Model (entities) and how the model can be fed into the Entity Registry Model Repository (ERMR). This session showed how the PERICLES tools can contribute to automated model instance creation and automatic updating.
The 'PERICLES in practice' sessions presented specific outcomes of the PERICLES project set in an example workflow, combining tools to accomplish a goal defined by practitioners and derived from real life challenges they experience in their field of work.
Approach and outcome of the Biodiversity Virtual e-Laboratory (BioVeL) projectAlex Hardisty
Describes what we set out to do, what we achieved, and some of the lessons learnt during the BioVeL project. This presentation was given at the BioVeL final event "BioVeL In Practice and In Future", Paris, 13th November 2014
The Implementation of the International Geo Sample Number in CSIRO: Experienc...Anusuriya Devaraju
In 2014 the Commonwealth Scientific and Industrial Research Organisation (CSIRO) began to implement the International Geo Sample Number (IGSN) to allow unambiguous identification of physical samples and data derived from these samples. In this paper we describe the requirements for the implementation of persistent identifiers for physical samples in the organisation and technical solutions we developed to meet these requirements.
BioCatalogue talk by Carole Goble. She outlines in these slides the reasons behind the BioCatalogue project. And present the BioCatalogue and its goals.
A presentation given by Manjula Patel (UKOLN) at the Repository Curation Environments (RECURSE) Workshop held at the 4th International Digital Curation Conference, Edinburgh, 1st December 2008,
http://www.dcc.ac.uk/events/dcc-2008/programme/
German Conference on Bioinformatics 2021
https://gcb2021.de/
FAIR Computational Workflows
Computational workflows capture precise descriptions of the steps and data dependencies needed to carry out computational data pipelines, analysis and simulations in many areas of Science, including the Life Sciences. The use of computational workflows to manage these multi-step computational processes has accelerated in the past few years driven by the need for scalable data processing, the exchange of processing know-how, and the desire for more reproducible (or at least transparent) and quality assured processing methods. The SARS-CoV-2 pandemic has significantly highlighted the value of workflows.
This increased interest in workflows has been matched by the number of workflow management systems available to scientists (Galaxy, Snakemake, Nextflow and 270+ more) and the number of workflow services like registries and monitors. There is also recognition that workflows are first class, publishable Research Objects just as data are. They deserve their own FAIR (Findable, Accessible, Interoperable, Reusable) principles and services that cater for their dual roles as explicit method description and software method execution [1]. To promote long-term usability and uptake by the scientific community, workflows (as well as the tools that integrate them) should become FAIR+R(eproducible), and citable so that author’s credit is attributed fairly and accurately.
The work on improving the FAIRness of workflows has already started and a whole ecosystem of tools, guidelines and best practices has been under development to reduce the time needed to adapt, reuse and extend existing scientific workflows. An example is the EOSC-Life Cluster of 13 European Biomedical Research Infrastructures which is developing a FAIR Workflow Collaboratory based on the ELIXIR Research Infrastructure for Life Science Data Tools ecosystem. While there are many tools for addressing different aspects of FAIR workflows, many challenges remain for describing, annotating, and exposing scientific workflows so that they can be found, understood and reused by other scientists.
This keynote will explore the FAIR principles for computational workflows in the Life Science using the EOSC-Life Workflow Collaboratory as an example.
[1] Carole Goble, Sarah Cohen-Boulakia, Stian Soiland-Reyes,Daniel Garijo, Yolanda Gil, Michael R. Crusoe, Kristian Peters, and Daniel Schober FAIR Computational Workflows Data Intelligence 2020 2:1-2, 108-121 https://doi.org/10.1162/dint_a_00033.
Similar to The Internet of Samples: IGSN in Action (20)
This slide deck provides an update on the development of the Astromaterials Data System, a project funded by NASA to ensure the long-term accessibility and utility of lab analytical data acquired on astromaterials samples curated at the Johnson Space Center, including samples collected on the moon during the Apollo missions and meteorites collected in Antarctica.
Presentation about geochemical research data access and publication provided to the Australian Geochemistry Network by Kerstin Lehnert of EarthChem and the Astromaterials Data System
Boosting Data Science in Geochemistry: We Need Global Geochemical Data Standa...Kerstin Lehnert
Presentation at AGU Fall Meeting 2018: Large-scale, global geochemical data syntheses like EarthChem and GEOROC have, for nearly two decades, inspired and made possible a vast range of scientific studies and new discoveries, facilitating the analysis and mining of geochemical data and creating new paradigms in geochemical data analysis such as statistical geochemistry. These syntheses provide easy access to fully integrated compilations of thousands of datasets (‘data fusion’) with millions of geochemical measurements that are accompanied by comprehensive and harmonized metadata for context and provenance to search, filter, sort, and evaluate the data.
The syntheses have been assembled and maintained through manual labor by data managers, who extract data and metadata from text, tables, and supplements of publications for inclusion in the databases, a time-consuming task due to the multitude of data formats, units, normalizations, vocabularies, etc., i.e. lack of best practices for geochemical data reporting. In order to support and advance future science endeavors that rely on access to and analysis of large volumes of geochemical data, we need to develop and implement global standards for geochemical data that not only make geochemical data FAIR (Findable, Accessible, Interoperable, Re-usable), but ready for data fusion. As more geochemical data systems are emerging at national, programmatic, and subdomain levels in response to Open Access policies and science needs, standard protocols for exchanging geochemical data among these systems will need to be developed, implemented, and governed.
Critical is the alignment with existing standards such as the Semantic Sensor Network (SSN) ontology, a recent joint W3C and OGC standard that standardizes description of sensors, observation, sampling, and actuation, with sufficient flexibility to allow details of these elements to be defined in different domains. New initiatives within the International Council for Science and CODATA are working towards coordinating the International Science Unions to identify and endorse the more authoritative standards (including vocabularies and ontologies). These initiatives present a timely opportunity for geochemical data to ensure that they are born ‘connected’ within and across disciplines.
Looking at the past of infrastructure development for research data in the context of infrastructure development patterns and experiences from the evolution of the IEDA data facility to inform future pathways and developments. A major focus of the lecture is on the FAIR principles and the issues surrounding reusability of data.
Presentation that describes the experiences and insights of the IEDA data facility gained during the >10 years of building cyberinfrastructure for a long-tail community geochemistry
Making Small Data BIG (UT Austin, March 2016)Kerstin Lehnert
Presentation given at the Texas Advanced Computing Center. It describes the potential of re-using small data for new science, achievements and the challenges to make small data re-usable.
Interdisciplinary Data Resources for Volcanology at the IEDA (Interdisciplina...Kerstin Lehnert
Presentation given at the EGU 2015 General Assembly in session "Methods for Understanding Volcanic Hazards and Risks" (NH2.2), describing EarthChem data systems that make accessible and synthesize geochemical data of volcanic rocks and gases, and the System for Earth Sample Registration that catalogs sample metadata and provides persistent unique sample identifiers (International Geo Sample Number IGSN). It also mentions EarthChem's plans and ongoing work to link geochemical data with other volcanological databases, and the IEDA data rescue initiative.
Lehnert: Making Small Data Big, IACS, April2015Kerstin Lehnert
Seminar presentation at the Institute for Advanced Computational Science at Stony Brook University, April 9, 2015, describing achievements and challenges of data infrastructure in a long-tail science domain with the example of geochemistry.
iSamples Research Coordination Network (C4P Webinar)Kerstin Lehnert
The iSamples (Internet of Samples in the Earth Sciences) Research Coordination Network is part of EarthCube and focuses on the integration of physical samples and collections into digital data infrastructure in the Earth sciences. This presentation summarizes the activities of the iSamples RCN and presents results from a major community survey about sharing and management of physical samples that was conducted as part of the RCN.
MoonDB: Restoration & Synthesis of Planetary Geochemical DataKerstin Lehnert
This presentation explains the MoonDB project that will restore and synthesize geochemical and petrological data acquired on lunar samples over more than 4 decades. The project is a collaboration between the IEDA data facility (http://www.iedadata.org) at the Lamont-Doherty Earth Observatory of Columbia University and the Astromaterials Acquisition and Curation Office (AACO) at Johnson Space Center (JSC).
This presentation was part of a workshop of IEDA (http://www.iedadata.org) at the AGU (American Geophysical Union) Fall Meeting 2013 in San Francisco that was intended as an introduction to the topic of data publication.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
Toxic effects of heavy metals : Lead and Arsenicsanjana502982
Heavy metals are naturally occuring metallic chemical elements that have relatively high density, and are toxic at even low concentrations. All toxic metals are termed as heavy metals irrespective of their atomic mass and density, eg. arsenic, lead, mercury, cadmium, thallium, chromium, etc.
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
The ability to recreate computational results with minimal effort and actionable metrics provides a solid foundation for scientific research and software development. When people can replicate an analysis at the touch of a button using open-source software, open data, and methods to assess and compare proposals, it significantly eases verification of results, engagement with a diverse range of contributors, and progress. However, we have yet to fully achieve this; there are still many sociotechnical frictions.
Inspired by David Donoho's vision, this talk aims to revisit the three crucial pillars of frictionless reproducibility (data sharing, code sharing, and competitive challenges) with the perspective of deep software variability.
Our observation is that multiple layers — hardware, operating systems, third-party libraries, software versions, input data, compile-time options, and parameters — are subject to variability that exacerbates frictions but is also essential for achieving robust, generalizable results and fostering innovation. I will first review the literature, providing evidence of how the complex variability interactions across these layers affect qualitative and quantitative software properties, thereby complicating the reproduction and replication of scientific studies in various fields.
I will then present some software engineering and AI techniques that can support the strategic exploration of variability spaces. These include the use of abstractions and models (e.g., feature models), sampling strategies (e.g., uniform, random), cost-effective measurements (e.g., incremental build of software configurations), and dimensionality reduction methods (e.g., transfer learning, feature selection, software debloating).
I will finally argue that deep variability is both the problem and solution of frictionless reproducibility, calling the software science community to develop new methods and tools to manage variability and foster reproducibility in software systems.
Exposé invité Journées Nationales du GDR GPL 2024
3. Why re-use samples?
Samples are often expensive to collect (drilling, remote
locations).
Many samples are unique and irreplaceable.
Re-analysis augments utility of existing data.
4/13/2015 SPM1.36
4. Internet of Samples: How?
Access to digital representations of physical samples
Use of resolvable unique & persistent identifiers
IGSN: Architecture & governance for sample PIDs
SESAR (System for Earth Sample Registration): Software tools for
sample registration, sample metadata management, etc.
DESC: Shared cyberinfrastructure for sample and collection
management
iSamples, CODATATG, Belmont Forum, etc.: Communities of Practice
4/16/2015 EGU Shortcourse4
5. SESAR (www.geosamples.org)
System for Earth Sample Registration
AllocatingAgent for individual investigators, sample repositories,
and science programs
tools and services for users to catalog and manage sample
metadata (MySESAR)
personal (authenticated) workspace
metadata template creator
label creation & printing (including QR code)
transfer of sample ownership
web services for client systems
register sample metadata & obtain IGSNs
access to IGSN metadata
preservation & persistent access of sample metadata
Global Sample Catalog (harvest metadata from other AAs
4/16/2015 EGU Shortcourse5
6. IGSN RegistrationWorkflow
4/16/2015 EGU Shortcourse6
SESAR
IGSN eV
Registry
Sample Label
• Sample Name
• Location
• Sample type
• ….
User submits metadata
1
SESAR stores metadata,
creates IGSN
2
SESAR registers
IGSN with IGSN e.V.
3
IGSN e.V. confirms
uniqueness
4
IGSN:XYZ08H7JG
SESAR confirms
IGSN to user5
Use IGSN
6
Registrants
9. iSamples
Research Coordination Network to advance
access and re-use of physical samples through
use of innovative cyberinfrastructure
advance best practices, standards, & policies
for sample curation, distribution, attribution,
and citation
Plan the “Digital Environment for Sample
Curation” (DESC)
Cross-disciplinary coordination (BIO,
archeology, etc.)
International coordination
4/16/2015 EGU Shortcourse9
10. International Initiatives
CODATA Task Group “Physical Samples in the Digital Era”
http://www.codata.org/task-groups/management-of-physical-
objects
SciColl: Scientific Collections International (Consortium)
http://www.scicoll.org/
(“Samples of Planet Earth”: proposal as Belmont Forum
Action ofWP4 under development)
3/26/2015 NDS ‘Internet of Sample’10