The Global Biodiversity Informatics Outlook helps to focus effort and investment towards better understanding of life on Earth and our impacts upon it. It proposes a framework that will help harness the immense power of information technology and an open data culture to gather unprecedented evidence about biodiversity and to inform better decisions.
Open data policy in the Earth Sciences:Exploring the application across government and industry sectors
Resources for Future Generations Conference 2018
Journal: A survey on fault node detection and recovery mechanisms in wireless sensor network, S. Keerthana; J. C Miraclin Joyce Pamila, 2015 International Conference on Advanced Computing and Communication Systems,Year: 2015,Pages: 1 – 5 , IEEE Conferences
Open data policy in the Earth Sciences:Exploring the application across government and industry sectors
Resources for Future Generations Conference 2018
Journal: A survey on fault node detection and recovery mechanisms in wireless sensor network, S. Keerthana; J. C Miraclin Joyce Pamila, 2015 International Conference on Advanced Computing and Communication Systems,Year: 2015,Pages: 1 – 5 , IEEE Conferences
Biodiversity Informatics Horizons 2013 - Introduction and ScopeAlex Hardisty
Outlines the scope and expected outcomes of the conference event "Biodiversity Informatics Horizons 2013" (BIH2013) that took place in Rome, 3-6th September 2013
dkNET Webinar: Creating and Sustaining a FAIR Biomedical Data Ecosystem 10/09...dkNET
Abstract
In this presentation, Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health, will share the NIH’s vision for a modernized, integrated FAIR biomedical data ecosystem and the strategic roadmap that NIH is following to achieve this vision. Dr. Gregurick will highlight projects being implemented by team members across the NIH’s 27 institutes and centers and will ways that industry, academia, and other communities can help NIH enable a FAIR data ecosystem. Finally, she will weave in how this strategy is being leveraged to address the COVID-19 pandemic.
Presenter: Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health
dkNET Webinar Information: https://dknet.org/about/webinar
The Human Cell Atlas Data Coordination PlatformLaura Clarke
This presentation gives a brief summary of the Human Cell Atlas project and describes the data coordination platform which is being built to support it.
Scott Edmunds: GigaScience - a journal or a database? Lessons learned from th...GigaScience, BGI Hong Kong
Scott Edmunds talk at the HUPO congress in Geneva, September 6th 2011 on GigaScience - a journal or a database? Lessons learned from the Genomics Tsunami.
The BlueBRIDGE approach to collaborative researchBlue BRIDGE
Gianpaolo Coro, ISTI-CNR, at BlueBRIDGE workshop on "Data Management services to support stock assessement", held during the Annual ICES Science conference 2016
OpenTox - an open community and framework supporting predictive toxicology an...Barry Hardy
Presented at ACS Boston 2015 at a Session on the growing impact of Open Science chaired by Andy Lang and Tony Williams dedicated to the work, memory and legacy of JC Bradley and the work we carry forward!
One important goal of OpenTox is to support the development of an Open Standards-based predictive toxicology framework that provides a unified access to toxicological data and models. OpenTox supports the development of tools for the integration of data, for the generation and validation of in silico models for toxic effects, libraries for the development and integration of modelling algorithms, and scientifically sound validation and reporting routines.
The OpenTox Application Programming Interface (API) is an important open standards development for software development purposes. It provides a specification against which development of global interoperable toxicology resources by the broader community can be carried out. The use of OpenTox API-compliant web services to communicate instructions between linked resources with URI addresses supports the use of a wide variety of commands to carry out operations such as data integration, algorithm use, model building and validation. The OpenTox Framework currently includes, with its APIs, services for compounds, datasets, features, algorithms, models, ontologies, tasks, validation, reporting, investigations, studies, assays, and authentication and authorisation, which may be combined into multiple applications satisfying a variety of different user needs. As OpenTox creates a semantic web for toxicology, it should be an ideal framework for incorporating toxicology data, ontology and modelling developments, thus supporting both a mechanistic framework for toxicology and best practices in statistical analysis and computational modelling.
In this presentation I will review the recent OpenTox-based development of applications including the ToxBank data infrastructure supporting integrated analysis across biochemical, functional and omics datasets supporting the safety assessment goals of the SEURAT-1 program which aims to develop alternatives to animal testing.
Finally, I will provide an overview of the working group activities of the newly formed OpenTox Association which aim to progress the development of open source, data, standards and tools in this area.
Nci clinical genomics data sharing ncra sept 2016Warren Kibbe
Gave an update on the Cancer Research Data Ecosystem, the Genomic Data Commons, Cloud Pilots, incentives for data sharing in cancer research to the NCI Council of Research Advocates (NCRA) on Monday, September 26th, 2016
Presented at the:
Canadian Aviation Safety Collaboration Forum
International Civil Aviation Organization (ICAO)
Montreal, QC
January 23, 2019
This presentation was made in real-time while attending the Forum. The objective was to observe and listen, and share some examples outside of this community that may provide insight about data sharing models with a focus on governance.
CINECA webinar slides: Making cohort data FAIRCINECAProject
Cohort studies, which recruit groups of individuals who share common characteristics and follow them over a period of time, are a robust and essential method in biomedical research for understanding the links between risk factors and diseases. Through questionnaires, medical assessments, and other interactions, voluminous and complex data are collected about the study participants. While cohort studies present a treasure trove of data, the data is often not FAIR (findable, accessible, interoperable and reusable). First, due to the sensitive and private nature of medical information, cohort data are often access controlled. Due to the lack of information about the studies (metadata), often one needs to dig deep to know what data is available in a cohort study. Therefore, many cohort datasets suffer from the findable and accessible issues. Second, often data collection is performed with instruments and data specifications tailored to the study. As a result, combining data across cohorts, even ones with similar characteristics, is difficult, making interoperability and reusability a challenge. In this presentation, we will explore several informatics techniques, such as the use of ontology, to make cohort data more FAIR. We will also consider the implications of making cohort data more open and the ethical and governance issues associated with open science benefit sharing.
This webinar is part of the “How FAIR are you” webinar series and hackathon, which aim at increasing and facilitating the uptake of FAIR approaches into software, training materials and cohort data, to facilitate responsible and ethical data and resource sharing and implementation of federated applications for data analysis.
The CINECA webinar series aims to discuss ways to address common challenges and share best practices in the field of cohort data analysis, as well as distribute CINECA project results. All CINECA webinars include an audience Q&A session during which attendees can ask questions and make suggestions. Please note that all webinars are recorded and available for posterior viewing. CINECA webinars include an audience Q&A session during which attendees can ask questions and make suggestions.
This webinar took place on 17th February 2021 and is part of the CINECA webinar series.
For previous and upcoming CINECA webinars see:
https://www.cineca-project.eu/webinars
Biodiversity Informatics Horizons 2013 - Introduction and ScopeAlex Hardisty
Outlines the scope and expected outcomes of the conference event "Biodiversity Informatics Horizons 2013" (BIH2013) that took place in Rome, 3-6th September 2013
dkNET Webinar: Creating and Sustaining a FAIR Biomedical Data Ecosystem 10/09...dkNET
Abstract
In this presentation, Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health, will share the NIH’s vision for a modernized, integrated FAIR biomedical data ecosystem and the strategic roadmap that NIH is following to achieve this vision. Dr. Gregurick will highlight projects being implemented by team members across the NIH’s 27 institutes and centers and will ways that industry, academia, and other communities can help NIH enable a FAIR data ecosystem. Finally, she will weave in how this strategy is being leveraged to address the COVID-19 pandemic.
Presenter: Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health
dkNET Webinar Information: https://dknet.org/about/webinar
The Human Cell Atlas Data Coordination PlatformLaura Clarke
This presentation gives a brief summary of the Human Cell Atlas project and describes the data coordination platform which is being built to support it.
Scott Edmunds: GigaScience - a journal or a database? Lessons learned from th...GigaScience, BGI Hong Kong
Scott Edmunds talk at the HUPO congress in Geneva, September 6th 2011 on GigaScience - a journal or a database? Lessons learned from the Genomics Tsunami.
The BlueBRIDGE approach to collaborative researchBlue BRIDGE
Gianpaolo Coro, ISTI-CNR, at BlueBRIDGE workshop on "Data Management services to support stock assessement", held during the Annual ICES Science conference 2016
OpenTox - an open community and framework supporting predictive toxicology an...Barry Hardy
Presented at ACS Boston 2015 at a Session on the growing impact of Open Science chaired by Andy Lang and Tony Williams dedicated to the work, memory and legacy of JC Bradley and the work we carry forward!
One important goal of OpenTox is to support the development of an Open Standards-based predictive toxicology framework that provides a unified access to toxicological data and models. OpenTox supports the development of tools for the integration of data, for the generation and validation of in silico models for toxic effects, libraries for the development and integration of modelling algorithms, and scientifically sound validation and reporting routines.
The OpenTox Application Programming Interface (API) is an important open standards development for software development purposes. It provides a specification against which development of global interoperable toxicology resources by the broader community can be carried out. The use of OpenTox API-compliant web services to communicate instructions between linked resources with URI addresses supports the use of a wide variety of commands to carry out operations such as data integration, algorithm use, model building and validation. The OpenTox Framework currently includes, with its APIs, services for compounds, datasets, features, algorithms, models, ontologies, tasks, validation, reporting, investigations, studies, assays, and authentication and authorisation, which may be combined into multiple applications satisfying a variety of different user needs. As OpenTox creates a semantic web for toxicology, it should be an ideal framework for incorporating toxicology data, ontology and modelling developments, thus supporting both a mechanistic framework for toxicology and best practices in statistical analysis and computational modelling.
In this presentation I will review the recent OpenTox-based development of applications including the ToxBank data infrastructure supporting integrated analysis across biochemical, functional and omics datasets supporting the safety assessment goals of the SEURAT-1 program which aims to develop alternatives to animal testing.
Finally, I will provide an overview of the working group activities of the newly formed OpenTox Association which aim to progress the development of open source, data, standards and tools in this area.
Nci clinical genomics data sharing ncra sept 2016Warren Kibbe
Gave an update on the Cancer Research Data Ecosystem, the Genomic Data Commons, Cloud Pilots, incentives for data sharing in cancer research to the NCI Council of Research Advocates (NCRA) on Monday, September 26th, 2016
Presented at the:
Canadian Aviation Safety Collaboration Forum
International Civil Aviation Organization (ICAO)
Montreal, QC
January 23, 2019
This presentation was made in real-time while attending the Forum. The objective was to observe and listen, and share some examples outside of this community that may provide insight about data sharing models with a focus on governance.
CINECA webinar slides: Making cohort data FAIRCINECAProject
Cohort studies, which recruit groups of individuals who share common characteristics and follow them over a period of time, are a robust and essential method in biomedical research for understanding the links between risk factors and diseases. Through questionnaires, medical assessments, and other interactions, voluminous and complex data are collected about the study participants. While cohort studies present a treasure trove of data, the data is often not FAIR (findable, accessible, interoperable and reusable). First, due to the sensitive and private nature of medical information, cohort data are often access controlled. Due to the lack of information about the studies (metadata), often one needs to dig deep to know what data is available in a cohort study. Therefore, many cohort datasets suffer from the findable and accessible issues. Second, often data collection is performed with instruments and data specifications tailored to the study. As a result, combining data across cohorts, even ones with similar characteristics, is difficult, making interoperability and reusability a challenge. In this presentation, we will explore several informatics techniques, such as the use of ontology, to make cohort data more FAIR. We will also consider the implications of making cohort data more open and the ethical and governance issues associated with open science benefit sharing.
This webinar is part of the “How FAIR are you” webinar series and hackathon, which aim at increasing and facilitating the uptake of FAIR approaches into software, training materials and cohort data, to facilitate responsible and ethical data and resource sharing and implementation of federated applications for data analysis.
The CINECA webinar series aims to discuss ways to address common challenges and share best practices in the field of cohort data analysis, as well as distribute CINECA project results. All CINECA webinars include an audience Q&A session during which attendees can ask questions and make suggestions. Please note that all webinars are recorded and available for posterior viewing. CINECA webinars include an audience Q&A session during which attendees can ask questions and make suggestions.
This webinar took place on 17th February 2021 and is part of the CINECA webinar series.
For previous and upcoming CINECA webinars see:
https://www.cineca-project.eu/webinars
Reagan Moore, UNC-RENCI; Policy-based Data Management; RDAP11 Summit
The 2nd Research Data Access and Preservation (RDAP) Summit
An ASIS&T Summit
March 31-April 1, 2011 Denver, CO
In cooperation with the Coalition for Networked Information
http://asist.org/Conferences/RDAP11/index.html
Being FAIR: Enabling Reproducible Data ScienceCarole Goble
Talk presented at Early Detection of Cancer Conference, OHSU, Portland, Oregon USA, 2-4 Oct 2018, http://earlydetectionresearch.com/ in the Data Science session
Similar to Global Biodiversity Informatics Outlook (20)
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Enhancing Performance with Globus and the Science DMZGlobus
ESnet has led the way in helping national facilities—and many other institutions in the research community—configure Science DMZs and troubleshoot network issues to maximize data transfer performance. In this talk we will present a summary of approaches and tips for getting the most out of your network infrastructure using Globus Connect Server.
2. Focus Area: Culture
Making data sharing the norm
Ensuring data can be understood and used across systems
and across disciplines
Creating a stable data archiving infrastructure to ensure
no data are lost or mislaid
Creating a policy framework to encourage sharing and
reuse of biodiversity data
Benefiting from the expertise of the whole global
community
3. Focus Area: Data
Using data mining and semantic tools to turn
unstructured data into information
Accelerating the rate at which historic specimen-
based data are made discoverable
Making field data accessible and interoperable
Incorporating data arising from genomics
Harnessing automated monitoring technology to
provide planet-wide surveys
4. Focus Area: Evidence
Creating a network of expertise to manage
biodiversity data
Providing a stable and comprehensive catalogue of
all species
Making accessible all data about when and where an
organism has been recorded
Providing the framework to capture trait information
and interactions between species
Delivering access to all published biodiversity
knowledge
5. Focus Area: Understanding
Estimating biodiversity patterns from available
evidence
Using predictive modelling to assess status, trends
and impacts
Building virtual models from molecules to ecosystems
Giving access to information in ways that will
revolutionize how we understand biodiversity
Concentrating on the areas of greatest change and
uncertainty
The GBIO FrameworkThe framework includes four focus areas, each with a set of related components. Each focus area – and the individual components – can be progressed independently but investments in each area will complement and reinforce each other, delivering wide-reaching benefits to many stakeholders. Culture: This area focuses on practices and infrastructure for sharing data. Investments here will ensure that digital biodiversity data benefit the whole global community and that these data can be preserved and improved for future use. Data: This area addresses the need to transform historical biodiversity knowledge into usable digital forms and to ensure that all new measurements and observations automatically contribute to a global knowledge resource. Evidence: This area deals with organizing and assessing data from all historical and new sources to provide clear, consistent and comprehensive views of everything that has been recorded about any species or ecosystem. Understanding: This area builds models from recorded measurements and observations to support data-driven research and evidence-based planning. Key deliverables include assessments and indicators of the state of biodiversity.
Focus area A: CulturePutting the foundations in place to make biodiversity data an openly shared, freely available, connected resource. Components:• A1. Open access and reuse culture: make open sharing of data standard practice through public funding and other incentives and through proper attribution and recognition of primary data resources, data creators and curators, including individuals as well as institutions.• A2. Data standards: deliver a flexible set of data standards that support the reuse and interoperability of all biodiversity data.• A3. Persistent storage and archives: provide a distributed network of data repositories for all types of biodiversity data, along with consistent handling of metadata, identifiers, licences, tools and services.• A4. Policy incentives: ensure that public policies, legislation and funding initiatives at all scales combine to reinforce this strategy and support its individual components.• A5. Biodiversity knowledge network: create the technical infrastructure to support curation and annotation of data using the best-available community expertise, in a way that makes such curation immediately visible to future users as well as providing feedback to data holders. Progress: With the exception of data standards, where there has been a long-standing community effort to develop common vocabularies and structures, progress at the global scale has generally been limited in this area although individual countries have made significant advances towards making scientific data openly accessible.
Focus area B: DataMobilizing biodiversity data from all sources and organizing it in forms that can support large-scale analysis and modelling. Components:• B1. Published materials: developing mechanisms to extract the biodiversity data currently embedded in publications and other, multimedia formats and to provide them as freely available, standardized and structured information.• B2. Collections and specimens: developing and sharing more efficient techniques to accelerate the efforts to digitize and capture historic data from collections.• B3. Field surveys and observations: capturing all biodiversity observations, including sounds and images, and making them available as soon as they are made, or within a defined period.• B4. Sequences and genomes: capturing all relevant data from genomic activity, including vouchered reference sequences, environmental metagenomics, genetic variation and full genomes.• B5. Automated and remote-sensed observations: exploiting opportunities for automated and semiautomated recording and identification of species and populations from sources ranging from satellite images down to automated gene sequencing. Progress: This focus area is probably the most advanced of the four, with the community already undertaking significant digitization projects and most new information now automatically held in digital form. However, such is the scale of the task that future projects will need to use more automation and algorithmic techniques. Moreover, digitized information of all kinds needs to be standardized and given a structure that enables it to be automatically processed: the importance of this task should not be underestimated.
Focus area C: EvidenceProviding the tools to support consistent and comprehensive global discovery and use of data from all sources about the biodiversity of any defined area over time, covering all taxonomic groups. Components:• C1. Fitness-for-use and annotation: an efficient mechanism to enable amateurs, experts and automated tools to correct and annotate individual data elements to improve quality and their fitness to be used for particular purposes, and to ensure that these annotations have to be made only once.• C2. Taxonomic framework: a comprehensive, expert curated catalogue of species, including data on names, classification and phylogeny (evolutionary relatedness) and incorporating taxa lacking formal names.• C3. Integrated occurrence data: bringing together data from all sources to document the known occurrences of all species in time and space.• C4. Aggregated species trait data: providing the tools to bring together all available data on species traits and interactions and ensure it is held in forms suitable for use in digital analysis and modelling.• C5. Comprehensive knowledge access: making all published biodiversity knowledge linked and accessible through the rich indexing of biodiversity literature, data, multimedia and other resources, including presentation of the information as species pages and via web services. Progress: Some of this work is already on course and significant progress has been made, although further investment will be needed to complete the process. International collaboration has delivered key components for building the taxonomic framework (C2) and work on integrated occurrence data (C3) is already mature and operating at a global scale. Mechanisms for recording fitness-for-use and annotations (C1) are starting to be developed in pilot while many of the individual pieces needed for comprehensive knowledge access (C5) have been trialled in different projects, but not brought together in one coherent framework. Standards for species trait data and interactions are starting to be adopted but there is as yet no common infrastructure in place to capture and query such data in a consistent way.
Focus area D: UnderstandingUsing the combined biodiversity data from multiple sources to generate new information, inform policy and decision makers, and help educate wider society to improve the way we manage the Earth’s resources. Components:• D1. Multiscale spatial modelling: Integrate data collected across disciplines and combine them with remote-sensed and geographic information systems (GIS) datasets to create the fullest-possible picture of geographical species distributions.• D2. Trends and predictions: Integrate historical data and changes over time to create predictive modeling tools to support decision making, make biodiversity estimates and predict the potential impact of changing conditions anywhere on Earth.• D3. Modelling biological systems: Build virtual models – from the molecular level to whole ecosystems – to improve understanding of biological systems and integrate that knowledge into other models.• D4. Visualization and dissemination: Provide the tools to make biodiversity information accessible and understandable by diverse audiences, increasing biodiversity literacy among the public and policy makers.• D5. Prioritizing new data capture: Use accumulating data to identify and prioritize new opportunities for data capture and to provide a timely response to changes in biodiversity patterns. Progress: This is the least developed of the four focus areas, and one, modeling biological systems (D3) has not progressed much beyond the conceptual stage. The biodiversity informatics community will have to draw heavily on the knowledge of other disciplines — from climate modelling and socio-economics, to games development and web design.
The GBIO is intended to be a dynamic and interactive process. The website at www.biodiversityinformatics.org will be updated with community input on projects, solutions and funding sources relevant to each component. Please visit this website and share your ideas and expertise.