EUDAT Collaborative Data Infrastructure: Data Access and Re-use Service AreaEUDAT
The document summarizes services provided by EUDAT, a collaborative data infrastructure funded by the European Union. It describes EUDAT's B2ACCESS identity and access management service, which allows users to access EUDAT and other services using their existing credentials. It also summarizes EUDAT's B2FIND data discovery service, B2SHARE data sharing and preservation service, and B2DROP file sharing service. It outlines their key features and integration with each other. Future plans include further deployments, improved interoperability, and enhanced user experiences across EUDAT services.
Aggregation of cultural heritage datasets through the Web of DataNuno Freire
The existence of many digital libraries, maintained by different organizations, brings challenges to the discoverability of cultural heritage (CH) resources. Metadata aggregation is an approach where centralized efforts like Europeana facilitate their discoverability by collecting the resource’s metadata. Nowadays, CH institutions are increasingly applying technologies designed for the wider interoperability on the Web. In this context, we have identified the Schema.org vocabulary and linked data (LD) as potential technologies for innovating CH metadata aggregation. We present the results of an analysis using the case of the Europeana network of aggregators and data providers as basis. We have conducted a survey of the available linked data technology, and we defined a solution, which we have put into practice in a pilot implementation within the Europeana network. In this pilot, the National Library of The Netherlands fulfils the role of data provider, with the Dutch Digital Heritage Network, as national aggregator, supporting the provision of several datasets from the national library to Europeana. The metadata is published using LD practices, having Schema.org as the main vocabulary. The national library also implements all the necessary semantic web mechanisms, defined in our solution, for making the datasets discoverable and harvestable by Europeana. Our proposal involves the use of vocabularies for description of datasets, and their distributions, namely DCAT, VoID and Schema.org. Europeana implements the LD harvester side of the solution and applies it to harvest the Schema.org data from the national library.
EUDAT & OpenAIRE Webinar: How to write a Data Management Plan - July 14, 2016...EUDAT
| www.eudat.eu | 2nd Session: July 14, 2016.
In this webinar, Sarah Jones (DCC) and Marjan Grootveld (DANS) talked through the aspects that Horizon 2020 requires from a DMP. They discussed examples from real DMPs and also touched upon the Software Management Plan, which for some projects can be a sensible addition
The document discusses the EUDAT project's Data Preservation Service Area and its achievements and future plans. It summarizes that initially the service lacked metadata management, clear workflows, high-level data policies, and integration with EUDAT's Common Data Infrastructure (CDI). It is now producing features like the Data Policy Manager (DPM) for high-level data policies, improved authentication, and better documentation. Planned future work includes extending DPM's data policies, integrating a centralized PID registry, and improving metadata management, knowledge management, and interoperability. The objective is to improve DPM metadata, the PID catalog, and support FAIR data principles.
Exposing EO Linked (meta-)Data from OpenSearch CatalogueRaul Palma
This document discusses exposing Earth observation (EO) linked (meta-)data from OpenSearch catalogues. It provides background on linked data principles and publishing EO data as linked data. It proposes an approach to generate linked data from EO product metadata by implementing wrappers around APIs to transform requests and results into RDF in real-time. This allows querying REST APIs with SPARQL and exposing the results through a SPARQL endpoint without needing to store the data physically. The FedEO system provides a specific use case, federating access to multiple EO catalogues through its OpenSearch interface.
Open Data management is still not trivial nor sustainable - COMSODE results are here to bring automation to publication and management of Open Data in public institutions and companies. Presentation includes Open Data Ready standard proposal, three use cases and invitation for Horizon 2020 projects 2016.
NordForsk Open Access Reykjavik 14-15/8-2014:RdaNordForsk
The Research Data Alliance provides opportunities for global collaboration on data-related issues. It grew from the need to connect research computers and share data openly across technologies and borders. RDA works through Working and Interest Groups to develop standards and best practices around topics like data citation and metadata. Recent outputs include recommendations for data type registries and persistent identifier information types. RDA membership includes over 1,900 individuals from 83 countries and represents academia, government, and industry.
Aggregation of Linked Data A case study in the cultural heritage domainNuno Freire
Presented at IEEE BIGDATA 2018 conference - December 2018
A very large number of online cultural heritage (CH) resources is made available through numerous digital libraries. To address the difficulties of discoverability in CH, the common practice is metadata aggregation, where centralized efforts like Europeana facilitate discoverability by collecting the resources’ metadata. In the last years, the CH domain has invested in data models for Linked Data (LD) representation of CH metadata. LD, however, also has potential for innovating metadata aggregation. We present the results of a pilot case study within the Europeana Network. In this pilot, the National Library of The Netherlands plays the role of initial data provider, with the Dutch Digital Heritage Network the one of intermediary service providing datasets to Europeana. We analysed the requirements for an LD aggregation solution and defined a workflow that fulfils the same functional requirements as Europeana’s current solution. The workflow was put into practice within the pilot and led to the development of several software components for managing datasets, harvesting LD, data analysis and integration. Our analysis of the experience discusses the effort of adopting such an LD approach for data providers and aggregators, the expertise required by CH data analysts, and the supporting tools required for semantic data.
EUDAT Collaborative Data Infrastructure: Data Access and Re-use Service AreaEUDAT
The document summarizes services provided by EUDAT, a collaborative data infrastructure funded by the European Union. It describes EUDAT's B2ACCESS identity and access management service, which allows users to access EUDAT and other services using their existing credentials. It also summarizes EUDAT's B2FIND data discovery service, B2SHARE data sharing and preservation service, and B2DROP file sharing service. It outlines their key features and integration with each other. Future plans include further deployments, improved interoperability, and enhanced user experiences across EUDAT services.
Aggregation of cultural heritage datasets through the Web of DataNuno Freire
The existence of many digital libraries, maintained by different organizations, brings challenges to the discoverability of cultural heritage (CH) resources. Metadata aggregation is an approach where centralized efforts like Europeana facilitate their discoverability by collecting the resource’s metadata. Nowadays, CH institutions are increasingly applying technologies designed for the wider interoperability on the Web. In this context, we have identified the Schema.org vocabulary and linked data (LD) as potential technologies for innovating CH metadata aggregation. We present the results of an analysis using the case of the Europeana network of aggregators and data providers as basis. We have conducted a survey of the available linked data technology, and we defined a solution, which we have put into practice in a pilot implementation within the Europeana network. In this pilot, the National Library of The Netherlands fulfils the role of data provider, with the Dutch Digital Heritage Network, as national aggregator, supporting the provision of several datasets from the national library to Europeana. The metadata is published using LD practices, having Schema.org as the main vocabulary. The national library also implements all the necessary semantic web mechanisms, defined in our solution, for making the datasets discoverable and harvestable by Europeana. Our proposal involves the use of vocabularies for description of datasets, and their distributions, namely DCAT, VoID and Schema.org. Europeana implements the LD harvester side of the solution and applies it to harvest the Schema.org data from the national library.
EUDAT & OpenAIRE Webinar: How to write a Data Management Plan - July 14, 2016...EUDAT
| www.eudat.eu | 2nd Session: July 14, 2016.
In this webinar, Sarah Jones (DCC) and Marjan Grootveld (DANS) talked through the aspects that Horizon 2020 requires from a DMP. They discussed examples from real DMPs and also touched upon the Software Management Plan, which for some projects can be a sensible addition
The document discusses the EUDAT project's Data Preservation Service Area and its achievements and future plans. It summarizes that initially the service lacked metadata management, clear workflows, high-level data policies, and integration with EUDAT's Common Data Infrastructure (CDI). It is now producing features like the Data Policy Manager (DPM) for high-level data policies, improved authentication, and better documentation. Planned future work includes extending DPM's data policies, integrating a centralized PID registry, and improving metadata management, knowledge management, and interoperability. The objective is to improve DPM metadata, the PID catalog, and support FAIR data principles.
Exposing EO Linked (meta-)Data from OpenSearch CatalogueRaul Palma
This document discusses exposing Earth observation (EO) linked (meta-)data from OpenSearch catalogues. It provides background on linked data principles and publishing EO data as linked data. It proposes an approach to generate linked data from EO product metadata by implementing wrappers around APIs to transform requests and results into RDF in real-time. This allows querying REST APIs with SPARQL and exposing the results through a SPARQL endpoint without needing to store the data physically. The FedEO system provides a specific use case, federating access to multiple EO catalogues through its OpenSearch interface.
Open Data management is still not trivial nor sustainable - COMSODE results are here to bring automation to publication and management of Open Data in public institutions and companies. Presentation includes Open Data Ready standard proposal, three use cases and invitation for Horizon 2020 projects 2016.
NordForsk Open Access Reykjavik 14-15/8-2014:RdaNordForsk
The Research Data Alliance provides opportunities for global collaboration on data-related issues. It grew from the need to connect research computers and share data openly across technologies and borders. RDA works through Working and Interest Groups to develop standards and best practices around topics like data citation and metadata. Recent outputs include recommendations for data type registries and persistent identifier information types. RDA membership includes over 1,900 individuals from 83 countries and represents academia, government, and industry.
Aggregation of Linked Data A case study in the cultural heritage domainNuno Freire
Presented at IEEE BIGDATA 2018 conference - December 2018
A very large number of online cultural heritage (CH) resources is made available through numerous digital libraries. To address the difficulties of discoverability in CH, the common practice is metadata aggregation, where centralized efforts like Europeana facilitate discoverability by collecting the resources’ metadata. In the last years, the CH domain has invested in data models for Linked Data (LD) representation of CH metadata. LD, however, also has potential for innovating metadata aggregation. We present the results of a pilot case study within the Europeana Network. In this pilot, the National Library of The Netherlands plays the role of initial data provider, with the Dutch Digital Heritage Network the one of intermediary service providing datasets to Europeana. We analysed the requirements for an LD aggregation solution and defined a workflow that fulfils the same functional requirements as Europeana’s current solution. The workflow was put into practice within the pilot and led to the development of several software components for managing datasets, harvesting LD, data analysis and integration. Our analysis of the experience discusses the effort of adopting such an LD approach for data providers and aggregators, the expertise required by CH data analysts, and the supporting tools required for semantic data.
B2SHARE: Record lifecycle and HTTP API| www.eudat.eu | EUDAT
| www.eudat.eu | B2SHARE is a scientific data repository providing persistent storage and sharing data facilities. Building on the new Invenio 3.0 digital assets management platform, a new version of B2SHARE has been developed which is focused on an improved user experience. Answering the requests of the current user base, B2SHARE version 2 provides customizable metadata schemas and a simple but effective workflow for depositing user data, exposed in its RESTful HTTP API.
The presentation will introduce the B2SHARE service, its organizing principles and its basic operations. The metadata schemas and the dataset lifecycle, which are essentials in understanding the possibilities of the service, will be the main focus of the talk. The concrete output of the session can be a full paper expanding the presented topics.
Target Audience:Researchers of any scientific domain, which work with publishable data sets.
Research engagement in EUDAT| www.eudat.eu | EUDAT
| www.eudat.eu | EUDAT’s vision is to enable European researchers and practitioners from any research discipline to preserve, find, access, and process data in a trusted environment, as part of a Collaborative Data Infrastructure (CDI) conceived as a network of collaborating, cooperating centres, that combine community-specific data repositories with the permanence and persistence of some of Europe’s largest scientific data centres. EUDAT services are community driven solutions. This presentation describes the different ways EUDAT engages with the research communities
The document provides an introduction to PREMIS (Preservation Metadata: Implementation Strategies) and its application in audiovisual archives. It discusses the challenges of digital preservation and the need for preservation metadata to ensure long-term access. It then summarizes the key aspects of PREMIS, including the PREMIS Data Dictionary, its relationship to the OAIS reference model, the five interacting entities in the PREMIS data model, and issues around implementing PREMIS in archives.
PoolParty Semantic Suite - Solutions for Sustainable Development - weadapt.or...Semantic Web Company
weADAPT is a free online platform containing over 700 georeferenced climate adaptation case studies, 500 organizations, and 1000 articles that are semantically tagged and linked using Climate Tagger to allow for intelligent searching and connections between related information. Future steps include further linking cases and content between platforms, adding location data to articles, and developing a common knowledge pool with smart visualization and monitoring capabilities.
This document discusses the development of preservation workflows and policies for an audiovisual archive to help ensure the long-term accessibility and usability of its digital collections. It outlines key concepts for digital preservation such as integrity, authenticity and provenance metadata. It also describes the archive's OAIS-compliant processes for ingesting, storing and disseminating archival objects and the importance of preservation metadata. The project has helped establish foundations for digital lifecycle management at the archive.
Coupling HPC and Data Resources and services together - EUDAT Workshop at exd...EUDAT
Giuseppe Fiameni (CINECA)
The goal of this EUDAT workshop is to present the EUDAT services, the results of the collaboration activity achieved so far and deliver a hands-on on how to write a Data Management Plan or DMP. The DMP is a useful instrument for researchers to reflect on and communicate about the way they will deal with their data as it prompts them to think about how they will generate, analyse and share data during their research project and afterwards.
SafeNet is a 2-year Jisc funded project that aims to improve assured continuity of access to e-journal content paid for by UK higher education institutions. It has two main components: an entitlement registry to hold subscription data for institutions, and a private LOCKSS network to provide assurance that paid-for content will remain available. The presentation provides details on the core service components and governance structure. It also outlines the project's goal of data gathering - to empower libraries by minimizing duplication of work and focusing on what content libraries believe they are entitled to access. Assistance is being sought from several universities on exploring the range of subscriptions, challenges, and data available to make assertions for an entitlement registry.
EUDAT Research Data Management | www.eudat.eu | EUDAT
| www.eudat.eu | The presentation gives an introduction to Research Data Management, explaining why it is important to manage and share data.
November 2016
| www.eudat.eu | B2FIND - User training Version 07, June 2017: B2FIND is EUDAT’s simple, user friendly metadata catalogue allowing users to discover metadata from a wide range of scientific communities.
EUDAT-B2FIND is the cross-disciplinary discovery service of EUDAT that allows facetted search over distributed research data repositories. It indexes metadata from different research domains and memory areas within EOSC-hub and beyond. B2FIND follows a flexible approach for data providers, supporting various metadata schemas, harvesting methods, and formats through its ingestion workflow.
Interoperability is the key: repositories networks promoting the quality and ...Pedro Príncipe
Presentation from José Carvalho and Pedro Principe, University of Minho, at ETD 2019 Conference (22nd International Symposium on Electronic Theses and Dissertations), Porto, Nov 7, 2019.
How EUDAT services support FAIR data - IDCC 2017| www.eudat.eu | EUDAT
| www.eudat.eu | Welcome Overview of the EUDAT service suite and the FAIR principles.
Sarah Jones, Marjan Grootveld, Yann Le Franc - IDCC Conference, February 20, 2017
‘Development of a MODS-RDF Cataloguing Tool for the Digital Resources and Ima...CONUL Conference
The ADAPT Centre collaborated with Digital Resources and Imaging Services (DRIS) of Trinity College Dublin to develop a MODS-RDF cataloguing tool. The tool allows DRIS cataloguers to generate MODS and RDF metadata for digital collections in a user-friendly interface. Usability testing identified improvements and new requirements. The tool facilitates publishing library metadata as linked data on the semantic web to improve discovery and sharing of resources across institutions.
B2STAGE- how to shift large amounts of data| www.eudat.eu | EUDAT
| www.eudat.eu | B2STAGE is a reliable, efficient, light-weight and easy-to-use service to transfer research data sets between EUDAT storage resources and high-performance computing (HPC) workspaces.
‘How the partnership between the Library of Trinity College Dublin Digital Re...CONUL Conference
The ADAPT Centre is a research centre located in Ireland funded by Science Foundation Ireland and the European Regional Development Fund. It focuses on digital content technology and has 30 academics and 205 researchers across several universities. The centre partners with the Library of Trinity College Dublin to bring expertise in areas like multi-modal interfaces, content curation, and recommendation systems to improve discovery and engagement with scholarly content. Current projects include an app prototype for the digital repository and approaches to linked data publication. The partnership aims to advance research at both institutions.
II-SDV 2016 Michael Iarrobino - Improving Text Mining Results with Access to ...Dr. Haxel Consult
Life science companies increasingly rely on text mining to gain important insights from vast amounts of published information. But researchers struggle to get access to full-text articles for text mining. When they do get the full text they must contend with multiple formats and inconsistent license terms – all of which inhibit text mining efforts. In this presentation, we will describe the value in mining full-text scientific literature and outline the issues researchers face in accessing and licensing this content for commercial purposes. We will provide a walkthrough of Copyright Clearance Center’s (CCC) RightFind™ XML for Mining solution and contrast this with other approaches to solving these time-consuming content and licensing challenges. CCC is the parent organization of RightsDirect.
IPTC Semantic Web 2012 Spring Working GroupStuart Myles
The document discusses the IPTC Semantic Web Working Group's efforts to leverage semantic web technologies for representing news content. Their work focuses on developing models for the aspects of news to represent and formats for encoding the news using semantic web standards like RDFa, microformats and Linked Data. The group also provides updates on their rNews specification for adding metadata to HTML news articles and their work to publish IPTC NewsCodes and link them to other datasets according to Linked Data principles. They announce plans to further expand these linking efforts and to participate in the W3C semnews Community Group. The next meeting is scheduled for June 2012 in New York.
Denodo Platform 7.0: Redefine Analytics with In-Memory Parallel Processing an...Denodo
The document discusses modern data architectures using a logical data lake approach with data virtualization. It describes how the Denodo platform provides big data integrations by allowing Hadoop to be used as a data source, cache, and processing engine. This is demonstrated through an example showing how Denodo's cost-based optimizer and integration with SQL-on-Hadoop processing engines like Impala can optimize query performance. The document concludes that this approach can surface all company data without replication, improve governance, and leverage existing Hadoop cluster processing power controlled by Denodo's optimizer.
This document discusses resolving taxonomic discrepancies through the use of electronic catalogues of known organisms. It highlights discrepancies found between IndFauna, an electronic catalogue of known Indian fauna, and global databases like ITIS. Discrepancies include differences in taxonomic hierarchies, spelling differences, and homonymies. Electronic catalogues can help identify such issues by allowing cross-checking of taxonomic information across sources. Resolving discrepancies requires discussion with taxonomy experts and application of nomenclature rules. The development of IndFauna and comparison to other databases demonstrated its ability to resolve taxonomic discrepancies.
The document discusses the objectives of the GBIF MRTG which include surveying multimedia resource providers and making recommendations on social, technical, and long-term issues. It also mentions drafting a multimedia metadata schema for biodiversity and a global strategy for mobilizing observation, ecology, and multimedia data. Finally, it directs the reader to join the TDWG IMG mailing list to provide comments on a draft multimedia metadata schema.
B2SHARE: Record lifecycle and HTTP API| www.eudat.eu | EUDAT
| www.eudat.eu | B2SHARE is a scientific data repository providing persistent storage and sharing data facilities. Building on the new Invenio 3.0 digital assets management platform, a new version of B2SHARE has been developed which is focused on an improved user experience. Answering the requests of the current user base, B2SHARE version 2 provides customizable metadata schemas and a simple but effective workflow for depositing user data, exposed in its RESTful HTTP API.
The presentation will introduce the B2SHARE service, its organizing principles and its basic operations. The metadata schemas and the dataset lifecycle, which are essentials in understanding the possibilities of the service, will be the main focus of the talk. The concrete output of the session can be a full paper expanding the presented topics.
Target Audience:Researchers of any scientific domain, which work with publishable data sets.
Research engagement in EUDAT| www.eudat.eu | EUDAT
| www.eudat.eu | EUDAT’s vision is to enable European researchers and practitioners from any research discipline to preserve, find, access, and process data in a trusted environment, as part of a Collaborative Data Infrastructure (CDI) conceived as a network of collaborating, cooperating centres, that combine community-specific data repositories with the permanence and persistence of some of Europe’s largest scientific data centres. EUDAT services are community driven solutions. This presentation describes the different ways EUDAT engages with the research communities
The document provides an introduction to PREMIS (Preservation Metadata: Implementation Strategies) and its application in audiovisual archives. It discusses the challenges of digital preservation and the need for preservation metadata to ensure long-term access. It then summarizes the key aspects of PREMIS, including the PREMIS Data Dictionary, its relationship to the OAIS reference model, the five interacting entities in the PREMIS data model, and issues around implementing PREMIS in archives.
PoolParty Semantic Suite - Solutions for Sustainable Development - weadapt.or...Semantic Web Company
weADAPT is a free online platform containing over 700 georeferenced climate adaptation case studies, 500 organizations, and 1000 articles that are semantically tagged and linked using Climate Tagger to allow for intelligent searching and connections between related information. Future steps include further linking cases and content between platforms, adding location data to articles, and developing a common knowledge pool with smart visualization and monitoring capabilities.
This document discusses the development of preservation workflows and policies for an audiovisual archive to help ensure the long-term accessibility and usability of its digital collections. It outlines key concepts for digital preservation such as integrity, authenticity and provenance metadata. It also describes the archive's OAIS-compliant processes for ingesting, storing and disseminating archival objects and the importance of preservation metadata. The project has helped establish foundations for digital lifecycle management at the archive.
Coupling HPC and Data Resources and services together - EUDAT Workshop at exd...EUDAT
Giuseppe Fiameni (CINECA)
The goal of this EUDAT workshop is to present the EUDAT services, the results of the collaboration activity achieved so far and deliver a hands-on on how to write a Data Management Plan or DMP. The DMP is a useful instrument for researchers to reflect on and communicate about the way they will deal with their data as it prompts them to think about how they will generate, analyse and share data during their research project and afterwards.
SafeNet is a 2-year Jisc funded project that aims to improve assured continuity of access to e-journal content paid for by UK higher education institutions. It has two main components: an entitlement registry to hold subscription data for institutions, and a private LOCKSS network to provide assurance that paid-for content will remain available. The presentation provides details on the core service components and governance structure. It also outlines the project's goal of data gathering - to empower libraries by minimizing duplication of work and focusing on what content libraries believe they are entitled to access. Assistance is being sought from several universities on exploring the range of subscriptions, challenges, and data available to make assertions for an entitlement registry.
EUDAT Research Data Management | www.eudat.eu | EUDAT
| www.eudat.eu | The presentation gives an introduction to Research Data Management, explaining why it is important to manage and share data.
November 2016
| www.eudat.eu | B2FIND - User training Version 07, June 2017: B2FIND is EUDAT’s simple, user friendly metadata catalogue allowing users to discover metadata from a wide range of scientific communities.
EUDAT-B2FIND is the cross-disciplinary discovery service of EUDAT that allows facetted search over distributed research data repositories. It indexes metadata from different research domains and memory areas within EOSC-hub and beyond. B2FIND follows a flexible approach for data providers, supporting various metadata schemas, harvesting methods, and formats through its ingestion workflow.
Interoperability is the key: repositories networks promoting the quality and ...Pedro Príncipe
Presentation from José Carvalho and Pedro Principe, University of Minho, at ETD 2019 Conference (22nd International Symposium on Electronic Theses and Dissertations), Porto, Nov 7, 2019.
How EUDAT services support FAIR data - IDCC 2017| www.eudat.eu | EUDAT
| www.eudat.eu | Welcome Overview of the EUDAT service suite and the FAIR principles.
Sarah Jones, Marjan Grootveld, Yann Le Franc - IDCC Conference, February 20, 2017
‘Development of a MODS-RDF Cataloguing Tool for the Digital Resources and Ima...CONUL Conference
The ADAPT Centre collaborated with Digital Resources and Imaging Services (DRIS) of Trinity College Dublin to develop a MODS-RDF cataloguing tool. The tool allows DRIS cataloguers to generate MODS and RDF metadata for digital collections in a user-friendly interface. Usability testing identified improvements and new requirements. The tool facilitates publishing library metadata as linked data on the semantic web to improve discovery and sharing of resources across institutions.
B2STAGE- how to shift large amounts of data| www.eudat.eu | EUDAT
| www.eudat.eu | B2STAGE is a reliable, efficient, light-weight and easy-to-use service to transfer research data sets between EUDAT storage resources and high-performance computing (HPC) workspaces.
‘How the partnership between the Library of Trinity College Dublin Digital Re...CONUL Conference
The ADAPT Centre is a research centre located in Ireland funded by Science Foundation Ireland and the European Regional Development Fund. It focuses on digital content technology and has 30 academics and 205 researchers across several universities. The centre partners with the Library of Trinity College Dublin to bring expertise in areas like multi-modal interfaces, content curation, and recommendation systems to improve discovery and engagement with scholarly content. Current projects include an app prototype for the digital repository and approaches to linked data publication. The partnership aims to advance research at both institutions.
II-SDV 2016 Michael Iarrobino - Improving Text Mining Results with Access to ...Dr. Haxel Consult
Life science companies increasingly rely on text mining to gain important insights from vast amounts of published information. But researchers struggle to get access to full-text articles for text mining. When they do get the full text they must contend with multiple formats and inconsistent license terms – all of which inhibit text mining efforts. In this presentation, we will describe the value in mining full-text scientific literature and outline the issues researchers face in accessing and licensing this content for commercial purposes. We will provide a walkthrough of Copyright Clearance Center’s (CCC) RightFind™ XML for Mining solution and contrast this with other approaches to solving these time-consuming content and licensing challenges. CCC is the parent organization of RightsDirect.
IPTC Semantic Web 2012 Spring Working GroupStuart Myles
The document discusses the IPTC Semantic Web Working Group's efforts to leverage semantic web technologies for representing news content. Their work focuses on developing models for the aspects of news to represent and formats for encoding the news using semantic web standards like RDFa, microformats and Linked Data. The group also provides updates on their rNews specification for adding metadata to HTML news articles and their work to publish IPTC NewsCodes and link them to other datasets according to Linked Data principles. They announce plans to further expand these linking efforts and to participate in the W3C semnews Community Group. The next meeting is scheduled for June 2012 in New York.
Denodo Platform 7.0: Redefine Analytics with In-Memory Parallel Processing an...Denodo
The document discusses modern data architectures using a logical data lake approach with data virtualization. It describes how the Denodo platform provides big data integrations by allowing Hadoop to be used as a data source, cache, and processing engine. This is demonstrated through an example showing how Denodo's cost-based optimizer and integration with SQL-on-Hadoop processing engines like Impala can optimize query performance. The document concludes that this approach can surface all company data without replication, improve governance, and leverage existing Hadoop cluster processing power controlled by Denodo's optimizer.
This document discusses resolving taxonomic discrepancies through the use of electronic catalogues of known organisms. It highlights discrepancies found between IndFauna, an electronic catalogue of known Indian fauna, and global databases like ITIS. Discrepancies include differences in taxonomic hierarchies, spelling differences, and homonymies. Electronic catalogues can help identify such issues by allowing cross-checking of taxonomic information across sources. Resolving discrepancies requires discussion with taxonomy experts and application of nomenclature rules. The development of IndFauna and comparison to other databases demonstrated its ability to resolve taxonomic discrepancies.
The document discusses the objectives of the GBIF MRTG which include surveying multimedia resource providers and making recommendations on social, technical, and long-term issues. It also mentions drafting a multimedia metadata schema for biodiversity and a global strategy for mobilizing observation, ecology, and multimedia data. Finally, it directs the reader to join the TDWG IMG mailing list to provide comments on a draft multimedia metadata schema.
The document summarizes recommendations from the GBIF GSAP-NHC Task Group on improving the digitization and publication of natural history collection data. It recommends that GBIF facilitate discovery of non-digital collection resources, increase efficiency and quality of data capture, and improve global infrastructure for publishing digitized collection data. Specifically, it calls for GBIF to publicize non-digital metadata, assess the scale of undigitized specimens, support technological innovations for digitization, and strengthen hosting and identification of published data.
The Global Biodiversity Information Facility (GBIF) is an international initiative that aims to provide open access to biodiversity data. It currently has 50 participating countries and 40 international organizations contributing data. GBIF develops standards and tools to help aggregate and publish biodiversity data from various sources. This includes species occurrences, names, taxonomic information, and more. GBIF's data portal allows users to search, browse, and access over 181 million georeferenced records. GBIF is working to integrate biodiversity data on a global scale and support areas like conservation and monitoring of biodiversity trends.
The document summarizes the development of an electronic catalogue (IndFauna) of known Indian fauna. It discusses the importance of developing electronic catalogues to consolidate information on biodiversity. IndFauna collates baseline information on over 93% of India's approximately 90,000 known faunal species. The catalogue demonstrates collaboration between domain experts and IT managers to overcome taxonomic challenges and enable sustainable management and conservation of biotic resources.
The document discusses technologies and infrastructure for publishing biodiversity data from environmental impact assessments (EIA). It covers the types and formats of EIA biodiversity data, tools for data capture and digitization, platforms for data discovery and publishing, ensuring data quality, and hosting data centers to facilitate long-term archiving and publishing of EIA biodiversity data.
The document discusses the Global Biodiversity Resources Discovery System (GBRDS) registry which aims to create a single indexed registry of biodiversity data sources. It outlines activities to develop the registry including building a flexible database, user interface, and API. The registry will integrate distributed biodiversity data and enable discovery of institutions, collections, services, and more.
Adoption of Persistent Identifiers for Biodiversity InformaticsPhil Cryer
The GBIF data portal is a focal point in the flow of biodiversity data. The feedback and data cleaning tools provided through the portal influence the quality of data being published by providers. GBIF should place the use and re-use of identifiers as a high priority in assessing the quality of data. GBIF should move to a position where it mandates the use of identifiers and well known vocabularies for all data accepted by the portal.
Citation information: http://academic.research.microsoft.com/Publication/13158741/adoption-of-persistent-identifiers-for-biodiversity-informatics
PDF Download: http://imsgbif.gbif.org/File/retrieve.php?PATH=4&FILE=2efc20187e6ad3dd828bbeadaa1040e6&FILENAME=LGTGReportDraft.pdf&TYPE=application/pdf
GBIF web services for biodiversity data, for USDA GRIN, Washington DC, USA (2...Dag Endresen
Presentation of GBIF and the sharing of biodiversity data with web services. USDA GRIN Beltsville Washington DC, 13th December 2005. GBIF is the Global Biodiversity Information Facility for free and open access to biodiversity data.
Presentation of information needs and tehcnical options for a Global Forest Information Service GFIS.
Authors: Saarikko J., R. Päivinen, T. Richards and M. Sini.
Event: Forestry Information Systems 2000 workshop, 16-20 May, Hyytiälä, Finland organized by the JOINT FAO/ECE/ILO - COMMITTEE ON FOREST TECHNOLOGY, MANAGEMENT AND TRAINING.
B2SAFE is a robust service that allows repositories to implement data management policies across administrative domains in a trustworthy manner. It offers an abstraction layer for large-scale heterogeneous storage, protects against data loss, allows optimized access, and enables compute-intensive analysis. B2SAFE is designed to execute auditable policy rules and use persistent identifiers to increase trust in data reuse by ensuring ownership rights and replicating data across sites for safekeeping.
The document summarizes the Global ARD Web RING, which aims to organize agricultural research and development information online. It does this by creating a registry and map of existing information services, gateways, and websites to facilitate access to networks of digital information. The Web RING will categorize and link featured services according to standards and specifications to promote interoperability. The overall goal is to create a network of integrated information services that meet the needs of the agricultural research community.
Presented by Tony Mathys at a Current Issues and Applications of the Geospatial Technologies Lecture, Department of Geography and Environment, Aberdeen University, 24 February 2012
In an expert webinar on April 15th 2020 we discussed (in Finnish) how the FAIR data principles affect service development in RDM services. I presented some relevant outputs from the FAIRsFAIR project. These are the slides (in English). The webinar will be published on the fairdata.fi service site https://www.fairdata.fi/koulutus/koulutuksen-tallenteet/
B2FIND is an interdisciplinary discovery portal that allows users to search for and find research data collections from various sources using faceted search options. It harvests and indexes metadata from EUDAT data centers, community repositories, and other sources using various protocols, and provides a unified search interface across data from different scientific domains and metadata standards. Search results can be filtered by location, time period, and other textual facets. Data providers can make their research data collections discoverable on B2FIND by defining mappings between their metadata and the B2FIND schema.
Inspire Helsinki 2019 - Keynote Bart De LathouwerHannaHorppila
The Inspire Helsinki 2019 event brought together around 170 people from 29 countries to foster discussion and new ideas on how to realise the full potential of spatial data. The three-day event featured data challenges, practical hands-on workshops and future-oriented keynote presentations. The event was summed up in a panel discussion, in which perspectives on tackling remaining challenges were brought up.
The document summarizes a keynote presentation by Bart De Lathouwer, President of the Open Geospatial Consortium (OGC). It provides an overview of OGC, including that it is a global consortium that develops open standards for location data and serves as a forum for communities to address interoperability issues. It discusses OGC's role in connecting people, communities, technology, and decision making through spatial data standards. It also summarizes OGC's history of developing standards like WMS, efforts to align with other organizations like W3C, and current focus on developing modular API standards like OGC API - Features to make spatial data more accessible and usable on the web.
The Inspire Helsinki 2019 event brought together around 170 people from 29 countries to foster discussion and new ideas on how to realise the full potential of spatial data. The three-day event featured data challenges, practical hands-on workshops and future-oriented keynote presentations. The event was summed up in a panel discussion, in which perspectives on tackling remaining challenges were brought up.
The document discusses the Research Data Alliance (RDA) and its process for having its technical specifications identified and approved as ICT Technical Specifications by the European Commission. The RDA works with various stakeholders to develop technical specifications that enhance data sharing and interoperability. Its specifications undergo an open review process before being submitted for identification. Previously identified RDA specifications include recommendations for data citation, repository requirements, and workflows for data publishing. The identification process involves review by the European Multi Stakeholder Platform and the European Commission to ensure specifications meet requirements for adoption in European public procurement.
Ataas2016 - Big data hadoop and map reduce - new age tools for aid to test...Agile Testing Alliance
Big Data - Hadoop and MapReduce - new age tools for aid to testing and QA BigData with its slew of technologies and terms has been the most talked about area in last couple of years. This has evolved in Big Data Science, Analytics and now on the IoT and automation side. There is a need for testers and QA team to not only get used to this new age digital transformation area but at the same time embrace the technology to their own advantage. We have experimented and successfully used Big Data Technologies – Hadoop and MapReduce for a recent testing engagement. The actual application was implemented using classic technologies like CentOS and C++. Testing team implemented Hadoop and MapReduce to help in quick turnaround for the testing.
This document provides an overview of a geospatial metadata and spatial data workshop held at the University of Oxford. The workshop covered topics such as metadata standards, application profiles, geospatial metadata tools and portals for sharing spatial data and metadata. Hands-on sessions demonstrated how to create metadata using the Geodoc Metadata Editor tool and access spatial data repositories through the Go-Geo portal and ShareGeo open data portal.
GBIF registry (GBRDS), at European Nodes meeting in Alicante, Spain (10 March...Dag Endresen
Regional NODES meeting of Europe 2010. Presentation of the Global Biodiversity Resources Discovery System (GBRDS, under development) for the NODES. How do we the NODES want the GBRDS to look like. What do we the NODES wish/need the GBRDS to be.
http://www.gbif.org/
http://gbrds.gbif.org/
http://code.google.com/p/gbif-registry/
The document summarizes a workshop on geospatial metadata and spatial data. It discusses the importance of metadata for managing and sharing spatial datasets, providing key information about the data. It also covers metadata standards like FGDC, ISO 19115, and application profiles. The workshop includes presentations on the UK Academic Geospatial Metadata Application Profile and tools for creating metadata like the Geodoc Metadata Editor and Go-Geo portal.
Slides of the AIMS webinar on the Conceptual Design of TAPipedia, introducing initial version of the Design for public feedback & comments.
http://aims.fao.org/activity/blog/new-webinarsaims%E2%80%9Cdesigning-tapipedia-information-sharing-platform-capacity-development
The document discusses establishing a National Digital Repository System (NDRS) in India using a harvesting model. It analyzes different technical models (centralized, distributed, harvesting), and recommends adopting the harvesting model. The harvesting model would involve individual institutional repositories exposing their metadata using OAI-PMH for a central searchable server to harvest and provide enhanced discovery services. Benefits of the NDRS for various stakeholders are discussed. Current scenarios of institutional repositories in India and potential organizations to contribute to the proposed NDRS are also outlined.
This document provides an overview of a geospatial metadata and spatial data workshop. It discusses the importance of metadata for discovering and managing spatial datasets. It introduces common geospatial metadata standards like FGDC, ISO 19115, and INSPIRE and the concept of application profiles. The document outlines tools and resources for UK academics to create and publish metadata, including the UK AGMAP profile, Geodoc editor, GoGeo portal, and ShareGeo repository. Hands-on sessions demonstrate using these resources to generate metadata and access open spatial data.
Similar to Gbrds Workshop Sept09 Metadata Identifiers (20)
Conceptualising Framework for Local Biodiversity Heritage Sites (LBHS): A Bio...Vishwas Chavan
This document proposes a conceptual framework for establishing Local Biodiversity Heritage Sites (LBHS) in Maharashtra, India based on a social-ecological model. It discusses how the Biological Diversity Act of 2002 allows local communities to designate biodiversity-rich areas as heritage sites. The framework identifies potential LBHS in two habitats: sacred groves, which are forest patches traditionally protected for their cultural and ecological values; and rocky plateaus, which support unique biodiversity through indigenous management practices. The document argues LBHS can preserve genetic resources, species, ecosystems, knowledge, culture and traditions as a legacy for future generations.
State Biodiversity Boards: Towards Better GovernanceVishwas Chavan
India’s Biological Diversity Act, 2002, and the three-tier
implementation mechanism of the National Biodiversity Authority (NBA), the State Biodiversity Board (SBB), the Union Territory Biodiversity Council (UTBC) and the Biodiversity Management Committee (BMC) is close to two decades old. However, our collective and compounding national progress is much less than satisfactory. One of the major reasons is lack of empowerment
of the SBBs, the UTBCs and resultantly passive functioning of the BMCs. Bottom-upward empowerment of BMCs to SBBs and UTBCs is crucial in order to achieve the National Biodiversity Targets (NBT) and other national biodiversity conservation and sustainable development ambitions. In this article, author proposes a five pillared work program that can help empower
the SBBs and UTBCs that can result in vibrant and optimally governing BMCs. Some or all of the activities mentioned in this article may have been initiated or implemented by few SBBs and UTBCs. However, author calls for coordinated and performance evaluation mechanism being developed and steered by SBBs and UTBC to achieve the national goal of development inclusive biodiversity conservation.
Exploring the future of scholarly publishing of biodiversity dataVishwas Chavan
Little more than decade back biodiversity data publishing was opportunistic and secondary spin-off activity of the biodiversity research and conservation management chain. Today, the Global Biodiversity Information Facility facilitate free and open access to over 420 million primary biodiversity data records contributed by publishers across the globe. This is an outcome of a growing realization that free and open access to biodiversity data is crucial to take informed decisions and actions for sustainable use of biotic resources and conservation of biodiversity areas. In recent past use of biodiversity data in research, conservation and management activities is on rise. However, users often complain about the low degree of ‘fitness-for-use’ of the accessible data. Most of the times potential use of data is hampered because of lack of adequate metadata, that can demonstrate the fintness-for-use of a given dataset.
To overcome this an appropriate incentivisation mechanism is essential, that can provide due credit and acknowledgement to a research groups for their efforts in authoring good metadata. In recent past a concept of ‘scholarly data publishing’ is being talked about where in both data and metadata undergo peer-review similar to other scientific publications. Pensoft publishing has launched a fresh data only journal called ‘Biodiversity Data Journal, and accepts data papers in six of its other journal titles. European aquatic biodiversity community through EU funded project ‘BioFresh’ has engaged with editors of 29 aquatic biodiversity journals to being accepting data papers. GBIF node in Columbia and South Africa are planning to kick start a journal that will publish data papers. Recently, Nature Publishing Group has announced a peer-reviewed data publishing only journal called ‘Scientific Data’. These developments announce the arrival of the new data publishing era ‘Scholarly Data Publishing’. Biodiversity science and biodiversity informatics stands to gain a lot by being on the forefront of this tide.
The document discusses GBIF's 2010-2011 work programme highlights related to improving content for science and society. It outlines GBIF's approach to focus on community needs, expand content coverage to include multimedia and observations, and increase relevance through facilitating the flow of data and information to scientific publications and decisions. It also analyzes current coverage and content biases and trends to help guide GBIF's science focus in 2011.
This document discusses data citation mechanisms and services for primary biodiversity data. It outlines the need for data citation to provide recognition for data producers and publishers. An ideal data citation framework would address social, technical, and policy issues to incentivize all stakeholders. Core technical components would include persistent identifiers, a data citation mechanism, and a data usage index. The document reviews the history calling for data citation standards and proposes requirements for an effective data citation model, including attributing roles across data production and publication. It also examines challenges in developing data citation practices.
The document summarizes the recommendations of the GBIF Governing Board's Global Strategy and Action Plan for Mobilization of Natural History Collections Data task group. The task group recommends that GBIF facilitate discovery of non-digital collection resources, increase efficiency of data capture and quality of digitized specimens, and improve infrastructure for publishing digitized collection data globally.
The document summarizes the state of data publishing through the Global Biodiversity Information Facility (GBIF) network. It finds that while data records are increasing, the rate of increase is declining. Developing regions contribute the most data, with the Avian Knowledge Network as the single largest data publisher. Over 2.4 billion records have been identified by GBIF participants but only around 800 million are accessible digitally and participants have committed to publishing less than 25% of available records by 2010. There remains a need for more strategic and planned approaches to data discovery and publishing with an emphasis on both local and global efforts.
This article discusses Balanophora, a rare and endangered plant found in North East India. It belongs to the family Balanophoraceae. The 15 species in the genus are native to the Old World Tropics. Most species are parasites on tree roots and are found in dense forests in the Himalayan region. They have underground inflorescences that rupture and emerge above ground. The plants are dioecious. Balanophora is listed as an endangered species under Indian law and prohibited from export due to its rarity. The article provides a brief description of the plant's rhizome, scapes, and reproductive structure.
This document provides an overview of bioinformatics education in India. It discusses how bioinformatics education has evolved from short workshops to formal degree programs over time. A key development was the establishment of the Biotechnology Information System network in the 1980s by the Department of Biotechnology, which helped develop bioinformatics infrastructure and training programs in India. The document then describes the current landscape of bioinformatics education in India, including a case study of the master's program in bioinformatics at the University of Pune. It concludes by noting that many universities and institutions now offer bioinformatics education at various levels to train students for careers in this growing field.
The document summarizes recommendations from the GBIF GSAP-NHC Task Group on improving the publishing of natural history collections data. It recommends that GBIF facilitate access to information about non-digital collections, work to increase the efficiency of digitizing specimen data and enhance data quality, and improve the global infrastructure for publishing digitized collections data.
The document discusses a meeting agenda between GBIF (Global Biodiversity Information Facility) and Elsevier to discuss opportunities for collaboration around data publishing and sharing biodiversity data. Some key points discussed in the agenda include GBIF's role in facilitating open access to biodiversity data, its data publishing framework to encourage data mobilization and sharing, and potential areas of collaboration around simultaneous publishing of data and scholarly articles.
The document discusses GBIF's (Global Biodiversity Information Facility) goals of facilitating open access to biodiversity data worldwide to support scientific research. GBIF shares over 200 million biodiversity records through data publishers and resources. The document proposes a Data Publishing Framework to improve data mobilization and cultural acceptance of open data sharing. It describes challenges to the framework and its potential impacts, such as increased data usage and quality through incentives like data papers and a Data Usage Index.
The document discusses the need for a Global Biodiversity Resources Discovery System (GBRDS) to address challenges in discovering biodiversity data. It proposes that GBRDS would act as a registry and discovery service to facilitate finding biodiversity information resources. GBRDS would provide an integrated 'yellow pages' reference for all biodiversity data by reconciling distributed resources and allowing meaningful discovery of data and services in a distributed manner. The document outlines how GBRDS could empower discovery of biodiversity data resources.
The document discusses two task groups convened by the Global Biodiversity Information Facility (GBIF) to provide recommendations around metadata and identifiers. The Metadata Implementation Framework Task Group recommended that GBIF develop a distributed metadata catalog system to describe biodiversity data using controlled vocabularies. The LSID-GUID Task Group recommended that GBIF encourage, support and advise on the use of identifiers like LSIDs and URIs to interconnect data and provide identifier resolution services.
The document discusses technical issues and opportunities for improving the Global Biodiversity Information Facility's (GBIF) registry and portals for discovering biodiversity resources. It analyzes GBIF's past use of UDDI registry and data portal, and outlines challenges in developing a new graph-based registry model to better represent the network of institutions, collections, and relationships. The new registry aims to improve discoverability through associating automated and human-generated metadata, uniquely identifying resources, and defining services and vocabularies.
The document proposes developing the Global Biodiversity Resources Discovery System (GBRDS) to address the challenge of discovering distributed biodiversity data and information resources. The GBRDS would consist of a registry to inventory publishers, institutions, datasets and services, and discovery services to search these resources. It would provide a "map" of all biodiversity information to enable discovery. The GBRDS is envisioned as the core of next generation biodiversity informatics infrastructure and aims to become a unified global entry point for discovering biodiversity resources by December 2010.
Global biodiversity data is critical for conservation, policymaking, and scientific research. However, most data is held by small, isolated publishers and is difficult to access. The Global Biodiversity Information Facility (GBIF) aims to mobilize this "small data" by creating common data standards and tools to publish data through its Integrated Publishing Toolkit. This allows data to be discovered through GBIF's portal and used for applications like predicting climate change impacts and invasive species spread. GBIF calls on all data holders to publish their data openly through its framework to build a comprehensive global resource for biodiversity data.
Gender and Mental Health - Counselling and Family Therapy Applications and In...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
How to Make a Field Mandatory in Odoo 17Celine George
In Odoo, making a field required can be done through both Python code and XML views. When you set the required attribute to True in Python code, it makes the field required across all views where it's used. Conversely, when you set the required attribute in XML views, it makes the field required only in the context of that particular view.
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumMJDuyan
(𝐓𝐋𝐄 𝟏𝟎𝟎) (𝐋𝐞𝐬𝐬𝐨𝐧 𝟏)-𝐏𝐫𝐞𝐥𝐢𝐦𝐬
𝐃𝐢𝐬𝐜𝐮𝐬𝐬 𝐭𝐡𝐞 𝐄𝐏𝐏 𝐂𝐮𝐫𝐫𝐢𝐜𝐮𝐥𝐮𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐏𝐡𝐢𝐥𝐢𝐩𝐩𝐢𝐧𝐞𝐬:
- Understand the goals and objectives of the Edukasyong Pantahanan at Pangkabuhayan (EPP) curriculum, recognizing its importance in fostering practical life skills and values among students. Students will also be able to identify the key components and subjects covered, such as agriculture, home economics, industrial arts, and information and communication technology.
𝐄𝐱𝐩𝐥𝐚𝐢𝐧 𝐭𝐡𝐞 𝐍𝐚𝐭𝐮𝐫𝐞 𝐚𝐧𝐝 𝐒𝐜𝐨𝐩𝐞 𝐨𝐟 𝐚𝐧 𝐄𝐧𝐭𝐫𝐞𝐩𝐫𝐞𝐧𝐞𝐮𝐫:
-Define entrepreneurship, distinguishing it from general business activities by emphasizing its focus on innovation, risk-taking, and value creation. Students will describe the characteristics and traits of successful entrepreneurs, including their roles and responsibilities, and discuss the broader economic and social impacts of entrepreneurial activities on both local and global scales.
Temple of Asclepius in Thrace. Excavation resultsKrassimira Luka
The temple and the sanctuary around were dedicated to Asklepios Zmidrenus. This name has been known since 1875 when an inscription dedicated to him was discovered in Rome. The inscription is dated in 227 AD and was left by soldiers originating from the city of Philippopolis (modern Plovdiv).
1. GLOBAL BIODIVERSITY INFORMATION FACILITY Metadata in context of GBRDS Éamonn Ó Tuama GBRDS Workshop, Copenhagen, 17-18 Sept 2009
2. - Task groups Outline - Metadata task group recommendations - LSID-GUID task group recommendations - Overview: where metadata fits in - Role of metadata
5. Why metadata? William K. Michener, Meta-information concepts for ecological data management, Ecological Informatics, Volume 1, Issue 1, January 2006, Pages 3-7, ISSN 1574-9541, DOI: 10.1016/j.ecoinf.2005.08.004. (http://www.sciencedirect.com/science/article/B7W63-4HJRS57-3/2/ea2e08412c6776456f540e66983546c0) Information about datasets deteriorates over time!
8. Data providers register their data and services in the GBIF UDDI Registry UDDI lists “business” information and binding template, i.e, the URL by which the provider installation can be accessed All further metadata are derived via DiGIR /TAPIR requests No separate metadata catalogue with dedicated client for searching or browsing Current GBIF metadata handling
9.
10.
11. MIFTG: summary The Global Biodiversity Information Facility (GBIF) aspires to … become a major provider of discovery and access services for a wide variety of biodiversity data types. A distributed metadata catalog system that describes and makes accessible general information on datasets of primary biodiversity data is recognised as an essential component of GBIF to achieve this objective.
12. MIFTG: recommendations Controlled vocabularies R53. Providers should use controlled vocabularies in any metadata field for which an appropriate vocabulary exists, and should use a multi-lingual thesaurus when appropriate R54. The GBIF vocabularies registry is a valuable service, but should be extended to include a canonical identifier for each vocabulary, and should work to be consistent with other vocabulary registries (e.g., oasis, info, srw)
13. MIFTG: recommendations Metadata specifications R6. Metadata should be able to describe multiple types of primary biodiversity data. R7. Metadata should support data discovery, interpretation, and analytical reuse R8. Metadata should support search/browse by space, time, taxa, and theme R9. Metadata should support search/browse by name of provider/name of organization R10. Metadata should support search by related publications
14. MIFTG: recommendations Metadata catalog system recommendations R27. The metadata catalog system must support multiple metadata models natively. R28. The metadata catalog system must be able to return the original contributed metadata object. R29. The metadata catalog system must support unique versioning of metadata and data objects using globally unique identifiers to differentiate revisions. R42. The metadata catalog system should register with one or more node registries to advertise services available. . R44. The metadata catalog system should provide attribution and branding for original metadata providers. .
15. MIFTG: recommendations Network Architecture Recommendations R20. GBIF should build a distributed system of regional nodes, each containing a replica of all metadata. R21. Each regional node must replicate metadata to other regional nodes when record changes occur using a GBIF-prescribed replication protocol. R22. Each regional node should also provide a harvesting interface that exposes metadata via their unique identifiers. R25. GBIF needs a registry to maintain list of regional nodes and their relevant service endpoints.
16.
17. LSID-GUID Task Group: summary Effective identification of data objects is essential for linking the world’s biodiversity data. If GBIF is to enable the exchange of biodiversity data it must promote identifier adoption through: - education, training, outreach - leadership - practical services
18. LGTG: recommendations Recommendation 4 GBIF should encourage, support and advise on the use of appropriate identifier technologies, in particular LSIDs and HTTP URIs , but not impose a requirement for one at the expense of the other. GBIF should provide specific advice for the issuing and use of LSIDs and for HTTP URIs.
19.
20.
21.
22.
23. LGTG: recommendations Recommendation 10 GBIF should provide services to support identifier resolution, redirection, metadata hosting, and caching.
25. LGTG: recommendations Recommendation 12 GBIF should extend the role of its data portal by hosting resources related to the use of identifiers, such as the TDWG vocabularies.