Challenges in Enabling Mixed Media Scholarly Research with Multi-Media Data i...roelandordelman.nl
Presentation at the Digital Humanities 2018 Conference, Mexico City, on the development of the Media Suite, an online research environment that facilitates scholarly research using large multimedia collections maintained at archives, libraries and knowledge institutions. The Media Suite unlocks the data on the collection level, item level, and segment level, provides tools that are aligned with the scholarly primitives (discovery, annotation, comparison, linking), and has a 'workspace' for storing personal mixed media collections and annotations, and to do advanced analysis using Jupyter Notebooks and NLP tools.
See the notes for the narrative that goes with the slides.
The Dryad Digital Repository: Published evolutionary data as part of the gre...Todd Vision
The Dryad Digital Repository aims to publish evolutionary data as part of the greater data ecosystem. Its goals are to publish data reported in biological literature, promote data reuse, and ensure responsible long-term data stewardship through a consortium of journals. Archiving data at publication is most effective for reuse and preservation. A survey found most researchers agree data should be publicly accessible, though opinions varied on whether it should be required or voluntary. Lessons learned indicate the importance of journals in data publication, the value of shared public repositories, and achieving a balance of benefit and burden for data authors.
Data reuse and scholarly reward: understanding practice and building infrastr...Todd Vision
This document discusses a study analyzing data reuse across scientific repositories. It finds that open gene expression data receives 69% more citations on average than closed data. It also finds patterns of data reuse vary across repositories, with references managers, blogs, bookmarks, and social networks all being venues for citations beyond traditional publications. The document advocates moving beyond traditional metrics like citations and journal impact factors to more fully capture the impact of open data sharing.
This presentation was provided by Patricia Brennan of The National Library of Medicine, during the NISO event "Defining The Library: The Mission, Role and Community," held on March 13, 2019.
Keynote presented to KE workshop held in conjunction with the release of the report "A Surfboard for Riding the Wave
Towards a four country action programme on research data": http://www.knowledge-exchange.info/Default.aspx?ID=469
This presentation was provided by Adam Rusbridge of EDINA during a NISO webinar on the topic of Providing Access: Ensuring What Libraries Have Licensed is What Users Can Reach on Feb 8, 2017
Challenges in Enabling Mixed Media Scholarly Research with Multi-Media Data i...roelandordelman.nl
Presentation at the Digital Humanities 2018 Conference, Mexico City, on the development of the Media Suite, an online research environment that facilitates scholarly research using large multimedia collections maintained at archives, libraries and knowledge institutions. The Media Suite unlocks the data on the collection level, item level, and segment level, provides tools that are aligned with the scholarly primitives (discovery, annotation, comparison, linking), and has a 'workspace' for storing personal mixed media collections and annotations, and to do advanced analysis using Jupyter Notebooks and NLP tools.
See the notes for the narrative that goes with the slides.
The Dryad Digital Repository: Published evolutionary data as part of the gre...Todd Vision
The Dryad Digital Repository aims to publish evolutionary data as part of the greater data ecosystem. Its goals are to publish data reported in biological literature, promote data reuse, and ensure responsible long-term data stewardship through a consortium of journals. Archiving data at publication is most effective for reuse and preservation. A survey found most researchers agree data should be publicly accessible, though opinions varied on whether it should be required or voluntary. Lessons learned indicate the importance of journals in data publication, the value of shared public repositories, and achieving a balance of benefit and burden for data authors.
Data reuse and scholarly reward: understanding practice and building infrastr...Todd Vision
This document discusses a study analyzing data reuse across scientific repositories. It finds that open gene expression data receives 69% more citations on average than closed data. It also finds patterns of data reuse vary across repositories, with references managers, blogs, bookmarks, and social networks all being venues for citations beyond traditional publications. The document advocates moving beyond traditional metrics like citations and journal impact factors to more fully capture the impact of open data sharing.
This presentation was provided by Patricia Brennan of The National Library of Medicine, during the NISO event "Defining The Library: The Mission, Role and Community," held on March 13, 2019.
Keynote presented to KE workshop held in conjunction with the release of the report "A Surfboard for Riding the Wave
Towards a four country action programme on research data": http://www.knowledge-exchange.info/Default.aspx?ID=469
This presentation was provided by Adam Rusbridge of EDINA during a NISO webinar on the topic of Providing Access: Ensuring What Libraries Have Licensed is What Users Can Reach on Feb 8, 2017
Closing the scientific literature access gap with CORE - how to gain free acc...Nancy Pontika
Presented during the International Open Access Week 2020 for the Kerala Library Association, October 21, 2020.
The presentation is about CORE, a global harvester of open access scientific content and the CORE services on content discovery, managing content and access to raw data.
BioDBCore: Current Status and Next DevelopmentsPascale Gaudet
The document discusses BioDBCore, a collaborative project aimed at gathering and standardizing metadata about biological databases. It provides an overview of BioDBCore's goals of improving data integration, encouraging standards, and maximizing resources. BioDBCore is led by Pascale Gaudet and Philippe Rocca-Serra and implemented on the BioSharing website. The document outlines the BioDBCore descriptors for databases and provides an example entry for the dictyBase database. It discusses maintaining and expanding BioDBCore records with the help of database providers and journals.
A Data Citation Roadmap for Scholarly Data RepositoriesLIBER Europe
This document summarizes a webinar about developing a roadmap for data citation in scholarly data repositories. The webinar discussed recommendations for repositories to support unique, persistent identifiers for datasets that resolve to landing pages containing human- and machine-readable metadata. It also covered tracking data citations between repositories and publishers, and next steps to publish an updated recommendations paper incorporating feedback from the community. The webinar was presented by experts from DataCite, Harvard University and other organizations to facilitate best practices for data citation.
Fedora is an open source digital repository system that is flexible, durable, and standards-based. It is developed and supported by a thriving community to store, preserve, and provide access to digital objects. Fedora repositories can handle both simple and complex use cases and content models. Examples of Fedora implementations include institutional repositories, research data repositories, digital archives and special collections, and manuscript collections.
Institutional repositories play an important role in making scholarly research openly accessible and increasing its impact. They house research outputs like journal articles, theses, and other works. Repositories enhance the visibility of the hosting institution and facilitate further research by providing access to latest information. There are currently over 1,400 repositories worldwide adhering to interoperability standards so their contents can be indexed together. Repositories provide benefits like increased citations and downloads for authors, and give institutions metrics to assess their research programs. For these reasons, more Indian institutions are establishing their own open access repositories.
THOR Workshop - Data Publishing ElsevierMaaike Duine
Elsevier supports researchers in sharing their data through several programs and services:
1. A data-linking program connects articles to over 60 domain-specific data repositories through in-article data accession numbers and banners.
2. Mendeley Data is Elsevier's research data repository, allowing researchers to store, share and publish research data with a DOI and link to related articles.
3. In-article data visualization tools display plot data from supplementary materials in journals, allowing readers to access, explore and download underlying data.
SEAD Prototype: Data Curation and Preservation for Sustainability ScienceSEAD
The SEAD prototype aims to enable data curation and preservation for sustainability science research by providing tools for ingesting, annotating, visualizing, and preserving heterogeneous research data. It integrates three components: Active Curation and Research (ACR) for data management and curation, VIVO for networking and analytics of research outputs, and Virtual Archive (VA) for long-term data publication, preservation, and discovery. The prototype is being tested by curating a 1.6 terabyte dataset from the National Center for Earth Surface Dynamics involving transfer of data and metadata between the three SEAD components.
re3data.org – Registry of Research Data RepositoriesHeinz Pampel
Heinz Pampel | GFZ German Research Centre for Geosciences, LIS
Maxi Kindling | Humboldt-Universität zu Berlin, Berlin School of Library and Information Science Frank Scholze | Karlsruhe Institute of Technology, KIT Library
RDA-Deutschland-Treffen 2015| Potsdam, November 26, 2015
From metadata to data curation: the role of libraries in data exchangeLIBER Europe
This document summarizes a presentation on the role of libraries in data exchange. It discusses several projects by LIBER and the European Research Infrastructure to promote data sharing and preservation. A survey found that libraries need to develop skills and strategies for data citation, curation, and preservation. Moving forward, libraries should advocate for digital preservation policies, invest in workforce training, and help educate researchers on data management best practices. Overall, the presentation argues that libraries have an important role to play in supporting open data and research reproducibility.
The Dryad Digital Repository: Published data as part of the greater data ecos...Hilmar Lapp
Presented at the M3 and Biosharing Special Interest Group (SIG) meeting at ISMB 2010 in Boston, MA: http://gensc.org/gc_wiki/index.php/M3_%26_BioSharing
Dataverse in the Universe of Data by Christine L. Borgmandatascienceiqss
Data repositories are much more than "black boxes" where data go in but may never come out. Rather, they are situated in communities, with contributors, users, reusers, and repository staff who may engage actively or passively with participants. This talk will explore the roles that Dataverse plays – or could play – in individual communities.
This document summarizes Mercè Crosas's presentation on the expanding dataverse and advances in data publishing. It discusses the growth of digital data and need for data citation, repositories, and metadata to make data discoverable, accessible, and reusable. The Dataverse software provides a framework for publishing data across different repository types. Recent improvements allow for rigorous data citation compliant with principles, rich metadata, support for public and restricted data, and publication workflows. Future areas of focus include integration with other systems, support for sensitive data, and expanding data citation and APIs.
February 18 2015 NISO Virtual Conference
Scientific Data Management: Caring for Your Institution and its Intellectual Wealth
Network Effects: RMap Project
Sheila M. Morrissey, Senior Researcher, ITHAKA
Connecting Dataverse with the Research Life CycleMerce Crosas
This document outlines how data repositories can connect with the research lifecycle. It discusses Dataverse, an open source data repository platform developed by IQSS at Harvard University. It describes several application programming interfaces (APIs) that can be used for interoperating with Dataverse, including OAI-PMH, SWORD, search, and data access APIs. It also lists several integrations that are implemented or in progress to connect Dataverse with other systems like the Open Science Framework (OSF) and support functions like depositing data management plans, data curation, assigning DOIs, and more.
Creating a sustainable business model for a digital repository: the Dryad exp...ASIS&T
Creating a sustainable business model for a digital repository: the Dryad experience
Peggy Schaeffer
Datadryad.org
Presentation at Research Data Access & Preservation Summit
22 March 2012
Preservation of Research Data: Dataverse / Archivematica Integration by Allan...datascienceiqss
Scholars Portal, a program of the Ontario Council of University Libraries (OCUL), provides the technical infrastructure to store, preserve, and provide access to shared digital library collections in Ontario - including hosting a local instance of Dataverse since 2011. As part of a national project known as Portage (a project of the Canadian Association of Research Libraries), Scholars Portal is partnering with Artefactual Systems, Dataverse, the University of British Columbia, the University of Alberta, and others, to integrate Dataverse with preservation software Archivematica. When completed, this project will facilitate the long-term preservation of research data according to the Open Archival Information System (OAIS) Reference Model.
GSmith Springer Nature Data policies and practices: HKU Open Data and Data Pu...GrahamSmith646206
Supporting research data across Springer Nature: joining up policy and practice. Slides from Graham Smith (Research Data Manager, Springer Nature) at HKU Open Data and Data Publishing Seminar, 25th October 2021.
The document discusses open access and open archives. It begins by explaining the origins of open access in the creation of the HEP-TH database by physicist Paul Ginsparg in 1991. It then defines open access as providing free online access to peer-reviewed articles, reports, and other materials with no licensing restrictions. The rest of the document discusses the benefits of open access repositories and archives, how to start and maintain an institutional repository, different types of repositories including subject-based repositories, standards and initiatives like OAI and Budapest Open Access Initiative, and the Directory of Open Access Journals.
Media Suite: Unlocking Archives for Mixed Media Scholarly Research roelandordelman.nl
This document summarizes the CLARIAH Media Suite, which aims to unlock institutional multimedia collections for mixed media scholarly research. It provides access to over 1.8 million radio/TV items, 60 million newspaper pages, and other collections. The Media Suite connects these collections through a shared virtual workspace that allows scholars to discover, access, annotate, analyze, and export data. It supports scholarly primitives like distant and close reading. However, the document notes some issues like complex interfaces, lack of robust analysis pipelines, and need for manual metadata work. Overall, the Media Suite brings analytical tools to archived data and enables new forms of mixed media research, though further development is still needed.
Where is the opportunity for libraries in the collaborative data infrastructure?LIBER Europe
Presentation by Susan Reilly at Bibsys2013 on the opportunties for libraries and their role in the collaborative data infrastructure. Looks at data sharing, authentication, preservation and advocacy.
This document summarizes an presentation about opportunities for data exchange and optimizing data sharing conditions. It discusses several projects by LIBER, including Europeana which aims to make cultural content available online. It notes that with proper infrastructure, researchers can collaborate on shared data sets across locations. However, challenges include authentication, skills, and managing large amounts of data being generated. Overall, the presentation argues that data sharing can advance scientific inquiry if barriers are addressed and key stakeholders work together.
Closing the scientific literature access gap with CORE - how to gain free acc...Nancy Pontika
Presented during the International Open Access Week 2020 for the Kerala Library Association, October 21, 2020.
The presentation is about CORE, a global harvester of open access scientific content and the CORE services on content discovery, managing content and access to raw data.
BioDBCore: Current Status and Next DevelopmentsPascale Gaudet
The document discusses BioDBCore, a collaborative project aimed at gathering and standardizing metadata about biological databases. It provides an overview of BioDBCore's goals of improving data integration, encouraging standards, and maximizing resources. BioDBCore is led by Pascale Gaudet and Philippe Rocca-Serra and implemented on the BioSharing website. The document outlines the BioDBCore descriptors for databases and provides an example entry for the dictyBase database. It discusses maintaining and expanding BioDBCore records with the help of database providers and journals.
A Data Citation Roadmap for Scholarly Data RepositoriesLIBER Europe
This document summarizes a webinar about developing a roadmap for data citation in scholarly data repositories. The webinar discussed recommendations for repositories to support unique, persistent identifiers for datasets that resolve to landing pages containing human- and machine-readable metadata. It also covered tracking data citations between repositories and publishers, and next steps to publish an updated recommendations paper incorporating feedback from the community. The webinar was presented by experts from DataCite, Harvard University and other organizations to facilitate best practices for data citation.
Fedora is an open source digital repository system that is flexible, durable, and standards-based. It is developed and supported by a thriving community to store, preserve, and provide access to digital objects. Fedora repositories can handle both simple and complex use cases and content models. Examples of Fedora implementations include institutional repositories, research data repositories, digital archives and special collections, and manuscript collections.
Institutional repositories play an important role in making scholarly research openly accessible and increasing its impact. They house research outputs like journal articles, theses, and other works. Repositories enhance the visibility of the hosting institution and facilitate further research by providing access to latest information. There are currently over 1,400 repositories worldwide adhering to interoperability standards so their contents can be indexed together. Repositories provide benefits like increased citations and downloads for authors, and give institutions metrics to assess their research programs. For these reasons, more Indian institutions are establishing their own open access repositories.
THOR Workshop - Data Publishing ElsevierMaaike Duine
Elsevier supports researchers in sharing their data through several programs and services:
1. A data-linking program connects articles to over 60 domain-specific data repositories through in-article data accession numbers and banners.
2. Mendeley Data is Elsevier's research data repository, allowing researchers to store, share and publish research data with a DOI and link to related articles.
3. In-article data visualization tools display plot data from supplementary materials in journals, allowing readers to access, explore and download underlying data.
SEAD Prototype: Data Curation and Preservation for Sustainability ScienceSEAD
The SEAD prototype aims to enable data curation and preservation for sustainability science research by providing tools for ingesting, annotating, visualizing, and preserving heterogeneous research data. It integrates three components: Active Curation and Research (ACR) for data management and curation, VIVO for networking and analytics of research outputs, and Virtual Archive (VA) for long-term data publication, preservation, and discovery. The prototype is being tested by curating a 1.6 terabyte dataset from the National Center for Earth Surface Dynamics involving transfer of data and metadata between the three SEAD components.
re3data.org – Registry of Research Data RepositoriesHeinz Pampel
Heinz Pampel | GFZ German Research Centre for Geosciences, LIS
Maxi Kindling | Humboldt-Universität zu Berlin, Berlin School of Library and Information Science Frank Scholze | Karlsruhe Institute of Technology, KIT Library
RDA-Deutschland-Treffen 2015| Potsdam, November 26, 2015
From metadata to data curation: the role of libraries in data exchangeLIBER Europe
This document summarizes a presentation on the role of libraries in data exchange. It discusses several projects by LIBER and the European Research Infrastructure to promote data sharing and preservation. A survey found that libraries need to develop skills and strategies for data citation, curation, and preservation. Moving forward, libraries should advocate for digital preservation policies, invest in workforce training, and help educate researchers on data management best practices. Overall, the presentation argues that libraries have an important role to play in supporting open data and research reproducibility.
The Dryad Digital Repository: Published data as part of the greater data ecos...Hilmar Lapp
Presented at the M3 and Biosharing Special Interest Group (SIG) meeting at ISMB 2010 in Boston, MA: http://gensc.org/gc_wiki/index.php/M3_%26_BioSharing
Dataverse in the Universe of Data by Christine L. Borgmandatascienceiqss
Data repositories are much more than "black boxes" where data go in but may never come out. Rather, they are situated in communities, with contributors, users, reusers, and repository staff who may engage actively or passively with participants. This talk will explore the roles that Dataverse plays – or could play – in individual communities.
This document summarizes Mercè Crosas's presentation on the expanding dataverse and advances in data publishing. It discusses the growth of digital data and need for data citation, repositories, and metadata to make data discoverable, accessible, and reusable. The Dataverse software provides a framework for publishing data across different repository types. Recent improvements allow for rigorous data citation compliant with principles, rich metadata, support for public and restricted data, and publication workflows. Future areas of focus include integration with other systems, support for sensitive data, and expanding data citation and APIs.
February 18 2015 NISO Virtual Conference
Scientific Data Management: Caring for Your Institution and its Intellectual Wealth
Network Effects: RMap Project
Sheila M. Morrissey, Senior Researcher, ITHAKA
Connecting Dataverse with the Research Life CycleMerce Crosas
This document outlines how data repositories can connect with the research lifecycle. It discusses Dataverse, an open source data repository platform developed by IQSS at Harvard University. It describes several application programming interfaces (APIs) that can be used for interoperating with Dataverse, including OAI-PMH, SWORD, search, and data access APIs. It also lists several integrations that are implemented or in progress to connect Dataverse with other systems like the Open Science Framework (OSF) and support functions like depositing data management plans, data curation, assigning DOIs, and more.
Creating a sustainable business model for a digital repository: the Dryad exp...ASIS&T
Creating a sustainable business model for a digital repository: the Dryad experience
Peggy Schaeffer
Datadryad.org
Presentation at Research Data Access & Preservation Summit
22 March 2012
Preservation of Research Data: Dataverse / Archivematica Integration by Allan...datascienceiqss
Scholars Portal, a program of the Ontario Council of University Libraries (OCUL), provides the technical infrastructure to store, preserve, and provide access to shared digital library collections in Ontario - including hosting a local instance of Dataverse since 2011. As part of a national project known as Portage (a project of the Canadian Association of Research Libraries), Scholars Portal is partnering with Artefactual Systems, Dataverse, the University of British Columbia, the University of Alberta, and others, to integrate Dataverse with preservation software Archivematica. When completed, this project will facilitate the long-term preservation of research data according to the Open Archival Information System (OAIS) Reference Model.
GSmith Springer Nature Data policies and practices: HKU Open Data and Data Pu...GrahamSmith646206
Supporting research data across Springer Nature: joining up policy and practice. Slides from Graham Smith (Research Data Manager, Springer Nature) at HKU Open Data and Data Publishing Seminar, 25th October 2021.
The document discusses open access and open archives. It begins by explaining the origins of open access in the creation of the HEP-TH database by physicist Paul Ginsparg in 1991. It then defines open access as providing free online access to peer-reviewed articles, reports, and other materials with no licensing restrictions. The rest of the document discusses the benefits of open access repositories and archives, how to start and maintain an institutional repository, different types of repositories including subject-based repositories, standards and initiatives like OAI and Budapest Open Access Initiative, and the Directory of Open Access Journals.
Media Suite: Unlocking Archives for Mixed Media Scholarly Research roelandordelman.nl
This document summarizes the CLARIAH Media Suite, which aims to unlock institutional multimedia collections for mixed media scholarly research. It provides access to over 1.8 million radio/TV items, 60 million newspaper pages, and other collections. The Media Suite connects these collections through a shared virtual workspace that allows scholars to discover, access, annotate, analyze, and export data. It supports scholarly primitives like distant and close reading. However, the document notes some issues like complex interfaces, lack of robust analysis pipelines, and need for manual metadata work. Overall, the Media Suite brings analytical tools to archived data and enables new forms of mixed media research, though further development is still needed.
Where is the opportunity for libraries in the collaborative data infrastructure?LIBER Europe
Presentation by Susan Reilly at Bibsys2013 on the opportunties for libraries and their role in the collaborative data infrastructure. Looks at data sharing, authentication, preservation and advocacy.
This document summarizes an presentation about opportunities for data exchange and optimizing data sharing conditions. It discusses several projects by LIBER, including Europeana which aims to make cultural content available online. It notes that with proper infrastructure, researchers can collaborate on shared data sets across locations. However, challenges include authentication, skills, and managing large amounts of data being generated. Overall, the presentation argues that data sharing can advance scientific inquiry if barriers are addressed and key stakeholders work together.
Networked Science, And Integrating with DataverseAnita de Waard
This document discusses the growing interconnectedness of research data and tools in a networked science environment. It summarizes Elsevier's current and potential future connections to the Dataverse platform, including exporting data from the Hivebench ELN to Dataverse, linking articles to datasets in Dataverse through frameworks like Scholix, indexing Dataverse through Elsevier's data search tools, and tracking metrics on Dataverse datasets through analytics platforms like PlumX. The author expresses interest in further strengthening integration between these systems to advance open sharing of research data.
The role of libraries in data exchangeLIBER Europe
The document discusses the role of libraries in data exchange and management. It summarizes the results of a survey of over 800 librarians on their current practices and skills related to supporting data management and exchange. Key findings include:
- 81% of respondents see demand from researchers for support in data exchange, but current library support is not meeting this demand.
- Only 32% of libraries believe they have the necessary skills currently, but 53% are working to develop these skills.
- The best place for underlying research data is agreed to be in official data repositories and archives.
- Developing subject expertise and integrating data management training into professional education are seen as important ways for libraries to build their skills
Scientific Data is a new category of publication that provides detailed descriptions of scientifically valuable datasets to improve data reproducibility and reuse, with descriptors covering topics like methods, data records, and technical validation. These descriptors undergo a peer review process to assess completeness, consistency, integrity, and experimental rigor. The publication is hosted on Nature.com and aims to improve data discoverability, curation, and peer review through machine-readable metadata and clear links between data, descriptors, and related research papers.
This document discusses libraries and data services at Purdue University. It introduces the Purdue University Research Repository (PURR), an online platform for sharing and archiving research data. PURR provides resources for data management planning, a place to publish datasets with DataCite DOIs, and long-term archiving. The document outlines PURR's features and collaboration between Purdue Libraries, the Office of the Vice President for Research, and Information Technology. It also discusses the importance of data citation, identifiers, and linking data to publications. Finally, it suggests ways libraries can get involved, such as talking to researchers, understanding the data landscape, and providing data literacy instruction.
New Metaphors: Data Papers and Data CitationsJohn Kunze
The document discusses new metaphors for data papers and data citations. It notes that metaphors are pervasive in thought and language, and digital objects like files and folders are based on metaphors. It then provides an overview of the California Digital Library and how their environment and focus has changed from preservation to include curation and support for data producers. Forces like rising journal costs, increased research publication and declining budgets create structural problems for libraries. The document advocates a practical incremental approach to the complex problem of data curation, including initiatives like DataONE and the use of data papers and citations.
SciDataCon 2014 Data Papers and their applications workshop - NPG Scientific ...Susanna-Assunta Sansone
Part of the SciDataCon14 workshop on "Data Papers and their applications" run by myself and Brian Hole to help attendees understand current data-publishing journals and trends and help them understand the editorial processes on NPG's Scientific Data and Ubiquity's Open Health Data.
How can we ensure research data is re-usable? The role of Publishers in Resea...LEARN Project
How can we ensure research data is re-usable? The role of Publishers in Research Data Management, by Catriona MacCallum. 2nd LEARN Workshop, Vienna, 6th April 2016
DataCite and Campus Data Services
Paul Bracke, Associate Dean for Digital Programs and Information Services, Purdue University
Research libraries are increasingly interested in developing data services for their campuses. There are many perspectives, however, on how to develop services that are responsive to the many needs of scientists; sensitive to the concerns of scientists who are not always accustomed to sharing their data; and that are attractive to campus administrators. This presentation will discuss the development of campus-based data services programs, the centrality of data citation to these efforts, and the ways in which engagement with DataCite can enhance local programs.
OSFair2017 Workshop | Building a global knowledge commons - ramping up reposi...Open Science Fair
Eloy Rodrigues, Petr Knoth & Kathleen Shearer showcase the conceptual model for this vision, as well as the role and functions of repositories within this model.
Workshop title: Building a global knowledge commons - ramping up repositories to support widespread change in the ecosystem
Workshop abstract:
The extensive international deployment of repository systems in higher education and research institutions, as well as scholarly communities, provides the foundation for a distributed, globally networked infrastructure for scholarly communication. This distributed network of repositories can and should be a powerful tool to promote the transformation of the scholarly communication ecosystem. However, repository platforms are still using technologies and protocols designed almost twenty years ago, before the boom of the web and the dominance of Google, social networking, semantic web and ubiquitous mobile devices. In April 2016, the Confederation of Open Access Repositories (COAR) launched a working group to help identify new functionalities and technologies for repositories and develop a road map for their adoption. For the past several months, the group has been working to define a vision for repositories and sketch out the priority user stories and scenarios that will help guide the development of new functionalities. The results of this work will be available in the summer of 2017.
This workshop will present the functionalities and technologies for the next generation of repositories and reflect on how these functionalities will be adopted into the existing software platforms. In addition, participants will discuss the important implications for the network layers, and how repositories will uniformly interact with the networks to provide value added services on top of their content.
DAY 3 - PARALLEL SESSION 6 & 7
http://www.opensciencefair.eu/workshops/parallel-day-3-1/building-a-global-knowledge-commons-ramping-up-repositories-to-support-widespread-change-in-the-ecosystem
Fair sample and data access -David Van enckevortData Science NIH
This document discusses making biobank data and samples FAIR (Findable, Accessible, Interoperable, and Reusable).
It explains the four FAIR principles and provides examples of how to apply each one. To make resources findable, they need unique and persistent identifiers, rich metadata, and to be findable through other systems. To make them accessible, they need to be retrievable using open standards. To make them interoperable, standardized languages and vocabularies should be used. And to make them reusable, they need to be richly described and released with clear usage terms and provenance.
The document recommends three steps to make samples and data FAIR: include sufficient metadata using common
David Van Enckevort - FAIR sample and data access DataSciSIG
David van Enckevort from the University of Groningen describes FAIR Sample and Data Access in Biobanking and Biorepositories.
This talk was sponsored by the NIH Data Science Special Interest Group and part of a webinar panel on June 23, 2017 on Global Biobanking and Access to Specimens.
This document discusses making biobank data and samples FAIR (Findable, Accessible, Interoperable, and Reusable).
It explains the four FAIR principles and provides examples of how to apply each one. To make resources findable, they need unique and persistent identifiers, rich metadata, and to be discoverable through other systems. To make them accessible, they need to be retrievable using open standards. To make them interoperable, standards for knowledge representation like ontologies should be used. And to make them reusable, they need to be richly described and released with clear usage terms and provenance.
The document recommends three steps to make samples and data FAIR: include sufficient metadata using
Open Science, Open Data: towards a new transparent and reproducible ecosystemLIBER Europe
Presented at the Preforma Open Source Workshop 8 April 2016
As a library membership organization, LIBER works on addressing Open Science barriers. Standardisation of file formats can really help in overcoming some of these barriers: it enables us to process and preserve data in a controlled way, it helps ensure that outputs are really open and accessible in the long term and it improves interoperability of new tools and services. Making sure data is stored in a controlled way and can be (re) used today and in the future is an important element in Open Science. We see this as not only a technical challenge but also a social one: awareness, trust and community building is needed in order to ensure uptake of these standards. Libraries therefore have a valuable role to play in the development of good research data management throughout all phases of the Open Data lifecycle.
This document discusses engaging researchers in research data management (RDM) through data reference interviews. It provides an overview of EDINA and the University of Edinburgh Data Library and their roles in assisting researchers. It then describes the data reference interview process, highlighting the importance of understanding the researcher's field and data. Recommendations are provided for interviewing researchers and tools for assessing data are introduced. The document concludes by discussing the University's RDM strategy and engagement tools.
Similar to A research passport: library requirements (20)
LIBER Europe Covid-19 Research Libraries Survey - December 2020LIBER Europe
This document presents the results of a LIBER COVID-19 survey categorized by country and institution groups. It divides respondent institutions into three categories: Category A includes Western European countries, Category B includes Central and Eastern European countries, and Category C includes Southeastern European and Eastern European countries. The document consists of a series of graphs comparing survey responses across the different categories of institutions regarding the impact of the COVID-19 pandemic.
LIBER Webinar: Turning FAIR Data Into RealityLIBER Europe
These slides relate to a LIBER Webinar given on 23 April 2018. Turning FAIR Data Into Reality — Progress and Plans from the European Commission FAIR Data Expert Group.
In this webinar, Simon Hodson, Executive Director of CODATA and Chair of the FAIR Data Expert Group, and Sarah Jones, Associate Director at the Digital Curation Centre and Rapporteur, reported on the Group’s progress.
Copyright Reform: EU Legislative Process & LIBER AdvocacyLIBER Europe
LIBER's Copyright & Legal Matters Working Group met in Helsinki on 7 December 2017. This presentation, outlining the EU legislative process on copyright reform and LIBER advocacy, was given at the meeting by Helena Lovegrove, LIBER's Advocacy Adviser.
Applying Bourdieu's Field Theory to MLS Curricula Development. Charlotte Nord...LIBER Europe
The document discusses applying Pierre Bourdieu's field theory concept to analyze the changing positions of research librarians within university structures over time. It presents field theory concepts such as fields, doxa, habitus, and forms of capital. Diagrams show how positions within the university and library fields have changed, with research librarians previously higher in cultural capital now lower. Reasons for this include changes in client needs and other library staff professionalizing. It suggests ways for research librarians to reclaim prestige by ensuring services' value and combining domain knowledge with client needs. Finally, it outlines a new flexible master's program to help research librarians specialize in areas like project management, bibliometrics and data management
Growing a Culture for Change at The University of Manchester Library. Penny H...LIBER Europe
The University of Manchester Library underwent a culture change process to improve their strategy and leadership. Their initial strategy saw over 100 projects but lacked staff involvement which led to disconnect and resistance. To improve, they held meetings to get staff feedback and have staff self-elect involvement in developing a new strategy. For the new strategy, 30 staff were involved across 4 themes linked to the university's goals, compared to just 3 staff previously. Lessons learned included the importance of empowering staff, maintaining involvement, and regularly checking in with staff.
Enabling the Exchange and use of Data in AgricultureLIBER Europe
This presentation by Imma Subirats was part of the "Research Data Support Meets Disciplines: Opportunities & Challenges" workshop at LIBER's 2017 Annual Conference in Patras, Greece. For more information, see www.libereurope.eu
GDPR - Thoughts on the EU Data Protection Regulation, Research and LibrariesLIBER Europe
This presentation by Jonas Holm was part of the "Research Data Support Meets Disciplines: Opportunities & Challenges" workshop at LIBER's 2017 Annual Conference in Patras, Greece. For more information, see www.libereurope.eu
Research Data Services and Data Collections: Library Synergies for Economic R...LIBER Europe
This presentation by Thomas Bourke was part of the "Research Data Support Meets Disciplines: Opportunities & Challenges" workshop at LIBER's 2017 Annual Conference in Patras, Greece. For more information, see www.libereurope.eu
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
A research passport: library requirements
1. A European Research
Passport: Library
Requirements
Susan Reilly
AAA Study Workshop, Brussels, 12th July, 2012
susan.reilly@kb.nl
@skreilly
2. Contents
• LIBER: enriching the researcher experience
• Environment
• Research behaviours
• Requirements & Challenges
3. LIBER: enriching the researcher experience
• Largest network of European reseach libraries: 400 in over 40 countries
• LIBER aims to represent the interests of European research libraries,
their universities and their researchers. It promotes in particular
-efficient information services
-access to research information, in any form whatsoever
-innovative end-user services for teaching, learning and research
-preservation of cultural heritage
-efficient and effective management of research libraries
4. LIBER: enriching the researcher experience
• Scholarly Communication and Research Infrastructures
• Reshaping the Research Library
5. LIBER: enriching the researcher experience
Content
Europeana Libraries
Europeana Newspapers
Policy
MEDOANET
Infrastructure
APARSEN
ODE
DART Europe
6. Environment
• Increasing amount of
digitised and born digital
content in libraries
• Increasing emphasis on
Open Access publications
and data- institutional
repositories
• Change in research
behaviour, increase in cross
disciplinary research
• Decreasing budgets,
emphasis on ROI
7. Data Sharing
Availability
Findability
Interpretability
Reusability
Citability
Curation
Preservation
Source: ODE 2011 Survey of LIBER libraries
9. The Data
Publication Pyramid (1) Data
contained and
explained within
the article
(2) Further data
explanations in
any kind of
supplementary (3) Data
files to articles referenced from
the article and
held in data
centers and
(4) Data
repositories
publications,
describing
available
datasets
(5) Data in
drawers and
on disks at the
institute
10. Linking Data to Publications
The best place for underlying data is in official data repositories and archives 84% agree
Publications should always contain links to the underlying research data 74%
Data archives should have a system in place for persistent identifiers that properly support citation of 74%
datasets
Research journals should have much stricter editorial policies on data availability 64%
Underlying data should be part of the peer review process 54%
Underlying data should be cited separately in the reference list 46%
Publishers and editors should only accept in supplements the summary datasets that are of direct 26%
relevance to the article
There are not sufficient trustworthy data archives available for authors to deposit their data 62.5%
11. Researchers
• Changes in information
seeking behaviour
• Growth in systematic collaboration
• More complex research (cross disciplinary)
• Less time for research (occurs in bursts)
• Some barriers:
• Lack of linking between archives
• Lack of standardisation of online databases
and archives
• Difficulties in data linking arising from the lack of standardisation,
and the inconsistencies in quality, reliability, and functionality
across different resources
14. Collaboration
“Networked science is on the rise, the researcher is no longer
working alone in his office, he is working virtually with other
researchers from around the world. For them it is important
that they can use the same software and share and reuse
the same content related objects, in a trusted environment.”
Heinke Neuroth, Head of Innovation, Goettingen State &
University Library
15. Licensed Content
“Sharing disk space and
software is the easy part but
this also requires a new
approach to licencing of e-
resources. Some poorer
institutions do not always have
access to the same e-
resources”
Morag Watson, Librarian
16. Requirements
• Simplified and seemless access to information resources
• Simplified licencing agreements/more open access publications
• Support collaborative research and new technologies to
exploit information resources
• Insitutional credential (for the near future)
• Support Open scholarly communications
• depositing of articles and data
• curation/preservation (versioning,
author identification)
• trust
17. Further Reading
• ODE/APARSEN Publications
http://www.alliancepermanentaccess.org/index.php/community/current-
projects/ode/
• Ian Rowlands, David Nicholas, Peter Williams, Paul Huntington, Maggie
Fieldhouse, Barrie Gunter, Richard Withey, Hamid R. Jamali, Tom Dobrowolski,
Carol Tenopir, (2008) "The Google generation: the information behaviour of the
researcher of the future", Aslib Proceedings, Vol. 60 Iss: 4, pp.290 - 310
• Susan K. Reilly, Marian Lefferts, Martin Moyle Collaboration to Build a
Meaningful Connection Between Library Content and the Researcher
• New Review of Information Networking Vol. 17, Iss. 1, 2012
• Most images from Europeana Libraries