The document discusses how to cost data curation and research data management. It defines data curation as the active management of digital items over the long term to ensure they remain secure, discoverable and accessible. Research data management involves storing, accessing and preserving data produced from an investigation over its entire lifecycle. The document recommends using activity-based costing to identify all direct and indirect activities associated with data curation and research data management. It also emphasizes the importance of defining what is being costed and choosing an appropriate costing methodology.
Certifying and Securing a Trusted Environment for Health Informatics Research...Jisc
The document discusses the certification and securing of a trusted environment for health informatics research data at the University of Dundee. It provides an overview of the Health Informatics Centre, its research data management platform, safe haven architecture, and ISO27001 certification. The platform standardizes data extraction and release, adds metadata and quality checks. A safe haven uses pseudonymized data and virtual environments prevent data from leaving. ISO27001 certification provides governance and reduces documentation through standardized information security practices.
Dr. Tito Castillo discusses challenges with data discovery and sharing at University College London Hospitals (UCLH) due to their multiple proprietary clinical systems with undocumented data and data warehouses. To address this, UCLH is taking a standards-based approach using models like DDI and SDMX to document metadata and map their processes. The goal is to enable better data access, sharing, and reuse to support research programmes and new models of care while respecting governance and privacy.
This document provides an overview of a webinar on digital curation and research data management for universities. The webinar covers an introduction to digital curation, the benefits and drivers for research data management, current initiatives in UK universities, and the role of libraries in supporting research data management. Libraries are increasingly involved in developing institutional policies, providing training, and advising researchers on writing data management plans and sharing data. The webinar highlights training opportunities for librarians to develop skills in research data management and digital curation.
A presentation offering an introduction to managing and sharing research data given at the Czech Open Science days as part of the EC-funded FOSTER project.
This document discusses licensing research data for reuse. It begins by providing a scenario where a user has downloaded a dataset but is unsure what they can do with the data due to licensing. It then discusses that licensing is critical to enabling data reuse and citation. It provides information on AusGOAL, the Australian open access and licensing framework, and notes it is recommended for data publishing by ANDS partners. It also includes links to licensing guides and FAQs. In summary, the document emphasizes the importance of data licensing for enabling reuse and outlines Australia's recommended licensing system.
Presentation given at the European Research Council workshop on research data management and sharing in Brussels on 18th-19th September 2014. The presentation covers the benefits and drivers for RDM, points to relevant tools and resources and closes with some open questions for discussion.
The document discusses how to cost data curation and research data management. It defines data curation as the active management of digital items over the long term to ensure they remain secure, discoverable and accessible. Research data management involves storing, accessing and preserving data produced from an investigation over its entire lifecycle. The document recommends using activity-based costing to identify all direct and indirect activities associated with data curation and research data management. It also emphasizes the importance of defining what is being costed and choosing an appropriate costing methodology.
Certifying and Securing a Trusted Environment for Health Informatics Research...Jisc
The document discusses the certification and securing of a trusted environment for health informatics research data at the University of Dundee. It provides an overview of the Health Informatics Centre, its research data management platform, safe haven architecture, and ISO27001 certification. The platform standardizes data extraction and release, adds metadata and quality checks. A safe haven uses pseudonymized data and virtual environments prevent data from leaving. ISO27001 certification provides governance and reduces documentation through standardized information security practices.
Dr. Tito Castillo discusses challenges with data discovery and sharing at University College London Hospitals (UCLH) due to their multiple proprietary clinical systems with undocumented data and data warehouses. To address this, UCLH is taking a standards-based approach using models like DDI and SDMX to document metadata and map their processes. The goal is to enable better data access, sharing, and reuse to support research programmes and new models of care while respecting governance and privacy.
This document provides an overview of a webinar on digital curation and research data management for universities. The webinar covers an introduction to digital curation, the benefits and drivers for research data management, current initiatives in UK universities, and the role of libraries in supporting research data management. Libraries are increasingly involved in developing institutional policies, providing training, and advising researchers on writing data management plans and sharing data. The webinar highlights training opportunities for librarians to develop skills in research data management and digital curation.
A presentation offering an introduction to managing and sharing research data given at the Czech Open Science days as part of the EC-funded FOSTER project.
This document discusses licensing research data for reuse. It begins by providing a scenario where a user has downloaded a dataset but is unsure what they can do with the data due to licensing. It then discusses that licensing is critical to enabling data reuse and citation. It provides information on AusGOAL, the Australian open access and licensing framework, and notes it is recommended for data publishing by ANDS partners. It also includes links to licensing guides and FAQs. In summary, the document emphasizes the importance of data licensing for enabling reuse and outlines Australia's recommended licensing system.
Presentation given at the European Research Council workshop on research data management and sharing in Brussels on 18th-19th September 2014. The presentation covers the benefits and drivers for RDM, points to relevant tools and resources and closes with some open questions for discussion.
The document discusses open data and data sharing, including defining open data, the benefits of open data, overcoming barriers to opening data such as concerns about scooping and sensitive data, best practices for making data open through formats, licensing and description, and the role of research databases and data citation in promoting open data.
An introduction to Research Data Management and Data Management Planning for research managers and administrators. The presentation was given at the Open University on 18th July 2013.
This presentation was provided by Melissa Levine of the University of Michigan during a NISO Virtual Conference on the topic of data curation, held on Wednesday, August 31, 2016
Strand 1: Connecting research and researchers: An introduction to ORCID by Ed...OAbooks
ORCID is an open, non-profit organization that provides a registry of unique researcher identifiers and aims to link researchers to their professional activities such as publications, datasets, and more. The presentation discusses the problems ORCID aims to address like linking researchers across databases and improving discoverability. It outlines ORCID's mission, benefits to the research community, how the ORCID registry works, privacy considerations, integration opportunities, growth since launch, international usage, members, support available, and how to join ORCID.
This document provides guidance on research data management and developing data management plans. It discusses why managing research data is important, including making research easier to conduct, avoiding accusations of fraud or bad science, and getting credit for data produced. The document outlines what is involved in research data management and considerations for sharing and preserving data, such as file formats, documentation, and standards. It emphasizes the importance of data management planning and provides tips on developing plans to meet funder requirements.
Simon Hodson discusses key aspects of open science including open access to research outputs, FAIR data principles, and engaging society. Open science requires addressing technical, funding, skills, and mindset challenges. While data created with public funds should be open by default, legitimate exceptions exist for commercial interests, privacy, and security. Criteria for data appraisal, selection and preservation need input from disciplines. Barriers to data sharing include concerns over misuse and lack of credit, while benefits include advancing research and building institutional reputation. Open science governance is needed to balance openness with other priorities like intellectual property, and define roles and responsibilities among stakeholders.
Presentation given by Sarah Jones at a seminar run by LSHTM on 6th November 2012. http://www.lshtm.ac.uk/newsevents/events/2012/11/developing-data-management-expertise-in-research---half-day-event
Paul Jeffreys - Research Integrity: Institutional ResponsibilityJisc
This document summarizes a presentation given at a research integrity conference about the actions the University of Oxford is taking to meet its responsibilities regarding research data management. The university recognizes data management as important for ensuring research integrity and is coordinating various digital services through developing policies, overseeing data management, addressing funding, and creating a university-wide research data catalogue and repository. While still in early stages, the university aims to provide sustainable data services and ensure long-term access to and integrity of research data.
Management of research data specifically for Engineering and Physical Science. Delivered by Stuart Macdonald at the "Support for Enhancing Research Impact" meeting at the University of Edinburgh on 22 June 2016.
This slide shows the set of task groups established under the aegis of the RDA/NISO Privacy Implications of Research Data Sets Interest Group; it was used during the NISO Symposium held on September 11, 2016 in conjunction with International Data Week events in Denver, Colorado.
The document outlines a 23 Things program for research data management training, which releases weekly activities and has monthly webinars, and provides a calendar of events and list of coordinators for the program at UWA.
Mind the Gap: Reflections on Data Policies and PracticeLizLyon
UKOLN is supported by the Mind the Gap project which reflects on data policies and practices. The document discusses the current state of data practices in institutions, challenges around open science and data sharing, and the need for improved data policies, planning tools, and codes of conduct to help researchers with issues like data storage, sharing, and long-term preservation. It also explores how emerging technologies and areas like genomics, personalized medicine, and citizen science will impact future data practices and policies.
Supporting Research Data Management at the University of StirlingLisa Haddow
The Digital Curation Centre (DCC) provides support to universities to help them manage research data. This includes tools to assess data needs and risks, plan data management, and develop policies. The DCC can help universities develop data management strategies, provide training to researchers, and pilot tools. Its goal is to build research data management capacity across UK higher education. The DCC is working intensively with 18 universities to increase capabilities in these areas over the next year.
Presentation given to EC project officers as part of workshops run by the FOSTER (foster open science) project. The presentation covers the Horizon 2020 open data pilot.
An overview of the LSHTM Research Data Management Policy, outlining the motivations for its introduction, obligations that need to be met and the support available
Presentación de Joy Davidson, Digital Curation Centre (UK) en FOSTER event: Data Management Plan and Social Impact of Research. Universitat Jaume I, 27 mayo 2016
Libraries and Research Data Management – What Works? Lessons Learned from the...LIBER Europe
This presentation by Dr Birgit Schmidt was given at the Scholarly Communication and Research Infrastructures Steering Committee Workshop. The workshop title was Libraries and Research Data Management – What Works?
Providing support and services for researchers in good data governanceRobin Rice
The University of Edinburgh provides support and services to help researchers with good data governance. This includes a research data policy, research data service with various tools across the data lifecycle, and a data safe haven for sensitive data. The research data service offers centralized storage, version control, collaboration tools, and repositories for sharing data openly or long-term retention. Training and outreach aim to educate researchers on topics like data management plans, sensitive data, and GDPR compliance.
This presentation introduced participants to the DC 101 course and was given at the Digital Curation and Preservation Outreach and Capacity Building Workshop in Belfast on September 14-15 2009.
http://www.dcc.ac.uk/events/workshops/digital-curation-and-preservation-outreach-and-capacity-building-workshop
The document discusses open data and data sharing, including defining open data, the benefits of open data, overcoming barriers to opening data such as concerns about scooping and sensitive data, best practices for making data open through formats, licensing and description, and the role of research databases and data citation in promoting open data.
An introduction to Research Data Management and Data Management Planning for research managers and administrators. The presentation was given at the Open University on 18th July 2013.
This presentation was provided by Melissa Levine of the University of Michigan during a NISO Virtual Conference on the topic of data curation, held on Wednesday, August 31, 2016
Strand 1: Connecting research and researchers: An introduction to ORCID by Ed...OAbooks
ORCID is an open, non-profit organization that provides a registry of unique researcher identifiers and aims to link researchers to their professional activities such as publications, datasets, and more. The presentation discusses the problems ORCID aims to address like linking researchers across databases and improving discoverability. It outlines ORCID's mission, benefits to the research community, how the ORCID registry works, privacy considerations, integration opportunities, growth since launch, international usage, members, support available, and how to join ORCID.
This document provides guidance on research data management and developing data management plans. It discusses why managing research data is important, including making research easier to conduct, avoiding accusations of fraud or bad science, and getting credit for data produced. The document outlines what is involved in research data management and considerations for sharing and preserving data, such as file formats, documentation, and standards. It emphasizes the importance of data management planning and provides tips on developing plans to meet funder requirements.
Simon Hodson discusses key aspects of open science including open access to research outputs, FAIR data principles, and engaging society. Open science requires addressing technical, funding, skills, and mindset challenges. While data created with public funds should be open by default, legitimate exceptions exist for commercial interests, privacy, and security. Criteria for data appraisal, selection and preservation need input from disciplines. Barriers to data sharing include concerns over misuse and lack of credit, while benefits include advancing research and building institutional reputation. Open science governance is needed to balance openness with other priorities like intellectual property, and define roles and responsibilities among stakeholders.
Presentation given by Sarah Jones at a seminar run by LSHTM on 6th November 2012. http://www.lshtm.ac.uk/newsevents/events/2012/11/developing-data-management-expertise-in-research---half-day-event
Paul Jeffreys - Research Integrity: Institutional ResponsibilityJisc
This document summarizes a presentation given at a research integrity conference about the actions the University of Oxford is taking to meet its responsibilities regarding research data management. The university recognizes data management as important for ensuring research integrity and is coordinating various digital services through developing policies, overseeing data management, addressing funding, and creating a university-wide research data catalogue and repository. While still in early stages, the university aims to provide sustainable data services and ensure long-term access to and integrity of research data.
Management of research data specifically for Engineering and Physical Science. Delivered by Stuart Macdonald at the "Support for Enhancing Research Impact" meeting at the University of Edinburgh on 22 June 2016.
This slide shows the set of task groups established under the aegis of the RDA/NISO Privacy Implications of Research Data Sets Interest Group; it was used during the NISO Symposium held on September 11, 2016 in conjunction with International Data Week events in Denver, Colorado.
The document outlines a 23 Things program for research data management training, which releases weekly activities and has monthly webinars, and provides a calendar of events and list of coordinators for the program at UWA.
Mind the Gap: Reflections on Data Policies and PracticeLizLyon
UKOLN is supported by the Mind the Gap project which reflects on data policies and practices. The document discusses the current state of data practices in institutions, challenges around open science and data sharing, and the need for improved data policies, planning tools, and codes of conduct to help researchers with issues like data storage, sharing, and long-term preservation. It also explores how emerging technologies and areas like genomics, personalized medicine, and citizen science will impact future data practices and policies.
Supporting Research Data Management at the University of StirlingLisa Haddow
The Digital Curation Centre (DCC) provides support to universities to help them manage research data. This includes tools to assess data needs and risks, plan data management, and develop policies. The DCC can help universities develop data management strategies, provide training to researchers, and pilot tools. Its goal is to build research data management capacity across UK higher education. The DCC is working intensively with 18 universities to increase capabilities in these areas over the next year.
Presentation given to EC project officers as part of workshops run by the FOSTER (foster open science) project. The presentation covers the Horizon 2020 open data pilot.
An overview of the LSHTM Research Data Management Policy, outlining the motivations for its introduction, obligations that need to be met and the support available
Presentación de Joy Davidson, Digital Curation Centre (UK) en FOSTER event: Data Management Plan and Social Impact of Research. Universitat Jaume I, 27 mayo 2016
Libraries and Research Data Management – What Works? Lessons Learned from the...LIBER Europe
This presentation by Dr Birgit Schmidt was given at the Scholarly Communication and Research Infrastructures Steering Committee Workshop. The workshop title was Libraries and Research Data Management – What Works?
Providing support and services for researchers in good data governanceRobin Rice
The University of Edinburgh provides support and services to help researchers with good data governance. This includes a research data policy, research data service with various tools across the data lifecycle, and a data safe haven for sensitive data. The research data service offers centralized storage, version control, collaboration tools, and repositories for sharing data openly or long-term retention. Training and outreach aim to educate researchers on topics like data management plans, sensitive data, and GDPR compliance.
This presentation introduced participants to the DC 101 course and was given at the Digital Curation and Preservation Outreach and Capacity Building Workshop in Belfast on September 14-15 2009.
http://www.dcc.ac.uk/events/workshops/digital-curation-and-preservation-outreach-and-capacity-building-workshop
This document provides an introduction to research data management for geoscience PhD students. It defines research data and different data types. It discusses the importance of managing data throughout its lifecycle for efficient and valid research. It outlines funder requirements, university policies, and activities involved in good research data management like data planning, documentation, storage, sharing and preservation.
Advanced Analytics and Machine Learning with Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/32c6TnG
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spent most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Attend this webinar and learn:
- How data virtualization can accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- How popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc. integrate with Denodo
- How you can use the Denodo Platform with large data volumes in an efficient way
- About the success McCormick has had as a result of seasoning the Machine Learning and Blockchain Landscape with data virtualization
This document provides an overview of research data management best practices. It discusses developing a data management plan, data storage and backup options, data sharing and preservation, and compliance with funder policies. Key points covered include writing a DMP, using structured file formats, having robust backup processes, choosing a trusted repository, and making data FAIR (findable, accessible, interoperable, and reusable).
A Logical Architecture is Always a Flexible Architecture (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3joZa0a
The current data landscape is fragmented, not just in location but also in terms of processing paradigms: data lakes, IoT architectures, NoSQL, and graph data stores, SaaS applications, etc. are found coexisting with relational databases to fuel the needs of modern analytics, ML, and AI. The physical consolidation of enterprise data into a central repository, although possible, is both expensive and time-consuming. A logical data warehouse is a modern data architecture that allows organizations to leverage all of their data irrespective of where the data is stored, what format it is stored in, and what technologies or protocols are used to store and access the data.
Watch this session to understand:
- What is a logical data warehouse and how to architect one
- The benefits of logical data warehouse – speed with agility
- Customer use case depicting logical architecture implementation
- Organizations are facing challenges with rapidly growing data volumes, increasing compliance needs, and fixed budgets. On average, 70% of data is unused after 90 days.
- Lab Data Management Suite offers 4 ranges of products to address these challenges: 1) data analysis and decision making, 2) online data backup, 3) archiving solutions, and 4) secure document sharing solutions.
- The suite provides automated, scalable data management to reduce costs, improve compliance, and ensure business continuity and disaster recovery through reliable data backup and recovery.
Reagan Moore, UNC-RENCI; Policy-based Data Management; RDAP11 Summit
The 2nd Research Data Access and Preservation (RDAP) Summit
An ASIS&T Summit
March 31-April 1, 2011 Denver, CO
In cooperation with the Coalition for Networked Information
http://asist.org/Conferences/RDAP11/index.html
TDWI Checklist Report: Active Data ArchivingRainStor
The document discusses best practices for active data archiving. It recommends embracing modern archiving platforms and practices to address problems with traditional archives. A modern archive should serve compliance needs through immutable, auditable data storage, while also enabling analytics through online access. It should scale to large volumes of structured and unstructured data from various sources and support roles-based security and multi-tenancy. The archive's primary tier should be a robust database for online querying and exploration of archived data.
Linking HPC to Data Management - EUDAT Summer School (Giuseppe Fiameni, CINECA)EUDAT
EUDAT and PRACE joined forces to help research communities gain access to high quality managed e-Infrastructures whose resources can be connected together to enable cross-utilization use cases and make them accessible without any technical barrier. The capability to couple data and compute resources together is considered one of the key factors to accelerate scientific innovation and advance research frontiers. The goal of this session was to present the EUDAT services, the results of the collaboration activity achieved so far and delivers a hands-on on how to write a Data Management Plan or DMP. The DMP is a useful instrument for researchers to reflect on and communicate about the way they will deal with their data. It prompts them to think about how they will generate, analyse and share data during their research project and afterwards.
Visit: https://www.eudat.eu/eudat-summer-school
The document discusses sharing research data through open data platforms. It describes the CGIAR as uniquely positioned to collect agricultural data worldwide and argues that most CGIAR data should be archived and shared to increase its value. However, data archiving across CGIAR centers is currently poor. The document then discusses using the Dataverse platform to improve data sharing. Dataverse allows researchers to publish, share, cite, and analyze data. It also facilitates making data available while giving credit to data authors and institutions.
[DSC Europe 23] Milos Solujic - Data Lakehouse Revolutionizing Data Managemen...DataScienceConferenc1
We will dive into modern data management approaches that have become prevalent and popular across many industries, built on top of good old data lakes: Lakehouse. Here are some of the most common problems that are being solved with this novel approach: Data Silos Demolished: Discover how organizations are breaking down data silos that have plagued them for decades, unifying structured and unstructured data from diverse sources. Inefficient Data Processing: We'll unveil real-world examples of how inefficient data processing can grind productivity to a halt and explore how Data Lakehouses provide a powerful solution while improving governance and security. Real-time Analytics: Learn how modern businesses are striving to achieve real-time analytics and the role Data Lakehouses play in achieving this. Have one data copy that will serve BI, Reporting, and ML workloads
A presentation given as part of the DC101 training course run by the DCC at Oxford University in June 2010. The course provided data management guidance for researchers.
Advanced Analytics and Machine Learning with Data Virtualization (India)Denodo
Watch full webinar here: https://bit.ly/3dMN503
Advanced data science techniques, like machine learning, have proven an extremely useful tool to derive valuable insights from existing data. Platforms like Spark, and complex libraries for R, Python, and Scala put advanced techniques at the fingertips of the data scientists. However, these data scientists spend most of their time looking for the right data and massaging it into a usable format. Data virtualization offers a new alternative to address these issues in a more efficient and agile way.
Watch this session to learn how companies can use data virtualization to:
- Create a logical architecture to make all enterprise data available for advanced analytics exercise
- Accelerate data acquisition and massaging, providing the data scientist with a powerful tool to complement their practice
- Integrate popular tools from the data science ecosystem: Spark, Python, Zeppelin, Jupyter, etc
Virtual Research Environments supporting tailor-made data management service...Blue BRIDGE
Presented by Pasquale Pagano of CNR at the BlueBRIDGE Workshop at SeaTech Week 2016 in Brest, France. http://www.bluebridge-vres.eu/events/join-bluebridge-10th-biennial-sea-tech-week-brest-france
Data grids are an emerging technology that enables the formation of sharable collections from data distributed across multiple storage resources. The integrated Rule Oriented Data System (iRODS) is a data grid developed by the DICE Center at UNC-CH. The iRODS data grid enforces management policies that control properties of the collection. Examples of policies include retention, disposition, distribution, replication, metadata extraction, time-dependent access controls, data processing, data redaction, and integrity checking. Policies can be defined that automate administrative functions (file migration and replication) and that validate assessment criteria (authenticity, integrity, chain of custody). iRODS is used to build data sharing environments, digital libraries, and preservation environments. The iRODS data grid is used at UNC-CH to support the Carolina Digital Repository, the LifeTime Library for the School of Information and Library Science, data grids for the Renaissance Computing Institute (RENCI), collaborations within North Carolina, and both national and international data sharing. At RENCI, the TUCASI data grid supports shared collections between UNC-CH, Duke, and NCSU. The RENCI data grid is federated with ten other data grids including the National Climatic Data Center, the Texas Advanced Computing Center data grid, and the Ocean Observatories Initiative data grid. International applications include the CyberSKA Square Kilometer Array for radio astronomy and the French National Institute for Nuclear Physics and Particle Physics. The collections that are assembled may contain hundreds of millions of files, and petabytes of data. A specific goal is the integration of institutional repositories with the national data infrastructure that is being assembled under the NSF DataNet program. The software is available as an open source distribution from http://irods.diceresearch.org.
Rebecca Grant - Archiving and Digital Preservation (Figshare Fest)dri_ireland
Presentation given by Rebecca Grant, Digital Archivist with Digital Repository of Ireland, part of a workshop on Digital Archiving and Digital Preservation held as part of Figshare Fest in London, May 12th 2016. Figshare is an online digital repository where researchers can preserve and share their research outputs, including figures, datasets, images, and videos. Its annual Figshare Fest is a chance to gather together institutional clients, advocates and friends to talk about open research.
Site up an open access-ICAR
Institutional Repository-Hardware, Software, Policies and Personnel.
ICAR Initiatives
Under NATP Project – Integrated National Agricultural Resources Information System INARIS (Rai et. Al., 2007). A Central Data warehouse (CWD) of agricultural resources was established at IASRI
This project having collaborations with 13 other organizations of ICAR.
In this view 13 different data marts were designed.
This Project was available under this link (http://agdw.iasri.res.in)
My outlook Country should have agri-search engine
Agri-Search Engine should be developed in country to aggregate information from the internet and provide it to farmers in meaningful manner through using ICT tools.
Agri-Search Engine be coordinated with Govt. of India’s Agricultural Websites to monitor each website per day.
Similar to Research Data Management, Challenges and Tools - Per Öster (20)
Research Data in an Open Science World - Prof. Dr. Eva Mendez, uc3mLEARN Project
This document summarizes a presentation by Prof. Eva Méndez on research data in an open science world from the perspective of a young EU university. The presentation discusses the changing landscape of open science brought about by exponential data growth, new technologies, and public demands for transparency. It addresses challenges for research data management, including skills gaps and lack of standards. The presentation also examines roles and responsibilities for universities to support open science, such as developing infrastructures and policies to incentivize data sharing and changing research cultures. Overall, the document outlines both the opportunities and challenges of open science for research data and universities.
This document summarizes a presentation about using the LEARN project's research data management policy and guidance. The LEARN project involved 5 partners across Europe working from 2015-2017 to develop best practices for RDM. It conducted workshops, published case studies and a toolkit. The presentation discusses developing an RDM policy, including understanding the progression from taboos to principles to policies to rules. It provides an example outline for a model RDM policy covering aspects like responsibilities and validity. The goal is to produce guidance that research institutions can tailor to their own needs to enhance coordination and alignment of RDM practices.
The webinar presentation summarizes the LEARN Toolkit project which developed best practices for research data management. It includes 23 case studies organized into 8 sections covering topics like policies, advocacy, costs, roles and responsibilities. The project produced a model research data management policy and guidance document to help institutions develop their own policies. It engaged stakeholders through workshops around Europe and Latin America to align policies and terminology. The materials from the project, including the model policy, are published in the LEARN Toolkit which aims to support research organizations in improving their research data management.
Presented by Ms Diane Quarless, Director, ECLAC subregional headquarters for the Caribbean, at the LEARN Caribbean Research Data Workshop. http://learn-rdm.eu/en/workshops/eclac-mini-workshops/3rd-mini-workshop
Presented by Ms Bernadette Lewis, Secretary General, Caribbean Telecommunications Union at the LEARN Caribbean Research Data Workshop. http://learn-rdm.eu/en/workshops/eclac-mini-workshops/3rd-mini-workshop
Open Data in a Big World by Fernando Ariel López LEARN Project
Este documento presenta los principios de datos abiertos desarrollados por un grupo de trabajo de cuatro organizaciones científicas internacionales. Describe las responsabilidades de científicos, instituciones, editores, financiadores, asociaciones profesionales y bibliotecas para implementar y promover la apertura de datos. También cubre los límites a la apertura de datos y prácticas como la citación, interoperabilidad y reutilización de datos abiertos.
Este documento describe los objetivos y tipos de datos de un centro de datos. El centro busca recopilar y sistematizar datos existentes en Colombia para mejorar la calidad e acceso a la información. Provee acceso libre a datos administrativos de entidades gubernamentales y de investigación realizada por profesores. El centro también busca documentar los metadatos de acuerdo a estándares internacionales y dar a conocer los datos a través de boletines y su plataforma.
Gestion de datos para la investigacion: el caso peruano by Edward Mezones, Su...LEARN Project
Este documento describe la gestión de datos para la investigación en Perú. Explica que el gobierno peruano ha promovido la apertura de datos gubernamentales a través de portales de datos abiertos y la Ley 30035, que establece un repositorio nacional digital de acceso abierto llamado ALICIA. También describe cómo SUSALUD, la agencia de seguro de salud de Perú, ha integrado su repositorio de datos abiertos a ALICIA para facilitar la investigación en salud.
Jennifer Schaus and Associates hosts a complimentary webinar series on The FAR in 2024. Join the webinars on Wednesdays and Fridays at noon, eastern.
Recordings are on YouTube and the company website.
https://www.youtube.com/@jenniferschaus/videos
Bharat Mata - History of Indian culture.pdfBharat Mata
Bharat Mata Channel is an initiative towards keeping the culture of this country alive. Our effort is to spread the knowledge of Indian history, culture, religion and Vedas to the masses.
karnataka housing board schemes . all schemesnarinav14
The Karnataka government, along with the central government’s Pradhan Mantri Awas Yojana (PMAY), offers various housing schemes to cater to the diverse needs of citizens across the state. This article provides a comprehensive overview of the major housing schemes available in the Karnataka housing board for both urban and rural areas in 2024.
Presentation by Julie Topoleski, CBO’s Director of Labor, Income Security, and Long-Term Analysis, at the 16th Annual Meeting of the OECD Working Party of Parliamentary Budget Officials and Independent Fiscal Institutions.
Presentation by Rebecca Sachs and Joshua Varcie, analysts in CBO’s Health Analysis Division, at the 13th Annual Conference of the American Society of Health Economists.
How To Cultivate Community Affinity Throughout The Generosity JourneyAggregage
This session will dive into how to create rich generosity experiences that foster long-lasting relationships. You’ll walk away with actionable insights to redefine how you engage with your supporters — emphasizing trust, engagement, and community!
The Power of Community Newsletters: A Case Study from Wolverton and Greenleys...Scribe
YOU WILL DISCOVER:
The engaging history and evolution of Wolverton and Greenleys Town Council's newsletter
Strategies for producing a successful community newsletter and generating income through advertising
The decision-making process behind moving newsletter design from in-house to outsourcing and its impacts
Dive into the success story of Wolverton and Greenleys Town Council's newsletter in this insightful webinar. Hear from Mandy Shipp and Jemma English about the newsletter's journey from its inception to becoming a vital part of their community's communication, including its history, production process, and revenue generation through advertising. Discover the reasons behind outsourcing its design and the benefits this brought. Ideal for anyone involved in community engagement or interested in starting their own newsletter.
Research Data Management, Challenges and Tools - Per Öster
1. CSC – Suomalainen tutkimuksen, koulutuksen, kulttuurin ja julkishallinnon ICT-osaamiskeskus
Research Data Management,
Challenges andTools
Per Öster, CSC – IT Center for Science Ltd
2. • Drivers
o Number of devices
o Number of communicating apps
o Number of users
2.6.20172
10% of UK
power
consumption
due to ICT
1/3
network
1/3
devices
1/3
datacentres
Amount of data
3. 6/2/173
Analysis
Publication
ReviewConceptualisation
Data
gathering
Open
access
Scientific
blogs Collaborative
bibliographies
Alternative
Reputation
systems
Citizens
science
Open
code
Open
workflows
Open
annotation
Open
data
Pre-
print
Data-
intensive
2!
Sci-
starter.com
Runmycode.
org
ArXiv
Roar.eprints.
org
Impact Story
Altmetric.com
Mendeley.com
Academia.edu
Researchgate.com
Openannotation.org
Datadryad.org
Myexperiment.org
Figshare.com
An#emerging#
ecosystem#of#
services#and#
standards#
It's real!
The DCC Curation
Lifecycle Model
Description and
Representation Information
Preservation Planning
Community Watch and
Participation
Curate and Preserve
Conceptualise
Create or Receive
Appraise and Select
Ingest
Preservation Action
Store
Access, Use and Reuse
Transform
Assign administrative, descriptive, technical, structural and preservation metadata, using appropriate standards, to ensure adequate description and control over the long-term. Collect and assign representation information required to understand
and render both the digital material and the associated metadata.
Plan for preservation throughout the curation lifecycle of digital material. This would include plans for management and administration of all curation lifecycle actions.
Maintain a watch on appropriate community activities, and participate in the development of shared standards, tools and suitable software.
Be aware of, and undertake management and administrative actions planned to promote curation and preservation throughout the curation lifecycle.
Conceive and plan the creation of data, including capture method and storage options.
Create data including administrative, descriptive, structural and technical metadata. Preservation metadata may also be added at the time of creation.
Receive data, in accordance with documented collecting policies, from data creators, other archives, repositories or data centres, and if required assign appropriate metadata.
Evaluate data and select for long-term curation and preservation. Adhere to documented guidance, policies or legal requirements.
Transfer data to an archive, repository, data centre or other custodian. Adhere to documented guidance, policies or legal requirements.
Undertake actions to ensure long-term preservation and retention of the authoritative nature of data. Preservation actions should ensure that data remains authentic, reliable and usable while maintaining its integrity. Actions include data cleaning,
validation, assigning preservation metadata, assigning representation information and ensuring acceptable data structures or file formats.
Store the data in a secure manner adhering to relevant standards.
Ensure that data is accessible to both designated users and reusers, on a day-to-day basis. This may be in the form of publicly available published information. Robust access controls and authentication procedures may be applicable.
Create new data from the original, for example
- By migration into a different format.
- By creating a subset, by selection or query, to create newly derived results, perhaps for publication.
www.dcc.ac.uk
info@dcc.ac.uk
The Curation Lifecycle
The DCC Curation Lifecycle Model provides a graphical high level overview of the stages required for successful curation and preservation of data from initial conceptualisation or receipt. The model can be used to plan activities within an organisation or consortium to
ensure that all necessary stages are undertaken, each in the correct sequence. The model enables granular functionality to be mapped against it; to define roles and responsibilities, and build a framework of standards and technologies to implement. It can help with
the process of identifying additional steps which may be required, or actions which are not required by certain situations or disciplines, and ensuring that processes and policies are adequately documented.
Data, any information in binary digital form, is at the centre of the Curation Lifecycle. This includes:
- Simple Digital Objects are discrete digital items; such as textual files, images or sound files, along with their related identifiers and metadata.
- Complex Digital Objects are discrete digital objects, made by combining a number of other digital objects, such as websites.
Structured collections of records or data stored in a computer system.
Full Lifecycle Actions
Sequential Actions
Data (Digital Objects or Databases)
Occasional Actions
Dispose
Reappraise
Migrate
Dispose of data, which has not been selected for long-term curation and preservation in accordance with documented policies, guidance or legal requirements. Typically data may be transferred to another archive, repository, data centre or
other custodian. In some instances data is destroyed. The data’s nature may, for legal reasons, necessitate secure destruction.
Return data which fails validation procedures for further appraisal and reselection.
Migrate data to a different format. This may be done to accord with the storage environment or to ensure the data’s immunity from hardware or software obsolescence.
Digital Objects
Databases
5. Secure Compute Clouds
Supporting sample
logistics
• Federated Authentication
• Authorization
• Dataset registry
• Data transfer hub
• Policy and Legal Framework
Services and
Coordination
High speed encrypted
data transfer
GridFTP/Globus/Aspera
Secure data access remote API
( GA4GH )
Sequencing centers
Data
Users
EGA
at
Data Archiving
Bringing users
to data
Data Generation
Managing Access
Data Owner
Data Access Agreement
Data Access Committee
Data Request
Authorization Management Tools
( EGA and CSC REMS )
5
6. From field measurements to open data
2.6.20176
Questions:
Sensitive data?
Requirements on Authentication and authorization?
7. Instrument
Measuring PCs:
Raw data
at the stations
File servers at
stations:
Raw data and
field diaries,
cal documents
File servers
in Helsinki:
Raw and
intermediate
data,
documents,
scripts
SMEAR database:
Processed data
in Helsinki
ICOS, EBAS,...
databases:
Near real time
and processed data
outside UH
Routine data processing =
(- unit conversion)
- calibration correction
- quality check, gapfilling
- averaging over space or time
SMEAR
data flow
A/D conversion
unit conversion
IDA (CSC data
service):
Raw data &
document archive,
database datasets
Field
documentation
Researchers,
Data processing
server
Feedback on
data quality
Metadata
Metadata
Metadata
2.6.20177
https://avaa.tdata.fi/web/smart/smear
8. Support in All Phases of Research Process
20168
Plan
Customer Portal
Experts
Guides
Websites
Training
Service Desk
Produce
& Collect
Data
International
resources
Modelling
Software
Supercomputers
Analyse
Cloud Services
Training
Data science
Computing
Software
Store
B2SAFE
B2SHARE
HPC Archive
IDA
Databases
Research long-
term preservation
(LTP)
Share &
Publish
AVAA
B2DROP
B2SHARE
Databank
Etsin
Funet FileSender
11. ARTICLE 2: DISCLAIMER
1. The Service is provided "as is" and the Provider disclaims any and all
representations and warranties, whether express or implied, including;- but
not limited to;- implied warranties of title, merchantability, fitness for any
particular purpose or non-infringement. The Provider does not promise any
specific results, effects or outcome from the use of the Service.
2. …
3. The Provider reserves the right to change, reduce, interrupt or discontinue
the Service or parts of it at any time.
4. No one has a right to use the Service; the Provider reserves the right to
exclude certain Users.
12. Are the commercial services sufficient?
• Nice complement but can not serve as the fundamental infrastructure for research
data of national and international interest
• Need for publicly funded and operated infrastructure
2.6.201712
e-Science Data Factory
13. EUDAT CDI
“The EUDAT Collaborative Data
Infrastructure is a defined data model
and a set of technical standards and
policies adopted by European research
data centres and community data
repositories to create a single European
e-infrastructure of interoperable data
services.”
“To date, over 20 major European research organizations,
data and computing centres have signed an agreement to
sustain the EUDAT – pan European collaborative data
infrastructure for the next 10 years giving the birth to
the EUDAT Collaborative Data Infrastructure”
www.eudat.eu
14.
15. Findable
– assign persistent IDs, provide rich metadata, register in a
searchable resource...
Accessible
– Retrievable by their ID using a standard protocol, metadata
remain accessible even if data aren’t...
Interoperable
– Use formal, broadly applicable languages, use standard
vocabularies, qualified references...
Reusable
– Rich, accurate metadata, clear licences, provenance, use of
community standards...
www.force11.org/group/fairgroup/fairprinciples
EUDAT is FAIR
17. B2FIND: multi-disciplinary metadata catalogue
now: common metadata catalogue, harvesting across
all CDI data, single point for data discoverability;
in development: aim to improve with agreed basic
metadata for all data objects.
B2HANDLE: policy-based prefix & PID management
now: common PID mechanism across all CDI data;
in development: aim to improve with agreed common
schema and behaviour.
B2SHARE: research data repository
now: full, tailored metadata support for data deposits.
B2SAFE: policy-driven data management
in development: aim to introduce a common data
model promoting metadata extraction and processing.
EUDAT & Findable
F
18. B2STAGE: data staging service
B2SHARE: research data repository
B2SAFE: policy-driven data management
now: EUDAT presents data through common Internet protocols and
APIs, http and gridftp;
in development: aim to improve with a single http API for all services
and data.
EUDAT & Accessible
A
19. B2HANDLE: policy-based prefix & PID management
now: common PID mechanism across all CDI data;
in development: aim to improve with agreed common
schema and behaviour.
B2STAGE: data staging service
B2SHARE: research data repository
B2SAFE: policy-driven data management
in development: single http API for all services and data
(interoperability of data services, if not data!).
B2FIND: multi-disciplinary metadata catalogue
in development: agreed basic metadata for all data
objects (a degree of metadata interop).
EUDAT & Interoperable
I
20. B2SHARE: research data repository
B2SAFE: policy-driven data management
now: encourage use of CC BY v 3 as common open data licence;
encourage open formats where we have any influence.
EUDAT & Reusable
R
21. Acknowledgment
• Tommi Nyrönen, ELIXIR-Finland Head of Node
• Mikael Linden, CSC
• TimmoVesala, INAR RI, Helsinki University
• Damien Lecarpentier, EUDAT Project Director
2.6.201721