This document summarizes a workshop on data management. It outlines the typical research lifecycle including proposal planning, project start up, data collection, analysis, sharing, and end of project. It discusses support for researchers within areas like data mining, curation, and preservation. It also discusses support from outside through infrastructure, policy, and best practices. Finally, it identifies 9 key skills gaps for librarians in advising researchers on data management tasks.
This document provides an overview of research data management and what it will mean for the School of Engineering at the University of Lincoln. It discusses the need to manage research data due to funder and institutional requirements. It also describes benefits like reducing duplicated work and supporting collaboration. The document then provides examples of research data sharing benefits. It outlines support available, including from data archives, the Digital Curation Centre, JISC and UKOLN. Finally, it discusses JISC's Managing Research Data program and projects around research data management infrastructure and planning.
Scientific discovery and innovation in an era of data-intensive science
William (Bill) Michener, Professor and Director of e-Science Initiatives for University Libraries, University of New Mexico; DataONE Principal Investigator
The scope and nature of biological, environmental and earth sciences research are evolving rapidly in response to environmental challenges such as global climate change, invasive species and emergent diseases. Scientific studies are increasingly focusing on long-term, broad-scale, and complex questions that require massive amounts of diverse data collected by remote sensing platforms and embedded environmental sensor networks; collaborative, interdisciplinary science teams; and new tools that promote scientific data preservation, discovery, and innovation. This talk describes the challenges facing scientists as they transition into this new era of data intensive science, presents current solutions, and lays out a roadmap to the future where new information technologies significantly increase the pace of scientific discovery and innovation.
This document provides an overview of data mining. It defines data mining as extracting meaningful information from large data sets. It describes the typical data mining process, which includes problem definition, data gathering/preparation, model building/evaluation, and knowledge deployment. It also outlines several common data mining techniques like neural networks, clustering, decision trees, and support vector machines. Finally, it discusses applications of data mining in business, science, security, marketing, and spatial data analysis.
Where is the opportunity for libraries in the collaborative data infrastructure?LIBER Europe
Presentation by Susan Reilly at Bibsys2013 on the opportunties for libraries and their role in the collaborative data infrastructure. Looks at data sharing, authentication, preservation and advocacy.
This document discusses challenges in scientific data management and proposes a framework called SciFrame to address them. It notes issues like dealing with large volumes of data from diverse sources, balancing implicit and explicit descriptions, and preserving information as data moves between acquisition, transformation, and publication. SciFrame aims to improve data discovery, extraction, and transfer through semantic-aware schemas, workflow inspection, adaptive databases, and database descriptors that consider storage approaches.
Presentation Title: Grand Challenges and Big Data: Implications for Public Participation in Scientific Research
Presenter: William Michener, Professor and PI/Director of DataONE, University Libraries, University of New Mexico
This document provides an overview of research data management and what it will mean for the School of Engineering at the University of Lincoln. It discusses the need to manage research data due to funder and institutional requirements. It also describes benefits like reducing duplicated work and supporting collaboration. The document then provides examples of research data sharing benefits. It outlines support available, including from data archives, the Digital Curation Centre, JISC and UKOLN. Finally, it discusses JISC's Managing Research Data program and projects around research data management infrastructure and planning.
Scientific discovery and innovation in an era of data-intensive science
William (Bill) Michener, Professor and Director of e-Science Initiatives for University Libraries, University of New Mexico; DataONE Principal Investigator
The scope and nature of biological, environmental and earth sciences research are evolving rapidly in response to environmental challenges such as global climate change, invasive species and emergent diseases. Scientific studies are increasingly focusing on long-term, broad-scale, and complex questions that require massive amounts of diverse data collected by remote sensing platforms and embedded environmental sensor networks; collaborative, interdisciplinary science teams; and new tools that promote scientific data preservation, discovery, and innovation. This talk describes the challenges facing scientists as they transition into this new era of data intensive science, presents current solutions, and lays out a roadmap to the future where new information technologies significantly increase the pace of scientific discovery and innovation.
This document provides an overview of data mining. It defines data mining as extracting meaningful information from large data sets. It describes the typical data mining process, which includes problem definition, data gathering/preparation, model building/evaluation, and knowledge deployment. It also outlines several common data mining techniques like neural networks, clustering, decision trees, and support vector machines. Finally, it discusses applications of data mining in business, science, security, marketing, and spatial data analysis.
Where is the opportunity for libraries in the collaborative data infrastructure?LIBER Europe
Presentation by Susan Reilly at Bibsys2013 on the opportunties for libraries and their role in the collaborative data infrastructure. Looks at data sharing, authentication, preservation and advocacy.
This document discusses challenges in scientific data management and proposes a framework called SciFrame to address them. It notes issues like dealing with large volumes of data from diverse sources, balancing implicit and explicit descriptions, and preserving information as data moves between acquisition, transformation, and publication. SciFrame aims to improve data discovery, extraction, and transfer through semantic-aware schemas, workflow inspection, adaptive databases, and database descriptors that consider storage approaches.
Presentation Title: Grand Challenges and Big Data: Implications for Public Participation in Scientific Research
Presenter: William Michener, Professor and PI/Director of DataONE, University Libraries, University of New Mexico
This presentation discusses managing research data through the data life cycle. It begins with an overview of the research life cycle and embedding the data life cycle within it. Key aspects of data management are then covered, including why manage data, ethical and legal issues, requirements for data sharing and retention, and creating a data management plan. The rest of the presentation delves into each stage of the data life cycle, providing best practices for data collection, organization, security, storage, documentation, processing, analysis, and long-term preservation or sharing. File formats, metadata, repositories, and bibliographic resources are also addressed.
The Johns Hopkins University Data Management Services conducted an investigation into its first year of operation to better understand how to develop services from a social science perspective. They found that (1) data curation needs were emerging as a "social movement" in response to digital data production, (2) policies from funders like the NSF were helping to formalize data sharing requirements, and (3) academic libraries were starting to expand services to support data management and curation. Based on these trends and investigating university cultures, the JHU Data Management Services developed consulting, archiving, and planning support services tailored to researchers' data practices.
The document provides information about MANTRA, a free online course for research data management created by the University of Edinburgh. MANTRA teaches best practices for managing research data through open educational modules aligned with the research data lifecycle. It is available for reuse and repurposing under an open license. The course covers topics like data planning, organization, documentation, storage, security, and sharing.
This document provides an introduction and overview of the DBM630: Data Mining and Data Warehousing course. It outlines the course syllabus, textbooks, assessment tasks, schedule, prerequisites, and provides a high-level introduction to data mining and data warehousing concepts including definitions, processes, applications and evolution of database technologies.
This document provides an introduction and overview of the DBM630: Data Mining and Data Warehousing course. It outlines the course syllabus, textbooks, assessment tasks, schedule, prerequisites, and introduces concepts related to data mining and data warehousing including definitions, processes, applications, and evolution of database technology. The goal of the course is to teach students about data warehousing, data mining techniques such as association rule mining, classification, clustering, and current trends in the field.
Libby Bishop, Ethics Of Data Sharing Ncess Jun 09 Finala.carusi
This document summarizes an overview of ethical frameworks for sharing and reusing qualitative research data presented at a workshop. It discusses the role of archives in facilitating ethical data sharing and building trust. Formal procedures for sharing confidential research data, such as obtaining informed consent and restricting access, are described. The need to consider duties to others beyond direct research participants in the ethical debate is also highlighted.
The document is a chapter from a textbook on data mining written by Akannsha A. Totewar, a professor at YCCE in Nagpur, India. It provides an introduction to data mining, including definitions of data mining, the motivation and evolution of the field, common data mining tasks, and major issues in data mining such as methodology, performance, and privacy.
SEAD is a NSF DataNet project that aims to provide cyberinfrastructure for long tail data in sustainability science research. It develops tools for active and social curation of data including an Active Curation Repository (ACR) and VIVO profiles. It also creates a Virtual Archive to facilitate long-term access and preservation of datasets across multiple institutional repositories. The presentation provides an overview of SEAD's approach and highlights pilots with the National Center for Earth Surface Dynamics, including ingesting their data collections into the ACR and Virtual Archive and building a social network in VIVO.
I shall provide a summary of JISC work in the area of ‘Big Data’. My primary focus will be on how to manage the huge amount of research data produced in UK Universities. I shall cover the history of JISC interventions to improve research data management and look at next steps. I shall touch on some other areas of work like ‘Digging into Data’ and web archiving which also deal with ‘big data’.
1. Find all frequent itemsets of length 1 by scanning the database to count item occurrences.
2. Iteratively generate candidate itemsets of length k from frequent itemsets of length k-1, and prune unpromising candidates using the Apriori property.
3. Scan the database to determine truly frequent itemsets.
4. Generate association rules from frequent itemsets by adding items to the antecedent and consequent of rules if they meet minimum confidence.
This document discusses knowledge discovery in databases (KDD) through the LON-CAPA online educational system. [1] It defines KDD and data mining, describing the tasks, methods, and applications of KDD. [2] The goals are to obtain predictive models of students, help students and instructors use resources more effectively, and provide information to increase student learning. [3] It then discusses the KDD process and data mining methods like classification, clustering, and dependency modeling that can be applied to discover knowledge from educational data.
The document discusses the growing trend of big data and how cloud storage is a viable option for enterprise data storage needs. It notes that while cloud storage adoption has been slow, offerings continue to mature to handle larger data volumes, varieties, and velocities. The document recommends that organizations prepare their storage environments, evaluate emerging big data solutions, and rationalize their data to take advantage of next generation cloud-based storage architectures optimized for big data.
The flexibility of Apache Hadoop is one of its biggest assets – enabling businesses to generate value from data that was previously considered too expensive to be stored and processed in traditional databases – but also results in Hadoop meaning different things to different people. In this session 451 Research’s Matt Aslett will explore the impact that Hadoop is having on the traditional data processing landscape, examining the expanding ecosystem of vendors and their relationships with Apache Hadoop, investigating the increasing variety of Hadoop use-cases, and exploring adoption trends around the world.
Data Management for Librarians: An IntroductionGarethKnight
The document provides an introduction to data management for librarians, outlining key concepts such as the research data lifecycle, challenges in managing digital data over time, best practices for organizing, documenting, and storing data, and resources for data management support. Common problems include difficulty locating, accessing, and understanding data in the long run without proper planning and preservation strategies. The role of librarians is to educate researchers on best practices and provide support and training resources.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
CNI Fall 2011 Meeting Presentation Margaret Hedstrom & Robert McDonald (Dec. ...SEAD
SEAD is a new NSF-funded project that aims to provide sustainable data services for sustainability science research. It will integrate existing technologies and tools to address the needs of researchers working on "long tail" sustainability problems. SEAD is in its initial phase of developing prototypes and will not be ready to accept data until after October 2012. It is a collaboration between researchers at the University of Michigan, Indiana University, University of Illinois, and Rensselaer Polytechnic Institute.
Research Data Management for Researchers: Module 1: Intro to Data, Metadata a...Glen Newton
This document provides an introduction to research data management. It defines key concepts like research, research data, and the research data lifecycle. It discusses the importance of data sharing and outlines benefits such as enabling new research, reducing duplication, and providing credit to researchers. The document notes that most research data disappears over time unless properly managed. It also explains that research data can be complex with multiple researchers, data types, formats and standards involved. Metadata is described as important data about data. The challenges of preserving complex and transformed data through archiving are also covered.
A sponsored supplement produced for Jisc on how researchers can cope with the data deluge of modern research techniques. Published by Times Higher Education on 25 November 2009
This document provides a summary of the Libra 2.x project to replace the University of Virginia's institutional repository software. It describes the approval of the project in November 2014 due to limitations of the original software. An analysis in 2015 recommended adopting Penn State's Sufia platform for open access and ETD repositories and Harvard's Dataverse for a data repository. The document outlines subsequent work by a working group on usability testing, policies, configuration, and technical setup for the Libra Data repository, which went live in January 2016.
Virginia Data Management Bootcamp: Building the Research Data Community of Pr...Sherry Lake
This document summarizes the Virginia Data Management Bootcamp, a collaborative data education initiative held annually since 2013 among several Virginia universities. It provides details on the planning, logistics, content, and assessments of the bootcamp. According to participant feedback, the hands-on sessions were most useful but some topics could have been covered in more depth. Organizers aim to expand participation to more institutions and offer additional workshops throughout the year, as well as biennial large-scale collaborations and other collaborative efforts to support the growing Virginia data management community of practice.
This presentation discusses managing research data through the data life cycle. It begins with an overview of the research life cycle and embedding the data life cycle within it. Key aspects of data management are then covered, including why manage data, ethical and legal issues, requirements for data sharing and retention, and creating a data management plan. The rest of the presentation delves into each stage of the data life cycle, providing best practices for data collection, organization, security, storage, documentation, processing, analysis, and long-term preservation or sharing. File formats, metadata, repositories, and bibliographic resources are also addressed.
The Johns Hopkins University Data Management Services conducted an investigation into its first year of operation to better understand how to develop services from a social science perspective. They found that (1) data curation needs were emerging as a "social movement" in response to digital data production, (2) policies from funders like the NSF were helping to formalize data sharing requirements, and (3) academic libraries were starting to expand services to support data management and curation. Based on these trends and investigating university cultures, the JHU Data Management Services developed consulting, archiving, and planning support services tailored to researchers' data practices.
The document provides information about MANTRA, a free online course for research data management created by the University of Edinburgh. MANTRA teaches best practices for managing research data through open educational modules aligned with the research data lifecycle. It is available for reuse and repurposing under an open license. The course covers topics like data planning, organization, documentation, storage, security, and sharing.
This document provides an introduction and overview of the DBM630: Data Mining and Data Warehousing course. It outlines the course syllabus, textbooks, assessment tasks, schedule, prerequisites, and provides a high-level introduction to data mining and data warehousing concepts including definitions, processes, applications and evolution of database technologies.
This document provides an introduction and overview of the DBM630: Data Mining and Data Warehousing course. It outlines the course syllabus, textbooks, assessment tasks, schedule, prerequisites, and introduces concepts related to data mining and data warehousing including definitions, processes, applications, and evolution of database technology. The goal of the course is to teach students about data warehousing, data mining techniques such as association rule mining, classification, clustering, and current trends in the field.
Libby Bishop, Ethics Of Data Sharing Ncess Jun 09 Finala.carusi
This document summarizes an overview of ethical frameworks for sharing and reusing qualitative research data presented at a workshop. It discusses the role of archives in facilitating ethical data sharing and building trust. Formal procedures for sharing confidential research data, such as obtaining informed consent and restricting access, are described. The need to consider duties to others beyond direct research participants in the ethical debate is also highlighted.
The document is a chapter from a textbook on data mining written by Akannsha A. Totewar, a professor at YCCE in Nagpur, India. It provides an introduction to data mining, including definitions of data mining, the motivation and evolution of the field, common data mining tasks, and major issues in data mining such as methodology, performance, and privacy.
SEAD is a NSF DataNet project that aims to provide cyberinfrastructure for long tail data in sustainability science research. It develops tools for active and social curation of data including an Active Curation Repository (ACR) and VIVO profiles. It also creates a Virtual Archive to facilitate long-term access and preservation of datasets across multiple institutional repositories. The presentation provides an overview of SEAD's approach and highlights pilots with the National Center for Earth Surface Dynamics, including ingesting their data collections into the ACR and Virtual Archive and building a social network in VIVO.
I shall provide a summary of JISC work in the area of ‘Big Data’. My primary focus will be on how to manage the huge amount of research data produced in UK Universities. I shall cover the history of JISC interventions to improve research data management and look at next steps. I shall touch on some other areas of work like ‘Digging into Data’ and web archiving which also deal with ‘big data’.
1. Find all frequent itemsets of length 1 by scanning the database to count item occurrences.
2. Iteratively generate candidate itemsets of length k from frequent itemsets of length k-1, and prune unpromising candidates using the Apriori property.
3. Scan the database to determine truly frequent itemsets.
4. Generate association rules from frequent itemsets by adding items to the antecedent and consequent of rules if they meet minimum confidence.
This document discusses knowledge discovery in databases (KDD) through the LON-CAPA online educational system. [1] It defines KDD and data mining, describing the tasks, methods, and applications of KDD. [2] The goals are to obtain predictive models of students, help students and instructors use resources more effectively, and provide information to increase student learning. [3] It then discusses the KDD process and data mining methods like classification, clustering, and dependency modeling that can be applied to discover knowledge from educational data.
The document discusses the growing trend of big data and how cloud storage is a viable option for enterprise data storage needs. It notes that while cloud storage adoption has been slow, offerings continue to mature to handle larger data volumes, varieties, and velocities. The document recommends that organizations prepare their storage environments, evaluate emerging big data solutions, and rationalize their data to take advantage of next generation cloud-based storage architectures optimized for big data.
The flexibility of Apache Hadoop is one of its biggest assets – enabling businesses to generate value from data that was previously considered too expensive to be stored and processed in traditional databases – but also results in Hadoop meaning different things to different people. In this session 451 Research’s Matt Aslett will explore the impact that Hadoop is having on the traditional data processing landscape, examining the expanding ecosystem of vendors and their relationships with Apache Hadoop, investigating the increasing variety of Hadoop use-cases, and exploring adoption trends around the world.
Data Management for Librarians: An IntroductionGarethKnight
The document provides an introduction to data management for librarians, outlining key concepts such as the research data lifecycle, challenges in managing digital data over time, best practices for organizing, documenting, and storing data, and resources for data management support. Common problems include difficulty locating, accessing, and understanding data in the long run without proper planning and preservation strategies. The role of librarians is to educate researchers on best practices and provide support and training resources.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
CNI Fall 2011 Meeting Presentation Margaret Hedstrom & Robert McDonald (Dec. ...SEAD
SEAD is a new NSF-funded project that aims to provide sustainable data services for sustainability science research. It will integrate existing technologies and tools to address the needs of researchers working on "long tail" sustainability problems. SEAD is in its initial phase of developing prototypes and will not be ready to accept data until after October 2012. It is a collaboration between researchers at the University of Michigan, Indiana University, University of Illinois, and Rensselaer Polytechnic Institute.
Research Data Management for Researchers: Module 1: Intro to Data, Metadata a...Glen Newton
This document provides an introduction to research data management. It defines key concepts like research, research data, and the research data lifecycle. It discusses the importance of data sharing and outlines benefits such as enabling new research, reducing duplication, and providing credit to researchers. The document notes that most research data disappears over time unless properly managed. It also explains that research data can be complex with multiple researchers, data types, formats and standards involved. Metadata is described as important data about data. The challenges of preserving complex and transformed data through archiving are also covered.
A sponsored supplement produced for Jisc on how researchers can cope with the data deluge of modern research techniques. Published by Times Higher Education on 25 November 2009
This document provides a summary of the Libra 2.x project to replace the University of Virginia's institutional repository software. It describes the approval of the project in November 2014 due to limitations of the original software. An analysis in 2015 recommended adopting Penn State's Sufia platform for open access and ETD repositories and Harvard's Dataverse for a data repository. The document outlines subsequent work by a working group on usability testing, policies, configuration, and technical setup for the Libra Data repository, which went live in January 2016.
Virginia Data Management Bootcamp: Building the Research Data Community of Pr...Sherry Lake
This document summarizes the Virginia Data Management Bootcamp, a collaborative data education initiative held annually since 2013 among several Virginia universities. It provides details on the planning, logistics, content, and assessments of the bootcamp. According to participant feedback, the hands-on sessions were most useful but some topics could have been covered in more depth. Organizers aim to expand participation to more institutions and offer additional workshops throughout the year, as well as biennial large-scale collaborations and other collaborative efforts to support the growing Virginia data management community of practice.
1. The document discusses best practices for managing research data over the data life cycle, from collection through sharing and archiving. It provides tips for organizing, documenting, and storing data in sustainable file formats and naming conventions. Following best practices helps ensure usability, reproducibility, and long-term access to research data.
2. Specific best practices covered include using consistent organization, standardized naming and formats, descriptive filenames, quality assurance, scripting for processing, documenting file contents, and choosing open file formats. The document also addresses data security, backup, and storage considerations.
3. Managing data properly is important for reuse and sharing data with others now or in the future. Scripting helps capture data workflows for reproducibility.
Using a Case Study to Teach Data Management to LibrariansSherry Lake
This document outlines the agenda and learning objectives for a workshop on research data management for libraries. The workshop uses a case study approach and hands-on activities to teach librarians best practices for data collection, organization, documentation, backup/storage, and sharing/preservation. The goal is to prepare librarians to teach researchers about data management and illustrate opportunities for library involvement in the area. Based on a survey after the workshop, most attendees felt their expectations were met or exceeded, and they found the hands-on case study activities and practical tips to be most useful.
Documentation and Metdata - VA DM BootcampSherry Lake
This document discusses documentation and metadata for research data. It begins with an overview of why documentation is important at different stages of the research data lifecycle from collection through archiving. Key elements to document include how the data was created, its content and structure, who created and maintains it, and how it can be accessed and cited. The document then discusses common documentation formats like readmes, data dictionaries, and codebooks. It also introduces metadata as structured information that describes resources and explains common metadata standards and tools for creating structured metadata files. Exercises guide creating documentation in these formats for a weather dataset example.
This document discusses creating a data management plan. It explains that a data management plan is a comprehensive plan for managing research data throughout a project's lifecycle and briefly describing how data will be shared per a funder's policy. It provides an overview of key elements to include in a plan such as file formats, organization, sharing, and preservation. The document also reviews funder requirements and available tools to create plans, noting they can be tailored to different funders' guidelines.
Introduction to DMPTool2. Originally released in 2011, the DMPTool provides a free step-by-step wizard, detailed guidance, and links to general and institutional resources to walk a researcher through the process of generating a comprehensive data management plan tailored to specific funder requirements.
This webinar will demonstrate some of the original features of the tool, as well as the new features in DMPTool2, which include institutional customizations and researcher collaborations.
This document summarizes a workshop on preparing data management plans. It discusses what data management entails, why it is important, and what components are typically included in a data management plan. Key points covered include an overview of the data lifecycle and how plans help ensure research can be replicated, preserved, and shared. The workshop also demonstrates how to create a data management plan using the DMPTool, a online tool that guides users through the process.
This document summarizes recent federal mandates requiring open access to publications and data resulting from federally funded scientific research. It discusses a 2013 White House memo requiring federal agencies spending over $100 million annually on research to develop public access plans. It also outlines policies from agencies like NIH, NSF, and NOAA requiring data management plans and sharing of published results and supporting data. Stakeholder responses to these mandates like the CHORUS publishing initiative and the SHARE academic consortium proposal are also summarized.
DMPTool2 demo for DMPTool-DMPonline Workshop IDCC 2014Sherry Lake
The document discusses updates and improvements to the DMPTool and DMPonline platforms. It notes that DMPTool was originally released in 2011 and was self-funded. DMPonline, funded by additional sources, includes expanded functionality for researchers and administrators, including collaborative plan creation, review capabilities, and institutional templates. Upcoming improvements include co-ownership of plans, expanded administrative roles, and self-service admin functions. The speaker encourages using the platforms' resources to advance data stewardship education.
This document provides information about developing a data management plan for grant proposals. It discusses the goals of the workshop which are to learn about data management planning, available resources, develop a draft plan, and receive feedback. It then covers what good data management involves, who requires data management plans, examples of requirements from agencies like NSF, and parts of a generic data management plan. Finally, it discusses resources available for creating plans like the DMPTool.
This document provides information about a webinar on environmental scanning to identify important stakeholders on campus for data management. Participants must call in for audio and can ask questions in the chat. The webinar will cover goals of environmental scanning, doing a scan based on a data management plan, resources, conducting an institutional scan, and supporting the research lifecycle through collaborations and partnerships.
The document summarizes the Data Management Planning Tool (DMPTool), an online tool that helps researchers create data management plans. It discusses the goals of providing a simple way for researchers to create plans for their funders and offering institution-specific resources. The summary describes the tool's increasing participation from universities and future plans to improve functionality, sustainability, and community involvement through grants and an open source model.
This document summarizes data management and sharing policies in the US and Canada. It outlines that major US funding agencies like the NSF and NIH require data management plans and sharing of results. Canadian agencies like CIHR and SSHRC also have data archiving policies. Other groups developing policies include domain-specific professional organizations and journals. Library organizations in both countries are working to help institutions support these policies through initiatives like the ARL/DLF E-Science Institute.
The document discusses the importance of managing research data. It notes that data management saves time, makes long-term data preservation easier, and supports sharing data with others. Data sharing is now required by most major funding agencies and academic journals. The document provides examples of problems caused by poor data management practices and outlines the key components of a data management plan, such as describing the data, file formats, sharing and archiving policies, and responsibilities. Researchers are encouraged to seek help from scientific consulting services for creating data management plans.
This document discusses re-tooling library staff and resources to support research data management. It describes the Scientific Data Consulting Group model developed at the University of Virginia Library, which involved conducting stakeholder analysis, prioritizing data interviews and preparing data management plans. It also outlines models from other universities, such as Purdue and Johns Hopkins, and discusses training librarians through workshops and data interviews. The document emphasizes that investment in staff and services is critical to providing effective research data management support.
Data management involves organizing and storing large amounts of information from various online sources. Links to data management resources include a Zotero group for sharing citations and references, as well as a site exploring the use of social networks for collaborating on data sets. Effective data management allows researchers to locate, access, and analyze information from different online locations and platforms.
This document outlines best practices for creating research data. [1] It recommends using consistent data organization with standardized formats and descriptive file names. [2] Researchers should perform quality assurance checks and use scripted programs to analyze data while keeping notes. [3] All aspects of data collection and analysis should be thoroughly documented. Following these practices will improve data usability, sharing, and reproducibility.
The document summarizes the DMPTool, an online tool that helps researchers create data management plans. It provides a step-by-step wizard to generate DMPs. The tool aims to 1) provide a simple way for researchers to create DMPs required by funding agencies and 2) provide institution-specific resources to help manage data. It is accessed through institutional login and provides customized help text, links, and answers. Usage has grown significantly since launch. Future work includes adding funders, functionality, and integrating with other systems to help coordinate data management.
Funder requirements for Data Management PlansSherry Lake
This document discusses funder requirements for data management and sharing. It notes that major funders like the National Science Foundation (NSF) and National Institutes of Health (NIH) require applicants to submit a data management plan. These plans describe how research data will be organized, preserved, and shared. The document provides details on what funders expect to see in a data management plan, including a description of the data, metadata standards, data access and sharing policies, and plans for long-term data preservation. It also lists other funders that require applicants to have a data management or sharing plan.
Assessment and Planning in Educational technology.pptxKavitha Krishnan
In an education system, it is understood that assessment is only for the students, but on the other hand, the Assessment of teachers is also an important aspect of the education system that ensures teachers are providing high-quality instruction to students. The assessment process can be used to provide feedback and support for professional development, to inform decisions about teacher retention or promotion, or to evaluate teacher effectiveness for accountability purposes.
The simplified electron and muon model, Oscillating Spacetime: The Foundation...RitikBhardwaj56
Discover the Simplified Electron and Muon Model: A New Wave-Based Approach to Understanding Particles delves into a groundbreaking theory that presents electrons and muons as rotating soliton waves within oscillating spacetime. Geared towards students, researchers, and science buffs, this book breaks down complex ideas into simple explanations. It covers topics such as electron waves, temporal dynamics, and the implications of this model on particle physics. With clear illustrations and easy-to-follow explanations, readers will gain a new outlook on the universe's fundamental nature.
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
Thinking of getting a dog? Be aware that breeds like Pit Bulls, Rottweilers, and German Shepherds can be loyal and dangerous. Proper training and socialization are crucial to preventing aggressive behaviors. Ensure safety by understanding their needs and always supervising interactions. Stay safe, and enjoy your furry friends!
বাংলাদেশের অর্থনৈতিক সমীক্ষা ২০২৪ [Bangladesh Economic Review 2024 Bangla.pdf] কম্পিউটার , ট্যাব ও স্মার্ট ফোন ভার্সন সহ সম্পূর্ণ বাংলা ই-বুক বা pdf বই " সুচিপত্র ...বুকমার্ক মেনু 🔖 ও হাইপার লিংক মেনু 📝👆 যুক্ত ..
আমাদের সবার জন্য খুব খুব গুরুত্বপূর্ণ একটি বই ..বিসিএস, ব্যাংক, ইউনিভার্সিটি ভর্তি ও যে কোন প্রতিযোগিতা মূলক পরীক্ষার জন্য এর খুব ইম্পরট্যান্ট একটি বিষয় ...তাছাড়া বাংলাদেশের সাম্প্রতিক যে কোন ডাটা বা তথ্য এই বইতে পাবেন ...
তাই একজন নাগরিক হিসাবে এই তথ্য গুলো আপনার জানা প্রয়োজন ...।
বিসিএস ও ব্যাংক এর লিখিত পরীক্ষা ...+এছাড়া মাধ্যমিক ও উচ্চমাধ্যমিকের স্টুডেন্টদের জন্য অনেক কাজে আসবে ...
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
हिंदी वर्णमाला पीपीटी, hindi alphabet PPT presentation, hindi varnamala PPT, Hindi Varnamala pdf, हिंदी स्वर, हिंदी व्यंजन, sikhiye hindi varnmala, dr. mulla adam ali, hindi language and literature, hindi alphabet with drawing, hindi alphabet pdf, hindi varnamala for childrens, hindi language, hindi varnamala practice for kids, https://www.drmullaadamali.com
How to Add Chatter in the odoo 17 ERP ModuleCeline George
In Odoo, the chatter is like a chat tool that helps you work together on records. You can leave notes and track things, making it easier to talk with your team and partners. Inside chatter, all communication history, activity, and changes will be displayed.
1. Sherry Lake
July 31, 2012 University of Florida Data Management Workshop
2. Research Life Cycle
Data Re- Data Deposit
Discovery Use
Archive
Proposal Project Data Data Data End of
Planning Start Up Collection Analysis Sharing Project
Writing
Re-
Purpose
Data Life Cycle
3.
4. Research Life Cycle
Support Within
Data Mining Data Curation
&
Data Preservation
Search Data Re- Data Deposit
Discovery Use
Archive
Proposal Project Data Data Data End of
Planning Start Up Collection Analysis Sharing Project
Writing
DMP DM Planning Data Publication
Consulting Storage Rights &
Grant Writing Re- Restriction
& Planning Purpose s
Data Processing Data Life Cycle
HPC/Visualizatio
n Metadata &
Tool Documentatio
Development n
5. Research Life Cycle
Outside Support
Data Mining Data Curation
&
Data Preservation
Search Data Re- Data Deposit
Policy
Discovery Use
Archive
Proposal
Planning
Project Infrastructure
Data Data Data End of
Start Up Collection Analysis Sharing Project
Writing
DMP DM Planning Data Publication
Consulting
Grant Writing Researcher Re-
Code of Practice Storage Rights &
Restriction
& Planning Purpose s
Data Processing Data Life Cycle
HPC/Visualizatio
n Metadata &
Tool Documentatio
Development n
9. 9 skills gaps
1. Ability to advise on preserving research
outputs
2. Knowledge to advise on data management and
curation
3. Knowledge on complying with funder
mandates, including open access
4. Knowledge to advise on potential data
manipulation tools
5. Knowledge to advise on data mining
(Auckland, 2012)
10. 9 skills gaps
6. Knowledge to advocate, and advise on, the use
of metadata
7. Ability to advise on the preservation of project
records
8. Knowledge of sources of research funding
9. Skills to develop metadata schema
(Auckland, 2012)
11. Research Life Cycle
Support Within
Data Mining Data Curation
&
Data Preservation
Search Data Re- Data Deposit
Discovery Use
Archive
Proposal Project Data Data Data End of
Planning Start Up Collection Analysis Sharing Project
Writing
DMP DM Planning Data Publication
Consulting Storage Rights &
Grant Writing Re- Restriction
& Planning Purpose s
Data Processing Data Life Cycle
HPC/Visualizatio
n Metadata &
Tool Documentatio
Development n
12. References
Auckland, M (2012), Re-skilling for Research, Research Libraries UK (RLUK)
report http://www.rluk.ac.uk/content/re-skilling-research
Lyon, L. (2012). The Informatics Transform: Re-Engineering Libraries for the
Data Decade. International Journal of Digital Curation, 7(1), 126–138.
doi:10.2218/ijdc.v7i1.220
Pryor, G., & Donnelly, M. (2009). Skilling Up to Do Data: Whose Role, Whose
Responsibility, Whose Career? International Journal of Digital Curation, 4(2), 158–
170. doi:10.2218/ijdc.v4i2.105
Editor's Notes
Research Data Services are defined as services that address the full data lifecycle, including the data management plan, digital curation (selection, preservation, maintenance, and archiving) and metadata creation and conversion.
Here are 4 roles that are involved directly in the day-to-day DM. The bubbles are the “skills” needed.If I am proposing that the Librarian be involved in the whole data lifecycle, then there are sResearchers: what skills should a librarian have? What services does UF already have? Should library fill the gaps?Pryor, G., & Donnelly, M. (2009). Skilling Up to Do Data: Whose Role, Whose Responsibility, Whose Career? International Journal of Digital Curation, 4(2), 158–170. doi:10.2218/ijdc.v4i2.105
Let’s look more closely at the skills needed along our Research Life Cycle.Then there are over-arching skills: marketing, raising awareness and user trainingOther more detailed services such as data format conversion & transformationData Search – like Lit serarch
Library and researcher can’t do this along.What is needed across the whole life cycle… across all of research are policies (funders, institution, discipline) and codes of practice (Open Data, Open Access)In addition to infrastructure and training.
There are other partners at every instutitioin who have roles in data management – policy, IT, VPR, Office of Grant awards.Libraries shouldn’t have to do it all. Look around your institution to see who else provides the “services” that are needed to support the research lifecycle.Lyon, L. (2012). The Informatics Transform: Re-Engineering Libraries for the Data Decade. International Journal of Digital Curation, 7(1), 126–138. doi:10.2218/ijdc.v7i1.220Pg 131Part 1 of Table 1. Research data management, the library and institutional stakeholders. Partnership approach – Library & institutional stakeholders 7 roles, responsibilities, requirements, relationshipsDirector (Leadership)Data Librarian (Advocacy)Repo managers (discovery)IT/ computing (Storage)
Lyon, L. (2012). The Informatics Transform: Re-Engineering Libraries for the Data Decade. International Journal of Digital Curation, 7(1), 126–138. doi:10.2218/ijdc.v7i1.220Pg 132Part 2 of Table 1. Research data management, the library and institutional stakeholders. Research and development (CRIS)Faculty training centers (training)PVC Research (policy)
According to Auckland, even if we narrow down the skills to those that are related to the “traditional” roles of librarians, there are still gaps that need to be addressed to truly support data management.2012 Re-skilling for Research report identified a skills gapRe-skilling for Research In January 2012 in the UK RLUK (Research Libraries UK) published a major report by Mary Auckland on the changing needs of researchers and the effect on the subject/liaison role within libraries. Research practices and activities are changing and evolving, research support provided by libraries must evolve with them. In terms of what libraries are currently offering the, Re-skilling for Research report found a Skills gap 9 areas:
The 9 areas identified as having potentially the most significant skills gap are: The Ability to advise on preserving research outputsKnowledge to advise on data management and curation, including ingest, discovery, access, dissemination, preservation, and portability Knowledge to support researchers in complying with the various mandates of funders, including open access requirementsKnowledge to advise on potential data manipulation tools used in the disciplineKnowledge to advise on data miningKnowledge to advocate, and advise on, the use of metadata Ability to advise on the preservation of project records e.g. Knowledge of sources of research funding to assist researchers to identify potential funders Skills to develop metadata schema, and advise on discipline/subject standards and practices, for individual research projects Auckland, M (2012), Re-skilling for Research, Research Libraries UK (RLUK) report http://www.rluk.ac.uk/content/re-skilling-research
Auckland, M (2012), Re-skilling for Research, Research Libraries UK (RLUK) report http://www.rluk.ac.uk/content/re-skilling-research
Let’s look more closely at the skills needed along our Research Life Cycle.Then there are over-arching skills: marketing, raising awareness and user trainingOther more detailed services such as data format conversion & transformationData Search – like Lit serarch