Presentation on legal and ethical issues in open access to research data given at the RECODE early career researcher workshop, University of Sheffield 14-15th May 2015
Librarian RDM Training: Ethics and copyright for research dataRobin Rice
This document provides an overview of ethics and copyright as they relate to research data management. It discusses ethical requirements for collecting human subject data, including obtaining consent and protecting privacy and confidentiality. Certain types of research may be exempt from ethics review. Intellectual property rights can apply to research data depending on the level of creativity in the data's collection and organization. Data licensing is an alternative to asserting copyright that allows explicitly defining how data can be used.
In 2010, President Obama said, “We're going to aggressively protect our intellectual property. Our single greatest asset is the innovation and the ingenuity and creativity of the American people...It is essential to our prosperity and it will only become more so in this century.” This 21-day project supports the work of librarians and archivists who protect the rights of intellectual property (IP) license holders while ensuring that the wider population is able to access and benefit from their efforts and ingenuity. Through meetings and discussions with public and private sector professionals, this project will explore the broad range of Intellectual Property Rights (IPR) protections, including copyright, patents, and trademarks, as well as their legal requirements for implementation and enforcement. The project will also explore the roles non-governmental organizations, including associations and universities, play in IPR protection. The participants will also liaise with library professionals to discuss best practices in library and archives management.
Tensions between intellectual property and knowledge discovery in the digital...LIBER Europe
Presentation on text and data mining, its uses, and issues to be addressed to increase its uptake: copyright, open access. Introduces a new European open research infrastructure, OpenMinTed, which will enable TDM and tha capturing of TMD outputs. Ex Libris session, #WLIC15 Cape Town, 18 August 2015.
Legal Issues in Applied Technology discusses four key laws - privacy, copyright, patent, and evidence - as they relate to new technologies. Privacy laws regulate the collection and sharing of personal information and have two categories of issues: those due to security vulnerabilities and those regarding unlawful surveillance. Copyright law governs ownership and use of creative works. Patent law regulates rights to inventions. Evidence law governs proof of facts in legal proceedings. As technology advances rapidly, associated legal issues will have difficulty keeping pace.
This document discusses security and privacy concerns regarding computer forensics applications. It provides an overview of the history and development of computer forensics as a field, both in law enforcement and academia. It also outlines common computer forensics techniques and tools, emerging curricula in university programs, and important considerations regarding privacy and security of digital evidence.
This document discusses the use of RFID (radio-frequency identification) technology in public libraries. It provides an overview of what RFID is and how it can benefit libraries by allowing for faster checkout, self-checkout kiosks, and better materials tracking. However, some concerns are also raised about privacy issues and the high costs of implementing an RFID system. The document reviews experiences with RFID from libraries in California and references several additional sources on the topic.
Librarian RDM Training: Ethics and copyright for research dataRobin Rice
This document provides an overview of ethics and copyright as they relate to research data management. It discusses ethical requirements for collecting human subject data, including obtaining consent and protecting privacy and confidentiality. Certain types of research may be exempt from ethics review. Intellectual property rights can apply to research data depending on the level of creativity in the data's collection and organization. Data licensing is an alternative to asserting copyright that allows explicitly defining how data can be used.
In 2010, President Obama said, “We're going to aggressively protect our intellectual property. Our single greatest asset is the innovation and the ingenuity and creativity of the American people...It is essential to our prosperity and it will only become more so in this century.” This 21-day project supports the work of librarians and archivists who protect the rights of intellectual property (IP) license holders while ensuring that the wider population is able to access and benefit from their efforts and ingenuity. Through meetings and discussions with public and private sector professionals, this project will explore the broad range of Intellectual Property Rights (IPR) protections, including copyright, patents, and trademarks, as well as their legal requirements for implementation and enforcement. The project will also explore the roles non-governmental organizations, including associations and universities, play in IPR protection. The participants will also liaise with library professionals to discuss best practices in library and archives management.
Tensions between intellectual property and knowledge discovery in the digital...LIBER Europe
Presentation on text and data mining, its uses, and issues to be addressed to increase its uptake: copyright, open access. Introduces a new European open research infrastructure, OpenMinTed, which will enable TDM and tha capturing of TMD outputs. Ex Libris session, #WLIC15 Cape Town, 18 August 2015.
Legal Issues in Applied Technology discusses four key laws - privacy, copyright, patent, and evidence - as they relate to new technologies. Privacy laws regulate the collection and sharing of personal information and have two categories of issues: those due to security vulnerabilities and those regarding unlawful surveillance. Copyright law governs ownership and use of creative works. Patent law regulates rights to inventions. Evidence law governs proof of facts in legal proceedings. As technology advances rapidly, associated legal issues will have difficulty keeping pace.
This document discusses security and privacy concerns regarding computer forensics applications. It provides an overview of the history and development of computer forensics as a field, both in law enforcement and academia. It also outlines common computer forensics techniques and tools, emerging curricula in university programs, and important considerations regarding privacy and security of digital evidence.
This document discusses the use of RFID (radio-frequency identification) technology in public libraries. It provides an overview of what RFID is and how it can benefit libraries by allowing for faster checkout, self-checkout kiosks, and better materials tracking. However, some concerns are also raised about privacy issues and the high costs of implementing an RFID system. The document reviews experiences with RFID from libraries in California and references several additional sources on the topic.
RDAP 15: Providing access to restricted data in our institutions ASIS&T
Research Data Access and Preservation Summit, 2015
Minneapolis, MN
April 22-23, 2015
Part of “Beyond metadata: Supporting non-standardized documentation to facilitate data reuse”
Sarah Pickle, CLIR/DLF Social Science Data Curation Fellow, Penn State University Libraries
ICT, Development & Privacy: Exploring the Debate and Evaluating Argumentative...Ilse Oosterlaken
A presentation on a research proposal which unfortunately did not get funding. This was the abstract included with th grant application:
Privacy, while prominent in Western debates on ICT, receives little attention in the context of ‘ICT for Development’ (ICT4D) research and practice. Arguments in favour or against paying systematic attention to privacy in ICT4D – e.g. “the income-poor are especially vulnerable to privacy-related harms” (pro), or “imposing privacy concerns is a form of cultural imperialism” (against) – have not yet been articulated in any detail.
In this research I will explore the argumentative landscape, expanding the methods of analytical discourse evaluation and argument mapping to a new domain and intercultural context. Based on focus groups with ICT4D scholars, practitioners and stakeholders I will sketch ‘argument maps’. Semi-structured interviews will provide further empirical information about actual reasoning concerning privacy in this context – allowing for a refinement of argument maps.
Key arguments will subsequently be assessed on their strength/plausibility. This requires a thorough normative-ethical analysis, drawing on both Computer Ethics and Development Ethics, and considering empirical realities as studied by ICT4D scholars – three disciplines which currently have little overlap. One contribution to the academic literature is providing the first steps towards such interdisciplinary analysis.
More practically, a pilot will explore how to facilitate deliberations on privacy within ICT4D practice.
This document discusses digital forensics. Digital forensics is a branch of forensic science that involves the recovery and investigation of material found in digital devices, often related to computer crimes. It has expanded to cover all devices capable of storing digital data. Digital forensics investigations aim to search for, detect, recover, and preserve electronic evidence, often for legal purposes. There are three main types of digital forensic analysis: media analysis, code analysis, and network analysis. A major limitation is encryption, which disrupts initial examination where evidence may be located.
Presentation by Robin Rice, Data Librarian, University of Edinburgh. Part of a roundtable discussion at a workshop for 'Scotland's National Collections and the Digital Humanities,' a knowledge-exchange project hosted at the University of Edinburgh. 2 May 2014. http://www.blogs.hss.ed.ac.uk/archives-now/
The document discusses the need for organizations to create digital preservation policies and procedures to protect their digital collections. It notes that most institutions currently collect born-digital materials without plans for their long-term preservation. The document defines digital preservation as the active management of digital records to preserve accurate access over time through policies, metadata, and file maintenance. It recommends that institutions draft a high-level digital preservation policy that commits to preservation, defines its scope and stakeholders, and outlines goals, challenges, and staff roles. Examples of existing policies are provided.
Defensive Publication - Patexia IP Matters WebinarPatexia Inc.
This document discusses defensive publications and how they can be used to limit bad software patents. It explains that defensive publications create explicit prior art, giving patent examiners more context about the state of technology. This helps prevent claims of invention that are not truly new or non-obvious. The document provides guidelines for writing a defensive publication, noting it should describe how an idea works. It also discusses how Linux Defenders can help get defensive publications in front of patent examiners through the IP.com digital notary service during prior art searches.
This document discusses censorship, intellectual property, and privacy in relation to the internet. It begins by defining censorship as the attempt to suppress material deemed offensive. It then discusses challenges of internet censorship due to its decentralized nature. Issues around privacy and intellectual property like peer-to-peer file sharing are also examined. Various organizations that advocate for internet freedom and openness are mentioned. Ethical views on these topics from philosophers like John Stuart Mill are referenced.
LIBER fostering Open Science and Knowledge DiscoveryLIBER Europe
Presentation by Kristiina Hormia Poutanen, LIBER President. Delivered at 25th Anniversary Conference of The National Repository Library of Finland
Kuopio 22th of May 2015. Content is cc-by.
Peter Doorn - Data Protection Issues for Trusted Digital Repositories: Chall...dri_ireland
Response of IAG member Prof. Doorn presented during Session 2 of the workshop "Data Protection Issues for Trusted Digital Repositories: Challenges and Solutions", which primarily examined the challenges facing data repositories in meeting ethical and legal requirements of archiving research data. This was presented on the 16th of January, 2014 at the "Data Protection Issues for Trusted Digital Repositories: Challenges and Solutions" at the Royal Irish Academy.
This document compares and contrasts the FAIR and GDPR frameworks for managing research data. It summarizes the key principles of each: FAIR focuses on making data findable, accessible, interoperable and reusable, while GDPR prioritizes privacy by default and design. The document notes challenges for researchers working with human subject data under these differing paradigms and expresses concerns that researchers may not receive adequate support to share such data or take advantage of open science and data science innovations if GDPR restrictions are too strict. It concludes by questioning which framework, FAIR or GDPR, will have more influence over human subject research going forward.
This document discusses copyright considerations for research data. It explains that primary research data and secondary data may be protected by copyright or database rights. Whether copyright protects research data depends on how the data is created, shared, and reused. Special considerations include determining who owns the copyright - the researcher, university, or research funder. Permission should always be checked when planning to share or reuse data due to variations in copyright laws between jurisdictions. Resources are provided to help researchers understand copyright as it relates to sharing and using research data, including on social media platforms.
Tutorial for ACM Multimedia 2016, given together with Gerald Friedland, with contributions from Julia Bernd and Yiannis Kompatsiaris. The presentation covered an introduction to the problem of disclosing personal information through multimedia sharing, the associated security risks, methods for conducting multimodla inferences and technical frameworks that could help alleviate such risks.
Copyright law faces challenges in the digital age. As information is increasingly digitized and shared online, copyright protections can come into conflict with fundamental human rights like privacy, education, free expression, and access to information. The document discusses how copyright was historically designed for physical works but now extensively impacts digital works and behaviors. Several problems are outlined where copyright restrictions may inhibit universal human rights on the internet if not properly balanced. The future of the internet and role of copyright on it will depend on how it is framed - as a tool for institutions, an environment for the public, or market for commercial interests.
Open Access: Data Protection, Storage and SharingOpenExeter
The document discusses open access data protection, storage, and sharing. It addresses how to ensure fair processing and security of personal data when making it openly accessible. Key points include obtaining participant consent, anonymizing data when possible, using technical security measures like encryption, and establishing a "dark archive" for confidential data not suitable for open access. The document provides guidance on what information to share with research participants and considers challenges around anonymization. It also briefly outlines exemptions to freedom of information requests.
A Lifecycle Approach to Information PrivacyMicah Altman
The document discusses challenges in privacy across the lifecycle of data from collection to dissemination and proposes taking a lifecycle approach. It analyzes how concepts like differential privacy could address issues raised at different stages and questions that approach generates regarding legal and technical issues. The goal is to advance interdisciplinary research at the intersection of law, social science, public policy, data collection methods, data management, statistics, and computer science.
Cyberethics is the study of moral, legal, and social issues involving cybertechnology and its impact. Cybertechnology refers to computing and communications devices like computers, networks, and the Internet. The history of cyberethics includes four phases defined by the evolving technologies from standalone machines to today's converged technologies. While some argue cyberethics issues are not unique, others believe new issues have emerged due to the scale and scope enabled by cybertechnology, like the ability to stalk multiple victims globally without leaving home.
The document discusses several topics related to ethics and privacy when dealing with information technology:
1. It introduces four common ethical frameworks used to evaluate decisions: utilitarian, rights-based, fairness-based, and common good approaches.
2. It outlines some fundamental tenets of ethics like responsibility, accountability, and liability that are important in a corporate environment.
3. It identifies four general categories of ethical issues related to IT: privacy, accuracy, property, and accessibility. It provides examples of issues that fall under each category.
4. It focuses specifically on privacy issues, outlining concerns around electronic surveillance, personal information in databases, information shared online, and international differences in privacy laws and standards.
This document discusses several legal and ethical issues related to e-marketing, including:
1. Software piracy is a major problem, with 40% of software being pirated in 2001 resulting in $10.7 billion in losses.
2. Ethics concerns what is right and wrong, while law is a public endeavor created by legislatures. There is often a close relationship between ethics and law.
3. Intellectual property on the internet is protected through copyright, patent, and trademark laws, though these protections are still developing for digital contexts.
A Revolution in Open Science: Open Data and the Role of Libraries (Professor ...LIBER Europe
This document discusses the opportunities and challenges of open science and open data. It argues that openly sharing scientific data and findings has significant benefits, including enabling faster scientific progress, deterring fraud, and supporting citizen science. However, for data to be truly open and useful to others, it needs to be accessible, intelligible, assessable, and reusable. The document also examines the roles and responsibilities of different stakeholders in working towards more open and reproducible science. This includes changing incentives for scientists, strategic funding for technical solutions from funders, and exploring how institutions like libraries and learned societies can help address the challenges of managing and making sense of the growing volume of research data.
This document discusses intellectual property in the context of information technology. It covers topics such as the value of intellectual property, how the internet has impacted copying and distribution, different types of intellectual property like patents, copyrights, and trade secrets. It also discusses debates around intellectual property protection including arguments around incentivizing innovation versus allowing works to enter the public domain. Open source software is presented as an alternative intellectual property model.
This slide is the final presentation for the Green IT course I have followed. This short course is all about how you can do to preserve, and avoid abusing the environment and having the sustainable development. In this presentation, my group show the terminology "The Internet of Things" and all benefits as well as its impacts to the human beings and the world. Feel free to give comments and information to me. Thank you and enjoy :)
RDAP 15: Providing access to restricted data in our institutions ASIS&T
Research Data Access and Preservation Summit, 2015
Minneapolis, MN
April 22-23, 2015
Part of “Beyond metadata: Supporting non-standardized documentation to facilitate data reuse”
Sarah Pickle, CLIR/DLF Social Science Data Curation Fellow, Penn State University Libraries
ICT, Development & Privacy: Exploring the Debate and Evaluating Argumentative...Ilse Oosterlaken
A presentation on a research proposal which unfortunately did not get funding. This was the abstract included with th grant application:
Privacy, while prominent in Western debates on ICT, receives little attention in the context of ‘ICT for Development’ (ICT4D) research and practice. Arguments in favour or against paying systematic attention to privacy in ICT4D – e.g. “the income-poor are especially vulnerable to privacy-related harms” (pro), or “imposing privacy concerns is a form of cultural imperialism” (against) – have not yet been articulated in any detail.
In this research I will explore the argumentative landscape, expanding the methods of analytical discourse evaluation and argument mapping to a new domain and intercultural context. Based on focus groups with ICT4D scholars, practitioners and stakeholders I will sketch ‘argument maps’. Semi-structured interviews will provide further empirical information about actual reasoning concerning privacy in this context – allowing for a refinement of argument maps.
Key arguments will subsequently be assessed on their strength/plausibility. This requires a thorough normative-ethical analysis, drawing on both Computer Ethics and Development Ethics, and considering empirical realities as studied by ICT4D scholars – three disciplines which currently have little overlap. One contribution to the academic literature is providing the first steps towards such interdisciplinary analysis.
More practically, a pilot will explore how to facilitate deliberations on privacy within ICT4D practice.
This document discusses digital forensics. Digital forensics is a branch of forensic science that involves the recovery and investigation of material found in digital devices, often related to computer crimes. It has expanded to cover all devices capable of storing digital data. Digital forensics investigations aim to search for, detect, recover, and preserve electronic evidence, often for legal purposes. There are three main types of digital forensic analysis: media analysis, code analysis, and network analysis. A major limitation is encryption, which disrupts initial examination where evidence may be located.
Presentation by Robin Rice, Data Librarian, University of Edinburgh. Part of a roundtable discussion at a workshop for 'Scotland's National Collections and the Digital Humanities,' a knowledge-exchange project hosted at the University of Edinburgh. 2 May 2014. http://www.blogs.hss.ed.ac.uk/archives-now/
The document discusses the need for organizations to create digital preservation policies and procedures to protect their digital collections. It notes that most institutions currently collect born-digital materials without plans for their long-term preservation. The document defines digital preservation as the active management of digital records to preserve accurate access over time through policies, metadata, and file maintenance. It recommends that institutions draft a high-level digital preservation policy that commits to preservation, defines its scope and stakeholders, and outlines goals, challenges, and staff roles. Examples of existing policies are provided.
Defensive Publication - Patexia IP Matters WebinarPatexia Inc.
This document discusses defensive publications and how they can be used to limit bad software patents. It explains that defensive publications create explicit prior art, giving patent examiners more context about the state of technology. This helps prevent claims of invention that are not truly new or non-obvious. The document provides guidelines for writing a defensive publication, noting it should describe how an idea works. It also discusses how Linux Defenders can help get defensive publications in front of patent examiners through the IP.com digital notary service during prior art searches.
This document discusses censorship, intellectual property, and privacy in relation to the internet. It begins by defining censorship as the attempt to suppress material deemed offensive. It then discusses challenges of internet censorship due to its decentralized nature. Issues around privacy and intellectual property like peer-to-peer file sharing are also examined. Various organizations that advocate for internet freedom and openness are mentioned. Ethical views on these topics from philosophers like John Stuart Mill are referenced.
LIBER fostering Open Science and Knowledge DiscoveryLIBER Europe
Presentation by Kristiina Hormia Poutanen, LIBER President. Delivered at 25th Anniversary Conference of The National Repository Library of Finland
Kuopio 22th of May 2015. Content is cc-by.
Peter Doorn - Data Protection Issues for Trusted Digital Repositories: Chall...dri_ireland
Response of IAG member Prof. Doorn presented during Session 2 of the workshop "Data Protection Issues for Trusted Digital Repositories: Challenges and Solutions", which primarily examined the challenges facing data repositories in meeting ethical and legal requirements of archiving research data. This was presented on the 16th of January, 2014 at the "Data Protection Issues for Trusted Digital Repositories: Challenges and Solutions" at the Royal Irish Academy.
This document compares and contrasts the FAIR and GDPR frameworks for managing research data. It summarizes the key principles of each: FAIR focuses on making data findable, accessible, interoperable and reusable, while GDPR prioritizes privacy by default and design. The document notes challenges for researchers working with human subject data under these differing paradigms and expresses concerns that researchers may not receive adequate support to share such data or take advantage of open science and data science innovations if GDPR restrictions are too strict. It concludes by questioning which framework, FAIR or GDPR, will have more influence over human subject research going forward.
This document discusses copyright considerations for research data. It explains that primary research data and secondary data may be protected by copyright or database rights. Whether copyright protects research data depends on how the data is created, shared, and reused. Special considerations include determining who owns the copyright - the researcher, university, or research funder. Permission should always be checked when planning to share or reuse data due to variations in copyright laws between jurisdictions. Resources are provided to help researchers understand copyright as it relates to sharing and using research data, including on social media platforms.
Tutorial for ACM Multimedia 2016, given together with Gerald Friedland, with contributions from Julia Bernd and Yiannis Kompatsiaris. The presentation covered an introduction to the problem of disclosing personal information through multimedia sharing, the associated security risks, methods for conducting multimodla inferences and technical frameworks that could help alleviate such risks.
Copyright law faces challenges in the digital age. As information is increasingly digitized and shared online, copyright protections can come into conflict with fundamental human rights like privacy, education, free expression, and access to information. The document discusses how copyright was historically designed for physical works but now extensively impacts digital works and behaviors. Several problems are outlined where copyright restrictions may inhibit universal human rights on the internet if not properly balanced. The future of the internet and role of copyright on it will depend on how it is framed - as a tool for institutions, an environment for the public, or market for commercial interests.
Open Access: Data Protection, Storage and SharingOpenExeter
The document discusses open access data protection, storage, and sharing. It addresses how to ensure fair processing and security of personal data when making it openly accessible. Key points include obtaining participant consent, anonymizing data when possible, using technical security measures like encryption, and establishing a "dark archive" for confidential data not suitable for open access. The document provides guidance on what information to share with research participants and considers challenges around anonymization. It also briefly outlines exemptions to freedom of information requests.
A Lifecycle Approach to Information PrivacyMicah Altman
The document discusses challenges in privacy across the lifecycle of data from collection to dissemination and proposes taking a lifecycle approach. It analyzes how concepts like differential privacy could address issues raised at different stages and questions that approach generates regarding legal and technical issues. The goal is to advance interdisciplinary research at the intersection of law, social science, public policy, data collection methods, data management, statistics, and computer science.
Cyberethics is the study of moral, legal, and social issues involving cybertechnology and its impact. Cybertechnology refers to computing and communications devices like computers, networks, and the Internet. The history of cyberethics includes four phases defined by the evolving technologies from standalone machines to today's converged technologies. While some argue cyberethics issues are not unique, others believe new issues have emerged due to the scale and scope enabled by cybertechnology, like the ability to stalk multiple victims globally without leaving home.
The document discusses several topics related to ethics and privacy when dealing with information technology:
1. It introduces four common ethical frameworks used to evaluate decisions: utilitarian, rights-based, fairness-based, and common good approaches.
2. It outlines some fundamental tenets of ethics like responsibility, accountability, and liability that are important in a corporate environment.
3. It identifies four general categories of ethical issues related to IT: privacy, accuracy, property, and accessibility. It provides examples of issues that fall under each category.
4. It focuses specifically on privacy issues, outlining concerns around electronic surveillance, personal information in databases, information shared online, and international differences in privacy laws and standards.
This document discusses several legal and ethical issues related to e-marketing, including:
1. Software piracy is a major problem, with 40% of software being pirated in 2001 resulting in $10.7 billion in losses.
2. Ethics concerns what is right and wrong, while law is a public endeavor created by legislatures. There is often a close relationship between ethics and law.
3. Intellectual property on the internet is protected through copyright, patent, and trademark laws, though these protections are still developing for digital contexts.
A Revolution in Open Science: Open Data and the Role of Libraries (Professor ...LIBER Europe
This document discusses the opportunities and challenges of open science and open data. It argues that openly sharing scientific data and findings has significant benefits, including enabling faster scientific progress, deterring fraud, and supporting citizen science. However, for data to be truly open and useful to others, it needs to be accessible, intelligible, assessable, and reusable. The document also examines the roles and responsibilities of different stakeholders in working towards more open and reproducible science. This includes changing incentives for scientists, strategic funding for technical solutions from funders, and exploring how institutions like libraries and learned societies can help address the challenges of managing and making sense of the growing volume of research data.
This document discusses intellectual property in the context of information technology. It covers topics such as the value of intellectual property, how the internet has impacted copying and distribution, different types of intellectual property like patents, copyrights, and trade secrets. It also discusses debates around intellectual property protection including arguments around incentivizing innovation versus allowing works to enter the public domain. Open source software is presented as an alternative intellectual property model.
This slide is the final presentation for the Green IT course I have followed. This short course is all about how you can do to preserve, and avoid abusing the environment and having the sustainable development. In this presentation, my group show the terminology "The Internet of Things" and all benefits as well as its impacts to the human beings and the world. Feel free to give comments and information to me. Thank you and enjoy :)
This document provides an overview of open data and data sharing in archaeology. It discusses intellectual property rights related to archaeological data, copyright issues, open access and open data principles. The benefits of data sharing are outlined, such as increased visibility, verification of research and enabling new collaborations. However, barriers to sharing such as lack of incentives and concerns about misuse are also covered. The document provides guidance on using open licenses, depositing data in repositories, and attributing data to maximize the benefits of open sharing practices.
This document discusses key issues related to privacy and the internet. It outlines different approaches to defining and protecting privacy, noting that privacy means different things to different people and cultures. It also discusses the trade-offs associated with privacy regulation and the challenge of controlling information online given factors like digitization, ubiquitous networks, and the user generation of large amounts of content. The document advocates for an alternative approach focused on education, empowerment, and targeted enforcement rather than anticipatory regulation.
This document discusses legal and ethical issues related to data sharing. It covers rights and copyright regarding data, how to address ethics when sharing personal data under GDPR, and obtaining consent from participants. Guidelines are provided for discovering and accessing shared data from repositories. Questions about data sharing are welcomed.
workshop session delivered alongside 'Making your thesis legal' workshop in July and September 2013 to PhD, MPhil, DrPh students who are completing their thesis. Discusses standards for sharing data, issues that need addressing, formats, data protection, usability, licenses
Slides Part 02 Copyright Law for Digital teaching and Learning May 2014Darius Whelan
Creative Commons is a non-profit organization that offers copyright licenses to allow sharing and use of creative work. They provide standardized click-through licenses that are free to use. Their licenses allow creators to choose various levels of copyright protection for their work from full copyright to public domain. Creative Commons licenses help express a preference for sharing creative works while still providing attribution to the original creators. Copyright law protects the expression of ideas but not the underlying ideas or facts themselves. Certain uses of copyrighted material may be allowed, such as incidental inclusion in another work. Open access policies aim to provide unrestricted access to scholarly works. Free and open source software licenses also govern how software and content can be shared and modified.
Data Mining and Big Data Challenges and Research OpportunitiesKathirvel Ayyaswamy
The document discusses 10 challenging problems in data mining research. It summarizes each problem with 1-2 paragraphs explaining the challenges. Some of the key problems discussed include developing a unifying theory of data mining, scaling up for high dimensional and streaming data, mining complex relationships from interconnected data, ensuring privacy and security of data, and dealing with non-static and unbalanced data. The document advocates that more research is needed to address these issues and better integrate data mining with database systems and domain knowledge.
Similar to Ethical and legal issues in making research data open (20)
ESR spectroscopy in liquid food and beverages.pptxPRIYANKA PATEL
With increasing population, people need to rely on packaged food stuffs. Packaging of food materials requires the preservation of food. There are various methods for the treatment of food to preserve them and irradiation treatment of food is one of them. It is the most common and the most harmless method for the food preservation as it does not alter the necessary micronutrients of food materials. Although irradiated food doesn’t cause any harm to the human health but still the quality assessment of food is required to provide consumers with necessary information about the food. ESR spectroscopy is the most sophisticated way to investigate the quality of the food and the free radicals induced during the processing of the food. ESR spin trapping technique is useful for the detection of highly unstable radicals in the food. The antioxidant capability of liquid food and beverages in mainly performed by spin trapping technique.
The binding of cosmological structures by massless topological defectsSérgio Sacani
Assuming spherical symmetry and weak field, it is shown that if one solves the Poisson equation or the Einstein field
equations sourced by a topological defect, i.e. a singularity of a very specific form, the result is a localized gravitational
field capable of driving flat rotation (i.e. Keplerian circular orbits at a constant speed for all radii) of test masses on a thin
spherical shell without any underlying mass. Moreover, a large-scale structure which exploits this solution by assembling
concentrically a number of such topological defects can establish a flat stellar or galactic rotation curve, and can also deflect
light in the same manner as an equipotential (isothermal) sphere. Thus, the need for dark matter or modified gravity theory is
mitigated, at least in part.
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...Leonel Morgado
Current descriptions of immersive learning cases are often difficult or impossible to compare. This is due to a myriad of different options on what details to include, which aspects are relevant, and on the descriptive approaches employed. Also, these aspects often combine very specific details with more general guidelines or indicate intents and rationales without clarifying their implementation. In this paper we provide a method to describe immersive learning cases that is structured to enable comparisons, yet flexible enough to allow researchers and practitioners to decide which aspects to include. This method leverages a taxonomy that classifies educational aspects at three levels (uses, practices, and strategies) and then utilizes two frameworks, the Immersive Learning Brain and the Immersion Cube, to enable a structured description and interpretation of immersive learning cases. The method is then demonstrated on a published immersive learning case on training for wind turbine maintenance using virtual reality. Applying the method results in a structured artifact, the Immersive Learning Case Sheet, that tags the case with its proximal uses, practices, and strategies, and refines the free text case description to ensure that matching details are included. This contribution is thus a case description method in support of future comparative research of immersive learning cases. We then discuss how the resulting description and interpretation can be leveraged to change immersion learning cases, by enriching them (considering low-effort changes or additions) or innovating (exploring more challenging avenues of transformation). The method holds significant promise to support better-grounded research in immersive learning.
Mending Clothing to Support Sustainable Fashion_CIMaR 2024.pdfSelcen Ozturkcan
Ozturkcan, S., Berndt, A., & Angelakis, A. (2024). Mending clothing to support sustainable fashion. Presented at the 31st Annual Conference by the Consortium for International Marketing Research (CIMaR), 10-13 Jun 2024, University of Gävle, Sweden.
Current Ms word generated power point presentation covers major details about the micronuclei test. It's significance and assays to conduct it. It is used to detect the micronuclei formation inside the cells of nearly every multicellular organism. It's formation takes place during chromosomal sepration at metaphase.
Immersive Learning That Works: Research Grounding and Paths ForwardLeonel Morgado
We will metaverse into the essence of immersive learning, into its three dimensions and conceptual models. This approach encompasses elements from teaching methodologies to social involvement, through organizational concerns and technologies. Challenging the perception of learning as knowledge transfer, we introduce a 'Uses, Practices & Strategies' model operationalized by the 'Immersive Learning Brain' and ‘Immersion Cube’ frameworks. This approach offers a comprehensive guide through the intricacies of immersive educational experiences and spotlighting research frontiers, along the immersion dimensions of system, narrative, and agency. Our discourse extends to stakeholders beyond the academic sphere, addressing the interests of technologists, instructional designers, and policymakers. We span various contexts, from formal education to organizational transformation to the new horizon of an AI-pervasive society. This keynote aims to unite the iLRN community in a collaborative journey towards a future where immersive learning research and practice coalesce, paving the way for innovative educational research and practice landscapes.
Sexuality - Issues, Attitude and Behaviour - Applied Social Psychology - Psyc...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
The debris of the ‘last major merger’ is dynamically youngSérgio Sacani
The Milky Way’s (MW) inner stellar halo contains an [Fe/H]-rich component with highly eccentric orbits, often referred to as the
‘last major merger.’ Hypotheses for the origin of this component include Gaia-Sausage/Enceladus (GSE), where the progenitor
collided with the MW proto-disc 8–11 Gyr ago, and the Virgo Radial Merger (VRM), where the progenitor collided with the
MW disc within the last 3 Gyr. These two scenarios make different predictions about observable structure in local phase space,
because the morphology of debris depends on how long it has had to phase mix. The recently identified phase-space folds in Gaia
DR3 have positive caustic velocities, making them fundamentally different than the phase-mixed chevrons found in simulations
at late times. Roughly 20 per cent of the stars in the prograde local stellar halo are associated with the observed caustics. Based
on a simple phase-mixing model, the observed number of caustics are consistent with a merger that occurred 1–2 Gyr ago.
We also compare the observed phase-space distribution to FIRE-2 Latte simulations of GSE-like mergers, using a quantitative
measurement of phase mixing (2D causticality). The observed local phase-space distribution best matches the simulated data
1–2 Gyr after collision, and certainly not later than 3 Gyr. This is further evidence that the progenitor of the ‘last major merger’
did not collide with the MW proto-disc at early times, as is thought for the GSE, but instead collided with the MW disc within
the last few Gyr, consistent with the body of work surrounding the VRM.
The cost of acquiring information by natural selectionCarl Bergstrom
This is a short talk that I gave at the Banff International Research Station workshop on Modeling and Theory in Population Biology. The idea is to try to understand how the burden of natural selection relates to the amount of information that selection puts into the genome.
It's based on the first part of this research paper:
The cost of information acquisition by natural selection
Ryan Seamus McGee, Olivia Kosterlitz, Artem Kaznatcheev, Benjamin Kerr, Carl T. Bergstrom
bioRxiv 2022.07.02.498577; doi: https://doi.org/10.1101/2022.07.02.498577
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...Sérgio Sacani
Context. With a mass exceeding several 104 M⊙ and a rich and dense population of massive stars, supermassive young star clusters
represent the most massive star-forming environment that is dominated by the feedback from massive stars and gravitational interactions
among stars.
Aims. In this paper we present the Extended Westerlund 1 and 2 Open Clusters Survey (EWOCS) project, which aims to investigate
the influence of the starburst environment on the formation of stars and planets, and on the evolution of both low and high mass stars.
The primary targets of this project are Westerlund 1 and 2, the closest supermassive star clusters to the Sun.
Methods. The project is based primarily on recent observations conducted with the Chandra and JWST observatories. Specifically,
the Chandra survey of Westerlund 1 consists of 36 new ACIS-I observations, nearly co-pointed, for a total exposure time of 1 Msec.
Additionally, we included 8 archival Chandra/ACIS-S observations. This paper presents the resulting catalog of X-ray sources within
and around Westerlund 1. Sources were detected by combining various existing methods, and photon extraction and source validation
were carried out using the ACIS-Extract software.
Results. The EWOCS X-ray catalog comprises 5963 validated sources out of the 9420 initially provided to ACIS-Extract, reaching a
photon flux threshold of approximately 2 × 10−8 photons cm−2
s
−1
. The X-ray sources exhibit a highly concentrated spatial distribution,
with 1075 sources located within the central 1 arcmin. We have successfully detected X-ray emissions from 126 out of the 166 known
massive stars of the cluster, and we have collected over 71 000 photons from the magnetar CXO J164710.20-455217.
8. Ethical concerns – Part II
• Unintended consequences and misinterpretation
• Publicly available data: social media
• Dual use
9. Human Genome Project
Bermuda principles
1. Automatic release of sequence assemblies
larger than 1 kb (preferably within 24 hours).
2. Immediate publication of finished annotated
sequences.
3. Aim to make the entire sequence freely
available in the public domain for both
research and development in order to
maximise benefits to society.
One of the success stories of open research data is the research on Malaria
“Open Source Malaria is discovering and developing new medicines for the treatment of malaria, a terrible disease that kills about 2000 people each day. The project's innovation is to use open source principles in drug discovery: Anyone can take part at any level of the project, all data and ideas are freely shared and there will be no patents. The project therefore operates without secrecy. About 100 people have so far contributed from all over the world - you can too.
Traditionally research into new medicines is secretive. In part this is because the pharmaceutical industry needs patents in order to be financially sustainable. But the secrecy is a weakness because it can lead to expensive duplication of effort and different groups of scientists not learning from each other. Open Source Malaria removes the secrecy so that the research can be carried out more efficiently and the public and patient groups can become involved in the work. In much the same way as open source software has made such great products we all use (like the Firefox web browser and the Android operating system), we think that open source malaria can produce high quality, low-cost medicines that could help a lot of people.To date four sets of molecules have been pursued in the OSM consortium. The starting points came from big pharmaceutical companies like GlaxoSmithKline and Pfizer. The OSM community then worked to improve these compounds by making them more active against the malaria parasite and by making them longer-lasting in the body. The current series, Series 4, is able to cure mice of malaria. This means that we're close to taking a compound into clinical trials where it could be used by people. If that is successful we would have achieved something really new - we would have taken a public domain compound all the way to market where it could be made at low cost by generics manufacturers. If we can do this then we can start asking whether we could envisage a whole new pharmaceutical industry based on open source principles. Could we use open source pharma to discover new medicines for heart disease, or Alzheimer's? How about cancer, or medicines to fight superbugs? We won't know until we try. It would be fantastic if you could help do that.” From: http://thinkable.org/submission/2055
https://www.youtube.com/watch?t=105&v=gCOokjOiVTc
Why make research data open?
There seems to be a consensus, especially amongst policy makers that open access to research data is a good thing; a desirable goal that the research community should pursue.
Ideas about open access to research data tend to build on a set of norms, deriving from Mertonian norms:
Research data is a public or common good
Verify and reproduce
Minimize costs
Collaboration
Open access would result in a number of benefits:
Reinforces scientific inquiry,
diversity of analysis and opinions,
increased public understanding of and trust in science,
stimulate business activity,
helps to solve global challenges, etc.
A pair of computer scientists famously proved this point by combing movie recommendations found on the Internet Movie Database with the Netflix data, and they learned that people could quite easily be picked from the Netflix data.
“On October 2, 2006 Netflix, the “world’s largest online movie rental service,” publicly released one hundred million records revealing how nearly a half-million of its users had rated movies from December 1999 to December 2005.90 In each record, Netflix disclosed the movie rated, the rating assigned (from one to five stars), and the rate of the rating. Netflix first anonymized the records,
removing identifying information like usernames, but assigning a unique user identifier to preserve rating-to-rating continuity. Thus, researchers could tell that user 1337 had rated Gattaca a 4 on March 3, 2003, and Minority Report a 5 on November 10, 2003. Netflix had a specific profit motive for releasing these records. Netflix thrives by being able to make accurate movie recommendations; if Netflix knows, for example, that people who liked Gattaca will also like The Lives of Others, it can make recommendations that keep its customers coming back to the website. To improve its recommendations, Netflix released the hundred million records to launch what it called the “Netflix Prize,” a prize that took almost three years to claim. The first team that used the data to significantly improve on Netflix’s recommendation algorithm would win one million dollars. As with the AOL release, researchers have hailed the Netflix Prize data release as a great boon for research, and many have used the competition to refine or develop important statistical theories.
Two weeks after the data release, researchers from the University of Texas, Arvind Narayanan and Professor Vitaly Shmatikov, announced that“an attacker who knows only a little bit about an individual subscriber can easily identify this subscriber’s record if it is present in the [Netflix Prize] dataset, or, at the very least, identify a small set of records which include the
subscriber’s record.”97 In other words, it is surprisingly easy to reidentify people in the database and thus discover all of the movies they have rated with only a little outside knowledge about their movie-watching preferences. The resulting research paper is brimming with startling examples of the ease with which someone could reidentify people in the database, and has been celebrated and cited as surprising and novel to computer scientists.98 If an adversary—the term used by computer scientists99—knows the precise ratings a person in the database has assigned to six obscure movies,100 and nothing else, he will be able to identify that person 84 percent of the time.101 If he knows approximately when (give or take two weeks) a person in the database has rated six movies, whether or not they are obscure, he can identify the person 99 percent of the time.102 In fact, knowing when ratings were assigned turns out to be so powerful that knowing only two movies a rating user has viewed (with the precise ratings and the rating dates give or take three days), an adversary can reidentify 68 percent of the users.103
From:
Ohm, Paul, Broken Promises of Privacy: Responding to the Surprising Failure of Anonymization (August 13, 2009). UCLA Law Review, Vol. 57, p. 1701, 2010; U of Colorado Law Legal Studies Research Paper No. 9-12. Available at SSRN: http://ssrn.com/abstract=1450006
“First, we can immediately find his political orientation based on his strong opinions about “Power and Terror: Noam Chomsky in Our Times” and “Fahrenheit 9/11.” Strong guesses about his religious views can be made based on his ratings on “Jesus of Nazareth” and “The Gospel of John”. He did not like “Super Size Me” at all; perhaps this implies something about his physical size? Both items that we found with predominantly gay themes, “Bent” and “Queer as folk” were rated one star out of five. He is a cultish follower of “Mystery Science Theater 3000”. This is far from all we found about this one person, but having made our point, we will spare the reader further lurid details. “
From: http://arxivblog.com/?p=142
Another example: William Weld
“The Massachusetts Group Insurance Commission had a bright idea back in the mid-1990s—it decided to release "anonymized" data on state employees that showed every single hospital visit. The goal was to help researchers, and the state spent time removing all obvious identifiers such as name, address, and Social Security number. But a graduate student in computer science saw a chance to make a point about the limits of anonymization.Latanya Sweeney requested a copy of the data and went to work on her "reidentification" quest. It didn't prove difficult. Law professor Paul Ohm describes Sweeney's work:At the time GIC released the data, William Weld, then Governor of Massachusetts, assured the public that GIC had protected patient privacy by deleting identifiers. In response, then-graduate student Sweeney started hunting for the Governor’s hospital records in the GIC data. She knew that Governor Weld resided in Cambridge, Massachusetts, a city of 54,000 residents and seven ZIP codes. For twenty dollars, she purchased the complete voter rolls from the city of Cambridge, a database containing, among other things, the name, address, ZIP code, birth date, and sex of every voter. By combining this data with the GIC records, Sweeney found Governor Weld with ease. Only six people in Cambridge shared his birth date, only three of them men, and of them, only he lived in his ZIP code. In a theatrical flourish, Dr. Sweeney sent the Governor’s health records (which included diagnoses and prescriptions) to his office”.
From: http://arstechnica.com/tech-policy/2009/09/08/your-secrets-live-online-in-databases-of-ruin/
Providing open access to research data makes it increasingly difficult to maintain the confidentiality and privacy of research subjects. The standard practice in human subject research, as prescribed by current ethical codes and international laws, is to anonymise and de-identify the data that research participants provide and to properly inform them about how the data will be used and by whom. However, anonymisation of data does not suffice to mitigate the risk for all data sets. As a result of technological advances and the availability of increasingly more digital data sets, anonymisation can be more easily undone, for instance by combining and integrating different data sets. Furthermore, in some cases, measures or strategies to preserve confidentiality can be reverse-engineered.
In some research projects, anonymisation is not even possible because the data content enables identification and resists effective obfuscation. For instance, data from ethnographic studies of particular communities (e.g., transcribed interviews or field notes) may contain descriptions of practices and people that could be easily used to identify specific individuals. A concern is that the removal of identifying characteristics of research subjects may compromise the meaning, integrity and quality of the data. Even if effective anonymisation is technically feasible, research participants may still feel uncomfortable:
“The data that I collect is interviews with people. I have not engaged with open context because I deal with people who are doing illegal acts! Under my ethics review, I provide anonymity. To anonymise the data and to put it on an open context, adds another level. I doubt that many of my participants would agree if one of the stipulations was that I put my data into an open context”. (Ethical editorial reviewer, Archaeology)
A response to the privacy risks can be to explain to research participants the extent to which subsequent use can be effectively anticipated and to ensure, so far as possible, that the principles that apply to the governance of the data are consistent with the prevailing privacy expectations. False assurance, and associated drawbacks, may also be avoided by not “overpromising”, i.e., being transparent and realistic about the possibilities of re-identification. Asking consent can be problematic, because it difficult to predict how the data maybe used. Moreover, it may affect recruitment strategies and have additional affects on how researchers design and conduct their investigations. It might deter individuals, but it may also provide new opportunities to empower research participants. For instance, identifiable data provides participants with more possibilities of keeping track of their data, and may enable important additional outcomes. For example, some health research initiatives may find additional markers, symptoms or characteristics that locate particular research participants in high-risk categories for specific diseases. In these cases, ethical research practice encourages that the individual in question be contacted with this information, which would not be possible unless identifiable information was stored and shared.
There are also ethical concerns in re-using data. You should always check what the limitations are on the use of data, but in some case it is difficult to know what the expectations are.
For instance, in the case of publicly available information on for example social media sites. Someone posting on a website may not appreciate his or her comments being used for scientific research.
Fullerton and Lee have identified some ethically questionable secondary uses of data from the Human Genome Diversity Panel (HGDP). The HGDP is a collection that contains human tissue samples from 51 different human populations that were originally donated by multiple independent researchers over a period of years. The samples are archived together with geographic location and the sex of the individual from whom the sample was taken. Fuller and Lee reviewed the secondary uses of this collection and found that whereas the majority of studies were in line with the original intent of the collection, some published studies “could be regarded as controversial, objectionable or potentially stigmatizing in their interpretation”. One publication that they reviewed used samples from the HGDP to support the findings of a study that examined genetic signatures of Jewish Ancestry in European Americans, concluding that Jewish people are genetically distinct. Fullerton and Lee argue that such studies may cause indirect harm to participants, as they may support potentially unfavourable conclusions about populations from which participants were drawn. It may lead to discrimination or stigmatisation within populations or communities.
Fullerton, Stephanie M, and Sandra S-J Lee, “Secondary uses and the governance of de-identified data: Lessons From the Human Genome Diversity Panel”, BMC Medical Ethics, Vol. 12, No. 16, 2011. http://www.biomedcentral.com/1472-6939/12/16
Ibid., p. 3.
In some instances the intended secondary use or misappropriation of research data may cause unacceptable damage or distress to individuals and groups, as well as to research and the scientific enterprise. It can harm or wrong research participants or other stakeholders, particularly when results are perceived to be manipulated or distorted or when data are used for purposes that research participants themselves find objectionable. An example is the secondary use of culturally sensitive samples and data, such as human remains. In particular, misinterpretation or misappropriation can offend communities and individuals. Unintended secondary use can damage identities, reputations and relationships between individuals, and may even endanger research subjects or sites.
Well certainly there are a myriad of First Nations people who may feel offended or compromised if the raw materials related to religious locations, remains etc., are made publicly available and consumable in the wrong fashion. […] if you are putting native artefacts on display on line, information about them online, it really comes down to a whole hodgepodge of historic questions regarding how each particular tribal entity has been treated politically and also what their particular cultural feelings are about such matters. For groups that have less sensitivity about the remains of the deceased, you have to remember that these really represent scores of different cultural sensibilities. (Editorial reviewer, Archaeology)
One concern is that the misinterpretation of publicly available medical health data by patients, for instance, can put these patients at risk. It often requires considerable knowledge and expertise to evaluate and interpret research data properly and to use it to decide on medical diagnosis and treatment. Unintended use can be particularly problematic when it involves personal data about research participants’ ethnic or racial origins, political opinions, sexuality, religious beliefs, criminal background, or physical or mental health. It may result in stigmatisation, discrimination or other kinds of harm. In addition, research participants may feel wronged or betrayed when their expectations about the use of their information do not match with intentions and practices of new studies.
Dual use:
Some data can be used for research that could produce knowledge, products or technologies that benefit society, but could also pose a threat to public health, agriculture, plants, animals, the environment or material. Such dual-use data present an ethical dilemma for data sharing and open access: do the benefits of providing access to research data outweigh the costs? Sharing data on a virus, for example, may facilitate research on an antidote, but people with ill intent may also use it to disrupt societies.
One of the better-known examples of the dual use dilemma was the publication of two manuscripts that reported on the details of laboratory experiments with the H5N1 avian flu virus. The manuscripts concluded that the virus had a greater potential to be transmitted between mammals, including humans, than previously thought. After various reviews, the journals Nature and Science decided to publish the articles, because they believed that the benefits of publishing outweighed the risks. After the publication scientists agreed to a one year moratorium “to provide time to explain the public-health benefits of this work, to describe the measures in place to minimise possible risks, and to enable organizations and governments around the world to review their policies (for example, on biosafety, biosecurity, oversight and communication) regarding these experiments”.
Committee on Research Standards and Practices to Prevent the Destructive Application of Biotechnology, Biotechnology Research in an Age of Terrorism, National Research Council, The National Academies Press, Washington, DC, 2004.
Royal Society, op. cit., 2012, p. 58
Fouchier, Ron A. M., Adolfo García-Sastre, Yoshihiro Kawaoka, and 37 co-authors, “H5N1 Virus: Transmission Studies Resume for Avian Flu”, Nature, Vol. 493, No. 609, 31 January 2013. DOI:10.1038/nature11858
Representatives from sequencing centres around the world meet in Bermuda to draft a set of principles for free and rapid access to Human Genome Project data. These become known as the “Bermuda Principles”.
Data they get and the data they give
Keeping it consistent with the terms of use: don’t use it unless you’re sure it’s ok
Are you entitled to pass the data on? Pass on mindful of IPR issues.
When the law
Licensing provides a useful way to address intellectual property issues as well as ethical issues such as commercialisation and misappropriation and misuse of scientific information. These licensing models include Creative Commons licenses (as the most commonly employed forms of licensing), and other Licenses such as Government Open Licenses. The creators of research data and/or the repositories in which they are stored may make use of licenses to establish clear conditions related to how the research data should be used, including, for example, attributing content to original researchers and restrictions on modifying data. However, whilst licensing presents solutions such as important protections for stakeholders, it also introduces pitfalls such as the availability of licensed material that does not fully comply with the European Commission’s definition of open access. Irrespective of this, licensing continues to be a commonly employed practical solution in the move towards open access to research data.
Archaeology, physics and clinical data all require some form of professional accreditation or other access management review in order to enable researchers to access data. This professional gate-keeping solution allows these disciplines to manage legal and ethical compliance in relation to open access to research data. Specifically, they serve to identify true “professionals” who will have expertise in research methods or legal requirements such as confidentiality, privacy, data protection and research ethics. This solution ensures that the data is used responsibly and any potential issues associated with misuse are identified and mitigated. It also serves as a mechanism for enforcement, whereby individuals who do not use data responsibly may not be “approved” a second time.
In relation to Bioengineering, a large, multi-national data bank of biological material uses the following strategy:
They would identify from our website, which data sets they want, because the data sets are listed, […] They would write the access request email, which is on our website […] And they would say, ‘I'm interested in these data, can you pass me onto the relevant data owners.’ And then there is a form to fill out, which might not be the same for each dataset. And you basically have to explain who you are and why you would to use the data and what you want to use it for, and that’s just passed onto the data owners. And if the committee says yes, then we give them access. So the data are encrypted, so we would send a creater user password. (Scientific services manager, Bioengineering).
These processes are particularly effective in meeting requirements around intellectual property rights, data protection, secondary or dual use of research material and commercialisation. Therefore, it prevents unethical usage of the research data and aims to achieve and maintain legal compliance.
, the use of editorial review mechanisms emerges as a useful tool in ensuring ethical data practice and legal compliance. Internal processes have been adopted amongst our case study participants as a solution to the publication of research data that may have resulted from unethical practices and/ or in a manner that may be contrary to applicable laws. However, the editorial review solution may also introduce new pitfalls not so dissimilar to those associated with access management as described above. By way of specific example, Open Context adhere to an editorial review process that involves participation from local governments:
So what we do is, before it even goes to open context, our data go through a cleaning process, where the sites are allocated to a grid in the grid system and then we scrub the coordinate data and any data that are considered sensitive by our state partners, which can potentially differ state to state, and then we put it up on open context. So the only location information relates to our grid. (Editorial reviewer, Archaeology)
Finally, the use of existing ethical and legal guidance instruments, such as checklists or professional codes of conduct are also employed by our case study participants as a solution to assist stakeholders in effectively evaluating their responsibilities. However, soft-law measures carry the potential for pitfalls to the extent that although they encourage ethical practices and legal compliance, they do not mandate them.
An ethical editorial reviewer in the Archaeology case study explains the adoption of soft-law measures by their organisation:
[W]e take a lot of our clues on the ethical front from various journals and other kinds of venues where people publish this kind of material routinely and most journals and publishing houses have ethical guidelines that they follow. And we look to them sometimes for clues, because it’s quite similar in many ways.