Across jurisdictions and domains (academia, government, business) there has been much recent attention paid to open forms of knowledge production (e.g., open-source software, open data/metadata, open infrastructures) and the creation of open digital repositories for the unrestricted sharing of data, publications and other resources. This paper focuses on the latter, documenting and critically examining 14 different funding streams, grouped into six classes (institutional, philanthropy, research, audience, service, volunteer), being pursued by open digital repositories to support their endeavours, with a particular focus on academic research data repositories. Whilst open digital repositories are free to access, they are not without significant cost to build and maintain, and unstable and cyclical funding poses considerable risks to their futures and the digital collections they hold. While the political and ethical debate concerning the merits of open access and open data is important, we argue that just as salient are concerns with respect to long-term, sustainable funding for the operation and maintenance of open access digital repositories.
An overview of the case studies reviewing the hybrid business models being used to sustain digital content in the public and not-for-profit sectors in the US, Europe and Egypt.
An overview of the case studies reviewing the hybrid business models being used to sustain digital content in the public and not-for-profit sectors in the US, Europe and Egypt.
Data Innovation Spaces are identified by BDVA as a key instrument to foster the Data-Driven Innovation in Europe. They provide innovation and experimentation environments where companies in their respective ecosystems could have their data-driven and AI-related products and solutions piloted, tested, and exploited before going to the market. BDVA launches every year a process to identify and recognize relevant initiatives in Europe that meet specific quality criteria in infrastructures, services, projects, and sectors of application, ecosystem and sustainability (BDVA i-Spaces call for labels).
Enabling Data-Intensive Science Through Data InfrastructuresLIBER Europe
These slides are from a talk given at LIBER's 42nd annual conference by Carlos Morais Pires of the European Commission.
In light of the current data deluge, and plans by the European Commission to harness this deluge through the implementation of e-infrastructures for data driven science under Horizon 2020, Pires issued a call to action to libraries to engage in the data infrastructure and bring their own unique, and now much needed competencies, to bear in bringing meaning to, and spreading the word about, data-driven science.
Roadmaps, Roles and Re-engineering: Developing Data Informatics Capability in...LIBER Europe
A presentation by Dr. Liz Lyon of the United Kingdom Office for Library and Information Networking, as given at LIBER's 42nd annual conference in Munich, Germany.
Toward distributed infrastructures for digital preservation: the roles of col...Michael Day
Presentation given at: Curating our Digital Scientific Heritage: a Global Collaborative Challenge, 3rd International Digital Curation Conference, Washington, D.C., December 11-13, 2007
EUDAT 3rd Conference: What's on the Horizon? - Kimmo Koski, Managing Director...EUDAT
| www.eudat.eu | EUDAT 3rd Conference Opening Session: What's on the Horizon? - Kimmo Koski, Managing Director CSC - IT Center for Science, Finland & EUDAT Co-ordinator - Wednesday 24th September 2014, Amsterdam, the Netherlands
OSFair2017 Workshop | Towards a Policy Framework for the European Open Scienc...Open Science Fair
Workshop title: Towards a Policy Framework for the European Open Science Cloud
Workshop abstract:
The workshop provides a hands on approach in relation both to the understanding of the EU open science policies and their application by related stakeholders. It will seek to explore, propose and test different aspects of policy documents created by and for different types of stakeholders (e.g. RPOs, funders, policy makers etc) in the context of EOSC. Drawing on the work by the EOSC policy work, the workshop invites participants to bring their own policies or work on model policies to develop a simple but comprehensive policy document tailored to their needs and conforming to the EU policy and legal framework.
It is useful to the broader Open Science community as it brings together services, stakeholders and policies and allows for a better understanding of the interaction between different constituencies.
DAY 2 - PARALLEL SESSION 3
Data Innovation Spaces are identified by BDVA as a key instrument to foster the Data-Driven Innovation in Europe. They provide innovation and experimentation environments where companies in their respective ecosystems could have their data-driven and AI-related products and solutions piloted, tested, and exploited before going to the market. BDVA launches every year a process to identify and recognize relevant initiatives in Europe that meet specific quality criteria in infrastructures, services, projects, and sectors of application, ecosystem and sustainability (BDVA i-Spaces call for labels).
Enabling Data-Intensive Science Through Data InfrastructuresLIBER Europe
These slides are from a talk given at LIBER's 42nd annual conference by Carlos Morais Pires of the European Commission.
In light of the current data deluge, and plans by the European Commission to harness this deluge through the implementation of e-infrastructures for data driven science under Horizon 2020, Pires issued a call to action to libraries to engage in the data infrastructure and bring their own unique, and now much needed competencies, to bear in bringing meaning to, and spreading the word about, data-driven science.
Roadmaps, Roles and Re-engineering: Developing Data Informatics Capability in...LIBER Europe
A presentation by Dr. Liz Lyon of the United Kingdom Office for Library and Information Networking, as given at LIBER's 42nd annual conference in Munich, Germany.
Toward distributed infrastructures for digital preservation: the roles of col...Michael Day
Presentation given at: Curating our Digital Scientific Heritage: a Global Collaborative Challenge, 3rd International Digital Curation Conference, Washington, D.C., December 11-13, 2007
EUDAT 3rd Conference: What's on the Horizon? - Kimmo Koski, Managing Director...EUDAT
| www.eudat.eu | EUDAT 3rd Conference Opening Session: What's on the Horizon? - Kimmo Koski, Managing Director CSC - IT Center for Science, Finland & EUDAT Co-ordinator - Wednesday 24th September 2014, Amsterdam, the Netherlands
OSFair2017 Workshop | Towards a Policy Framework for the European Open Scienc...Open Science Fair
Workshop title: Towards a Policy Framework for the European Open Science Cloud
Workshop abstract:
The workshop provides a hands on approach in relation both to the understanding of the EU open science policies and their application by related stakeholders. It will seek to explore, propose and test different aspects of policy documents created by and for different types of stakeholders (e.g. RPOs, funders, policy makers etc) in the context of EOSC. Drawing on the work by the EOSC policy work, the workshop invites participants to bring their own policies or work on model policies to develop a simple but comprehensive policy document tailored to their needs and conforming to the EU policy and legal framework.
It is useful to the broader Open Science community as it brings together services, stakeholders and policies and allows for a better understanding of the interaction between different constituencies.
DAY 2 - PARALLEL SESSION 3
Birgit Plietzsch “RDM within research computing support” SALCTG June 2013SALCTG
An overview of Research Data Management: the research process from developing ideas to preservation of data; funder perspectives, the impact on the wider service, Data Asset Frameworks, preservation and access, and cost implications.
Business Models for Digital Content: A Perspective from UK Higher Education a...EUscreen
Presentation by Catherine Grout about a perspective on business models for digital content from UK Higher Education and on the work of the Strategic Content Alliance at the Second EUscreen International Conference on Use and Creativity, which took place at the National Library of Sweden, Stockholm, on September 15-16, 2011.
How the Research Data Service supports Open Research (aka Open Science) at the University of Edinburgh. Abridged slides used for presentation to Open Access Scotland meeting in Edinburgh on Wednesday 27th of March 2019.
Data ecosystems: turning data into public valueSlim Turki, Dr.
Africa Information Highway Live Exchange #Session 7
8 October 2021
The AIH Live Exchange between the Africa Information Highway Team, partners and countries is a free monthly webinar hosted by the African Development Bank to discuss topics related to government data and statistics. This webinar series is the main platform for countries to share their experiences and best practices around open data including using their Open Data Platform of the AIH.
This session is co-organized with the Luxembourg Institute of Science and Technology (LIST) which is a mission-driven Research and Technology Organization (RTO) that develops advanced technologies and delivers innovative products and services to industry and society. These innovations can also be used to solve several societal challenges, particularly in the areas of the environment, security, education and culture, sustainable development, as well as the efficient use of resources.
Official statistical data are recognized as high-value datasets for the society and economy, to enrich research, inform decision making or develop new products and services. The use of these authoritative data sources contributes to building a society with more empowered people, better policies, more effective and accountable decision-making, greater participation and stronger democratic mechanisms.
Official statistics are produced to be used and re-used to make an impact on society through a higher degree of openness and transparency while ensuring confidentiality and, at the same time, providing equal access to information to citizens.
The value of data lies in its use and re-use. In this interactive webinar, you will learn new techniques to improve the use and re-use of your statistical data, going beyond the provision logic and adopting the ecosystem mindset. You will:
● Sharpen your capacity at identifying and engaging users and re-users and stakeholders (data ecosystem mapping)?
● Effectively tackle technical and organizational barriers to stimulate data use and re-use?
● Smartly orchestrate a self-sustainable data ecosystem to increase the impact of statistical data.
This session is an opportunity for Regional members countries to '' Sharpen their skills in making data used and re-used by developing an ecosystem mindset to effectively build sustainable community of users around their Open Data Platform thus promoting transparency and better decision-making”
Topics covered at the workshop address basic questions related to Research Data Management for open data, which include preparing a Research Data Management (RDM) plan, licensing data and intellectual property, metadata and contextual description (documentation), ethical and legal aspects of sharing sensitive or confidential data, anonymizing research data for reuse, data archiving and long-term preservation, and data security and storage.
Event: http://conferences.nib.si/AS2015/default.htm
Related material: http://conferences.nib.si/AS2015/BookAS15.pdf
Agencies such as the NSF and NIH require data management plans as part of research proposals and the Office of Science and Technology Policy (OSTP) is requiring federal agencies to develop plans to increase public access to results of federally funded scientific research. These slides explore sustainable data sharing models, including models for sharing restricted-use data. Demos of these models and tips for accessing public data access services are provided as well as resources for creating data management plans for grant applications.
Enabling Re-Use and Sustainability: The role of information infrastructure fu...Platforma Otwartej Nauki
Conference Opening Science to Meet Future Challenges, Warsaw, March 11, 2014, organized by Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw.
Similar to Funding models for open access digital repositories (20)
Talk presented at TILT, Tilburg University, Netherlands, 14th March 2019. Relates to the book: Cardullo, P., di Feliciantonio, C. and Kitchin, R. (eds) (2019, June, in press) The Right to the Smart City. Emerald.
Cities around the world are pursuing a smart cities agenda in which digital technologies are used to manage cities. In general, these initiatives are promoted and rolled-out by governments and corporations and enact various forms of top-down, technocratic governance and reproduce neoliberal governmentality. Despite calls for the smart city agenda to be more citizen-centric and bottom-up in nature, how this translates into policy and initiatives is still weakly articulated and practiced. Indeed, there is little meaningful engagement by key stakeholders with respect to rights, citizenship, social justice, commoning, civic participation, co-creation, ethics, and how the smart city might be productively reimagined and remade. This talk advocates for the Right to the Smart City and considers how to produce a genuinely humanizing smart urbanism, both with respect to setting out a normative vision for smart cities rooted in ideas of fairness, equity, care, democracy and the public good, and enacting this vision through citizen-centric tactics.
This paper examines issues that are impeding the roll-out of smart city initiative and challenges the ideas, ideals and ideology of smart cities as presently conceived.
Citizenship, social justice, and the Right to the Smart Cityrobkitchin
This presentation was delivered at the Right to the Smart City workshop at Maynooth University, Sept 5-6 2017. It sets out a set of questions and theoretical concepts for thinking through issues of citizenship, social justice, and the right to the smart city.
Being a ‘citizen’ in the smart city: Up and down the scaffold of smart citize...robkitchin
This paper discusses the issue of citizens’ participation and rights in the smart city. It does so by drawing on and extending Sherry Arnstein’s seminal work (1969) conceptualising participation in planning and renewal programmes. We argue that citizenship in the smart city is rooted in a pragmatic and paternalistic discourse and practice, rather than in theories around rights and citizenship. Promoters of smart cities, including those advocating a citizen-centric version, tend to conflate limited forms of engagement as a user or consumer of services with citizenship and rights. We develop a modified version of Arnstein’s ladder ― the ‘Scaffold of Smart Citizen Participation’ ― as a conceptual tool to unpack the diverse ways in which the smart city frames citizens and measure smart citizen inclusion, participation, and empowerment in Dublin, Ireland.
Data-driven urbanism (Amsterdam, Jan 2017)robkitchin
This talk details the shift from data-informed urbanism to data-driven urbanism, the use of urban big data and smart city technologies in urban governance, and outlines various concerns and critiques.
The ethics and risks of urban big data and smart citiesrobkitchin
This slidedeck provides a brief introduction to the ethics and risks associated with urban big data and smart cities and was presented at the launch of the Institute for Data, Systems and Society at MIT, Sept 2016
Over the past three decades city infrastructure and services have increasingly become digitally networked, programmable and data-driven. Moreover, citizens now regularly use mobile spatial media to mediate their spatial behavior and urban experiences and share information via crowdsourced platforms. As a result we are ever more living in the era of smart urbanism — city systems can be operationally managed dynamically using algorithms processing urban big data, citizens can access and contribute live information about the city, and planners and policy makers can redeploy new streams of data to model and plan the city with increasing granularity. The development of smart urbanism poses opportunities and challenges for urban planning, reshaping how we come to know and govern cities, and this talk will examine these drawing on research conducted in Boston and Dublin.
Big data, new epistemologies and paradigm shiftsrobkitchin
This presentation examines how the availability of Big Data, coupled with new data analytics, challenges established epistemologies across the sciences, social sciences and humanities, and assesses the extent to which they are engendering paradigm shifts across multiple disciplines.
The ethics of urban big data and smart citiesrobkitchin
This presentation discusses the ethical, privacy and security implications of urban big data and smart city technologies and forwards a set of potential solutions for minimizing harms.
This paper was presented at the 'Towards a Magna Carta for Data' workshop at the RDS in Dublin, Sept 17th. It discusses how considerations of the ethics of big data consist of much more than the issues of privacy and security that it often gets boiled down to, and argues that the various ethical issues related to big data are multidimensional and contested; vary in nature across domains, and which ethical philosophy is adopted matters to the deliberation over data rights.
The Real-Time City? Data-driven, networked urbanism and the production of sm...robkitchin
Keynote talk presented at IGU Urban conference in Dublin, August 9th. The paper discusses the transition from data-informed to data-driven, smart cities and the impact of such a transition on city governance and wider society.
Why the National Spatial Strategy failed and prospects for the National Plann...robkitchin
This talk delivered at the MacGill Summer School in Glenties, Donegal as part of a panel on the National Spatial Strategy and where next for spatial planning in Ireland. It sets out the history of spatial planning in Ireland and why the NSS failed and discusses the prospects for a new National Planning Framework
Big data and smart cities: Key data issuesrobkitchin
This presentation was delivered at the first meeting of the Irish Government Data Forum, July 14th 2015. It was designed to provide an overview of key data issues related to smart cities in order to set the scene for a discussion about the kinds of data issues the forum might explore across a range of domains.
Slides from the Privacy: Insights from Lawyers and Technologiest at Maynooth University, July 1st 2015. The talk argues that privacy is multidimensional in nature; notions and practices of privacy are changing rapidly; has all kinds of direct and indirect effects; technology and industry are running ahead of legislators; there is no teleological inevitability to the emerging privacy landscape; it is incumbent on states to address privacy issues and to find a balance with respect to interests of citizens, states and industry.
This slide set examines the contention that opening data is an inherently good thing - that the case for open data is an open and shut case. It sets out a contrary view that whilst open data is desirable, much more critical thinking is required as to what this means in practice and the possible negative implications of opening data, and calls for a wider debate about the relative merits and politics of open data and how we go about opening data.
Praxis and politics of urban data: Building the Dublin Dashboardrobkitchin
This paper was presented at the Association of American Geographers meeting in Chicago, April 22nd 2015.
This paper critically reflects on the building of the Dublin Dashboard (www.dublindashboard.ie) from the perspective of critical data studies. The Dashboard is a website that provides citizens, planners, policy makers and companies with an extensive set of data and data visualizations about Dublin City, including real-time information, indicator trends, inter and intra-urban benchmarking, interactive maps, the location of services, and a means to directly report issues to city authorities. The data used in the Dashboard is open and available for others to build their own apps. One member of the development team was an ethnographer who attended meetings, observed and discussed with key actors the creation of the Dashboard and its attendant praxis and politics up to the point of its launch in September 2014. This paper draws on that material to consider the contextual, contingent, iterative and relational unfolding of the Dashboard and the emergent politics of data and design. In so doing, it reveals the contested and negotiated politics of smart city initiatives.
Housing in Ireland: From Crisis to Crisisrobkitchin
This paper provides a macro analysis of Irish housing since 1991 and makes the argument that it has been perpetually in crisis and that we need to find a way to treat housing as a sector, not simply a market, and to create stability that will mitigate against boom/bust cycles.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Funding models for open access digital repositories
1. Rob Kitchin, National University of Ireland Maynooth
Sandra Collins, Royal Irish Academy
Dermot Frost, Trinity College Dublin
Funding models for open access
digital repositories
2. The future is data infrastructures
• Over the past two decades there has been a concerted effort to
produce a range of digital data infrastructures aimed at preserving
and sharing scientific and cultural data
• As well local and national instances there are large-scale EU
programmes such as European Strategy Forum on Research
Infrastructures (ESFRI) and e-Infrastructures Reflection Group (e-
IRG), and thematic large-scale European Research Infrastructure
Consortiums (ERICs) relating to supporting access to research data
in the humanities and social sciences, such as DARIAH (Digital
Research Infrastructure for the Arts and Humanities), CLARIN
(Common Language Resources and Technology Infrastructure), and
CESSDA (Council of European Social Science Data Archives), as well
as many others related to the sciences
• there is a commitment to making such repositories open access in
nature
3. Open access
• Open access in its purest form is “digital, online, free of
charge, and free of most copyright and licensing restrictions”
• It seeks to remove both “price barriers (subscriptions,
licensing fees, pay-per-view fees) and permission barriers
(most copyright and licensing restrictions)” (Suber 2013) so
that material is freely available “on the public internet” and
can be used for “any lawful purpose, without financial, legal,
or technical barriers other than those inseparable from
gaining access to the internet itself” (Budapest Open Access
Initiative 2002).
• Digital data and products paid for by the public purse are
seen as public goods and their sharing represents a public
good
4. Open access
• Rather than a standard form of open access being adopted a
range of different positions have emerged that take varying
stances on issues such as permission barriers, timing, who
pays and how for production
• Different models include:
– gratis OA (free of charge, but not free of copyright of licensing
restrictions)
– libre OA (free of charge and expressly permits uses beyond fair use)
– delayed OA (paid access initially, becoming open after a set time
period)
– (a) green and (b) gold OA (pay-for-production followed (a) delayed
publication in an open access repository, (b) immediate publication
5. Direct benefits Indirect benefits (costs avoided)
New research opportunities.
Scholarly communication/access to data.
Re-purposing and re-use of data.
Increasing research productivity.
Stimulating new network/collaborations.
Data available for teaching and student projects.
Knowledge transfer to industry.
Improves skills base.
Increasing productivity/economic growth.
Verification of research/research integrity.
Fulfilling mandate(s).
No re-creation/duplication of data.
No loss of future research opportunities.
Lower future preservation costs.
Re-purposing data for new audiences.
Re-purposing methodologies.
Use by new audiences.
Protecting return on earlier investment.
Tools and standards have potential to increase data
quality.
Reduces ad-hoc queries concerning data.
Near term benefits Long term benefits
Value to current researcher and students.
No data lost from researcher turnover.
Widens access where costs prohibitive for
researchers/institutions.
Short term re-use of well curated data.
Secure storage for data intensive research.
Availability of data underpinning publications.
Secures value to future researchers and students.
Adds value over time as collection grows and
develops critical mass.
Increases speed of research and time to realise
impacts.
Stimulates new research questions, especially
relating to linked and derived data.
Private benefits Public benefits
Benefits to sponsors/funder of research/archive.
Benefits to researchers and institutions.
Fulfil grant obligations.
Increased visibility/citation.
Commercialising research.
Input for future research.
Motivating new research.
Catalysing new companies and high skills
employment.
Transparency in research funding.
6. Funding dilemma
• Rhetoric of state and supra-national agencies suggests that open access
digital data repositories will be financed by the public purse through core
grants
• The reality for many repositories is somewhat different
• Many national archives receive c.70% (+/-20%) of their funding directly
from the state, making up difference through other funding mechanisms
• Repositories that are not national in status are less likely to secure
sufficient on-going state subventions and are therefore under more
pressure to identify and source other funding streams
• Consequently, whilst there is significant pressure being exerted on
institutions to make their digital data collections open for wider use at no
cost, many open access digital data repositories lack a sustainable
financial model, placing their collections at risk
• One thing is clear: open access is not cost free, involving significant
labour, service and technology costs
7. Our challenge
• Our interest in funding models for open access repositories stems from
our work building the Digital Repository of Ireland (DRI; www.dri.ie).
• The DRI is a national research infrastructure for the humanities and social
sciences that also serves as a trusted digital repository for the Irish GLAM
sector (Galleries, Libraries, Archives, and Museums).
• It is an open digital repository using open source software and open
metadata CC-BY licence, and advocates for open access, however the
content owners can set the rights and access conditions, with some data
being under copyright and access to sensitive social science data
restricted for legal reasons
• Access to the DRI collections is free to use
• DRI was initially funded for four years by the Irish Higher Education
Authority through its PRTLI Cycle 5.
• Core funding for DRI terminates in December 2015 with an expectation
that it will transfer to a new, as yet undefined financial model.
• Challenge: How do we raise revenue when we cannot charge for use?
8. Funding models
• In the remainder of the paper we provide a synoptic overview
of the 14 potential funding sources, organised into a basic
typology of six classes (institutional, philanthropy, research,
audience, service, volunteer), and assess their relative merits
• We then discuss potential risks arising from failing to find
sustainable funding models
9. Institutional
Core funded The state provides the core operational costs through a
subvention as with other state data services such as libraries,
national archives, statistical agencies, etc. (non-cyclical, not in own
hands)
Consortia
(membership)
model
Build a consortium that collectively owns the data, pools labour,
resources, and tools and facilitates capacity building, but charges
a membership fee to consortium members to cover shared value-
added services. (non-cyclical, buy-in)
Built-in costs at
source
When research grants are awarded by funders applicants must
build in the costs for archiving the data and associated outputs in
a repository at the end of the project. This funding is transferred
to the repository for any services rendered. (non-cyclical, not in
own hands)
Public/private
partnership
Public/private partnerships, with the public sector providing the
data and private companies providing finance and value-added
services for access and re-use rights. (potentially cyclical, weak
commercial value)
10. Philanthropy
Philanthropy/
corporate
sponsorship
Funding is sourced from philanthropic organisations as grants,
donations, endowments and/or corporate sponsorship. If an
endowment is sizable then core services can be funded from the
interest. The donations can also be used to leverage other
funding, for example, matched money from the state. This can
also be reversed, so that state funding is used to try and leverage
philanthropic funding/corporate sponsorship. (Cyclical; focus
specific projects not core activities)
11. Research
Research funded The majority of funding is generated through the sourcing of
research grants from national and international sources, with
overheads being used to subvent core services. (cyclical, project
focused not core activity)
12. Audience (individual user)
Freemium
product/service
Offers end-users a graded set of options, including a free-of-charge
option that includes basic elements (e.g., limited features or sampled
dataset), with more advanced, valuing adding options (e.g., special
formats, additional functionality, tools) being charged a fee. Opens
up the product/service to a wider, low-end market and more causal
use, whilst retaining paid, high-end product/service for more
specialised users. (cyclical, requires strong take-up of services)
Content licensing Make the data free for non-commercial re-use, but charge for-profit
re-users. (need to own content, cost-recovery)
Infrastructural razor
& blades
An initial inexpensive or free trial is offered for products/services
(razor) that encourages take-up and continued paid use (blades). It
might be that access is free through APIs, but that computational
usage is charged on a pay-as-you-go model, with the latter cross-
subsidizing the former. Might be combined with freemium. (lacks
open access ethos)
13. Service (institutional user)
Premium
product/service
Offers end-users a high-end product or a service that adds
value to data (e.g., derived data, tools or analysis) for
payment, either as fixed payment, recurrent fees or pay-per-
use, without using monopoly rights. This enables the data
producer to gain first-mover advantages in the marketing and
the sale of complementary goods. (cyclical, runs against open
access ethos, pays for service not core activities)
Pay per purpose Charge for services beyond data use, such as ingest, archiving,
consulting and training services. (cyclical, pays for service not
core activities)
Free with advertising Products/services are provided for free, but users receive
advertising when using the product/service (revenue
generating) or the products/services are provided by different
companies and branded as such to encourage use of their
other products/services (cross-subsidization). (need very high
visits to generate sizable income)
White-label
development/platform
licensing
A customised product/service is created for a client and
branded for their use, with that client paying a one-off fee or
subscription that includes maintenance and update costs.
(cyclical, pays for service not core activities)
14. Volunteer
Open source Offers end-users data products/services for free, with the
infrastructure maintained on a voluntary basis, including
crowdsourcing. (building crowd, continuity/motivation,
trust/assurance)
15. Risks
• Failing to secure a funding model or to create a robust and transparent
cost recovery model puts an open access repository at risk and creates
potential impediments to future investment in at least four ways:
1. the most significant risk, is that the repository closes because it cannot
cover its core costs and puts the digital collections it holds at risk
2. there is the risk of major reputational damage to those associated with
the repository and its original funders. Moreover, if the repository closes
it will be very difficult to resurrect due to trust issues
3. repositories do not just consume financial resources, but also enable
funds to be leveraged on their technologies and collections. A repository
closing would foreclose any such leveraging
4. the closure of a repository leads to a significant loss of consolidated
human resource expertise, stakeholder networks, technical
infrastructures, and the legal and policy frameworks developed.
Restarting would be just as expensive as first time
16. Conclusion
• In this paper we have sought to document and critically
examine the benefits, challenges and risks of funding open
access data repositories
• We have outlined and appraised 14 different potential
funding streams, grouped into six classes
• With the exception of core funding from a state agency, each
funding streams have associated issues, such as being cyclical,
creating new services rather than supporting the core
functions, or undermine the notion of an open, free resource
• How to fund repositories in the open access age is a
significant challenge
• We are pursuing a blended model in discussion with our
stakeholders to find the most suitable model