Researchers and archivists worldwide have begun to investigate the scholarly potential of web archiving as a complement to the study of the ‘live’ web. But the grounding of ethical research decisions in web archiving
projects is still murky. Archiving web artifacts is a challenge. The procedures require detailed decisions about inclusion and exclusion, meaning, and access. It forces us confront our definitions of public and private, of published and unpublished, and of producer and audience. This proposal offers an exploration of the ethical questions that arise during the building a web archiving project.
This is a preliminary document I've prepared as a support reference for an upcoming forum on assessment. Its intended audience is academic staff who are highly unfamiliar with wiki technology and what they can be used for in an educational context.
This is a preliminary document I've prepared as a support reference for an upcoming forum on assessment. Its intended audience is academic staff who are highly unfamiliar with wiki technology and what they can be used for in an educational context.
Paper presented at the SALIS Conference 2009 in Halifax N.S. Discusses the current state of play in the sector and suggests possible courses for the future.
This presentation will help you to build on your knowledge about Creative Commons by exploring in detail the principles of the licences, the conditions that underpin all the licence expressions, and the resulting licences and their characteristics.
This presentation will introduce you to the Creative Commons organisation; the licences; and the way in which application of those licences has facilitated some inspirational examples of sharing in the GLAM sector.
This is a presentation that I will be giving at the Oct. 22-24, 2009 Interntaional Association of Business Communicators Southern Region Conference in Houston. Please check my blog or twitter feed for additional comments on engagement and community.
Information technology and resources are an integral and indispensable part of the contemporary academic enterprise. In particular, technological advances have nurtured a new paradigm of data-intensive research. However, far too much of this activity still takes place in silos, to the detriment of open scholarly inquiry, integrity, and advancement. To counteract this tendency, the University of California Curation Center (UC3) has been developing and deploying a comprehensive suite of curation services that facilitate widespread data management, preservation, publication, sharing, and reuse. Through these services UC3 is engaging with new communities of use: in addition to its traditional stakeholders in cultural heritage memory organizations, e.g., libraries, museums, and archives, the UC3 service suite is now attracting significant adoption by research projects, laboratories, and individual faculty researchers. This webinar will present an introduction to five specific services – DMPTool, DataUp, EZID, Merritt, Web Archiving Service (WAS) – applicable to data curation throughout the scholarly lifecycle, two recent initiatives in collaboration with UC campuses, UC Berkeley Research Hub and UC San Francisco DataShare, and the ways in which they encourage and promote new communities of practice and greater transparency in scholarly research.
Courtney Mumma presenteerde op de studiedag Een web van webarchieven (17-11-2016, NCDD/Beeld en Geluid/Netwerk Digitaal Erfgoed) over de web crawling tools en services van the Internet Archive.
Improving user engagement in a data repository with web analyticsIUPUI
Presented at LITA Forum 2013
Abstract: A goal of data curation activities is to enable discovery and reuse of valuable data sets. How well repositories facilitate these activities is difficult to measure with existing metrics. In this presentation we will discuss how to utilize usage statistics from DSpace (Apache SOLR) and Google Analytics to better understand how researchers discover, access, and use datasets archived in an institutional repository. Our focus will be on data analysis to explore the information seeking needs and behavior of data repository users. Ultimately, this analytic approach will inform the outreach, marketing, and impact evaluation of data repositories.
Also available at: http://hdl.handle.net/1805/3665
Joining it all up: developing research-practice linkages in the UKHazel Hall
Seminar presentation on efforts to strengthen research-practice linkages in librarianship and information science in the UK since 2009 presented to the School of Business and Economics, Åbo Akademi University, Finland on Thursday 13th March 2014. There is a fuller report of my work visit to Finland at http://hazelhall.org/2014/03/17/social-media-and-public-libraries-a-doctoral-defence-in-finland/.
Paper presented at the SALIS Conference 2009 in Halifax N.S. Discusses the current state of play in the sector and suggests possible courses for the future.
This presentation will help you to build on your knowledge about Creative Commons by exploring in detail the principles of the licences, the conditions that underpin all the licence expressions, and the resulting licences and their characteristics.
This presentation will introduce you to the Creative Commons organisation; the licences; and the way in which application of those licences has facilitated some inspirational examples of sharing in the GLAM sector.
This is a presentation that I will be giving at the Oct. 22-24, 2009 Interntaional Association of Business Communicators Southern Region Conference in Houston. Please check my blog or twitter feed for additional comments on engagement and community.
Information technology and resources are an integral and indispensable part of the contemporary academic enterprise. In particular, technological advances have nurtured a new paradigm of data-intensive research. However, far too much of this activity still takes place in silos, to the detriment of open scholarly inquiry, integrity, and advancement. To counteract this tendency, the University of California Curation Center (UC3) has been developing and deploying a comprehensive suite of curation services that facilitate widespread data management, preservation, publication, sharing, and reuse. Through these services UC3 is engaging with new communities of use: in addition to its traditional stakeholders in cultural heritage memory organizations, e.g., libraries, museums, and archives, the UC3 service suite is now attracting significant adoption by research projects, laboratories, and individual faculty researchers. This webinar will present an introduction to five specific services – DMPTool, DataUp, EZID, Merritt, Web Archiving Service (WAS) – applicable to data curation throughout the scholarly lifecycle, two recent initiatives in collaboration with UC campuses, UC Berkeley Research Hub and UC San Francisco DataShare, and the ways in which they encourage and promote new communities of practice and greater transparency in scholarly research.
Courtney Mumma presenteerde op de studiedag Een web van webarchieven (17-11-2016, NCDD/Beeld en Geluid/Netwerk Digitaal Erfgoed) over de web crawling tools en services van the Internet Archive.
Improving user engagement in a data repository with web analyticsIUPUI
Presented at LITA Forum 2013
Abstract: A goal of data curation activities is to enable discovery and reuse of valuable data sets. How well repositories facilitate these activities is difficult to measure with existing metrics. In this presentation we will discuss how to utilize usage statistics from DSpace (Apache SOLR) and Google Analytics to better understand how researchers discover, access, and use datasets archived in an institutional repository. Our focus will be on data analysis to explore the information seeking needs and behavior of data repository users. Ultimately, this analytic approach will inform the outreach, marketing, and impact evaluation of data repositories.
Also available at: http://hdl.handle.net/1805/3665
Joining it all up: developing research-practice linkages in the UKHazel Hall
Seminar presentation on efforts to strengthen research-practice linkages in librarianship and information science in the UK since 2009 presented to the School of Business and Economics, Åbo Akademi University, Finland on Thursday 13th March 2014. There is a fuller report of my work visit to Finland at http://hazelhall.org/2014/03/17/social-media-and-public-libraries-a-doctoral-defence-in-finland/.
Laboratories around the world continue to generate immense amounts of data that are non-proprietary and of value to the community. If available these data could dramatically reduce costs by minimizing rework and ultimately facilitating faster research. High quality reference data collections of chemical compound dictionaries, properties and spectra have been generated over many decades. With the advent of social networking tools and platforms such as Wikipedia, the community has an opportunity to contribute. The ChemSpider platform hosted by the Royal Society of Chemistry is a compound centric database with associated data. Already populated with almost 25 million unique compounds the community can deposit and host their own data, and curate and annotate existing data including those generated in Open Notebook Science Efforts. This presentation will provide an overview of progress to date and outline the vision of this community platform for chemistry and ensuring the longevity of chemistry reference data.
web 2.0, library systems and the library systemlisld
The Web 2.0 environment is characterized by concentration and diffusion. Library services are not well matched to this environment: they are fragmented and difficult to mobilize in user workflows. This presentation analyzes this situation and suggests some directions.
Library Connect Webinar - Making the case for sharing with indicators of rese...Library_Connect
University of Bath’s Alex Ball talks about the indicators of research data impact. From the Oct. 22, 2015 Library Connect webinar, How to assist researchers in sharing research data: http://libraryconnect.elsevier.com/library-connect-webinars?commid=175949
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
8. ...it’s more than simply
saving websites
Technical choices Methodological choices
proprietary or open inclusion and exclusion
source software? of materials in the
archive
institutional storage
of archived merging of producer
materials or hire a and user roles
service?
transparency for reuse
open access or
14. Web archivists
produce a data resource
questions of maintenance and service
questions of representation
15. Ethics of Web Archiving
Meghan Dougherty - LUC
Kirsten A. Foot - UW
Steven M. Schneider - SUNYIT
Editor's Notes
\n
There is a growing recognition in digital scholarship and cultural heritage that digital culture (or the material culture of the web) is heritage worth preserving and understanding. \n\nResearch methods in web studies and web historiography (namely the practices used to stabilize the volatile material culture of the web in order to study it) are shifting as scholars and archivists collaborate.\n\nresearchers examining the web need collections of web material that are stabilized, documented, and accessible.\n\nWeb archiving, in part, aims to meet that need. Web archiving is a process by which we preserve the cultural material of the web. A number of different organizations in different countries are active in this pursuit. The mission and scope of each project depends on the individual organizational mission, but a few examples are... \nThe Internet Archive and it’s Wayback Machine. And it’s European counterpart the European Archive.\nThe Library of Congress Minerva Project, and projects in other National Archives and Libraries around Europe\nAnd there are smaller boutique style archivists - such as Webarchivist.org, LiWA, HanzoWeb that provide different services from project consultation, analysis tools, and collection stewardship. \nAnd project specific archives - Archipol in Groningen, DACHS in Leiden, and the dr.dk archive in Aarhus\n\nThe primary differences between these examples is scope - in terms of level of access, what’s included and excluded, and the systematic and strategic approach to archiving - all of the examples I just gave are influenced by different disciplinary methods and professional practices from Libraries and archives, to social science, humanities, and even service to corporate libraries. \n\nFor the most part though, technically, they all work generally like the Internet Archive and the Wayback Machine here: \nIn this screenshot, I went to Archive.org, searched for a specific URL (whitehouse.gov), and through their rendering tool - the Wayback machine - it shows me a list of all the “impressions” - essentially these are snapshots of this URL at different times. \n\nSo, a goal of web archiving is to stabilize the volatile material on the web. People misunderstand the web itself as an archive, when in fact content is constantly changing, appearing and disappearing with little documentation of those changes. This leaves us with a landscape constantly in the present - evaluating structural development or deterioration is difficult if what is out there reflects a permanent present, rather than examinable past or predictable future. \n\nLet me show you some examples of what I mean by that and what you get in these archives.... If you begin clicking through the Wayback machine, here’s what you’ll see...\n\n\n\n
example of an impression from 1996\n
and then in 2003\n
a little redesign from the 2003 site.\n
and errors with our collecting and rendering tools\n
and data retrieval errors. \n\nSo, even with this very quick introduction to web archiving, we can start anticipating some of the big problems for research ethics: \nwe have technical questions and methodological questions that speak to representativeness, inclusion and exclusion, interpretation, and sometimes even fabrication.\n
All of that is to say, there is more to web archiving than simply saving web sites. \n\nin the process of archiving, we have to make ethical choices - choices in technology and choices in methodology.\nSometimes those choices are able to be handled ahead of time, and sometimes (like we saw in this quick whitehouse example) we have to assess them and address them as our technology fails and forces these issues on us.\n\ntechnically, we must make choices about proprietary or open source software; whether we store archived objects at our institutions or whether we hire out a service; and once we determine those choices, we ultimately have to negotiate open or restricted access to the archive we’ve created. \n\nIn addition to those technical barriers, we struggle with methodological questions that end up reflected in the representations we create about the social phenomena that we collect and study. We struggle with questions of what we include and exclude when building an archive. We know that those choices trickle down to influence meaning made within the archives we build. And this extends beyond the first decisions that come to mind like “what sites do I capture?” - it also includes questions about whether or not to include links, and it involves questions of file type and the nature of a digital document living on the web. It also involves questions about search strategy - not only ethically and methodologically justifying what you want to capture, but then also justifying how you go about finding that material. \n\n\n
we can find some guidance from work in traditional archiving. Typical concerns there are privacy, security, and copyright. \n\nArchivists (and scholars using archives in research) are acutely aware of their choices \n- choices about inclusion and exclusion can violate privacy, or influence meaning by possibly protecting privacy too carefully;\n- copyright and fair use policies that guide access and use may vary across institutions; and \n- security... By archiving and aggregating, materials that may seem innocuous by themselves might pose threats only after being stabilized and collected with other objects.\n\nWeb archiving takes on a new flavor by being a merger of stewardship methods and scholarly methods - the way social science researchers tackle these questions differs from how archivists do. \n\nThese traditional guidelines can only take us so far because web archiving takes on special characteristics of online communication and collaboration. \n
When archiving material culture of the web, \nanonymity, scope and reproducibility take on new meaning:\n\nAnonymity: Anonymity can offer protection in many cases, but create integrity problems in others. Anonymity can make it “... difficult to develop a reliable history of experiences”\n\nScope: The immediacy, reach, and interactivity made possible by material culture that is built in and accessible through networked media change power structures — the scope of who and what each web actor can reach is vastly changed.\n\nReproducibility: Information and cultural artifacts can be reproduced and stabilized online without loss, and while having not been removed from evolving “live” web circulation. This kind of reproducibility changes expectations of permanence online — especially for users who use the web as temporary storage, or a space for works in progress.\n\nThese three overarching ethical issues confront the researcher-archivist repeatedly when building a web archive. \n
The basic web archiving operations include collecting, cataloging, and display. For each web archiving operation, there are ethical questions.\n\nFirst, in the collection stage involves the creation of the archive itself. The creative collection process includes procedures for notification, decisions of inclusion and exclusion and decisions about robot behavior or machine-generated data.\n\n Once collection is underway, cataloging begins. That process poses questions of interpretation and more challenges of interpreting machine-generated data. \n\nDisplay of the resulting archive can vary depending on the goals of the scholar-archivist. I may want to display portions of my archived artifacts to illustrate my research, but I might also consider enabling wholesale access to and reuse of my archive itself to support more studies.\n\n\n\n\n
But as a web archivist (no matter how small or inaccessible my archive might be), I can’t make these decisions in a vacuum - there are other actors and stakeholders involved.\n\nPrincipal investigators who value intellectual freedom bear a measure of social responsibility, and in some circumstances bear personal liability. IRBs maintain standards for social responsibility and manage institutional liability. Collection commissioners, agents, exhibitors, and site producers have a responsibility to protect intellectual property and prevent harm to their users. Law enforcement bodies may become involved with security concerns. The public has some stake too as these kinds of archives become valuable heritage resources. \n\nWho should be charged with making what decisions and on what bases?\n
New methods for research online are blurring the boundaries between researcher and participant, between producer and user, between scholar and archivist.\n\nI think this brings me to one last question that I hope to get some help from all of you on...\n
This one last question that I think is often overlooked (and reasonably so given all the other minutiae of designing and executing a web archiving project) is this:\nwhat do we do with the archive we’ve produced?\n\nin addition to scholarship, the researcher-archivist produces a potential research resource — an archive that they may choose to maintain and serve to others, or not.\nRather than being a research byproduct — a large data set that according to IRB standards will be disposed of or properly stored after research is complete — web archivists may aim to create a reusable resource, an artifact in and of itself.\n\nDo we have an ethical obligation to our disciplines, to our research communities to share that resource? ..to encourage it’s reuse? If so, how are those kinds of contributions valued in different fields?\n\nand even if that is a deeply personal question for each research group to answer individually, we still have to ask... \n\nwho provides it? who gets access to what materials? when?\nand who decides?\n