The ISSN Register was created by UNESCO and France in the 1970s to index and identify analogue and digital serial publications, i.e. journals, newspapers, magazines, and later, websites and databases worldwide to foster scientific cooperation at a time of Cold War between the Western and Soviet blocs. The technology available at that time did not allow wide access to this database which was made available to Member States and subscribers. Despite the emergence of the internet and the world wide web and the web 2.0, this situation continued until 2013, when the ROAD database of open access scientific resources was made available on the web. This movement to free up ISSN data has accelerated under the impetus of a new management and thanks to the support of member countries, leading to the opening of the ISSN Portal in 2018. Today, the ISSN Portal offers a suite of services to libraries, publishers and the information industry that aims to trace as accurately as possible the trajectory of serial publications from their birth to their long-term preservation. This presentation is intended to provide an overview of the progress made since the opening of ROAD in 2013 and an outline of the 2024 strategy.
A talk given at 'Taking the Long View: International Perspectives on E-Journal Archiving', a conference hosted by EDINA and ISSN IC at the University of Edinburgh, September 7th 2015.
The Collections UofT Repository and Enterprise Content ManagementKellliBee
The Collections UofT Repository and Enterprise Content Management - use cases from archivists' perspectives for the Islandora digital collections platform.
The document discusses the purposes of web archives according to various national web archives. It identifies four main purposes: 1) Preservation of digital cultural heritage, with examples of archives focused on preserving websites as part of a country's heritage. 2) Responding to the risk of "digital dark ages" by selecting websites to archive. 3) Allowing viewing of past versions of websites to see their evolution over time. 4) Supporting future research by providing a source of information about society and the development of the web.
The document summarizes a presentation given on the Collections UofT Repository and Enterprise Content Management. It introduces Collections UofT as a platform that takes an enterprise content management approach to managing digital projects and assets across the University of Toronto in a collaborative way. Several use cases are described, including the UTARMS archives, digitized Nouwen family photograph albums, and sharing metadata between repositories using OAI-PMH. Challenges with the system are discussed along with potential solutions.
The Needs of Archives: 16 (simple) rules for a better archival managementTom Cobbaert
This document provides guidance on archival management through 16 rules organized under 6 themes: context, appraisal, arrangement, preservation, digitization, and access. It emphasizes the importance of understanding the history and use of archives, following national retention policies, avoiding unnecessary conservation through good storage practices, using digitization to increase access while preserving originals, and developing transparent access policies. While archival management is complex, these rules provide guidelines to help archives fulfill their roles in supporting rights, governance, history, and policy.
Human Scale Web Collecting for Individuals and Institutions (Webrecorder Work...Anna Perricci
This is the main slide deck for a workshop at iPRES 2018 on human scale web collecting. A primary focus of the presentation was the use of Webrecorder.io, a free, open source web archiving tool available to all.
The ISSN Register was created by UNESCO and France in the 1970s to index and identify analogue and digital serial publications, i.e. journals, newspapers, magazines, and later, websites and databases worldwide to foster scientific cooperation at a time of Cold War between the Western and Soviet blocs. The technology available at that time did not allow wide access to this database which was made available to Member States and subscribers. Despite the emergence of the internet and the world wide web and the web 2.0, this situation continued until 2013, when the ROAD database of open access scientific resources was made available on the web. This movement to free up ISSN data has accelerated under the impetus of a new management and thanks to the support of member countries, leading to the opening of the ISSN Portal in 2018. Today, the ISSN Portal offers a suite of services to libraries, publishers and the information industry that aims to trace as accurately as possible the trajectory of serial publications from their birth to their long-term preservation. This presentation is intended to provide an overview of the progress made since the opening of ROAD in 2013 and an outline of the 2024 strategy.
A talk given at 'Taking the Long View: International Perspectives on E-Journal Archiving', a conference hosted by EDINA and ISSN IC at the University of Edinburgh, September 7th 2015.
The Collections UofT Repository and Enterprise Content ManagementKellliBee
The Collections UofT Repository and Enterprise Content Management - use cases from archivists' perspectives for the Islandora digital collections platform.
The document discusses the purposes of web archives according to various national web archives. It identifies four main purposes: 1) Preservation of digital cultural heritage, with examples of archives focused on preserving websites as part of a country's heritage. 2) Responding to the risk of "digital dark ages" by selecting websites to archive. 3) Allowing viewing of past versions of websites to see their evolution over time. 4) Supporting future research by providing a source of information about society and the development of the web.
The document summarizes a presentation given on the Collections UofT Repository and Enterprise Content Management. It introduces Collections UofT as a platform that takes an enterprise content management approach to managing digital projects and assets across the University of Toronto in a collaborative way. Several use cases are described, including the UTARMS archives, digitized Nouwen family photograph albums, and sharing metadata between repositories using OAI-PMH. Challenges with the system are discussed along with potential solutions.
The Needs of Archives: 16 (simple) rules for a better archival managementTom Cobbaert
This document provides guidance on archival management through 16 rules organized under 6 themes: context, appraisal, arrangement, preservation, digitization, and access. It emphasizes the importance of understanding the history and use of archives, following national retention policies, avoiding unnecessary conservation through good storage practices, using digitization to increase access while preserving originals, and developing transparent access policies. While archival management is complex, these rules provide guidelines to help archives fulfill their roles in supporting rights, governance, history, and policy.
Human Scale Web Collecting for Individuals and Institutions (Webrecorder Work...Anna Perricci
This is the main slide deck for a workshop at iPRES 2018 on human scale web collecting. A primary focus of the presentation was the use of Webrecorder.io, a free, open source web archiving tool available to all.
Web archiving is the process of collecting portions of the World Wide Web for preservation and future access. It involves using web crawlers like Heritrix to automatically collect and archive web pages and sites. Large organizations like the Internet Archive and national libraries participate in web archiving to preserve culturally and historically important web content. However, web archiving faces challenges due to the scale of the web, rapid changes, and intellectual property issues.
The document summarizes the Web@rchive Austria project. It discusses how the Austrian National Library archives websites on the internet, including major domains in Austria and important websites that change regularly. It notes some of the challenges in web archiving like the short lifespan of webpages and how content selection must be careful. Examples of archived websites are provided from different time periods to show the project history.
Archiving for Now and Later - workshop at Common Field Convening 2019Anna Perricci
- Webrecorder is open source software that allows users to archive web pages in an interactive format, preserving elements that crawlers often miss like dynamic content. It provides a free tool for "archiving at a human scale" by capturing pages intentionally.
- True archiving requires more than just saving files - it involves appraisal, description, preservation, and access. Webrecorder helps with this process by allowing users to manage and share their archived collections online.
- While useful for individual archiving needs, Webrecorder is also working to improve tools for long-term stewardship of collections and address challenges around ethics, sustainability, and preserving rapidly changing websites.
This document provides an overview of archival technologies presented at the 46th Annual Georgia Archives Institute on June 10-21, 2013. The presentation introduces various archival management tools like Archon and Archivists' Toolkit for managing archival collections. It also discusses digital collection management software such as CONTENTdm and Islandora. Emerging standards, formats and linked open data initiatives are also covered. The goal is to help archivists identify existing and new technologies that can help manage and provide access to archival materials.
This document summarizes a presentation on Memento and web archiving. It discusses Memento, which aims to make navigating archived web pages easy. It also introduces SiteStory, an alternative method of web archiving that archives pages as users access them rather than through crawling. Finally, it discusses applying Memento to linked data by creating archives of DBpedia versions.
Leslie Johnston: Challenges of Preserving Every Digital Format, 2012lljohnston
The document discusses some of the challenges the Library of Congress faces in collecting and preserving digital content. It receives content in a wide variety of formats from different programs and partners. These include digitized newspapers, web archives, audiovisual content, tweets, and electronic publications. The Library uses various strategies to help manage this complex task, such as file format standards, multiple copies in different locations, and partnerships with other institutions. However, the diversity of formats and sources means preserving every digital format is extremely challenging.
Web archiving challenges and opportunitiesAhmed AlSum
The document discusses challenges and opportunities in web archiving. It outlines the key stages in the web archiving lifecycle including selection of content, harvesting techniques, storage formats and infrastructure, ways to provide access, and the role of community. Specific challenges are discussed such as representing dynamic and social media content, optimizing storage solutions, and addressing limitations of current access interfaces. Opportunities exist in focusing collection efforts on underrepresented regions, leveraging existing archived data, and developing innovative services and tools to support researchers.
This document discusses two digital library software systems: Greenstone and DSpace.
[1] Greenstone and DSpace allow librarians to build their own digital collections and customize them for their needs. Both systems aim to make it easy for others to build comprehensive digital libraries.
[2] The document describes the key features and functions of each software, including advantages like being open source and customizable, as well as disadvantages like technical knowledge requirements.
[3] Options for integrating the two systems are explored, including using the OAI-PMH protocol, the METS standard, or developing a direct bridge between the software like the StoneD module.
This document discusses various strategies and resources for archiving internet content for research purposes. It describes several existing large-scale web archives like the Internet Archive and Common Crawl, as well as national and institutional archives. It also outlines how researchers can collect targeted web archives using open-source tools or subscription-based services.
Web 3.0 is the current stage of the internet focused on machine-facilitated understanding to provide a more intuitive experience through personalized and intelligent search. Web 4.0 will be an open, linked, and intelligent web that functions similarly to the human brain. The document also defines electronic resources as information stored digitally, including e-books, e-journals, databases, and digital libraries. It discusses the need for and types of electronic resources as well as some issues with licensing, intellectual property rights, and infrastructure.
An hour lecture with hands-on on how to install the GREENSTONE DIGITAL LIBRAY. The seminar was sponsored by Baguio-Benguet Librarians Association, Inc. and conducted at the University Of the Cordilleras Library on July 19 & 20, 2010
Ipres2013 panel: Web Archiving – Lessons and Potential. This presentation highlights the main lessons learned while developing the Portuguese Web Archive and its potential use as an infrastructure for research.
This document provides guidance on building a digital archive using free or low-cost tools. It recommends using Internet Archive for unlimited file hosting and LibGuides for a centralized access point. Specific steps are outlined for setting up a collection on Internet Archive, adding files, and publishing the archive through LibGuides. Budget, reliability, and ease of use of Internet Archive are also discussed.
This document discusses how libraries can use Web 2.0 and 3.0 technologies to share and control their data. It covers topics like social tagging, sharing bibliographic data through open licenses, and allowing reuse and remixing of content. The document also discusses emerging technologies like semantic web and microformats that could allow machines to better interpret library content. It encourages libraries to engage with users on social networks and consider how to provide mobile services.
Kris Carpenter Negulescu Gordon Paynter Archiving the National Web of New Zea...Future Perfect 2012
This document summarizes lessons learned from archiving the New Zealand web domain. It discusses the legal requirements to archive internet documents, two approaches used - selective and domain harvesting. Challenges include defining a national domain, harvest scope and shape, policies, infrastructure needs, assessing quality, sustainability and being responsive. Final thoughts are on New Zealand facing similar challenges to peers and benefits of collaboration between institutions.
Slides for Web Archiving in the Heritage and Archive SectorsAnna Perricci
Webrecorder is a free and open source web archiving tool that allows users to create high-fidelity interactive captures of web pages. It aims to make web archiving accessible to all. The tool captures pages as users see them, including interactive elements. Captures can be browsed like live web pages. Webrecorder seeks to improve functionality through automation, search capabilities, and curation tools. It also works to establish sustainable funding models and provide training to support web archiving practices.
Presentada en la Jornada Internacional sobre Archivos Web y Depósito Legal Electrónico, en la Biblioteca Nacional de España (BNE), el día 9 de julio de 2013.
"Creating and Maintaining Web Archives"
Presented by Joanne Archer (University of Maryland), Tessa Fallon (Columbia University), Abbie Grotke (Library of Congress), and Kate Odell (Internet Archive)
Website Archivability - Library of Congress NDIIPP Presentation 2015/06/03Vangelis Banos
Website Archivability (WA) captures the core
aspects of a website crucial in diagnosing
whether it has the potentiality to be archived
with completeness and accuracy.
Web archiving is the process of collecting portions of the World Wide Web for preservation and future access. It involves using web crawlers like Heritrix to automatically collect and archive web pages and sites. Large organizations like the Internet Archive and national libraries participate in web archiving to preserve culturally and historically important web content. However, web archiving faces challenges due to the scale of the web, rapid changes, and intellectual property issues.
The document summarizes the Web@rchive Austria project. It discusses how the Austrian National Library archives websites on the internet, including major domains in Austria and important websites that change regularly. It notes some of the challenges in web archiving like the short lifespan of webpages and how content selection must be careful. Examples of archived websites are provided from different time periods to show the project history.
Archiving for Now and Later - workshop at Common Field Convening 2019Anna Perricci
- Webrecorder is open source software that allows users to archive web pages in an interactive format, preserving elements that crawlers often miss like dynamic content. It provides a free tool for "archiving at a human scale" by capturing pages intentionally.
- True archiving requires more than just saving files - it involves appraisal, description, preservation, and access. Webrecorder helps with this process by allowing users to manage and share their archived collections online.
- While useful for individual archiving needs, Webrecorder is also working to improve tools for long-term stewardship of collections and address challenges around ethics, sustainability, and preserving rapidly changing websites.
This document provides an overview of archival technologies presented at the 46th Annual Georgia Archives Institute on June 10-21, 2013. The presentation introduces various archival management tools like Archon and Archivists' Toolkit for managing archival collections. It also discusses digital collection management software such as CONTENTdm and Islandora. Emerging standards, formats and linked open data initiatives are also covered. The goal is to help archivists identify existing and new technologies that can help manage and provide access to archival materials.
This document summarizes a presentation on Memento and web archiving. It discusses Memento, which aims to make navigating archived web pages easy. It also introduces SiteStory, an alternative method of web archiving that archives pages as users access them rather than through crawling. Finally, it discusses applying Memento to linked data by creating archives of DBpedia versions.
Leslie Johnston: Challenges of Preserving Every Digital Format, 2012lljohnston
The document discusses some of the challenges the Library of Congress faces in collecting and preserving digital content. It receives content in a wide variety of formats from different programs and partners. These include digitized newspapers, web archives, audiovisual content, tweets, and electronic publications. The Library uses various strategies to help manage this complex task, such as file format standards, multiple copies in different locations, and partnerships with other institutions. However, the diversity of formats and sources means preserving every digital format is extremely challenging.
Web archiving challenges and opportunitiesAhmed AlSum
The document discusses challenges and opportunities in web archiving. It outlines the key stages in the web archiving lifecycle including selection of content, harvesting techniques, storage formats and infrastructure, ways to provide access, and the role of community. Specific challenges are discussed such as representing dynamic and social media content, optimizing storage solutions, and addressing limitations of current access interfaces. Opportunities exist in focusing collection efforts on underrepresented regions, leveraging existing archived data, and developing innovative services and tools to support researchers.
This document discusses two digital library software systems: Greenstone and DSpace.
[1] Greenstone and DSpace allow librarians to build their own digital collections and customize them for their needs. Both systems aim to make it easy for others to build comprehensive digital libraries.
[2] The document describes the key features and functions of each software, including advantages like being open source and customizable, as well as disadvantages like technical knowledge requirements.
[3] Options for integrating the two systems are explored, including using the OAI-PMH protocol, the METS standard, or developing a direct bridge between the software like the StoneD module.
This document discusses various strategies and resources for archiving internet content for research purposes. It describes several existing large-scale web archives like the Internet Archive and Common Crawl, as well as national and institutional archives. It also outlines how researchers can collect targeted web archives using open-source tools or subscription-based services.
Web 3.0 is the current stage of the internet focused on machine-facilitated understanding to provide a more intuitive experience through personalized and intelligent search. Web 4.0 will be an open, linked, and intelligent web that functions similarly to the human brain. The document also defines electronic resources as information stored digitally, including e-books, e-journals, databases, and digital libraries. It discusses the need for and types of electronic resources as well as some issues with licensing, intellectual property rights, and infrastructure.
An hour lecture with hands-on on how to install the GREENSTONE DIGITAL LIBRAY. The seminar was sponsored by Baguio-Benguet Librarians Association, Inc. and conducted at the University Of the Cordilleras Library on July 19 & 20, 2010
Ipres2013 panel: Web Archiving – Lessons and Potential. This presentation highlights the main lessons learned while developing the Portuguese Web Archive and its potential use as an infrastructure for research.
This document provides guidance on building a digital archive using free or low-cost tools. It recommends using Internet Archive for unlimited file hosting and LibGuides for a centralized access point. Specific steps are outlined for setting up a collection on Internet Archive, adding files, and publishing the archive through LibGuides. Budget, reliability, and ease of use of Internet Archive are also discussed.
This document discusses how libraries can use Web 2.0 and 3.0 technologies to share and control their data. It covers topics like social tagging, sharing bibliographic data through open licenses, and allowing reuse and remixing of content. The document also discusses emerging technologies like semantic web and microformats that could allow machines to better interpret library content. It encourages libraries to engage with users on social networks and consider how to provide mobile services.
Kris Carpenter Negulescu Gordon Paynter Archiving the National Web of New Zea...Future Perfect 2012
This document summarizes lessons learned from archiving the New Zealand web domain. It discusses the legal requirements to archive internet documents, two approaches used - selective and domain harvesting. Challenges include defining a national domain, harvest scope and shape, policies, infrastructure needs, assessing quality, sustainability and being responsive. Final thoughts are on New Zealand facing similar challenges to peers and benefits of collaboration between institutions.
Slides for Web Archiving in the Heritage and Archive SectorsAnna Perricci
Webrecorder is a free and open source web archiving tool that allows users to create high-fidelity interactive captures of web pages. It aims to make web archiving accessible to all. The tool captures pages as users see them, including interactive elements. Captures can be browsed like live web pages. Webrecorder seeks to improve functionality through automation, search capabilities, and curation tools. It also works to establish sustainable funding models and provide training to support web archiving practices.
Presentada en la Jornada Internacional sobre Archivos Web y Depósito Legal Electrónico, en la Biblioteca Nacional de España (BNE), el día 9 de julio de 2013.
"Creating and Maintaining Web Archives"
Presented by Joanne Archer (University of Maryland), Tessa Fallon (Columbia University), Abbie Grotke (Library of Congress), and Kate Odell (Internet Archive)
Similar to Can you save the web? Web Archiving! (20)
Website Archivability - Library of Congress NDIIPP Presentation 2015/06/03Vangelis Banos
Website Archivability (WA) captures the core
aspects of a website crucial in diagnosing
whether it has the potentiality to be archived
with completeness and accuracy.
BlogForever Crawler: Techniques and algorithms to harvest modern weblogs Pres...Vangelis Banos
Blogs are a dynamic communication medium which has been
widely established on the web. The BlogForever project has
developed an innovative system to harvest, preserve, manage
and reuse blog content. This paper presents a key component
of the BlogForever platform, the web crawler. More
precisely, our work concentrates on techniques to automatically
extract content such as articles, authors, dates and
comments from blog posts. To achieve this goal, we introduce
a simple and robust algorithm to generate extraction
rules based on string matching using the blog’s web feed in
conjunction with blog hypertext. This approach leads to a
scalable blog data extraction process. Furthermore, we show
how we integrate a web browser into the web harvesting process
in order to support the data extraction from blogs with
JavaScript generated content.
The theory and practice of Website ArchivabilityVangelis Banos
The document discusses website archivability and presents CLEAR, a method for evaluating the archivability of websites. CLEAR assesses website attributes like accessibility, cohesion, metadata, performance, and standards compliance to determine an overall archivability score. It was developed to help automate quality assurance for web archives by providing credible, live measurements of how completely and accurately a website can be archived. The authors also describe a demonstration of CLEAR called ArchiveReady.com and discuss the potential impact of evaluating website archivability for web professionals and archive operators.
CLEAR: a Credible Live Evaluation Method of Website Archivability, iPRES2013Vangelis Banos
This document presents CLEAR, a method for evaluating the archivability of websites. CLEAR assesses website attributes like accessibility, cohesion, metadata, performance, and standards compliance by performing evaluations of facets within each attribute. It generates an archivability score on a scale of 0-100% for each facet and attribute, and an overall score for the website. The document demonstrates CLEAR's implementation in a web application called ArchiveReady.com and discusses its potential benefits for web archivists and professionals to improve web archiving practices and preserve websites effectively. It also outlines some limitations and directions for future work, such as differentially weighting facet evaluations.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
1. Can we save the web?
WEB ARCHIVING
Vangelis Banos
http://vbanos.gr/
Unconference, 9-10 Δεκεμβρίου 2013
2. Can we save the web?
• What do you mean?
• What is web archiving;
• The practical use of web archives.
• Making your own web archive.
3. What is the World Wide Web?
A huge collection of digital documents (websites) which are
stored on special computers (web servers),
interconnected with each other.
8. Why save the web?
1. More and more items are born digital only material!
2. Some websites contain unique data and valuable
information.
–
Users take action and make important decisions based
on this information.
3. The web is a live record of contemporary:
1.
2.
3.
4.
Society,
Culture,
Science,
Economy.
4. Responsibility to preserve the web.
5. Transparency is promoted by saving the web.
9. Isn’t the web already safe?
• The answer is: NOT really!
• Websites are in danger:
– Organisations that maintain them stop caring about
them,
– Organisations than maintain them cease to exist,
– Natural disasters destroy computer facilities (fires,
floods, storms, etc)
– Technical problems damage websites (bugs, computer
viruses, backup failures, hardware failures)
– Their data are tampered on purpose!!! for many
reasons (political, financial, crime, etc)
10. A major blog hosting company was shut down
by the U.S. Authorities
19. WEB ARCHIVING
The process of collecting portions of
the World Wide Web to ensure the
information is preserved in an
archive for future researchers,
historians, and the public.
20. Challenges
• How it is done technically?
• What should I choose to archive?
– The whole website? some pages? Some files only?
• What do I want to do with the web archive I’m
creating?
• Who will have access?
• Who is the owner of the web archive content?
21. Archiving web pages is a technical challenge
File(s)
Software
Hardware
RECORD
Generic file archiving operation
22. Archiving web pages is a technical challenge
File(s)
File(s)
Software
File(s)
File(s)
Software
Hardware
File(s)
Software
File(s)
File(s)
Web archiving operation
Website
23. How it is done?
• Possible web archiving targets:
–
–
–
–
Government websites, Educational institutions,
People’s suggestions, Currently popular websites,
Popular media, Big companies,
Special events
25. Who is working on web archiving?
Many important organisations work on
web archiving since 1996.
26. International Internet Preservation Consortium
• IIPC Members
–
–
–
–
–
National Libraries,
Academic Libraries,
Cultural Organisations,
Universities,
Software development companies
• Web Archiving Timeline
– http://timeline.webarchivists.org/
27. Obligation of the National Library
• According to UNESCO:
– «a national library is responsible for the
collection and storage of the national
cultural heritage».
• In Greece, accoding to law No.3149/03:
– «publishers or authors (when there is no
publisher) of any printed material, are
obliged to submit three copies of their work
to the National Library of Greece. This
obligation also includes audiovisual and epublishing material».
• What about the Greek web?
28. Bibliothèque nationale de France
2006: legal deposit extended to
“signs, signals, writings,
images, sounds or messages of
any kind communicated to the
public by electronic means”.
The goal is not to gather the «best of the web»,
but to preserve a collection representative of the web
at a certain date.
29. Can we save the web?
• What do you mean?
• What is web archiving?
• The practical use of web archives.
• Making your own web archive.
36. Making your own web archive
• Using HTTrack software (Open Source)
– Installation
– Practical advice
– Features
– Usage scenarios
• Archive http://2013.futurelibrary.gr/
• Archive http://www.auth.gr/
37. Things worth considering
• Set Limits
– Filters to define the file types you want to copy.
– Bandwidth limits & Connection limits to avoid overloading the
site you are archiving AND avoid saturating your library network.
– Time limits
• Check the size of the files you have downloaded.
• Plan for disk space according to your needs.
• Check target website copyrights. Are you allowed to:
– Archive for personal use?
– Archive for public use in library computers?
– Archive to publish on the web?
• If you are not sure, please ask the website owner before
beginning web archiving.
38. Scenario: create your own mini web
archive in your library on a shoestring.
• Equipment:
– Typical Windows computer with the biggest possible hard
disk. (The more ΤΒ, the better).
– Equal backup disk (e.g. External USB hard disk).
– DSL Internet connection.
– HTTRACK open source software
• Select important local websites.
• Get permissions from website owners if necessary.
• Setup a regular web archiving schedule (e.g. Once per
month).
• Provide information and access to the web archive in
your library’s local computers for the public.
39. Can we save the web?
YES WE CAN!
• Questions?
• Thank you for your attention
• Contact:
– Web: http://vbanos.gr
– Email: vbanos@gmail.com
– Twitter: @vbanos