This document discusses significant environment information (SEI) needed to ensure long-term usability of digital objects. SEI is defined as relationships between a digital object and related environment information, qualified by purpose and weights. A tool called the Pericles Extraction Tool collects SEI by monitoring digital object usage. An experiment uses the tool to record environment, events, and inferred documentation when resolving anomalies in spacecraft payload data. Future work includes improving dependency inference and defining significance weights.
Trees4Future general presentation June 2012Trees4Future
Trees4Future is an Integrative European Research Infrastructure project that aims to integrate, develop and improve major forest genetics and forestry research infrastructures.
Presented by Jean-Yves Vion-Dury at the PERICLES workshop 'From Semantics of Change to Change of Semantics', University of Borås, 19 May 2015.
http://www.hb.se/en/About-UB/Current/Events/Pericles-F2F/Workshop/
Overview of the world of geospatial metadata, and the role of the EDINA service GoGeo in creating, saving, and discovering it. Presented on 19 June 2014 by Tony Mathys in Aberdeen, Scotland.
PERICLES workshop (IDCC 2016) - Introduction to the PERICLES projectPERICLES_FP7
This presentation was delivered by Simon Waddington from project partner King's College London during the PERICLES workshop entitled 'Appraisal, Quality Assurance and Risk Assessment in the Data Continuum'. The half-day workshop took place at IDCC on 25 February 2016 and presented the PERICLES approach to appraisal and quality assurance of complex digital objects.
See more at: http://www.dcc.ac.uk/events/idcc16/workshops#Worskhop%209
PERICLES Process Compiler - ‘Eye of the Storm: Preserving Digital Content in ...PERICLES_FP7
This presentation was delivered by Noa Campos López and Marcel Hellkamp from PERICLES project partner Georg-August-Universität Göttingen (GWDG), at the interactive workshop ‘Eye of the Storm: Preserving Digital Content in an Ever-Changing World’ (Wellcome Collection Conference Centre, London, 2 December 2016).
This full-day event aimed at introducing and experimenting with the PERICLES model-driven approach demonstrating its usefulness for managing change in evolving digital ecosystems.
http://pericles-project.eu/
Trees4Future general presentation June 2012Trees4Future
Trees4Future is an Integrative European Research Infrastructure project that aims to integrate, develop and improve major forest genetics and forestry research infrastructures.
Presented by Jean-Yves Vion-Dury at the PERICLES workshop 'From Semantics of Change to Change of Semantics', University of Borås, 19 May 2015.
http://www.hb.se/en/About-UB/Current/Events/Pericles-F2F/Workshop/
Overview of the world of geospatial metadata, and the role of the EDINA service GoGeo in creating, saving, and discovering it. Presented on 19 June 2014 by Tony Mathys in Aberdeen, Scotland.
PERICLES workshop (IDCC 2016) - Introduction to the PERICLES projectPERICLES_FP7
This presentation was delivered by Simon Waddington from project partner King's College London during the PERICLES workshop entitled 'Appraisal, Quality Assurance and Risk Assessment in the Data Continuum'. The half-day workshop took place at IDCC on 25 February 2016 and presented the PERICLES approach to appraisal and quality assurance of complex digital objects.
See more at: http://www.dcc.ac.uk/events/idcc16/workshops#Worskhop%209
PERICLES Process Compiler - ‘Eye of the Storm: Preserving Digital Content in ...PERICLES_FP7
This presentation was delivered by Noa Campos López and Marcel Hellkamp from PERICLES project partner Georg-August-Universität Göttingen (GWDG), at the interactive workshop ‘Eye of the Storm: Preserving Digital Content in an Ever-Changing World’ (Wellcome Collection Conference Centre, London, 2 December 2016).
This full-day event aimed at introducing and experimenting with the PERICLES model-driven approach demonstrating its usefulness for managing change in evolving digital ecosystems.
http://pericles-project.eu/
PERICLES Domain Specific Modelling - ‘Eye of the Storm: Preserving Digital Co...PERICLES_FP7
This presentation was delivered by Stratos Kontopoulos and Panos Mitzias from PERICLES project partner CERTH/ITI at the interactive workshop ‘Eye of the Storm: Preserving Digital Content in an Ever-Changing World’ (Wellcome Collection Conference Centre, London, 2 December 2016).
This full-day event aimed at introducing and experimenting with the PERICLES model-driven approach demonstrating its usefulness for managing change in evolving digital ecosystems.
http://pericles-project.eu/
PERICLES Domain-specific ontological representations and ontology evolutionPERICLES_FP7
Presented by Stratos Kontopoulos at the PERICLES workshop 'From Semantics of Change to Change of Semantics', University of Borås, 19 May 2015.
http://www.hb.se/en/About-UB/Current/Events/Pericles-F2F/Workshop/
PERICLES workshop (IDCC 2016) - Policy and Quality Assurance in the Data Cont...PERICLES_FP7
This presentation was delivered by Fabio Corubolo from project partner University of Liverpool during the PERICLES workshop that took place at IDCC16 on 25 February. The half-day workshop presented the PERICLES approach to appraisal and quality assurance of complex digital objects.
http://www.dcc.ac.uk/events/idcc16/workshops#Worskhop%209
PERICLES Building Digital Ecosystem Models - ‘Eye of the Storm: Preserving Di...PERICLES_FP7
This presentation was delivered by Johannes Biermann and Anna Eggers from PERICLES project partner State and University Library Goettingen, at the interactive workshop ‘Eye of the Storm: Preserving Digital Content in an Ever-Changing World’ (Wellcome Collection Conference Centre, London, 2 December 2016).
This full-day event aimed at introducing and experimenting with the PERICLES model-driven approach demonstrating its usefulness for managing change in evolving digital ecosystems.
http://pericles-project.eu/
Presented by Christian Muller at the PERICLES workshop 'From Semantics of Change to Change of Semantics', University of Borås, 19 May 2015.
http://www.hb.se/en/About-UB/Current/Events/Pericles-F2F/Workshop/
This presentation was delivered by Simon Waddington from project partner King's College London during the PERICLES workshop entitled 'Appraisal, Quality Assurance and Risk Assessment in the Data Continuum'. The half-day workshop took place at IDCC on 25 February 2016 and presented the PERICLES approach to appraisal and quality assurance of complex digital objects.
See more at: http://www.dcc.ac.uk/events/idcc16/workshops#Worskhop%209
Automatic policy application and change management - Acting on Change 2016PERICLES_FP7
This presentation was delivered by Fabio Corubolo (University of Liverpool) at PERICLES final project conference 'Acting on Change: New Approaches and Future Practices in LTDP' (Wellcome Collection Conference Centre, London, 30 Nov -1 Dec 2016).
In this 'PERICLES in practice' workshop, Fabio Corubolo used the EUMETSAT example to demonstrate the change of a policy in a real life scenario applying the Policy Editor and the Entity Registry Model Repository (ERMR).
The PERICLES in practice workshops presented specific outcomes of the PERICLES project set in an example workflow, combining tools to accomplish a goal defined by practitioners and derived from real life challenges they experience in their field of work.
PERICLES Workflow for the automated updating of Digital Ecosystem Models with...PERICLES_FP7
This presentation was delivered by Anna Eggers (State and University Library Göttingen) and Fabio Corubolo (University of Liverpool) at PERICLES final project conference 'Acting on Change: New Approaches and Future Practices in LTDP' (Wellcome Collection Conference Centre, London, 30 Nov -1 Dec 2016).
This 'PERICLES in practice' session aimed at demonstrating how the PET and the EcoBuilder tools can be used to generate information to populate Digital Ecosystem Model (entities) and how the model can be fed into the Entity Registry Model Repository (ERMR). This session showed how the PERICLES tools can contribute to automated model instance creation and automatic updating.
The 'PERICLES in practice' sessions presented specific outcomes of the PERICLES project set in an example workflow, combining tools to accomplish a goal defined by practitioners and derived from real life challenges they experience in their field of work.
TERN's Siddeswara Guru presents on the Australian Ecosystem Science Cloud, which will provide the ecosystem science community improved access to shared data, tools, platforms and computing resources.
Stuart Macdonald talks about the Research Data Management programme at the University of Edinburgh Data Library, delivered at the ADP Workshop for Librarians: Open Research Data in Social Sciences and Humanities (ADP), Ljubljana, Slovenia, 18 June 2014
MindTrek2011 - ContextCapture: Context-based Awareness Cues in Status UpdatesVille Antila
Presentation of an experimental mobile application, which allows users to add different descriptions of context information to their Facebook status updates. The meaningfulness and the usage of different context descriptions were evaluated in a two-week user trial. The results show that the most frequently used awareness cues in the test setting were location, surroundings, friends and activity. The results also indicate that user-defined semantic abstractions of context items (e.g. “home”, “work”) were often more informative and useful than more accurate indicators (e.g. the address or the name of the place). We also found out that using shared context from friends in vicinity (e.g. identifying the people around) needs careful design to overcome the extended privacy implications.
Researchers require infrastructures that ensure a maximum of accessibility, stability and reliability to facilitate working with and sharing of research data. Such infrastructures are being increasingly summarised under the term Research Data Repositories (RDR). The project re3data.org – Registry of Research Data Repositories – began to index research data repositories in 2012 and offers researchers, funding organisations, libraries and publishers an overview of the heterogeneous research data repository landscape. In December 2014 re3data.org listed more than 1,030 research data repositories, which are described in detail using the re3data.org schema (http://dx.doi.org/10.2312/re3.003). Information icons help researchers to identify easily an adequate repository for the storage and reuse of their data. This talk describes the heterogeneous RDR landscape and presents a typology of institutional, disciplinary, multidisciplinary and project-specific RDR. Further, it outlines the features of re3data. org and it shows current developments for integration into data management planning tools and other services.
By the end of 2015 re3data.org and Databib (Purdue University, USA) will merge their services, which will then be managed under the auspices of DataCite. The aim of this merger is to reduce duplication of effort and to serve the research community better with a single, sustainable registry of research data repositories. The talk will present this organisational development as a best practice example for the development of international research information services.
Birgit Plietzsch “RDM within research computing support” SALCTG June 2013SALCTG
An overview of Research Data Management: the research process from developing ideas to preservation of data; funder perspectives, the impact on the wider service, Data Asset Frameworks, preservation and access, and cost implications.
A brief overview of the development and current workflows for Research Data Management at Imperial College London, presented to colleagues at the University of Copenhagen and Roskilde University in Denmark.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
UiPath Test Automation using UiPath Test Suite series, part 4
IPRES 2014 paper presentation: significant environment information for LTDP
1. SIGNIFICANT ENVIRONMENT
INFORMATION FOR LTDP
Fabio Corubolo, Adil Hasan – University of Liverpool
Anna Eggers, Jens Ludwig - Göttingen State University Library
Mark Hedges, Simon Waddington - King’s College London
This project has received funding from the European Union’s Seventh Framework
Programme for research, technological development and demonstration under
grant agreement no FP7-601138 PERICLES.
2. Objective and outline
• Aim: Ensure long term usability of Digital Objects (DO)
• Usability of Digital Object usually requires access to parts of its
environment
• Define a broad set of information (Environment information)
• Consider its significance (Significant environment information)
• Explore and test pragmatic methods to collect such information
3. Environment information definition
• All the entities (DOs, metadata, policies, rights, services, users,
etc.) useful to correctly access, render and use the DO.
Refinement:
• The information about the set of relationships between the
source DO and any related objects from its environment.
4. Environment for a DO
• Technical system information (OS, system architecture, etc.)
• DO metadata (descriptive, structural, technical)
• User, policy, process information (User BG knowledge, …)
• Information necessary to make use of the object including:
• Auxiliary data (e.g. calibration data for to support sensor data)
• External documentation (e.g. specifications, related documents)
• Implicit knowledge about what data is useful to use the DO (e.g. the user
knowledge about what is relevant and what not in the collection)
• More…
5. No object is an island, entire of itself
• Digital objects are used in a rich environment
Digital object
Ext. Metadata
Environment
Storage Digital object
6. Digital object information
• Rich and varied terminology
• The scope of each term is not
absolutely defined
• We are aiming to support
object use: use-centric view
• First broad - Environment
information: more or less all
that sits outside of the DO
8. Significant Environment Information (SEI)
• Use of a DO has a purpose
• The purpose gives a scope to the dependent environment
information
• Weights can express the importance for a specific purpose
(definition)
We define SEI as the set of relationships between a DO and its
environment information qualified with purpose and weights
9. How to collect and measure SEI?
• Observe the use of DOs – in different phases of lifecycle
• in the environment of creation and use
• Collect dependencies for use (relationships to other DOs)
• Measure significance e.g. based on frequency of use
• Different semantics and factors for significance weights (value,…) – WIP
• Weights will change in time
• Sheer curation: curation activities integrated in the use
workflow; lightweight and transparent
10. Pericles Extraction Tool (PET)
• Open source* framework - builds on the SEI concepts
• Uses a sheer curation approach – right time and place
• Generic, modular, domain agnostic
• Collection by observation – monitoring changes in time
• Snapshot of the system environment
• To observe unstructured workflows
• https://github.com/pericles-project/pet
* Release due soon, approved but waiting for final stamps
11. PET Architecture and modules
• Available and used system resources;
• File format identification and
checksums;
• Currently running processes;
• Event information (file and network)
from processes;
• Graphic configuration information;
• MS Office and PDF font
dependencies.
• Native commands
13. How to setup PET for a use scenario
• PET is installed, configured, started on the machine where the
DOs are used – stays in monitoring mode
• The profile (modules and configuration) are use case specific
• The user interacts normally with the DOs while PET collects SEI
in the background
• The environment information, DO events and changes are
collected for future use and analysis
14. General scenario for PET
1. Use PET to collect environment information when-where the
DOs are used, based on profiles
--- We are now here ---
2. Analyse the information collected to infer new relationships
(also SEI) between DOs - forming a graph structure
3. Assign weights to relationships based on the purpose and
significance – weighted graph
15. Experiment: use case description
• Fictional scenario, based on operations for ISS SOLAR payload
• Operator’s task: resolve anomalies
• Process: extensive search in the archived data + documents
• Issue: how to preserve implicit information, help with overload
• PET task: record SEI for a specific anomaly
• monitor environment, record significant events, infer documentation
useful to solve the anomaly
• SEI: to identify and debug a specific anomaly, that is the implicit
operator knowledge
16. Experimental results (1)
An anomaly is reported in an handover sheet
The operator proceeds with
documentation search and
consultation, all tracked by PET
17. Experimental results (2)
• Environment monitoring
• Events, extraction on occurrence of events
• Leads to dependency inference
• In future work we consider more complex issues
• ‘noise’ from multitask,
• careful analysis of collected data in the next phases
18. Conclusions, Future work
• Define Significant Environment Information (SEI) for object reuse
• Base for dependency graphs weighted on significance and purpose
• Explain ways to obtain SEI and significance weights
• Present the PET tool – to collect SEI
• Show experimental results - initial dependency collection
Future:
• Improve: filtering, dependency inference
• Work on definition and semantics for significance weights
• Use weighted dependency graphs to support appraisal
19. Thank you!
More information:
• https://github.com/pericles-project/pet
20. About the PERICLES project
• Promoting and enhancing reuse of information throughout the
content lifecycle taking account of evolving semantics
• Ensure availability and reuse of digital objects for the next
generations
• Extensions to current preservation and lifecycle models to
address the evolution of dynamic heterogeneous resources and
their dependencies
• Models capturing intent and interpretative context: key to
achieving “preservation by design”
21. Facts & Figures
• Collaborative FP7 project on digital preservation
• 12 million Euro, co-funded by the European Commission
• 11 partners: research institutions, IT development and
application domain
• 6 European countries
• Feb 2013 – Feb 2017
• Project website: http://www.pericles-project.eu
22. Consortium
COORDINATOR: King’ s College London – UK
ACADEMIC PARTNERS:
Hoegskolan i Borås – University of Borås – SE
Georg-August-Universität Göttingen – DE
University of Liverpool – UK
Centre for Research and Technology Hellas – GR
University of Edinburgh – UK
NON-ACADEMIC PUBLIC SECTOR ORGANISATIONS
Tate – UK
Belgian User Service and Operation Centre - B.USOC – BE
PRIVATE SECTOR ORGANISATIONS
Dotsoft – GR
Space Applications Services NV/SA (SpaceApps) – BE
Xerox Research Centre Europe - FR
Editor's Notes
WE want to collect important information that could be lost if not gathered at the right time.
Users and their interaction with DOs are also to be considered part of the environment!
This is just ONE vision on the different sets of data. I think it’s a reasonable one, but not for sure the absolute truth.
Environment here thought as ‘where data lives’
Environment does not necessarily have a structure (metadata has usually standards) and that can include a lot of not necessarily related information.
This is to say, it’s still not qualified as ‘data about the data’ but as ‘where the data lives’; so likely much broader.
Another definition of environment is ‘anything that is not the object’ that is to say the universe - the object.
PLEASE NOTE: THIS IS One example – based on one scenario, I prefer to give you a complete example in one scenario, but there are many possible scenarios that can be addressed by PET with proper configuration and modules.
I will now introduce briefly a synthetic scenario (fictional) inspired by the BUSOC mission operators use case
- Busoc operators are sometime facing the task of resolving anomalies, such as when some instrument does not respond as expected
the process they follow is guided by their knowledge of the domain and involves research on the archived documentation and operation data can include for example solutions from previous anomalies, telemetry, console logs, meeting notes, emails, etc.
Such data, although present in the storage, requires experience and its selection is a task that requires specific knowledge that is usually passed from operator to operator
- the issue we want to address is that of preserving the useful information that is in the use of specific documents from the large collection in order to solve the issue, and help the operators with the information overload.
the task the PET tool is trying to accomplish is to record the SEI for this use case, for a specific anomaly. This is done by monitoring the environment and recording significant events (via a PET profile) and from there allow the inferring of new dependencies
dependencies between anomalies and mission documentation, in order to preserve useful information that is otherwise not captured.
The SEI in this case is EI that will help to identify and debug a specific anomaly
we set up a specific PET profile that tracks the use of relevant software on specific files, using the PET software monitor; this enables us to have a trace of the documents that have been used at a given moment in time
At the same time, it is possible to observe the ‘handover sheet’ and track the reporting of an anomaly start and end times
The connection between the documentation track and the ‘handover sheet’ tracking can allow us to infer the ‘anomaly solving time span’ (indicated with a red line in Figure 4) and assume there is a dependency between the solution to the anomaly and the documentation that was used between the start and end of the anomaly.
In future work we will consider more complex issues that we have ignored in this simplified example, such as the ‘noise’ that can be reported by the event tracking. This ‘noise’ can be for example due to the fact that users often multitask, so there can be unrelated documentation that was used but not relevant to the anomaly solution, or documentation that was quickly opened and closed may also indicate in some cases that the document was not relevant. We will explore also ways to obtain a fine-grained tracking, as for example to include what pages have been consulted in a document. We are planning to dedicate effort to a more careful analysis of the collected data in the next phases.
In this paper we presented our work on determining what information is significant to collect, from the widest set of the Environment Information. We presented a definition of Significant Environment Information that takes into account the purpose of use of a DO, and can apply to relationships with significance weights. We also presented ways to determine significance weights and their relations to the DO lifecycle.
Finally, we presented the tool we are developing to collect such information, together with its methods of extraction, and showed experimental results to support the importance of such information. We believe the importance of the contribution also lies in the way that the information is collected, that is domain agnostic and aims at collection in the context of spontaneous workflows, with minimal input from the user and very limited assumption on the system structure and its infrastructure.
We plan to continue our work on exploring new methods of automated information collection, and improving the filtering and inference of dependencies. We also plan to explore and implement the methods for determining significance described in section 3.1, and look at the aspects of dependency graphs based on the purpose and significance weights that the tool will allow to infer.