This presentation introduces the Process Mining as the cutting-edge data analytics approach for discovering the real processes by analyzing the event logs, detecting the bottlenecks, and generating recommendations for enhancing the business performance.
This Slide Deck was presented at the annual international conference of itSMF Slovensko on May, 6th. in Bratislava. It gives an introduction into Process Mining as a new useful approach to discover real life processes in IT Service Management end everywhere else where processes are driven by tools providing log file information.
Many thanks to Anne Rozinat http://fluxicon.com for the graphs and information she provided to itSMF Austria. Many thanks to Celonis for providing a demo application.
Please recognize the further links and recommendations at the end of the presentation.
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Process Mining - Chapter 2 - Process Modeling and AnalysisWil van der Aalst
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Process Mining - Chapter 6 - Advanced Process Discovery_techniquesWil van der Aalst
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Object-Centric Processes - from cases to objects and relations… and beyondDirk Fahland
Through this tutorial-style presentation, I want to broaden the uptake of object-centric process mining in research and in practice. It introduces to the concept of object-centric processes, and highlights the core thinking and concepts that underly object-centric processes and explain what makes them effective in analyzing complex real-world processes.
The first part of the talk looks back at key ideas from academic research that led to object-centric process mining.
The second part first explains the basic ideas and techniques of object-centric process mining and the new kinds of process analysis that are enabled by it. We then take a look under the hood of object-centric process mining and look at the key data structures and operations that make it work.
In the third part, we show how these key ideas work for use cases that go far beyond object-centric process mining.
The talk gives pointers to ready-to-use Python libraries and public datasets and tutorials so that you can directly start doing research, development, and analysis in an object-centric approach.
This presentation introduces the Process Mining as the cutting-edge data analytics approach for discovering the real processes by analyzing the event logs, detecting the bottlenecks, and generating recommendations for enhancing the business performance.
This Slide Deck was presented at the annual international conference of itSMF Slovensko on May, 6th. in Bratislava. It gives an introduction into Process Mining as a new useful approach to discover real life processes in IT Service Management end everywhere else where processes are driven by tools providing log file information.
Many thanks to Anne Rozinat http://fluxicon.com for the graphs and information she provided to itSMF Austria. Many thanks to Celonis for providing a demo application.
Please recognize the further links and recommendations at the end of the presentation.
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Process Mining - Chapter 2 - Process Modeling and AnalysisWil van der Aalst
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Process Mining - Chapter 6 - Advanced Process Discovery_techniquesWil van der Aalst
Slides supporting the book "Process Mining: Discovery, Conformance, and Enhancement of Business Processes" by Wil van der Aalst. See also http://springer.com/978-3-642-19344-6 (ISBN 978-3-642-19344-6) and the website http://www.processmining.org/book/start providing sample logs.
Object-Centric Processes - from cases to objects and relations… and beyondDirk Fahland
Through this tutorial-style presentation, I want to broaden the uptake of object-centric process mining in research and in practice. It introduces to the concept of object-centric processes, and highlights the core thinking and concepts that underly object-centric processes and explain what makes them effective in analyzing complex real-world processes.
The first part of the talk looks back at key ideas from academic research that led to object-centric process mining.
The second part first explains the basic ideas and techniques of object-centric process mining and the new kinds of process analysis that are enabled by it. We then take a look under the hood of object-centric process mining and look at the key data structures and operations that make it work.
In the third part, we show how these key ideas work for use cases that go far beyond object-centric process mining.
The talk gives pointers to ready-to-use Python libraries and public datasets and tutorials so that you can directly start doing research, development, and analysis in an object-centric approach.
Slides of the tutorial on Multi-Dimensional Process Analysis shown at the BPM 2022 conference in Muenster, Germany.
Processes are complex phenomena that emerge from the interplay of human actors, materials, data, and machines. Process science develops effective methods and techniques for studying and improving processes. The BPM field has developed mature methods and techniques for studying and improving process executions from the control-flow perspective, and the limitations of control-flow focused thinking are well-known. Current research explores concepts from related disciplines to study behavioral phenomena “beyond” control-flow. However, it remains challenging to relate models and concepts of other behavioral phenomena to the dominant control-flow oriented paradigm.
This tutorial introduces several recently developed simple models that naturally describe behavior beyond control-flow, but are inherently compatible with control-flow oriented thinking. We discuss the Performance Spectrum to study performance patterns and their propagation over time, Event Knowledge Graphs to study networks of behavior over data objects and actors, and Proclets as a formal model for reasoning over control-flow, data object, queue and actor behavior. For each model, we discuss which phenomena can be studied, which insights can be gained, which tools are available, and to which other fields they relate.
https://doi.org/10.1007/978-3-031-16103-2_3
Artifacts and Databases - the Need for Event Relation Graphs and Synchronous ...Dirk Fahland
This was a work-in progress presentation given in Nov 2010 (uploaded here for historic records). It was part of the ACSI EU project for process mining over processes with multiple objects. The slides discuss an example of an order process over multiple objects and propose how to model this data in a graph (event relation graph) that is stored in a database and supported by a relational algebra over behavior (e.g. to join event traces of multiple objects into a case). The final part of the presentation discusses the synchronous proclet model ultimately formalized in https://doi.org/10.1007/978-3-030-21571-2_1
Describing, Discovering, and Understanding Multi-Dimensional ProcessesDirk Fahland
Processes are a key application area for formal models of concurrency. The most adopted model-driven techniques are centered around
describing and analyzing the control-flow of a well-structured process instance in isolation - within this single dimension one could argue the case to be "solved".
Unaddressed challenges in modeling and analysis arise where processes are not well-structured or not isolated from each other. In both cases a single process model can no longer adequately describe process behavior.
Taking recorded event data from such processes as a starting point, I will outline and develop a number of challenges and characteristics of such processes that can be observed in practice. I will discuss how the
behavior of such processes can be classified along different dimensions and outline a few fundamental concepts that complement concepts from Petri nets and allow to adequately describe behavior of
such processes.
Process Mining: Past, Present, and Open Challenges (AIST 2017 Keynote)Dirk Fahland
Since the first algorithms for automatically discovering process models from event logs have been proposed in the late 1990ies the problem of obtaining insights into processes by mining from event logs gained growing attention. By now, the field has grown into a maturing discipline and industry has begun adopting process mining in regular operations, supported by several commercial process mining solutions are available on the market.
In the early days of process mining, several algorithms for constructively discovering a process model from an event log were proposed, each algorithm pursuing unique principles for constructing a model. This first generation of process discovery techniques, which includes for instance the alpha-algorithm, paved the ground for process mining as research discipline. As these algorithms were applied in practice, new research challenges showed up, sparking new results in both pre-processing event data and evaluating process models on event logs. In particular the latter deepened the understanding of the challenges in process mining and established a reliable feedback mechanism in process mining in the form of conformance checking. This feedback mechanism enabled researching a second generation of process mining techniques addressing a large variety of problems such as quality guarantees for discovered models, including the data perspective in discovered models, or discovering temporal logic constraints. In particular, the inductive miner family was seen as a new milestone as it provided a systematic way to develop process discovery algorithms with reliable results. Yet again, as these more capable techniques are being applied to the growing and more detailed event data recorded in practice, further unsolved challenges arise.
In the first part of my talk I will draw an arc from the early days of process mining to the current state of the art in process mining – highlighting central techniques and their impact on later developments. In the second part of my talk, I will then turn to what kinds of event data and challenges are being found in practice today, how existing process mining techniques fail to address them, and thus which open challenges and opportunities the process mining field offers also for researchers from other domains.
Where did I go wrong? Explaining errors in process modelsDirk Fahland
This presentation shows how to reduce diagnostic information returned by general purpose model checkers (counter example paths) to essential parts that help understanding the error. The presentation has been given at the 12th International Conference on Business Process Management (BPM'14), September 2014 in Eindhoven.
Mining Branch-Time Scenarios From Execution LogsDirk Fahland
This presentation was given at the International Conference on Automated Software Engineering (ASE 2013) in Palo Alto, November 2013.
We describe a technique for automatically extracting specifications from execution traces of an application. The particular specification that we extract are scenarios in the form of conditional existential Live-Sequence Charts (LSC), which are similar to UML Sequence Diagrams.
The technique is implemented in a tool and was evaluated on two real-life event logs.
From Live Sequence Chart Specifications to Distributed ComponentsDirk Fahland
This document proposes a method to synthesize a decentralized system from a specification of Live Sequence Charts (LSCs). It introduces distributed LSCs (dLSCs) which use partial orders and local states instead of global states. A two-step process first synthesizes a Petri net structure from LSC main charts, then adds guards to tokens based on LSC precharts. This extracts autonomous components while preserving behaviors. A prototype tool demonstrates the approach on an emergency management example. Future work includes code generation and increasing dLSC expressiveness.
LSC Revisited - From Scenarios to Distributed ComponentsDirk Fahland
Scenario-based techniques such as Message Sequence Charts
(MSC) and Live Sequence Charts (LSC) are a technique to specify
behavior of complex, distributed systems in an intuitive manner,
particularly at early stages of system design. Despite its intuitive
nature, the technique poses some challenges. The most prominent is to
automatically synthesize an operational system model (a statechart or
a Petri net) from a given specification; the model can then serve as a
blue print for implementation in hard- and software. While MSC are
essentially too weak to specify complex systems, LSCs are too strong:
synthesis of components of a distributed system fails.
In my talk, I will reconsider the semantics of LSC-style scenarios
regarding expressive power, ability to specify distributed behaviors
and solving the synthesis problem. I will show that by changing the
interpretation of LSC from linear time to simple branching time
semantics, one obtains a simple, yet very expressive and intuitive
scenario-based specification language. By choosing partial orders
instead of sequential runs as semantic domain, one can faithfully
specify the behaviors of a distributed system. We call this notation
distributed LSC (dLSC). As the main result, I will present a complete
technique for synthesizing Petri net components from any given dLSC
specification, in polynomial time.
Remote seminar talk held in the Advanced Software Tools Research Seminar of S. Maoz and A. Yehudai at Tel Aviv University, January 7, 2013.
Repairing Process Models to Match RealityDirk Fahland
The document discusses repairing process models to improve conformance to event logs. It presents an approach for repairing models that involves aligning the log and model, identifying sublogs of events that cannot be replayed, and using these to add optional/remove activities or add subprocesses to the model. The approach was implemented in ProM and evaluated on a case study, showing it can effectively repair models while maintaining a low distance to the original model.
The document describes techniques for simplifying process models mined from event logs to make them more readable and understandable for users. It involves unfolding the mined model based on the event log to represent concurrency explicitly, then refolding and merging equivalent nodes. Implied places that do not restrict behavior can be removed. The resulting model has less complexity but the same behavior as the original model. Experimental results on benchmark logs show the techniques can significantly reduce model complexity while maintaining precision.
This talk was given by Dirk Fahland and Hajo A. Reijers at the BPM Roundtable at TU Eindhoven in July 2011. We presented first insights into how people model and the modeling outcome.
Behavioral Conformance of Artifact-Centric Process ModelsDirk Fahland
A talk help by Boudewijn van Dongen at the 14th International Conference on Business Information Systems (BIS 2011) in Poznan, Poland, June 2011. We present the problem of checking whether an artifact-centric process model conforms to process behavior observed in reality.
Many-to-Many: Interactions in Artifact-Centric ChoreographiesDirk Fahland
A talk given by Dirk Fahland at the 3rd Central European Workshop on Services and their Composition (ZEUS'11) in Karlsruhe, February 22, 2011. The talk explains behavioral phenomena in services choreographies where several service instances interact with each other.
Artifacts - Processes with Multiple InstancesDirk Fahland
How Artifacts allow to describe processes where multiple instances of data objects interact with each other. A talk given by Dirk Fahland in the group of David Harel at the Weizmann Institute of Science.
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
"$10 thousand per minute of downtime: architecture, queues, streaming and fin...Fwdays
Direct losses from downtime in 1 minute = $5-$10 thousand dollars. Reputation is priceless.
As part of the talk, we will consider the architectural strategies necessary for the development of highly loaded fintech solutions. We will focus on using queues and streaming to efficiently work and manage large amounts of data in real-time and to minimize latency.
We will focus special attention on the architectural patterns used in the design of the fintech system, microservices and event-driven architecture, which ensure scalability, fault tolerance, and consistency of the entire system.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Dandelion Hashtable: beyond billion requests per second on a commodity serverAntonios Katsarakis
This slide deck presents DLHT, a concurrent in-memory hashtable. Despite efforts to optimize hashtables, that go as far as sacrificing core functionality, state-of-the-art designs still incur multiple memory accesses per request and block request processing in three cases. First, most hashtables block while waiting for data to be retrieved from memory. Second, open-addressing designs, which represent the current state-of-the-art, either cannot free index slots on deletes or must block all requests to do so. Third, index resizes block every request until all objects are copied to the new index. Defying folklore wisdom, DLHT forgoes open-addressing and adopts a fully-featured and memory-aware closed-addressing design based on bounded cache-line-chaining. This design offers lock-free index operations and deletes that free slots instantly, (2) completes most requests with a single memory access, (3) utilizes software prefetching to hide memory latencies, and (4) employs a novel non-blocking and parallel resizing. In a commodity server and a memory-resident workload, DLHT surpasses 1.6B requests per second and provides 3.5x (12x) the throughput of the state-of-the-art closed-addressing (open-addressing) resizable hashtable on Gets (Deletes).
Slides of the tutorial on Multi-Dimensional Process Analysis shown at the BPM 2022 conference in Muenster, Germany.
Processes are complex phenomena that emerge from the interplay of human actors, materials, data, and machines. Process science develops effective methods and techniques for studying and improving processes. The BPM field has developed mature methods and techniques for studying and improving process executions from the control-flow perspective, and the limitations of control-flow focused thinking are well-known. Current research explores concepts from related disciplines to study behavioral phenomena “beyond” control-flow. However, it remains challenging to relate models and concepts of other behavioral phenomena to the dominant control-flow oriented paradigm.
This tutorial introduces several recently developed simple models that naturally describe behavior beyond control-flow, but are inherently compatible with control-flow oriented thinking. We discuss the Performance Spectrum to study performance patterns and their propagation over time, Event Knowledge Graphs to study networks of behavior over data objects and actors, and Proclets as a formal model for reasoning over control-flow, data object, queue and actor behavior. For each model, we discuss which phenomena can be studied, which insights can be gained, which tools are available, and to which other fields they relate.
https://doi.org/10.1007/978-3-031-16103-2_3
Artifacts and Databases - the Need for Event Relation Graphs and Synchronous ...Dirk Fahland
This was a work-in progress presentation given in Nov 2010 (uploaded here for historic records). It was part of the ACSI EU project for process mining over processes with multiple objects. The slides discuss an example of an order process over multiple objects and propose how to model this data in a graph (event relation graph) that is stored in a database and supported by a relational algebra over behavior (e.g. to join event traces of multiple objects into a case). The final part of the presentation discusses the synchronous proclet model ultimately formalized in https://doi.org/10.1007/978-3-030-21571-2_1
Describing, Discovering, and Understanding Multi-Dimensional ProcessesDirk Fahland
Processes are a key application area for formal models of concurrency. The most adopted model-driven techniques are centered around
describing and analyzing the control-flow of a well-structured process instance in isolation - within this single dimension one could argue the case to be "solved".
Unaddressed challenges in modeling and analysis arise where processes are not well-structured or not isolated from each other. In both cases a single process model can no longer adequately describe process behavior.
Taking recorded event data from such processes as a starting point, I will outline and develop a number of challenges and characteristics of such processes that can be observed in practice. I will discuss how the
behavior of such processes can be classified along different dimensions and outline a few fundamental concepts that complement concepts from Petri nets and allow to adequately describe behavior of
such processes.
Process Mining: Past, Present, and Open Challenges (AIST 2017 Keynote)Dirk Fahland
Since the first algorithms for automatically discovering process models from event logs have been proposed in the late 1990ies the problem of obtaining insights into processes by mining from event logs gained growing attention. By now, the field has grown into a maturing discipline and industry has begun adopting process mining in regular operations, supported by several commercial process mining solutions are available on the market.
In the early days of process mining, several algorithms for constructively discovering a process model from an event log were proposed, each algorithm pursuing unique principles for constructing a model. This first generation of process discovery techniques, which includes for instance the alpha-algorithm, paved the ground for process mining as research discipline. As these algorithms were applied in practice, new research challenges showed up, sparking new results in both pre-processing event data and evaluating process models on event logs. In particular the latter deepened the understanding of the challenges in process mining and established a reliable feedback mechanism in process mining in the form of conformance checking. This feedback mechanism enabled researching a second generation of process mining techniques addressing a large variety of problems such as quality guarantees for discovered models, including the data perspective in discovered models, or discovering temporal logic constraints. In particular, the inductive miner family was seen as a new milestone as it provided a systematic way to develop process discovery algorithms with reliable results. Yet again, as these more capable techniques are being applied to the growing and more detailed event data recorded in practice, further unsolved challenges arise.
In the first part of my talk I will draw an arc from the early days of process mining to the current state of the art in process mining – highlighting central techniques and their impact on later developments. In the second part of my talk, I will then turn to what kinds of event data and challenges are being found in practice today, how existing process mining techniques fail to address them, and thus which open challenges and opportunities the process mining field offers also for researchers from other domains.
Where did I go wrong? Explaining errors in process modelsDirk Fahland
This presentation shows how to reduce diagnostic information returned by general purpose model checkers (counter example paths) to essential parts that help understanding the error. The presentation has been given at the 12th International Conference on Business Process Management (BPM'14), September 2014 in Eindhoven.
Mining Branch-Time Scenarios From Execution LogsDirk Fahland
This presentation was given at the International Conference on Automated Software Engineering (ASE 2013) in Palo Alto, November 2013.
We describe a technique for automatically extracting specifications from execution traces of an application. The particular specification that we extract are scenarios in the form of conditional existential Live-Sequence Charts (LSC), which are similar to UML Sequence Diagrams.
The technique is implemented in a tool and was evaluated on two real-life event logs.
From Live Sequence Chart Specifications to Distributed ComponentsDirk Fahland
This document proposes a method to synthesize a decentralized system from a specification of Live Sequence Charts (LSCs). It introduces distributed LSCs (dLSCs) which use partial orders and local states instead of global states. A two-step process first synthesizes a Petri net structure from LSC main charts, then adds guards to tokens based on LSC precharts. This extracts autonomous components while preserving behaviors. A prototype tool demonstrates the approach on an emergency management example. Future work includes code generation and increasing dLSC expressiveness.
LSC Revisited - From Scenarios to Distributed ComponentsDirk Fahland
Scenario-based techniques such as Message Sequence Charts
(MSC) and Live Sequence Charts (LSC) are a technique to specify
behavior of complex, distributed systems in an intuitive manner,
particularly at early stages of system design. Despite its intuitive
nature, the technique poses some challenges. The most prominent is to
automatically synthesize an operational system model (a statechart or
a Petri net) from a given specification; the model can then serve as a
blue print for implementation in hard- and software. While MSC are
essentially too weak to specify complex systems, LSCs are too strong:
synthesis of components of a distributed system fails.
In my talk, I will reconsider the semantics of LSC-style scenarios
regarding expressive power, ability to specify distributed behaviors
and solving the synthesis problem. I will show that by changing the
interpretation of LSC from linear time to simple branching time
semantics, one obtains a simple, yet very expressive and intuitive
scenario-based specification language. By choosing partial orders
instead of sequential runs as semantic domain, one can faithfully
specify the behaviors of a distributed system. We call this notation
distributed LSC (dLSC). As the main result, I will present a complete
technique for synthesizing Petri net components from any given dLSC
specification, in polynomial time.
Remote seminar talk held in the Advanced Software Tools Research Seminar of S. Maoz and A. Yehudai at Tel Aviv University, January 7, 2013.
Repairing Process Models to Match RealityDirk Fahland
The document discusses repairing process models to improve conformance to event logs. It presents an approach for repairing models that involves aligning the log and model, identifying sublogs of events that cannot be replayed, and using these to add optional/remove activities or add subprocesses to the model. The approach was implemented in ProM and evaluated on a case study, showing it can effectively repair models while maintaining a low distance to the original model.
The document describes techniques for simplifying process models mined from event logs to make them more readable and understandable for users. It involves unfolding the mined model based on the event log to represent concurrency explicitly, then refolding and merging equivalent nodes. Implied places that do not restrict behavior can be removed. The resulting model has less complexity but the same behavior as the original model. Experimental results on benchmark logs show the techniques can significantly reduce model complexity while maintaining precision.
This talk was given by Dirk Fahland and Hajo A. Reijers at the BPM Roundtable at TU Eindhoven in July 2011. We presented first insights into how people model and the modeling outcome.
Behavioral Conformance of Artifact-Centric Process ModelsDirk Fahland
A talk help by Boudewijn van Dongen at the 14th International Conference on Business Information Systems (BIS 2011) in Poznan, Poland, June 2011. We present the problem of checking whether an artifact-centric process model conforms to process behavior observed in reality.
Many-to-Many: Interactions in Artifact-Centric ChoreographiesDirk Fahland
A talk given by Dirk Fahland at the 3rd Central European Workshop on Services and their Composition (ZEUS'11) in Karlsruhe, February 22, 2011. The talk explains behavioral phenomena in services choreographies where several service instances interact with each other.
Artifacts - Processes with Multiple InstancesDirk Fahland
How Artifacts allow to describe processes where multiple instances of data objects interact with each other. A talk given by Dirk Fahland in the group of David Harel at the Weizmann Institute of Science.
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
"$10 thousand per minute of downtime: architecture, queues, streaming and fin...Fwdays
Direct losses from downtime in 1 minute = $5-$10 thousand dollars. Reputation is priceless.
As part of the talk, we will consider the architectural strategies necessary for the development of highly loaded fintech solutions. We will focus on using queues and streaming to efficiently work and manage large amounts of data in real-time and to minimize latency.
We will focus special attention on the architectural patterns used in the design of the fintech system, microservices and event-driven architecture, which ensure scalability, fault tolerance, and consistency of the entire system.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Dandelion Hashtable: beyond billion requests per second on a commodity serverAntonios Katsarakis
This slide deck presents DLHT, a concurrent in-memory hashtable. Despite efforts to optimize hashtables, that go as far as sacrificing core functionality, state-of-the-art designs still incur multiple memory accesses per request and block request processing in three cases. First, most hashtables block while waiting for data to be retrieved from memory. Second, open-addressing designs, which represent the current state-of-the-art, either cannot free index slots on deletes or must block all requests to do so. Third, index resizes block every request until all objects are copied to the new index. Defying folklore wisdom, DLHT forgoes open-addressing and adopts a fully-featured and memory-aware closed-addressing design based on bounded cache-line-chaining. This design offers lock-free index operations and deletes that free slots instantly, (2) completes most requests with a single memory access, (3) utilizes software prefetching to hide memory latencies, and (4) employs a novel non-blocking and parallel resizing. In a commodity server and a memory-resident workload, DLHT surpasses 1.6B requests per second and provides 3.5x (12x) the throughput of the state-of-the-art closed-addressing (open-addressing) resizable hashtable on Gets (Deletes).
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
Session 1 - Intro to Robotic Process Automation.pdfUiPathCommunity
👉 Check out our full 'Africa Series - Automation Student Developers (EN)' page to register for the full program:
https://bit.ly/Automation_Student_Kickstart
In this session, we shall introduce you to the world of automation, the UiPath Platform, and guide you on how to install and setup UiPath Studio on your Windows PC.
📕 Detailed agenda:
What is RPA? Benefits of RPA?
RPA Applications
The UiPath End-to-End Automation Platform
UiPath Studio CE Installation and Setup
💻 Extra training through UiPath Academy:
Introduction to Automation
UiPath Business Automation Platform
Explore automation development with UiPath Studio
👉 Register here for our upcoming Session 2 on June 20: Introduction to UiPath Studio Fundamentals: https://community.uipath.com/events/details/uipath-lagos-presents-session-2-introduction-to-uipath-studio-fundamentals/
inQuba Webinar Mastering Customer Journey Management with Dr Graham HillLizaNolte
HERE IS YOUR WEBINAR CONTENT! 'Mastering Customer Journey Management with Dr. Graham Hill'. We hope you find the webinar recording both insightful and enjoyable.
In this webinar, we explored essential aspects of Customer Journey Management and personalization. Here’s a summary of the key insights and topics discussed:
Key Takeaways:
Understanding the Customer Journey: Dr. Hill emphasized the importance of mapping and understanding the complete customer journey to identify touchpoints and opportunities for improvement.
Personalization Strategies: We discussed how to leverage data and insights to create personalized experiences that resonate with customers.
Technology Integration: Insights were shared on how inQuba’s advanced technology can streamline customer interactions and drive operational efficiency.
zkStudyClub - LatticeFold: A Lattice-based Folding Scheme and its Application...Alex Pruden
Folding is a recent technique for building efficient recursive SNARKs. Several elegant folding protocols have been proposed, such as Nova, Supernova, Hypernova, Protostar, and others. However, all of them rely on an additively homomorphic commitment scheme based on discrete log, and are therefore not post-quantum secure. In this work we present LatticeFold, the first lattice-based folding protocol based on the Module SIS problem. This folding protocol naturally leads to an efficient recursive lattice-based SNARK and an efficient PCD scheme. LatticeFold supports folding low-degree relations, such as R1CS, as well as high-degree relations, such as CCS. The key challenge is to construct a secure folding protocol that works with the Ajtai commitment scheme. The difficulty, is ensuring that extracted witnesses are low norm through many rounds of folding. We present a novel technique using the sumcheck protocol to ensure that extracted witnesses are always low norm no matter how many rounds of folding are used. Our evaluation of the final proof system suggests that it is as performant as Hypernova, while providing post-quantum security.
Paper Link: https://eprint.iacr.org/2024/257
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
Overcoming the PLG Trap: Lessons from Canva's Head of Sales & Head of EMEA Da...
Process Mining for ERP Systems
1. Erik Nooijen,
Boudewijn v. Dongen, Dirk Fahland
Process Mining for ERP Systems
2. Process Discovery
process
event process
discovery
log model
algorithm
c1: A B C D E assumptions
c2: A C B D E • case = sequence of events of this case
c3: A F D E • cases are isolated:
event A in c1 happens only in c1 (and not in c2)
…
• cases of the same process
• one unique case id,
• each event associated to exactly one case id
PAGE 1
3. Typical Process in an ERP System
Manufacturer
Material A Material B
order
Material B Material B
product X order
Alice materials
ACME Inc.
Material B Material A
order
Material C Material C
product Y order
Bob
materials
Build to Order Mega Corp.
PAGE 2
4. n-to-m relations database
process
process
discovery
model
algorithm
id attributes time-stamp attributes ProductOrder Customer
poID cust. … created processed built shipped cust. address …
po1 Alice 30-08 9:22 30-08 13:12 01-09 15:12 03-09 10:15 Alice … …
po2 Bob 30-08 10:15 30-08 13:14 01-09 16:13 03-09 17:18 Bob … …
relations data attributes
OrderedMaterial id attributes MaterialOrder
poID moID type added moID suppl. … completed sent received
po1 mo3 B 30-08 13:13 mo3 ACME 30-08 13:15 30-08 14:15 01-09 9:05
po1 mo4 A 30-08 13:14 mo4 MEGA 30-08 13:17 30-08 16:12 01-09 10:13
po2 mo3 B 30-08 13:15
po2 mo4 C 30-08 13:16 relations
PAGE 3
5. Process Discovery for ERP Systems
process
process
discovery
model
algorithm
0..*
Customer
reality: data in a relational DB
ProductOrder - cust
1
-… • events stored as time-stamped
- poID
- cust attributes in tables
- created OrderedMat.
MaterialOrder
- processed - poID
- built 1
- moID
- moID • multiple primary keys
- shipped 1..* - supplier multiple notions of case
- type
1..* - completed
- added 1
- sent
- received • tables are related
one event related to
multiple cases
PAGE 4
6. Process Discovery for ERP Systems
process
process
discovery
model
algorithm
0..*
Customer
reality: data in a relational DB
ProductOrder - cust
1
-… • events stored as time-stamped
- poID
- cust attributes in tables
- created OrderedMat.
MaterialOrder
- processed - poID
- built 1
- moID
- moID • multiple primary keys
- shipped 1..* - supplier multiple notions of case
- type
1..* - completed
- added 1
- sent
- received • tables are related
one event related to
multiple cases
PAGE 5
7. Outline
process
model
related by
primary foreign-key
relations
decompose by primary keys
model f.
log f. discovery PO
log f. model f.
MO
PO MO
discovery
PAGE 6
8. Find Artifact Schemas
process
model
related by
primary foreign-key
relations
decompose by primary keys
model f.
log f. discovery PO
log f. model f.
MO
PO MO
discovery
PAGE 7
9. Step 0: discover database schema
document schema vs. actual schema identify
• column types (esp. time-stamped columns)
• primary keys
• foreign keys
various (non-trivial) techniques available
key discovery is NP-complete in the size of the
table(s)
result:
PAGE 8
10. Step 1: decompose schema into processes
= schema summarization find:
1. sets of
corresponding
tables
2. links between
those
ProductOrder MaterialOrder
PAGE 9
11. Automatic Schema Summarization
= group similar tables
through clustering
define a distance between
any 2 tables
• by relations
• by information content
tables that are close to
each other
same cluster
# of clusters: user input
PAGE 10
12. Automatic Schema Summarization
1. structural distance A
between tables 1
2 fanout: 1 = (2+0)/2
fanout ~ avg. # of child fanout: 1
records related to the fanout: 2
same parent record
A B A B A B
1 X 1 X 1 X
2 Y 1 Y 1 Y
2 Z
2 U
PAGE 11
13. Automatic Schema Summarization
1. structural distance A
between tables 1
2 fanout: 1
fanout ~ avg. # of child fanout: 1 m.fr: 2 = 1/ (1/2)
records related to the m.fr: 1 fanout: 2
same parent record m.fr: 1
A B A B A B
matched fraction ~ 1 X 1 X 1 X
1 / (fraction of records in 2 Y 1 Y 1 Y
parent with matching child 2 Z
record) 2 U
PAGE 12
14. Grouping by Clustering
1. structural distance
2. information distance
importance of each table
= entropy (is maximal if all
records are different)
distance: 2 tables with high
entropies large distance
3. weighted distance by
structure + information
4. k-means clustering: most important table of cluster
k clusters based on = table with least distance to all
key attribute of the cluster
weighted distance
PAGE 13
15. Artifact Schema Artifact Log
process
model
related by
primary foreign-key
relations
decompose by primary keys
model f.
log f. discovery PO
log f. model f.
MO
PO MO
discovery
PAGE 14
16. Log Extraction
cluster = set of related tables
+ primary key of most important table
case id
poID cust. … created processed built shipped
log f.
PO po1 Alice 30-08 9:22 30-08 13:12 01-09 15:12 03-09 10:15
po2 Bob 30-08 10:15 30-08 13:14 01-09 16:13 03-09 17:18
poID moID type added
po1 mo3 B 30-08 13:13
po1: po1 mo4 A 30-08 13:14
po2 mo3 B 30-08 13:15
po2: po2 mo4 C 30-08 13:16
PAGE 15
17. Log Extraction
cluster = set of related tables
+ primary key of most important table
case id
time-stamped attribute event
poID cust. … created processed built shipped
log f.
PO po1 Alice 30-08 9:22 30-08 13:12 01-09 15:12 03-09 10:15
po2 Bob 30-08 10:15 30-08 13:14 01-09 16:13 03-09 17:18
poID moID type added
po1 mo3 B 30-08 13:13
po1: (created, poID=po1, time=30-08 9:22, …) po1 mo4 A 30-08 13:14
po2 mo3 B 30-08 13:15
po2 mo4 C 30-08 13:16
PAGE 16
18. Log Extraction
cluster = set of related tables
+ primary key of most important table
case id
time-stamped attribute event
related attributes event attributes
poID cust. … created processed built shipped
log f.
PO po1 Alice 30-08 9:22 30-08 13:12 01-09 15:12 03-09 10:15
po2 Bob 30-08 10:15 30-08 13:14 01-09 16:13 03-09 17:18
poID moID type added
po1 mo3 B 30-08 13:13
po1: (created, poID=po1, time=30-08 9:22, cust.=Alice, …)po1 mo4 A 30-08 13:14
po2 mo3 B 30-08 13:15
po2 mo4 C 30-08 13:16
PAGE 17
19. Log Extraction
cluster = set of related tables
+ primary key of most important table
case id
time-stamped attribute event
related attributes event attributes
poID cust. … created processed built shipped
log f.
PO po1 Alice 30-08 9:22 30-08 13:12 01-09 15:12 03-09 10:15
po2 Bob 30-08 10:15 30-08 13:14 01-09 16:13 03-09 17:18
poID moID type added
po1 mo3 B 30-08 13:13
po1: (created, poID=po1, time=30-08 9:22, cust.=Alice, …)po1 mo4 A 30-08 13:14
(processed, poID=po1, time=30-08 13:12, …) po2 mo3 B 30-08 13:15
po2 mo4 C 30-08 13:16
PAGE 18
20. Log Extraction
cluster = set of related tables
+ primary key of most important table
case id
time-stamped attribute event
related attributes event attributes
poID cust. … created processed built shipped
log f.
PO po1 Alice 30-08 9:22 30-08 13:12 01-09 15:12 03-09 10:15
po2 Bob 30-08 10:15 30-08 13:14 01-09 16:13 03-09 17:18
poID moID type added
po1 mo3 B 30-08 13:13
po1: (created, poID=po1, time=30-08 9:22, cust.=Alice, …)po1 mo4 A 30-08 13:14
(processed, poID=po1, time=30-08 13:12, …) po2 mo3 B 30-08 13:15
(added, poID=po1, time=30-08 13:13, moID=mo3, …)po2 mo4 C 30-08 13:16
refers to artifact “MaterialOrder”
PAGE 19
21. Outline
process
model
compose by
primary foreign-key
relations
decompose by primary keys
model f.
log f. discovery order
log f. model f.
order
quote quote
discovery
PAGE 20
22. Resulting Model(s)
Product Order Material Order
1..*
added
create
completed
processed
added 1..* sent
built
received
shipped
(addded, poID=po1, …, moID=mo3)
PAGE 21
26. Open issues
performance
• key discovery: NP-complete in R (# of records)
• foreign key discovery: NP-complete in R2
• problem is in the “hard part” of NP
• sampling of data, domain knowledge, semi-automatic
requires good database structure
• proper relations, proper keys
• otherwise wrong clusters are formed
• events don’t get right attributes
• semi-automatic approach
events shared by multiple cases… working on it…
PAGE 25
27. Erik Nooijen,
Boudewijn v. Dongen, Dirk Fahland
Process Mining for ERP Systems