This document discusses using event logs to generate business process simulation models. It describes traditional discrete event simulation approaches that discover simulation models from event logs recorded by information systems. Deep learning techniques are also discussed that can generate traces without an explicit process model. The document suggests that combining discrete event simulation and deep learning may produce more accurate simulations, but challenges remain around validating such hybrid approaches and testing them in previously unseen scenarios. More research is needed before these data-driven simulation methods can reliably predict the effects of interventions.
Keynote: Building and Operating A Serverless Streaming Runtime for Apache Bea...Flink Forward
Apache Beam is Flink’s sibling in the Apache family of streaming processing frameworks. The Beam and Flink teams work closely together on advancing what is possible in streaming processing, including Streaming SQL extensions and code interoperability on both platforms.
Beam was originally developed at Google as the amalgamation of its internal batch and streaming frameworks to power the exabyte-scale data processing for Gmail, YouTube and Ads. It now powers a fully-managed, serverless service Google Cloud Dataflow, as well as is available to run in other Public Clouds and on-premises when deployed in portability mode on Apache Flink, Spark, Samza and other runners. Users regularly run distributed data processing jobs on Beam spanning tens of thousands of CPU cores and processing millions of events per second.
In this session, Sergei Sokolenko, Cloud Dataflow product manager, and Reuven Lax, the founding member of the Dataflow and Beam team, will share Google’s learnings from building and operating a global streaming processing infrastructure shared by thousands of customers, including:
safe deployment to dozens of geographic locations,
resource autoscaling to minimize processing costs,
separating compute and state storage for better scaling behavior,
dynamic work rebalancing of work items away from overutilized worker nodes,
offering a throughput-optimized batch processing capability with the same API as streaming,
grouping and joining of 100s of Terabytes in a hybrid in-memory/on-desk file system,
integrating with the Google Cloud security ecosystem, and other lessons.
Customers benefit from these advances through faster execution of jobs, resource savings, and a fully managed data processing environment that runs in the Cloud and removes the need to manage infrastructure.
Winter Simulation Conference 2021 - Process Wind Tunnel TalkSudhendu Rai
The talk associated with this presentation can be accessed at:
https://youtu.be/VXEVuXW9knU
Abstract
In this talk, we will introduce a simulation-based process improvement framework and methodology called the Process Wind Tunnel. We will describe this framework and introduce the underlying technologies namely process mapping and data collection, data wrangling, exploratory data analysis and visualization, process mining, discrete-event simulation optimization and solution implementation. We will discuss how Process Wind Tunnel framework was utilized to improve a critical business process namely, the post-execution trade settlement process. The work builds upon and generalizes the Lean Document Production solution (2008 Edelman finalist) for optimizing printshops to more general and complex business processes found within the insurance and financial services industry.
Process wind tunnel - A novel capability for data-driven business process imp...Sudhendu Rai
A talk I gave recently on data-driven process improvement methodology and techniques with applications and results from insurance and finance processes
Business Process Monitoring and MiningMarlon Dumas
Lecture delivered at the Second Latin-American Summer School in Business Process Management, Bogota, Colombia, 28 June 2017 - http://ii-las-bpm.uniandes.edu.co/
Keynote: Building and Operating A Serverless Streaming Runtime for Apache Bea...Flink Forward
Apache Beam is Flink’s sibling in the Apache family of streaming processing frameworks. The Beam and Flink teams work closely together on advancing what is possible in streaming processing, including Streaming SQL extensions and code interoperability on both platforms.
Beam was originally developed at Google as the amalgamation of its internal batch and streaming frameworks to power the exabyte-scale data processing for Gmail, YouTube and Ads. It now powers a fully-managed, serverless service Google Cloud Dataflow, as well as is available to run in other Public Clouds and on-premises when deployed in portability mode on Apache Flink, Spark, Samza and other runners. Users regularly run distributed data processing jobs on Beam spanning tens of thousands of CPU cores and processing millions of events per second.
In this session, Sergei Sokolenko, Cloud Dataflow product manager, and Reuven Lax, the founding member of the Dataflow and Beam team, will share Google’s learnings from building and operating a global streaming processing infrastructure shared by thousands of customers, including:
safe deployment to dozens of geographic locations,
resource autoscaling to minimize processing costs,
separating compute and state storage for better scaling behavior,
dynamic work rebalancing of work items away from overutilized worker nodes,
offering a throughput-optimized batch processing capability with the same API as streaming,
grouping and joining of 100s of Terabytes in a hybrid in-memory/on-desk file system,
integrating with the Google Cloud security ecosystem, and other lessons.
Customers benefit from these advances through faster execution of jobs, resource savings, and a fully managed data processing environment that runs in the Cloud and removes the need to manage infrastructure.
Winter Simulation Conference 2021 - Process Wind Tunnel TalkSudhendu Rai
The talk associated with this presentation can be accessed at:
https://youtu.be/VXEVuXW9knU
Abstract
In this talk, we will introduce a simulation-based process improvement framework and methodology called the Process Wind Tunnel. We will describe this framework and introduce the underlying technologies namely process mapping and data collection, data wrangling, exploratory data analysis and visualization, process mining, discrete-event simulation optimization and solution implementation. We will discuss how Process Wind Tunnel framework was utilized to improve a critical business process namely, the post-execution trade settlement process. The work builds upon and generalizes the Lean Document Production solution (2008 Edelman finalist) for optimizing printshops to more general and complex business processes found within the insurance and financial services industry.
Process wind tunnel - A novel capability for data-driven business process imp...Sudhendu Rai
A talk I gave recently on data-driven process improvement methodology and techniques with applications and results from insurance and finance processes
Business Process Monitoring and MiningMarlon Dumas
Lecture delivered at the Second Latin-American Summer School in Business Process Management, Bogota, Colombia, 28 June 2017 - http://ii-las-bpm.uniandes.edu.co/
Learning Accurate Business Process Simulation Models from Event Logs via Auto...Marlon Dumas
Paper presentation at the International Conference on Advanced Information Systems Engineering (CAiSE).
This paper presents an approach to automatically discover business process simulation models from event logs by combining process mining and deep learning techniques.
Paper available at: https://link.springer.com/chapter/10.1007/978-3-031-07472-1_4
Business Process Analytics: From Insights to PredictionsMarlon Dumas
Keynote talk at the 13th Baltic Conference on Databases and Information Systems, Trakai, Lithuania, 2 July 2018.
Abstract
Business process analytics is a body of methods for analyzing data generated by the execution of business processes in order to extract insights about weaknesses and improvement opportunities, both at the tactical and operational levels. Tactical process analytics methods (also known as process mining) allow us to understand how a given business process is actually executed, if and how its execution deviates with respect to expected or normative pathways, and what factors contribute to poor process performance or undesirable outcomes. Meantime, operational process analytics methods allow us to monitor ongoing executions of a business process in order to predict future states and undesirable outcomes at runtime (predictive process monitoring). Existing methods in this space allow us to predict, for example, which task will be executed next in a case, when, and who will perform it? When will an ongoing case complete? What will its outcome be and how can negative outcomes be avoided? This keynote will present a framework for conceptualizing business process analytics methods and applications. The talk will provide an overview of state-of-art methods and tools in the field and will outline open challenges and research opportunities.
Introduction to Business Process Analysis and RedesignMarlon Dumas
Special course on business process analysis and design delivered at University of Granada on 23-24 January 2014. The course covers qualitative and quantitative process analysis techniques and redesign heuristics. Based on the textbook Fundamentals of Business Process Management by Dumas et al.
These slides are from the Metrics-Based Process Mapping webinar delivered 09-29-2021.
Companion resources:
• View the recording - https://tkmg.com/webinars/metrics-based-process-mapping-3/
• Buy the book - https://tkmgacademy.com/products/metrics-based-process-mapping/
• Take the TKMG Academy course - https://tkmgacademy.com/courses/metrics-based-process-mapping/
C++ Is One Of The widely used programming language. Here is the complete presentation PPT notes of C++ programming language. hope it will be helpful to you.
Tieghi Anipla 20 04 2010 Come Possiamo Essere Sicuri Che Tutti Seguano Le Pro...Enzo M. Tieghi
How can you be sure that operators will follow SOPs Standard Operating Procedures? Using an Industrial WorkFlow management tool such as GEIP Procify WorkFlow.
This paper was presented at ANIPLA conference in Milano April 20th, 2010
Unbiased, Fine-Grained Description of Processes Performance from Event DataVadim Denisov
International Conference on Business Process Management 2018 (BPM2018 Sydney).
Performance is central to processes management and event data pro-vides the most objective source for analyzing and improving performance. Currentprocess mining techniques give only limited insights into performance by aggre-gating all event data for each process step. In this paper, we investigate processperformance of all process behaviors without prior aggregation. We propose theperformance spectrumas a simple model that maps all observed flows betweentwo process steps together regarding their performance over time. Visualizing theperformance spectrum of event logs reveals a large variety of very distinctpatternsof process performanceand performance variability that have not been describedbefore. We provide a taxonomy for these patterns and a comprehensive overviewof elementary and composite performance patterns observed on several real-lifeevent logs from business processes and logistics. We report on a case study whereperformance patterns were central to identify systemic, but not globally visibleprocess problems.
Mine Your Simulation Model: Automated Discovery of Business Process Simulatio...Marlon Dumas
Keynote talk by Marlon Dumas at the SIMULTECH 2021 conference. The talk gives an overview of ongoing research on automated construction of simulation models / digital twins from business process execution logs, including approaches that combine discrete event simulation with deep learning methods.
How GenAI will (not) change your business?Marlon Dumas
Not all new technology waves are the same. Some waves are vertical (3D printing, digital twins, blockchain) while others are horizontal (the PC in the 80s, the Web in the 90s). GenAI is a horizontal wave. The question is not if GenAI will impact my business, but what will be the scope of this impact. In this talk, we will go through a journey of collisions: GenAI colliding with customer service, clerical work, information search, content production, IT development, product design, and other knowledge work. A common thread to understand the impact of GenAI is to distinguish between descriptive use cases (search, summarize, expand, transcribe & translate) versus creative use.
Walking the Way from Process Mining to AI-Driven Process OptimizationMarlon Dumas
While generative AI grabs headlines, most organizations are yet to achieve continuous process improvement from predictive and prescriptive analytics.
Why? It’s largely about data, people, and a methodical approach to deploy AI to connect data and people. The good news is that if your organization has built a process mining capability, you are well placed to climb the ladder to achieve AI-driven process optimization. But to get there, you need a disciplined step-by-step approach along two tracks: a tactical management track and an operational management track.
First, it’s about predicting what will happen if you leave your process as-is, and what will happen if you implement a change in your process. At a tactical level, a predictive capability allows you to prioritize improvement opportunities. At an operational level, it allows you to predict issues, such as deadline violations. The challenges here are how to manage the inherent uncertainty of data-driven AI systems, and how to change your people and culture to manage processes proactively, rather than reactively. One thing is to deploy predictive dashboards, another entirely different thing is to get people to use them effectively to improve the processes.
Next, it’s about becoming preemptive: continuously optimizing your processes by leveraging streams of data-driven recommendations to trigger changes and actions. At the tactical level, this prescriptive capability allows you to implement the right changes to maximize competing KPIs. At the operational level, it means triggering interventions in your processes to “wow” customers and to meet SLAs in a cost-effective manner. The challenge here is how to help process owners, workers, and other stakeholders to understand the causes of performance issues and how the recommendations generated by the AI-driven optimization system will tackle those causes?
And finally, as an icing on the cake, generative AI allows you to produce improvement scenarios to adapt to external changes. Importantly, the transformative potential of generative AI in the context of process improvement does not come from its ability to provide question-and-answer interfaces to query data. It comes from its ability to support continuous process adaptation by generating and validating hypotheses based on a holistic view of your organization.
In this talk, we will discuss how organizations are driving sustainable business value by strategically layering predictive, prescriptive, and generative AI onto a process mining foundation, one brick at a time.
Industry keynote talk by Marlon Dumas at the 5th International Conference on Process Mining (ICPM'2023), Rome, Italy, 25 October 2023
More Related Content
Similar to Accurate and Reliable What-If Analysis of Business Processes: Is it Achievable?
Learning Accurate Business Process Simulation Models from Event Logs via Auto...Marlon Dumas
Paper presentation at the International Conference on Advanced Information Systems Engineering (CAiSE).
This paper presents an approach to automatically discover business process simulation models from event logs by combining process mining and deep learning techniques.
Paper available at: https://link.springer.com/chapter/10.1007/978-3-031-07472-1_4
Business Process Analytics: From Insights to PredictionsMarlon Dumas
Keynote talk at the 13th Baltic Conference on Databases and Information Systems, Trakai, Lithuania, 2 July 2018.
Abstract
Business process analytics is a body of methods for analyzing data generated by the execution of business processes in order to extract insights about weaknesses and improvement opportunities, both at the tactical and operational levels. Tactical process analytics methods (also known as process mining) allow us to understand how a given business process is actually executed, if and how its execution deviates with respect to expected or normative pathways, and what factors contribute to poor process performance or undesirable outcomes. Meantime, operational process analytics methods allow us to monitor ongoing executions of a business process in order to predict future states and undesirable outcomes at runtime (predictive process monitoring). Existing methods in this space allow us to predict, for example, which task will be executed next in a case, when, and who will perform it? When will an ongoing case complete? What will its outcome be and how can negative outcomes be avoided? This keynote will present a framework for conceptualizing business process analytics methods and applications. The talk will provide an overview of state-of-art methods and tools in the field and will outline open challenges and research opportunities.
Introduction to Business Process Analysis and RedesignMarlon Dumas
Special course on business process analysis and design delivered at University of Granada on 23-24 January 2014. The course covers qualitative and quantitative process analysis techniques and redesign heuristics. Based on the textbook Fundamentals of Business Process Management by Dumas et al.
These slides are from the Metrics-Based Process Mapping webinar delivered 09-29-2021.
Companion resources:
• View the recording - https://tkmg.com/webinars/metrics-based-process-mapping-3/
• Buy the book - https://tkmgacademy.com/products/metrics-based-process-mapping/
• Take the TKMG Academy course - https://tkmgacademy.com/courses/metrics-based-process-mapping/
C++ Is One Of The widely used programming language. Here is the complete presentation PPT notes of C++ programming language. hope it will be helpful to you.
Tieghi Anipla 20 04 2010 Come Possiamo Essere Sicuri Che Tutti Seguano Le Pro...Enzo M. Tieghi
How can you be sure that operators will follow SOPs Standard Operating Procedures? Using an Industrial WorkFlow management tool such as GEIP Procify WorkFlow.
This paper was presented at ANIPLA conference in Milano April 20th, 2010
Unbiased, Fine-Grained Description of Processes Performance from Event DataVadim Denisov
International Conference on Business Process Management 2018 (BPM2018 Sydney).
Performance is central to processes management and event data pro-vides the most objective source for analyzing and improving performance. Currentprocess mining techniques give only limited insights into performance by aggre-gating all event data for each process step. In this paper, we investigate processperformance of all process behaviors without prior aggregation. We propose theperformance spectrumas a simple model that maps all observed flows betweentwo process steps together regarding their performance over time. Visualizing theperformance spectrum of event logs reveals a large variety of very distinctpatternsof process performanceand performance variability that have not been describedbefore. We provide a taxonomy for these patterns and a comprehensive overviewof elementary and composite performance patterns observed on several real-lifeevent logs from business processes and logistics. We report on a case study whereperformance patterns were central to identify systemic, but not globally visibleprocess problems.
Mine Your Simulation Model: Automated Discovery of Business Process Simulatio...Marlon Dumas
Keynote talk by Marlon Dumas at the SIMULTECH 2021 conference. The talk gives an overview of ongoing research on automated construction of simulation models / digital twins from business process execution logs, including approaches that combine discrete event simulation with deep learning methods.
How GenAI will (not) change your business?Marlon Dumas
Not all new technology waves are the same. Some waves are vertical (3D printing, digital twins, blockchain) while others are horizontal (the PC in the 80s, the Web in the 90s). GenAI is a horizontal wave. The question is not if GenAI will impact my business, but what will be the scope of this impact. In this talk, we will go through a journey of collisions: GenAI colliding with customer service, clerical work, information search, content production, IT development, product design, and other knowledge work. A common thread to understand the impact of GenAI is to distinguish between descriptive use cases (search, summarize, expand, transcribe & translate) versus creative use.
Walking the Way from Process Mining to AI-Driven Process OptimizationMarlon Dumas
While generative AI grabs headlines, most organizations are yet to achieve continuous process improvement from predictive and prescriptive analytics.
Why? It’s largely about data, people, and a methodical approach to deploy AI to connect data and people. The good news is that if your organization has built a process mining capability, you are well placed to climb the ladder to achieve AI-driven process optimization. But to get there, you need a disciplined step-by-step approach along two tracks: a tactical management track and an operational management track.
First, it’s about predicting what will happen if you leave your process as-is, and what will happen if you implement a change in your process. At a tactical level, a predictive capability allows you to prioritize improvement opportunities. At an operational level, it allows you to predict issues, such as deadline violations. The challenges here are how to manage the inherent uncertainty of data-driven AI systems, and how to change your people and culture to manage processes proactively, rather than reactively. One thing is to deploy predictive dashboards, another entirely different thing is to get people to use them effectively to improve the processes.
Next, it’s about becoming preemptive: continuously optimizing your processes by leveraging streams of data-driven recommendations to trigger changes and actions. At the tactical level, this prescriptive capability allows you to implement the right changes to maximize competing KPIs. At the operational level, it means triggering interventions in your processes to “wow” customers and to meet SLAs in a cost-effective manner. The challenge here is how to help process owners, workers, and other stakeholders to understand the causes of performance issues and how the recommendations generated by the AI-driven optimization system will tackle those causes?
And finally, as an icing on the cake, generative AI allows you to produce improvement scenarios to adapt to external changes. Importantly, the transformative potential of generative AI in the context of process improvement does not come from its ability to provide question-and-answer interfaces to query data. It comes from its ability to support continuous process adaptation by generating and validating hypotheses based on a holistic view of your organization.
In this talk, we will discuss how organizations are driving sustainable business value by strategically layering predictive, prescriptive, and generative AI onto a process mining foundation, one brick at a time.
Industry keynote talk by Marlon Dumas at the 5th International Conference on Process Mining (ICPM'2023), Rome, Italy, 25 October 2023
Discovery and Simulation of Business Processes with Probabilistic Resource Av...Marlon Dumas
In the field of business process simulation, the availability of resources is captured by assigning a calendar to each resource, e.g., Monday-Friday 9:00-18:00. Resources are assumed to be always available to perform activities during their calendar. This assumption often does not hold due to interruptions, breaks, or because resources time-share across multiple processes. A simulation model that captures availability via crisp time slots (a resource is either on or off during a slot) does not capture these behaviors, leading to inaccuracies in the simulation output. This paper presents a simulation approach wherein resource availability is modeled probabilistically. In this approach, each availability time slot is associated with a probability, allowing us to capture, for example, that a resource is available on Fridays between 14:00-15:00 with 90% probability and between 17:00-18:00 with 50% probability. The paper proposes an algorithm to discover probabilistic availability calendars from event logs. An empirical evaluation shows that simulation models with probabilistic calendars discovered from event logs, replicate the temporal distribution of activity instances and cycle times of a process more closely than simulation models with crisp calendars.
This presentation was delivered at the 5th International Conference on Process Mining (ICPM'2023), Rome, Italy, October 2023.
The paper is available at: https://easychair.org/publications/preprint/Rz9g
Can I Trust My Simulation Model? Measuring the Quality of Business Process Si...Marlon Dumas
Business Process Simulation (BPS) is an approach to analyze the performance of business processes under different scenarios. For example, BPS allows us to estimate what would be the cycle time of a process if one or more resources became unavailable. The starting point of BPS is a process model annotated with simulation parameters (a BPS model). BPS models may be manually designed, based on information collected from stakeholders and empirical observations, or automatically discovered from execution data. Regardless of its origin, a key question when using a BPS model is how to assess its quality. In this paper, we propose a collection of measures to evaluate the quality of a BPS model w.r.t. its ability to replicate the observed behavior of the process. We advocate an approach whereby different measures tackle different process perspectives. We evaluate the ability of the proposed measures to discern the impact of modifications to a BPS model, and their ability to uncover the relative strengths and weaknesses of two approaches for automated discovery of BPS models. The evaluation shows that the measures not only capture how close a BPS model is to the observed behavior, but they also help us to identify sources of discrepancies.
Presentation delivered by David Chapela-Campa at the BPM'2023 conference, Utrecht, September 2023.
Business Process Optimization: Status and PerspectivesMarlon Dumas
For decades, business process optimization has been largely about art and craft (and sometimes wizardry). Apart from narrowly scoped approaches to optimize resource allocation (often assuming that workers behave like robots), a lot of business process optimization relies on high-level guidelines, with A/B testing for idea validation, which is hard to scale to complex processes. As a result, managers end up settling for a "good enough" process. Can we do more? In this talk, we review recent work on the use of high-fidelity simulation models discovered from execution data. The talk also explores the possibilities (and perils) that LLMs bring to the field of business process optimization.
This talk was delivered at the Workshop on Data-Driven Business Process Optimization at the BPM'2023 conference.
Learning When to Treat Business Processes: Prescriptive Process Monitoring wi...Marlon Dumas
Paper presentation at the 35th International Conference on Advanced Information Systems Engineering (CAiSE'2023).
Abstract.
Increasing the success rate of a process, i.e. the percentage of cases that end in a positive outcome, is a recurrent process improvement goal. At runtime, there are often certain actions (a.k.a. treatments) that workers may execute to lift the probability that a case ends in a positive outcome. For example, in a loan origination process, a possible treatment is to issue multiple loan offers to increase the probability that the customer takes a loan. Each treatment has a cost. Thus, when defining policies for prescribing treatments to cases, managers need to consider the net gain of the treatments. Also, the effect of a treatment varies over time: treating a case earlier may be more effective than later in a case. This paper presents a prescriptive monitoring method that automates this decision-making task. The method combines causal inference and reinforcement learning to learn treatment policies that maximize the net gain. The method leverages a conformal prediction technique to speed up the convergence of the reinforcement learning mechanism by separating cases that are likely to end up in a positive or negative outcome, from uncertain cases. An evaluation on two real-life datasets shows that the proposed method outperforms a state-of-the-art baseline.
Why am I Waiting Data-Driven Analysis of Waiting Times in Business ProcessesMarlon Dumas
Presentation of a research paper at the 35th International Conference on Advanced Information Systems Engineering (CAiSE) in Zaragoza Spain. The paper presents a classification of causes of waiting times in business processes and a method to automatically detect and quantify the presence of each of these causes in a business process recorded in an event log.
This talk introduces the concept of Augmented Business Process Management System: An ABPMS is a process-aware information system that relies on trustworthy AI technology to
reason and act upon data, within a set of restrictions, with the aim to continuously adapt and
improve a set of business processes with respect to one or more key performance indicators.
The talk describes the transition from existing process mining technology to AI-Augmented BPM as a pyramid, where predictive, prescriptive, conversational and reasoning capabilities are stacked up incrementally to reach the level of Augmented BPM.
Talk delivered at the AAAI'2023 Workshop on AI for Business Process Management.
Process Mining and Data-Driven Process SimulationMarlon Dumas
Guest lecture delivered at the - Institut Teknologi Sepuluh on 8 December 2022.
This lecture gives an overview of process mining and simulation techniques, and how the two can be used together in process improvement projects.
Modeling Extraneous Activity Delays in Business Process SimulationMarlon Dumas
This paper presents a technique to enhance the fidelity of business process simulation models by detecting unexplained (extraneous) delays from business process execution data, and modeling these delays in the simulation model, via timer events.
The presentation was delivered at the 4th International Conference on Process Mining (ICPM'2022).
Paper available at: https://arxiv.org/abs/2206.14051
Business Process Simulation with Differentiated Resources: Does it Make a Dif...Marlon Dumas
Existing methods for discovering business process simulation models from execution data (event logs) assume that all resources in a pool have the same performance and share the same availability calendars. This paper proposes a method for discovering simulation models, wherein each resource is treated as an individual entity, with its own performance and availability calendar. An evaluation shows that simulation models with differentiated resources more closely replicate the distributions of cycle times and the work rhythm in a process than models with undifferentiated resources. The paper is available at: https://link.springer.com/chapter/10.1007/978-3-031-16103-2_24
Prescriptive Process Monitoring Under Uncertainty and Resource ConstraintsMarlon Dumas
This paper presents an approach to trigger runtime interventions at runtime, in order to improve the success rate of a process, when the number of resources who can perform these interventions is limited.
The paper is available at: https://link.springer.com/chapter/10.1007/978-3-031-16171-1_13
The presentation delivered at the 20th International Conference on Business Process Management (BPM'2022), in Muenster, Germany, September 2022.
Slides of a lecture delivered at the First Process Mining Summer School in Aachen, Germany, July 2022.
This lecture introduces techniques in the area of "task mining" with an emphasis on Robotic Process Mining. Robotic Process Mining (RPM) is a family of techniques to discover repetitive routines that can be automated using Robotic Process Automation (RPA) technology, by analyzing interactions between
one or more workers and one or more software applications, during the performance of one or more tasks in a business process. In general, RPM techniques take as input logs of User Interactions (UI logs). These UI logs are recorded while workers interact with one or more applications, typically desktop applications. Based on these logs, RPM techniques produce specifications of one or more routines that can be automated using RPA or related tools.
Process Mining: A Guide for PractitionersMarlon Dumas
Paper presentation delivered at the Research Conference on Challenges in Information Science (RCIS 2022). The paper studies the following questions:
1) What are the most common use cases for process mining methods?
2) What business questions do process mining methods address?
Paper available at:
https://link.springer.com/chapter/10.1007/978-3-031-05760-1_16
Process Mining for Process Improvement.pptxMarlon Dumas
Presentation of a research paper at the 16th International Conference on Research Challenges in Information Science (RCIS). The paper presents the results of an empirical study on how practitioners use process mining to identify business process improvement opportunities. The paper is available at: https://link.springer.com/chapter/10.1007/978-3-031-05760-1_13
Data-Driven Analysis of Batch Processing Inefficiencies in Business ProcessesMarlon Dumas
Slides of a research paper presentation at the 16th International Conference on Research Challenges in Information Science (RCIS).
The research paper presents an approach to analyze event logs of business processes in order to identify batched activities and to analyze the waiting times caused by these activities.
Paper available at: https://link.springer.com/chapter/10.1007/978-3-031-05760-1_14
Optimización de procesos basada en datosMarlon Dumas
Ponencia en BPM Day Lima 2021.
En esta charla, hablaremos de métodos y aplicaciones emergentes en el ámbito de la optimización de procesos basada en datos. Hablaremos de avances en el área de la minería de procesos, de métodos de construcción de gemelos digitales de procesos y de métodos de monitoreo predictivo. Mostraremos por medio de ejemplos y casos de estudio, cómo estos métodos permiten guiar las iniciativas de transformación digital y de mejora continua de procesos, En particular, ilustraremos el uso de estos métodos para: (1) analizar el rendimiento de los procesos de negocio de manera a identificar fricciones y oportunidades de automatización; (2) predecir el impacto de cambios, y en particular, predecir el impacto de una iniciativa de automatización; (3) realizar predicciones sobre el rendimiento del proceso y ajustar la ejecución del proceso de manera a prevenir incumplimientos del SLA, quejas de clientes, y otros eventos indeseables.
Process Mining and AI for Continuous Process ImprovementMarlon Dumas
Talk delivered at BPM Day Rio Grande do Sul on 11 November 2021.
Abstract.
Process mining is a technology that marries methods from business process management and from data science, to support operational excellence and digital transformation. Process mining tools can transform data extracted from enterprise systems, into visualizations and reports that allow managers to improve organizational performance along different dimensions, such as efficiency, quality, and compliance. In this talk, we will give an overview of the capabilities of process mining tools, and we will illustrate the benefits of process mining via several case studies in the fields of insurance, manufacturing, and IT service management.
Prescriptive Process Monitoring for Cost-Aware Cycle Time ReductionMarlon Dumas
Paper presentation at the 3rd International Conference on Process Mining (ICPM), 4 November 2021.
The paper is available at: https://arxiv.org/abs/2105.07111
On the Road to AI-Infused Process ExecutionMarlon Dumas
Talk delivered at the ISSIP Discovery Summit: AI & Automation on 12 May 2021.
Abstract:
Traditionally, process automation and process monitoring have been living in two worlds. As AI technology reaches maturity, we are seeing a convergence between automation and monitoring. Data collected during the execution of business processes is fed into AI tools, which in turn drive the automation of these processes.
For example, predictive process monitoring techniques exploit events generated during the execution of a process (e.g. in an ERP or a BPMS) allow us to make predictions about the future states of a process. These predictions are then used to trigger interventions, to enhance the performance of the process. In the RPA world, UI-level events gathered on the background while workers perform their daily work are used to discover automatable routines that are then used to instrument RPA bots.
These and other applications of AI techniques are going to reshape the way we think about the BPM lifecycle. Monitoring and automation will be replaced by AI-Infused Process Execution.
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Accurate and Reliable What-If Analysis of Business Processes: Is it Achievable?
1. Marlon Dumas
Professor @ University of Tartu
Co-founder @ Apromore
With contributions by Manuel Camargo and Oscar González-Rojas
Baltic DB&IS’2022, 5 July 2022
6. 6
Process Credit Card
Accept Cash or Check
Identify payment method
18 / 8 min
20 / 5 min
5 / 2 min
Prepare package for customer
10 / 5 min
Cycle time
Processing / Waiting times
Costs x activity x resource …
Resource utilization
1h
5 min
38 min
10 min
The Traditional Answer: Business Process Simulation
11. Business Process Simulation: Assumptions
The process model is authoritative (always followed to the letter)
• No deviations
• No workarounds
The simulation parameters accurately reflect reality
• …whereas in reality, they are often guesstimates
A resource only works on one task instance at a time / a task is performed by one resource
• No multi-tasking / no multi-resource tasks (teamwork)
Resources have robotic behavior (eager resources consume work items in FIFO mode)
• No batching
• No tiredness effects, no interruptions, no distractions beyond “stochastic” ones
Undifferentiated resources
• Every resource in a pool has the same performance as others
No time-sharing outside the simulated process
• Resources fully dedicated to one process 12
12. End Result
Business process simulations based
on incomplete models,
guesstimates, and simplifying
assumptions are not faithful
adoption of business process
simulation is disappointing
3
14. Given
• one or more business processes, for which we
have:
• one or more process specifications and/or
• event logs generated by the execution of the
processes on top of one or more information
systems.
• one or more process performance measures of
interest (e.g. cycle time, resource cost)
• One or more changes to the process (interventions)
Predict
• Predict the values of the process performance
measures after the given interventions.
15. Non-Functional Requirements
16
Predictions accurate.
Accuracy may be measured e.g. via an error
between the predicted and the actual
performance measures after intervention.
Predictions should be accompanied by a
reliability estimate. In most cases, the
reliability is high.
Reliability could be captured, e.g. by
confidence intervals
21. Dataset Control-Flow
Similarity
(string-edit distance)
Temporal
Similarity
(timed-string edit distance)
Call centre 0.37 0.41
Pharmacy customer service 0.29 0.30
Purchase-to-Pay 0.55 0.57
Make-to-order manufacturing 0.65 0.69
Academic credentials recognition 0.32 0.29
Insurance claims handling 0.39 0.43
Loan Origination 0.41 0.42
Discover simulation
model
Simulate model
10 times
Evaluate Similarity
(mean string-edit
distance & timed-
string edit distance)
This Photo by Unknown Author is
licensed under CC BY-NC
22. We can try to fill
in the glass
• Discover and add batching behavior to simulation models
• .. prioritization
• … timers and external factors (not explicit in the data)
• etc.
Or perhaps we should look for another paradigm….
23
23. 24
How is going to continue this case until it is finished?
How long is this case still going to take until it is finished?
What is the next activity for this case?
When is this next activity going to take place?
…
Generate a set of traces (event log)
E1 E2 E3
Running case
24. 2
5
{T1 -> T2 -> T3}
{T1 -> T3 -> T3}
{T1 -> T2 -> T3}
{T1 -> T2 -> T3}
{T1 -> T2 -> T2}
{T1 -> T2 -> T3}
{T1 -> T2 -> T3}
{T1 -> T3 -> T2}
{T1 -> T2 -> T3}
{T1 -> T2 -> T3}
Event log
Pre-processing
• Timestamp
relativization
• Scaling
Continuous features
• Embedded dimensions
• N-grams extraction
• Discovery of roles
Discrete features
Generation &
Assessment
Selection method
• Arg. Max
• Random choice
Accuracy assessment
Model selection &
Training
Architecture selection
Training with
examples of size 0
• Specialized
• Shared
• Concatenated
26. • DDS Models (SIMOD) and DL
models have comparable
performance w.r.t. control-flow
similarity (CLFS)
• DL models sometimes clearly
outperform DDS models on
temporal metrics (MAE, ELS)
Could we combine them?
27. - Assumes undifferentiated resources with robotic
behavior
- Branches are selected using branching probabilities
Generative Deep Learning Methods
- Does not explicitly take into account resource
constraints
- Models resource availability via neural networks that
may capture non-linear availability functions
- Learns the case arrival process from data (univariate
or multivariate models)
- May take as input a process specification (helps with
interpretability)
- Models the case creation process via a probability
distribution
- Takes into account resource constraints
- No interpretable process specification
- Branching behavior modeled via neural networks (e.g.
LSTM) that may capture complex relations
- Models resource availability as calendars (possibly
discovered from historical data)
- May capture differentiated resources and robotic
behavior
Data-Driven (Discrete Event) Simulation
- Provides a natural mechanism for capturing the effect
of changes to the process
- Does not have a mechanism for capturing the effect
of changes to the process
28. Phase 1
A1 A2 A3 A5 A6
A4
Ꝺ1:
A2 A3 A4 A5
Ꝺ2:
Phase 2
A1 A2 A3 A5 A6
A4
A2 A3 A4 A5
Phase 3
e1- start e1- complete
e2- start
𝜎1 Ac1
Ac2
e2- complete
Waiting time predictive model
Features: Wait+Ac2+Cx+WIP+RO
Processing time predictive model
Features: Proc+Ac1+Cx+WIP+RO
A1 A2 A3 A5 A6
A4
A2 A3 A4 A5
Discovering a process model to
generate traces
Learning a time series generator to
determine when each trace starts
Deep-learning the processing time and
waiting time of each activity in a given trace
30. 34
Batch 1 Batch 2 Batch 3 Batch 4 Batch 5 Batch 6
• DeepSimulator can better estimate the impact of changes in the demand in settings where such
changes have been previously observed in the data.
31. 36
• The accuracy of DeepSimulator degrades when evaluated in a previously unobserved scenario (new
task is added to the process)
SIMOD DSIM SIMOD DSIM SIMOD DSIM
Version 1
BPI17W 971151 417572 0.02222 0.03593 3185 3647
BPI12W 660211 534341 0.11295 0.04853 515 458
CVS 1489252 467572 0.03213 0.00001 3380 849
Version 2
BPI17W 895524 290980 0.06438 0.03218 4528 3431
BPT12W 550266 524995 0.25888 0.22003 726 507
CVS 540112 246159 0.15674 0.05708 2453 1967
AS-IS WHAT-IF AS-IS WHAT-IF AS-IS WHAT-IF
CFM 7155 17546 22006 33137 0.15629 0.28762
CVS 283061 1040344 357717 1052255 0.31972 1.84601
Log
MAE RMSE SMAPE
Scenario
1
Scenario
2
Log
MAE EMD DTW
This Photo by Unknown Author
is licensed under CC BY-NC-ND
32. Wrap-Up
• There’s a long road ahead to constructing accurate and reliable
simulation models from event logs
• Combination of deep learning techniques & simulation promising, but
need to be further researched to become practically usable for what-
if analysis
• Extensions needed to support a wide range of interventions / changes
• Extensions needed to provide reliability estimates (for what-if analysis)
• More validation in large-scale scenarios
37
33. References
Limitations and pitfalls of traditional BP simulation
• van der Aalst: Business Process Simulation Survival Guide. In Handbook on Business
Process Management Vol. 1, 2015, 337-370
Data-Driven Simulation (discovering simulation models from logs)
• Rozinat et al. Discovering simulation models. Inf. Syst. 34(3): 305-327 (2009)
• Martin et al. The Use of Process Mining in Business Process Simulation Model Construction
- Structuring the Field. Bus. Inf. Syst. Eng. 58(1): 73-87
• Camargo et al. Automated discovery of business process simulation models from event
logs. Decis. Support Syst. 134:113284, 2020 https://arxiv.org/abs/2009.03567
• Pourbafrani et al. Extracting Process Features from Event Logs to Learn Coarse-Grained
Simulation Models. CAiSE 2021: 125-140
Data-Driven Simulation and Deep Learning
• Camargo et al. Discovering Generative Models from Event Logs: Data-driven Simulation vs
Deep Learning, PeerJ Computer Science, 7: e577, 2021 https://peerj.com/articles/cs-577/
• Camargo et al. Learning Accurate Business Process Simulation Models from Event Logs via
Automated Process Discovery and Deep Learning. CAiSE’2022
https://arxiv.org/abs/2103.11944