Keynote speech at KES 2022 on "Intelligent Systems for Process Mining". I introduce process mining, discuss why process mining tasks should be approached by using intelligent systems, and show a concrete example of this combination, namely (anticipatory) monitoring of evolving processes against temporal constraints, using techniques from knowledge representation and formal methods (in particular, temporal logics over finite traces and their automata-theoretic characterization).
How Fast Data Is Turned into Fast Information and Timely Action (OOW 2014)Lucas Jellema
Fast data is big data, continuously streaming in, from which information is to be learned in (near) real time. This session demonstrates how Oracle Event Processing is used to analyze live streams of data to find patterns, deviations, and aggregates. The findings are reported in the form of business events that are pushed in live dashboards to Oracle Business Activity Monitoring, which also evaluates business rules on the business events and takes action when required. Examples to be demonstrated in this session include car sensors, website traffic, Twitter feeds, and bank run detection. Oracle SOA Suite 12c, WebSockets, Oracle Application Development Framework (Oracle ADF) active data visualization tools components, and JMS are used to process, forward, and act.
OSA Con 2022 - Building Event Collection SDKs and Data Models - Paul Boocock ...Altinity Ltd
OSA Con 2022: Building Event Collection SDKs and Data Models
Paul Boocock - Snowplow
In this talk we'll go through how we have designed and built over 20 different SDKs to collect events from all sorts of applications (from web & mobile to IoT to server-side), allowing users to collect a rich event stream of data. Then we'll dive into, and demonstrate, the cross-warehouse downstream data models which aggregate the event stream into easy-to-consume data products for analytics, AI, composable CDP, recommendation engines, and many other use cases.
Beyond php - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Keynote speech at the Belgian Process Mining Research Day 2021. I discuss the open, critical challenge of data preparation in process mining, considering the case where the original event data are implicitly stored in (legacy) relational databases. This case covers the common situation where event data are stored inside the data layer of an ERP or CRM system. This is usually handled using manual, ad-hoc, error-prone ETL procedures. I propose instead to adopt a pipeline based on semantic technologies, in particular the framework of ontology-based data access (also known as virtual knowledge graph). The approach is code-less, and relies on three main conceptual steps: (1) the creation of a data model capturing the relevant classes, attributes, and associations in the domain of interest (2) the definition of declarative mappings from the source database to the data model, following the ontology-based data access paradigm (3) the annotation of the data model with indications on which classes/associations/attributes provide the relevant notions of case, events, event attributes, and event-to-case relation. Once this is done, the framework automatically extracts the event log from the legacy data. This makes extremely smooth to generate logs by taking multiple perspectives on the same reality. The approach has been operationalized in the onprom tool, which employs semantic web standard languages for the various steps, and the XES standard as the target format for the event logs.
Using neo4j for enterprise metadata requirementsNeo4j
Metadata is everywhere yet traditionally approaches to managing it have been disparate, siloed and often ineffective.
In this talk James will discuss the opportunities for using graph technology to address the fundamental challenges and questions of metadata management such as impact analysis, data lineage and definitions.
Data to Value are a Data Consultancy based in London that specialise in applying lean and agile techniques to complex data requirements. Connected Data is a particular focus for the firm which they see as the new frontier for data leaders.
James Phare has over 15 years experience of creating and leading data teams in various roles in Financial Services. Prior to cofounding Data Consultancy Data to Value he was Head of Information Management and Data Architecture at Man Group – one of the world’s largest Hedge funds. James started his career at Thomson Reuters after graduating in Economics from the University of York.
CS8075 - Data Warehousing and Data Mining (Ripped from Amazon Kindle eBooks b...vinoth raja
This document provides an overview of data warehousing and online analytical processing (OLAP). It discusses the overall architecture of a data warehouse including the data warehouse database, sourcing and transformation tools, metadata, access tools, administration, and delivery systems. It also covers building a data warehouse including business, design, technical, and implementation considerations. Database architectures for parallel processing like shared everything, shared disk, and shared-nothing architectures are described. Common data warehouse schemas like the star schema and multidimensional data model are explained. Finally, the document discusses the characteristics of OLAP systems and typical OLAP operations.
The document discusses various reporting tools in SAP that provide information and analysis on workflows, work items, and system load including the Workflow Information System, Work Item Analysis, Workflow Load Analysis, and workflow reports. These tools allow users to monitor the status of workflows, workload statistics, workflow performance metrics, and troubleshoot issues. SAP provides standard functionality for capturing critical workflow information through statistical analysis and centralized reporting resources.
The document provides instructions for a Microsoft Access 2003 training workshop. It includes an introduction, learning objectives, and overview of Access components like tables, forms, queries, and reports. The exercises section provides step-by-step instructions for creating tables, forms, queries, and reports in Access using wizards and manual processes. Participants will learn basic skills for designing and building an Access database.
How Fast Data Is Turned into Fast Information and Timely Action (OOW 2014)Lucas Jellema
Fast data is big data, continuously streaming in, from which information is to be learned in (near) real time. This session demonstrates how Oracle Event Processing is used to analyze live streams of data to find patterns, deviations, and aggregates. The findings are reported in the form of business events that are pushed in live dashboards to Oracle Business Activity Monitoring, which also evaluates business rules on the business events and takes action when required. Examples to be demonstrated in this session include car sensors, website traffic, Twitter feeds, and bank run detection. Oracle SOA Suite 12c, WebSockets, Oracle Application Development Framework (Oracle ADF) active data visualization tools components, and JMS are used to process, forward, and act.
OSA Con 2022 - Building Event Collection SDKs and Data Models - Paul Boocock ...Altinity Ltd
OSA Con 2022: Building Event Collection SDKs and Data Models
Paul Boocock - Snowplow
In this talk we'll go through how we have designed and built over 20 different SDKs to collect events from all sorts of applications (from web & mobile to IoT to server-side), allowing users to collect a rich event stream of data. Then we'll dive into, and demonstrate, the cross-warehouse downstream data models which aggregate the event stream into easy-to-consume data products for analytics, AI, composable CDP, recommendation engines, and many other use cases.
Beyond php - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Keynote speech at the Belgian Process Mining Research Day 2021. I discuss the open, critical challenge of data preparation in process mining, considering the case where the original event data are implicitly stored in (legacy) relational databases. This case covers the common situation where event data are stored inside the data layer of an ERP or CRM system. This is usually handled using manual, ad-hoc, error-prone ETL procedures. I propose instead to adopt a pipeline based on semantic technologies, in particular the framework of ontology-based data access (also known as virtual knowledge graph). The approach is code-less, and relies on three main conceptual steps: (1) the creation of a data model capturing the relevant classes, attributes, and associations in the domain of interest (2) the definition of declarative mappings from the source database to the data model, following the ontology-based data access paradigm (3) the annotation of the data model with indications on which classes/associations/attributes provide the relevant notions of case, events, event attributes, and event-to-case relation. Once this is done, the framework automatically extracts the event log from the legacy data. This makes extremely smooth to generate logs by taking multiple perspectives on the same reality. The approach has been operationalized in the onprom tool, which employs semantic web standard languages for the various steps, and the XES standard as the target format for the event logs.
Using neo4j for enterprise metadata requirementsNeo4j
Metadata is everywhere yet traditionally approaches to managing it have been disparate, siloed and often ineffective.
In this talk James will discuss the opportunities for using graph technology to address the fundamental challenges and questions of metadata management such as impact analysis, data lineage and definitions.
Data to Value are a Data Consultancy based in London that specialise in applying lean and agile techniques to complex data requirements. Connected Data is a particular focus for the firm which they see as the new frontier for data leaders.
James Phare has over 15 years experience of creating and leading data teams in various roles in Financial Services. Prior to cofounding Data Consultancy Data to Value he was Head of Information Management and Data Architecture at Man Group – one of the world’s largest Hedge funds. James started his career at Thomson Reuters after graduating in Economics from the University of York.
CS8075 - Data Warehousing and Data Mining (Ripped from Amazon Kindle eBooks b...vinoth raja
This document provides an overview of data warehousing and online analytical processing (OLAP). It discusses the overall architecture of a data warehouse including the data warehouse database, sourcing and transformation tools, metadata, access tools, administration, and delivery systems. It also covers building a data warehouse including business, design, technical, and implementation considerations. Database architectures for parallel processing like shared everything, shared disk, and shared-nothing architectures are described. Common data warehouse schemas like the star schema and multidimensional data model are explained. Finally, the document discusses the characteristics of OLAP systems and typical OLAP operations.
The document discusses various reporting tools in SAP that provide information and analysis on workflows, work items, and system load including the Workflow Information System, Work Item Analysis, Workflow Load Analysis, and workflow reports. These tools allow users to monitor the status of workflows, workload statistics, workflow performance metrics, and troubleshoot issues. SAP provides standard functionality for capturing critical workflow information through statistical analysis and centralized reporting resources.
The document provides instructions for a Microsoft Access 2003 training workshop. It includes an introduction, learning objectives, and overview of Access components like tables, forms, queries, and reports. The exercises section provides step-by-step instructions for creating tables, forms, queries, and reports in Access using wizards and manual processes. Participants will learn basic skills for designing and building an Access database.
Beyond php - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Excel can be used as a powerful business intelligence (BI) tool to analyze data and present actionable insights. It allows users to collect, organize, and analyze data from various sources. Excel provides capabilities for data discovery, dashboards, reporting, analysis, and predictive analytics to help users identify patterns and trends and make more informed business decisions. The document demonstrates how Excel can be used to analyze call center metrics data through features like pivot tables, charts, and dashboards to provide visualizations and insights.
Fast data is big data, continuously streaming in, from which information is to be learned in (near) real time. This session demonstrates how Oracle Event Processing is used to analyze live streams of data to find patterns, deviations, and aggregates. The findings are reported in the form of business events that are pushed in live dashboards to Oracle Business Activity Monitoring, which also evaluates business rules on the business events and takes action when required. Examples to be demonstrated in this session include car sensors, website traffic, Twitter feeds, and bank run detection. Oracle SOA Suite 12c, WebSockets, Oracle Application Development Framework (Oracle ADF) active data visualization tools components, and JMS are used to process, forward, and act.
Beyond PHP - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just writing PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Beyond php - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Topic: Discover deep insights with Salesforce Einstein Analytics and Discovery
ImpactSalesforceSaturday Session
by @newdelhisfdcdug
Speaker: Jayant Joshi
AGENDA
a. What is SFDC Einstein Analytics?
b. Let us build great Visualizations using Einstein Analytics
c. Discover Deep Insights with Einstein Discovery
d. Demo and QA
https://newdelhisfdcdug.com/salesforce-einstein-analytics-and-discovery/
Azure Stream Analytics : Analyse Data in MotionRuhani Arora
The document discusses evolving approaches to data warehousing and analytics using Azure Data Factory and Azure Stream Analytics. It provides an example scenario of analyzing game usage logs to create a customer profiling view. Azure Data Factory is presented as a way to build data integration and analytics pipelines that move and transform data between on-premises and cloud data stores. Azure Stream Analytics is introduced for analyzing real-time streaming data using a declarative query language.
This document discusses high-velocity big data analytics and describes how streaming data can be captured, processed, and analyzed in real-time to enable immediate action. It outlines an approach that assimilates structured and unstructured data from various sources, processes the data using distributed in-memory computing, correlates and enriches the real-time data records, and delivers results and alerts. Visual dashboards are used to view the real-time analytics and detect patterns, outliers, and trends in big data.
M|18 Analytics in the Real World, Case Studies and Use CasesMariaDB plc
This document provides examples of analytics use cases across various industries including finance, manufacturing, telecom, healthcare, life science, digital advertisement, high tech, and more. It also includes case studies on using MariaDB ColumnStore for genetic profiling, decision support, ad analytics, asset management, predictive maintenance, and trading analysis. Specific benefits highlighted include faster data ingestion, complex queries on large datasets, real-time monitoring and analysis, and predictive capabilities.
"Business and Manufacturing Intelligence" presented on September 9, 2016, by E2i at ComEd's Accelerate Tech event, Olivet Nazarene University, Bourbonnais, IL. Hosts included the Economic Alliance of Kankakee County, ONU Walker School of Engineering and the Kankakee County Chamber of Commerce.
This document provides an introduction to Sybase Event Stream Processing (ESP). ESP is a technology for analyzing streams of event data in real-time to derive continuous intelligence. It allows defining logic to combine data sources, compute values, detect patterns, and produce summaries. The ESP runtime continuously processes incoming event streams before storing to disk, enabling low latency analysis. It is not a replacement for databases but rather complements them by supporting continuous, real-time analysis of fast data streams. ESP uses data-flow programming where streams and operations on the data are defined as it flows from sources to outputs.
This project will analyze large existing data sets concerning the usage and maintenance of equipment at the South Carolina Department of Transportation. The effort will assess SCDOT repair shop capability needs, capacity, skills, and tools/equipment. The analysis is expected to yield correlations and insights about the equipment, so that it can be maintained more effectively and at lower cost.
This talk is about data science and statistics applied to flight safety in commercial aviation worldwide. In the introductory part, we will stress the importance of monitoring your flight data and show you some real records coming from flight data recorders (aircraft “black boxes”). We will then explain how the data is recorded, downloaded, analysed, converted to safety events and finally, validated by experts in the field - flight data analysts. Data aggregation across many flights will result with statistical images of safety risks in airlines’ operations. However, this valuable tool can turn into a deadly weapon if used negligently – we’ll support this claim with examples. We are convinced the audience will know about some of these traps, regardless of the industry they are coming from, but hopefully there will be something valuable to take home, too. In the second part, we are saying goodbye to the data analyst and the statistician – the two dominant guys from the first part of the presentation. However, a data scientist will stem from valuable experiences and domain knowledges of the two. This guy will walk the audience through three simple, but working examples. The first one is about how we can improve the accuracy of automated analysis by using historic data and a probabilistic, Bayesian approach. The second example is about finding novel safety risks in airlines’ operations by using simple principal component analysis. Lastly, we’ll use a Markov model to detect aircraft which have changed behaviour with respect to frequency of data downloads, so we collect as many flight data as possible. We will try to make this chat as interesting and as interactive as we can and are looking forward to meeting you at this fun and interesting conference!
This document provides a summary of Letgo's data platform. It discusses how Letgo ingests 500GB of data daily from various event types and processes over 530 million events daily with near real-time processing. It describes Letgo's data journey in moving to an event-driven architecture and storing data in S3. Key components of Letgo's data platform are discussed like stream processing for real-time user segmentation, batch processing for geo data enrichment, querying data using Spark SQL, and data exploitation tools. The document also provides some tips for orchestrating jobs and monitoring metrics.
Learning Accurate Business Process Simulation Models from Event Logs via Auto...Marlon Dumas
Paper presentation at the International Conference on Advanced Information Systems Engineering (CAiSE).
This paper presents an approach to automatically discover business process simulation models from event logs by combining process mining and deep learning techniques.
Paper available at: https://link.springer.com/chapter/10.1007/978-3-031-07472-1_4
Beyond PHP - It's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
This document discusses how advanced analytics and predictive modeling can help associations achieve strategic goals. It contrasts business intelligence with data science and explains how predictive modeling fits within an association's analytics framework. The document also provides examples of predictive models that associations could use, such as models to predict meeting attendance, purchasing likelihood, or long-term revenue from new members. Finally, it discusses how AFP has used predictive analytics to improve initiatives like customer journeys, onboarding programs, and community platforms.
Creating a Big data Strategy with Tactics for Quick ImplementationLewandog, Inc,
The document provides an overview of creating a big data strategy with quick implementation tactics. It discusses establishing centers of excellence, empowering users to access and analyze data, and transforming organizations to be data-driven. It also notes that only a small percentage of data is currently analyzed and highlights opportunities that additional analysis could provide through increased profits and insights.
Stork Webinar | Digital Transformation AssessmentStork
In dit webinar wordt aan de hand van voorbeelden uit de praktijk uitgelegd hoe een assessment kan helpen jouw positie op ladder van digitalisering binnen asset management te bepalen, om doelstellingen vast te leggen om tot een pragmatische aanpak te komen en hoe de gaten te overbruggen tussen doelstelling en praktijk met acties die reëel, praktisch en geprioriteerd zijn.
The document discusses challenges with modeling processes that involve multiple interacting objects. Conventional process modeling approaches encourage separating objects and focusing on one object type per process, which can lead to issues when objects interact. The document proposes modeling objects as first-class citizens and capturing relationships between objects to better represent real-world processes where objects corelate and influence each other. It provides examples of how conventional case-centric modeling can struggle to accurately capture a hiring process that involves interacting candidate, application, job offer and other objects.
Slides of our BPM 2022 paper on "Reasoning on Labelled Petri Nets and Their Dynamics in a Stochastic Setting", which received the best paper award at the conference. Paper available here: https://link.springer.com/chapter/10.1007/978-3-031-16103-2_22
Beyond php - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Excel can be used as a powerful business intelligence (BI) tool to analyze data and present actionable insights. It allows users to collect, organize, and analyze data from various sources. Excel provides capabilities for data discovery, dashboards, reporting, analysis, and predictive analytics to help users identify patterns and trends and make more informed business decisions. The document demonstrates how Excel can be used to analyze call center metrics data through features like pivot tables, charts, and dashboards to provide visualizations and insights.
Fast data is big data, continuously streaming in, from which information is to be learned in (near) real time. This session demonstrates how Oracle Event Processing is used to analyze live streams of data to find patterns, deviations, and aggregates. The findings are reported in the form of business events that are pushed in live dashboards to Oracle Business Activity Monitoring, which also evaluates business rules on the business events and takes action when required. Examples to be demonstrated in this session include car sensors, website traffic, Twitter feeds, and bank run detection. Oracle SOA Suite 12c, WebSockets, Oracle Application Development Framework (Oracle ADF) active data visualization tools components, and JMS are used to process, forward, and act.
Beyond PHP - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just writing PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Beyond php - it's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
Topic: Discover deep insights with Salesforce Einstein Analytics and Discovery
ImpactSalesforceSaturday Session
by @newdelhisfdcdug
Speaker: Jayant Joshi
AGENDA
a. What is SFDC Einstein Analytics?
b. Let us build great Visualizations using Einstein Analytics
c. Discover Deep Insights with Einstein Discovery
d. Demo and QA
https://newdelhisfdcdug.com/salesforce-einstein-analytics-and-discovery/
Azure Stream Analytics : Analyse Data in MotionRuhani Arora
The document discusses evolving approaches to data warehousing and analytics using Azure Data Factory and Azure Stream Analytics. It provides an example scenario of analyzing game usage logs to create a customer profiling view. Azure Data Factory is presented as a way to build data integration and analytics pipelines that move and transform data between on-premises and cloud data stores. Azure Stream Analytics is introduced for analyzing real-time streaming data using a declarative query language.
This document discusses high-velocity big data analytics and describes how streaming data can be captured, processed, and analyzed in real-time to enable immediate action. It outlines an approach that assimilates structured and unstructured data from various sources, processes the data using distributed in-memory computing, correlates and enriches the real-time data records, and delivers results and alerts. Visual dashboards are used to view the real-time analytics and detect patterns, outliers, and trends in big data.
M|18 Analytics in the Real World, Case Studies and Use CasesMariaDB plc
This document provides examples of analytics use cases across various industries including finance, manufacturing, telecom, healthcare, life science, digital advertisement, high tech, and more. It also includes case studies on using MariaDB ColumnStore for genetic profiling, decision support, ad analytics, asset management, predictive maintenance, and trading analysis. Specific benefits highlighted include faster data ingestion, complex queries on large datasets, real-time monitoring and analysis, and predictive capabilities.
"Business and Manufacturing Intelligence" presented on September 9, 2016, by E2i at ComEd's Accelerate Tech event, Olivet Nazarene University, Bourbonnais, IL. Hosts included the Economic Alliance of Kankakee County, ONU Walker School of Engineering and the Kankakee County Chamber of Commerce.
This document provides an introduction to Sybase Event Stream Processing (ESP). ESP is a technology for analyzing streams of event data in real-time to derive continuous intelligence. It allows defining logic to combine data sources, compute values, detect patterns, and produce summaries. The ESP runtime continuously processes incoming event streams before storing to disk, enabling low latency analysis. It is not a replacement for databases but rather complements them by supporting continuous, real-time analysis of fast data streams. ESP uses data-flow programming where streams and operations on the data are defined as it flows from sources to outputs.
This project will analyze large existing data sets concerning the usage and maintenance of equipment at the South Carolina Department of Transportation. The effort will assess SCDOT repair shop capability needs, capacity, skills, and tools/equipment. The analysis is expected to yield correlations and insights about the equipment, so that it can be maintained more effectively and at lower cost.
This talk is about data science and statistics applied to flight safety in commercial aviation worldwide. In the introductory part, we will stress the importance of monitoring your flight data and show you some real records coming from flight data recorders (aircraft “black boxes”). We will then explain how the data is recorded, downloaded, analysed, converted to safety events and finally, validated by experts in the field - flight data analysts. Data aggregation across many flights will result with statistical images of safety risks in airlines’ operations. However, this valuable tool can turn into a deadly weapon if used negligently – we’ll support this claim with examples. We are convinced the audience will know about some of these traps, regardless of the industry they are coming from, but hopefully there will be something valuable to take home, too. In the second part, we are saying goodbye to the data analyst and the statistician – the two dominant guys from the first part of the presentation. However, a data scientist will stem from valuable experiences and domain knowledges of the two. This guy will walk the audience through three simple, but working examples. The first one is about how we can improve the accuracy of automated analysis by using historic data and a probabilistic, Bayesian approach. The second example is about finding novel safety risks in airlines’ operations by using simple principal component analysis. Lastly, we’ll use a Markov model to detect aircraft which have changed behaviour with respect to frequency of data downloads, so we collect as many flight data as possible. We will try to make this chat as interesting and as interactive as we can and are looking forward to meeting you at this fun and interesting conference!
This document provides a summary of Letgo's data platform. It discusses how Letgo ingests 500GB of data daily from various event types and processes over 530 million events daily with near real-time processing. It describes Letgo's data journey in moving to an event-driven architecture and storing data in S3. Key components of Letgo's data platform are discussed like stream processing for real-time user segmentation, batch processing for geo data enrichment, querying data using Spark SQL, and data exploitation tools. The document also provides some tips for orchestrating jobs and monitoring metrics.
Learning Accurate Business Process Simulation Models from Event Logs via Auto...Marlon Dumas
Paper presentation at the International Conference on Advanced Information Systems Engineering (CAiSE).
This paper presents an approach to automatically discover business process simulation models from event logs by combining process mining and deep learning techniques.
Paper available at: https://link.springer.com/chapter/10.1007/978-3-031-07472-1_4
Beyond PHP - It's not (just) about the codeWim Godden
Most PHP developers focus on writing code. But creating Web applications is about much more than just wrting PHP. Take a step outside the PHP cocoon and into the big PHP ecosphere to find out how small code changes can make a world of difference on servers and network. This talk is an eye-opener for developers who spend over 80% of their time coding, debugging and testing.
This document discusses how advanced analytics and predictive modeling can help associations achieve strategic goals. It contrasts business intelligence with data science and explains how predictive modeling fits within an association's analytics framework. The document also provides examples of predictive models that associations could use, such as models to predict meeting attendance, purchasing likelihood, or long-term revenue from new members. Finally, it discusses how AFP has used predictive analytics to improve initiatives like customer journeys, onboarding programs, and community platforms.
Creating a Big data Strategy with Tactics for Quick ImplementationLewandog, Inc,
The document provides an overview of creating a big data strategy with quick implementation tactics. It discusses establishing centers of excellence, empowering users to access and analyze data, and transforming organizations to be data-driven. It also notes that only a small percentage of data is currently analyzed and highlights opportunities that additional analysis could provide through increased profits and insights.
Stork Webinar | Digital Transformation AssessmentStork
In dit webinar wordt aan de hand van voorbeelden uit de praktijk uitgelegd hoe een assessment kan helpen jouw positie op ladder van digitalisering binnen asset management te bepalen, om doelstellingen vast te leggen om tot een pragmatische aanpak te komen en hoe de gaten te overbruggen tussen doelstelling en praktijk met acties die reëel, praktisch en geprioriteerd zijn.
Similar to Intelligent Systems for Process Mining (20)
The document discusses challenges with modeling processes that involve multiple interacting objects. Conventional process modeling approaches encourage separating objects and focusing on one object type per process, which can lead to issues when objects interact. The document proposes modeling objects as first-class citizens and capturing relationships between objects to better represent real-world processes where objects corelate and influence each other. It provides examples of how conventional case-centric modeling can struggle to accurately capture a hiring process that involves interacting candidate, application, job offer and other objects.
Slides of our BPM 2022 paper on "Reasoning on Labelled Petri Nets and Their Dynamics in a Stochastic Setting", which received the best paper award at the conference. Paper available here: https://link.springer.com/chapter/10.1007/978-3-031-16103-2_22
Slides of the keynote speech on "Constraints for process framing in Augmented BPM" at the AI4BPM 2022 International Workshop, co-located with BPM 2022. The keynote focuses on the problem of "process framing" in the context of the new vision of "Augmented BPM", where BPM systems are augmented with AI capabilities. This vision is described in a manifesto, available here: https://arxiv.org/abs/2201.12855
Presentation (jointly with Claudio Di Ciccio) on "Declarative Process Mining", as part of the 1st Summer School in Process Mining (http://www.process-mining-summer-school.org). The Presentation summarizes 15 years of research in declarative process mining, covering declarative process modeling, reasoning on declarative process specifications, discovery of process constraints from event logs, conformance checking and monitoring of process constraints at runtime. This is done without ad-hoc algorithms, but relying on well-established techniques at the intersection of formal methods, artificial intelligence, and data science.
1. The document discusses representing business processes with uncertainty using ProbDeclare, an extension of Declare that allows constraints to have uncertain probabilities.
2. ProbDeclare models contain both crisp constraints that must always hold and probabilistic constraints that hold with some probability. This leads to multiple possible "scenarios" depending on which constraints are satisfied.
3. Reasoning involves determining which scenarios are logically consistent using LTLf, and computing the probability distribution over scenarios by solving a system of inequalities defined by the constraint probabilities.
Presentation on "From Case-Isolated to Object-Centric Processes - A Tale of Two Models" as part of the Hasselt University BINF Research Seminar Series (see https://www.uhasselt.be/en/onderzoeksgroepen-en/binf/research-seminar-series).
Invited seminar on "Modeling and Reasoning over Declarative Data-Aware Processes" as part of the KRDB Summer Online Seminars 2020 (https://www.inf.unibz.it/krdb/sos-2020/).
Presentation of the paper "Soundness of Data-Aware Processes with Arithmetic Conditions" at the 34th International Conference on Advanced Information Systems Engineering (CAiSE 2022). Paper available here: https://doi.org/10.1007/978-3-031-07472-1_23
Abstract:
Data-aware processes represent and integrate structural and behavioural constraints in a single model, and are thus increasingly investigated in business process management and information systems engineering. In this spectrum, Data Petri nets (DPNs) have gained increasing popularity thanks to their ability to balance simplicity with expressiveness. The interplay of data and control-flow makes checking the correctness of such models, specifically the well-known property of soundness, crucial and challenging. A major shortcoming of previous approaches for checking soundness of DPNs is that they consider data conditions without arithmetic, an essential feature when dealing with real-world, concrete applications. In this paper, we attack this open problem by providing a foundational and operational framework for assessing soundness of DPNs enriched with arithmetic data conditions. The framework comes with a proof-of-concept implementation that, instead of relying on ad-hoc techniques, employs off-the-shelf established SMT technologies. The implementation is validated on a collection of examples from the literature, and on synthetic variants constructed from such examples.
Presentation of the paper "Probabilistic Trace Alignment" at the 3rd International Conference on Process Mining (ICPM 2021). Paper available here: https://doi.org/10.1109/ICPM53251.2021.9576856
Abstract:
Alignments provide sophisticated diagnostics that pinpoint deviations in a trace with respect to a process model. Alignment-based approaches for conformance checking have so far used crisp process models as a reference. Recent probabilistic conformance checking approaches check the degree of conformance of an event log as a whole with respect to a stochastic process model, without providing alignments. For the first time, we introduce a conformance checking approach based on trace alignments using stochastic Workflow nets. This requires to handle the two possibly contrasting forces of the cost of the alignment on the one hand and the likelihood of the model trace with respect to which the alignment is computed on the other.
Presentation of the paper "Strategy Synthesis for Data-Aware Dynamic Systems with Multiple Actors" at the 7th International Conference on Principles of Knowledge Representation and Reasoning (KR 2020). Paper available here: https://proceedings.kr.org/2020/32/
Abstract: The integrated modeling and analysis of dynamic systems and the data they manipulate has been long advocated, on the one hand, to understand how data and corresponding decisions affect the system execution, and on the other hand to capture how actions occurring in the systems operate over data. KR techniques proved successful in handling a variety of tasks over such integrated models, ranging from verification to online monitoring. In this paper, we consider a simple, yet relevant model for data-aware dynamic systems (DDSs), consisting of a finite-state control structure defining the executability of actions that manipulate a finite set of variables with an infinite domain. On top of this model, we consider a data-aware version of reactive synthesis, where execution strategies are built by guaranteeing the satisfaction of a desired linear temporal property that simultaneously accounts for the system dynamics and data evolution.
Presentation of the paper "Extending Temporal Business Constraints with Uncertainty" at the 18th Int. Conference on Business Process Management (BPM 2020). Paper available here: https://doi.org/10.1007/978-3-030-58666-9_3
Abstract: Temporal business constraints have been extensively adopted to declaratively capture the acceptable courses of execution in a business process. However, traditionally, constraints are interpreted logically in a crisp way: a process execution trace conforms with a constraint model if all the constraints therein are satisfied. This is too restrictive when one wants to capture best practices, constraints involving uncontrollable activities, and exceptional but still conforming behaviors. This calls for the extension of business constraints with uncertainty. In this paper, we tackle this timely and important challenge, relying on recent results on probabilistic temporal logics over finite traces. Specifically, our contribution is threefold. First, we delve into the conceptual meaning of probabilistic constraints and their semantics. Second, we argue that probabilistic constraints can be discovered from event data using existing techniques for declarative process discovery. Third, we study how to monitor probabilistic constraints, where constraints and their combinations may be in multiple monitoring states at the same time, though with different probabilities.
Presentation of the paper "Extending Temporal Business Constraints with Uncertainty" at the CAiSE2020 Forum. The paper is available here: https://link.springer.com/chapter/10.1007/978-3-030-58135-0_8
Abstract: Conformance checking is a fundamental task to detect deviations between the actual and the expected courses of execution of a business process. In this context, temporal business constraints have been extensively adopted to declaratively capture the expected behavior of the process. However, traditionally, these constraints are interpreted logically in a crisp way: a process execution trace conforms with a constraint model if all the constraints therein are satisfied. This is too restrictive when one wants to capture best practices, constraints involving uncontrollable activities, and exceptional but still conforming behaviors. This calls for the extension of business constraints with uncertainty. In this paper, we tackle this timely and important challenge, relying on recent results on probabilistic temporal logics over finite traces. Specifically, we equip business constraints with a natural, probabilistic notion of uncertainty. We discuss the semantic implications of the resulting framework and show how probabilistic conformance checking and constraint entailment can be tackled therein.
Presentation of the paper "Modeling and Reasoning over Declarative Data-Aware Processes with Object-Centric Behavioral Constraints" at the 17th Int. Conference on Business Process Management (BPM 2019). Paper available here: https://link.springer.com/chapter/10.1007/978-3-030-26619-6_11
Abstract
Existing process modeling notations ranging from Petri nets to BPMN have difficulties capturing the data manipulated by processes. Process models often focus on the control flow, lacking an explicit, conceptually well-founded integration with real data models, such as ER diagrams or UML class diagrams. To overcome this limitation, Object-Centric Behavioral Constraints (OCBC) models were recently proposed as a new notation that combines full-fledged data models with control-flow constraints inspired by declarative process modeling notations such as DECLARE and DCR Graphs. We propose a formalization of the OCBC model using temporal description logics. The obtained formalization allows us to lift all reasoning services defined for constraint-based process modeling notations without data, to the much more sophisticated scenario of OCBC. Furthermore, we show how reasoning over OCBC models can be reformulated into decidable, standard reasoning tasks over the corresponding temporal description logic knowledge base.
Keynote speech at the 7th International Workshop on DEClarative, DECision and Hybrid approaches to processes ( DEC2H 2019) In conjunction with BPM 2019.
This is a talk about the combined modeling and reasoning techniques for decisions, background knowledge, and work processes.
The advent of the OMG Decision Model and Notation (DMN) standard has revived interest, both from academia and industry, in decision management and its relationship with business process management. Several techniques and tools for the static analysis of decision models have been brought forward, taking advantage of the trade-off between expressiveness and computational tractability offered by the DMN S-FEEL language.
In this keynote, I argue that decisions have to be put in perspective, that is, understood and analyzed within their surrounding organizational boundaries. This brings new challenges that, in turn, require novel, advanced analysis techniques. Using a simple but illustrative example, I consider in particular two relevant settings: decisions interpreted the presence of background, structural knowledge of the domain of interest, and (data-aware) business processes routing process instances based on decisions. Notably, the latter setting is of particular interest in the context of multi-perspective process mining. I report on how we successfully tackled key analysis tasks in both settings, through a balanced combination of conceptual modeling, formal methods, and knowledge representation and re
Presentation at "Ontology Make Sense", an event in honor of Nicola Guarino, on how to integrate data models with behavioral constraints, an essential problem when modeling multi-case real-life work processes evolving multiple objects at once. I propose to combine UML class diagrams with temporal constraints on finite traces, linked to the data model via co-referencing constraints on classes and associations.
The document discusses representing and querying norm states using temporal ontology-based data access (OBDA). It presents the QUEN framework which models norms and their state transitions declaratively on top of a relational database. QUEN has three layers: 1) an ontological layer representing norms, 2) a specification of norm state transitions in response to database events, and 3) a legacy relational database storing events. It demonstrates QUEN on an example of patient data access consent, modeling authorizations and their lifecycles. Norm state queries are answered directly over the database using the declarative specifications without materializing states.
Presentation ad EDOC 2019 on monitoring multi-perspective business constraints accounting for time and data, with a specific focus on the (unsolvable in general) problem of conflict detection.
1) The document discusses business process management and how conceptual modeling and process mining can help understand and improve digital enterprises.
2) Process mining techniques like process discovery from event logs, decision mining, and social network mining can provide insights into how processes are executed in reality.
3) Replay techniques can enhance process models with timing information and detect deviations to help align actual behaviors with expected behaviors.
Presentation at BPM 2019, focused on a data-aware extension of BPMN encompassing read-write and read-only data, and on SMT-techniques for effectively tackling parameterized verification of the resulting integrated models.
Presentation at BPM 2019, proposing object-centric behavioral constraints as a modeling approach to capture real-life processes with many-to-many and one-to-many relations, and bringing forward a formal semantics and automated reasoning techniques grounded in temporal description logics.
More from Faculty of Computer Science - Free University of Bozen-Bolzano (20)
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...AbdullaAlAsif1
The pygmy halfbeak Dermogenys colletei, is known for its viviparous nature, this presents an intriguing case of relatively low fecundity, raising questions about potential compensatory reproductive strategies employed by this species. Our study delves into the examination of fecundity and the Gonadosomatic Index (GSI) in the Pygmy Halfbeak, D. colletei (Meisner, 2001), an intriguing viviparous fish indigenous to Sarawak, Borneo. We hypothesize that the Pygmy halfbeak, D. colletei, may exhibit unique reproductive adaptations to offset its low fecundity, thus enhancing its survival and fitness. To address this, we conducted a comprehensive study utilizing 28 mature female specimens of D. colletei, carefully measuring fecundity and GSI to shed light on the reproductive adaptations of this species. Our findings reveal that D. colletei indeed exhibits low fecundity, with a mean of 16.76 ± 2.01, and a mean GSI of 12.83 ± 1.27, providing crucial insights into the reproductive mechanisms at play in this species. These results underscore the existence of unique reproductive strategies in D. colletei, enabling its adaptation and persistence in Borneo's diverse aquatic ecosystems, and call for further ecological research to elucidate these mechanisms. This study lends to a better understanding of viviparous fish in Borneo and contributes to the broader field of aquatic ecology, enhancing our knowledge of species adaptations to unique ecological challenges.
The debris of the ‘last major merger’ is dynamically youngSérgio Sacani
The Milky Way’s (MW) inner stellar halo contains an [Fe/H]-rich component with highly eccentric orbits, often referred to as the
‘last major merger.’ Hypotheses for the origin of this component include Gaia-Sausage/Enceladus (GSE), where the progenitor
collided with the MW proto-disc 8–11 Gyr ago, and the Virgo Radial Merger (VRM), where the progenitor collided with the
MW disc within the last 3 Gyr. These two scenarios make different predictions about observable structure in local phase space,
because the morphology of debris depends on how long it has had to phase mix. The recently identified phase-space folds in Gaia
DR3 have positive caustic velocities, making them fundamentally different than the phase-mixed chevrons found in simulations
at late times. Roughly 20 per cent of the stars in the prograde local stellar halo are associated with the observed caustics. Based
on a simple phase-mixing model, the observed number of caustics are consistent with a merger that occurred 1–2 Gyr ago.
We also compare the observed phase-space distribution to FIRE-2 Latte simulations of GSE-like mergers, using a quantitative
measurement of phase mixing (2D causticality). The observed local phase-space distribution best matches the simulated data
1–2 Gyr after collision, and certainly not later than 3 Gyr. This is further evidence that the progenitor of the ‘last major merger’
did not collide with the MW proto-disc at early times, as is thought for the GSE, but instead collided with the MW disc within
the last few Gyr, consistent with the body of work surrounding the VRM.
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxMAGOTI ERNEST
Although Artemia has been known to man for centuries, its use as a food for the culture of larval organisms apparently began only in the 1930s, when several investigators found that it made an excellent food for newly hatched fish larvae (Litvinenko et al., 2023). As aquaculture developed in the 1960s and ‘70s, the use of Artemia also became more widespread, due both to its convenience and to its nutritional value for larval organisms (Arenas-Pardo et al., 2024). The fact that Artemia dormant cysts can be stored for long periods in cans, and then used as an off-the-shelf food requiring only 24 h of incubation makes them the most convenient, least labor-intensive, live food available for aquaculture (Sorgeloos & Roubach, 2021). The nutritional value of Artemia, especially for marine organisms, is not constant, but varies both geographically and temporally. During the last decade, however, both the causes of Artemia nutritional variability and methods to improve poorquality Artemia have been identified (Loufi et al., 2024).
Brine shrimp (Artemia spp.) are used in marine aquaculture worldwide. Annually, more than 2,000 metric tons of dry cysts are used for cultivation of fish, crustacean, and shellfish larva. Brine shrimp are important to aquaculture because newly hatched brine shrimp nauplii (larvae) provide a food source for many fish fry (Mozanzadeh et al., 2021). Culture and harvesting of brine shrimp eggs represents another aspect of the aquaculture industry. Nauplii and metanauplii of Artemia, commonly known as brine shrimp, play a crucial role in aquaculture due to their nutritional value and suitability as live feed for many aquatic species, particularly in larval stages (Sorgeloos & Roubach, 2021).
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...Travis Hills MN
Travis Hills of Minnesota developed a method to convert waste into high-value dry fertilizer, significantly enriching soil quality. By providing farmers with a valuable resource derived from waste, Travis Hills helps enhance farm profitability while promoting environmental stewardship. Travis Hills' sustainable practices lead to cost savings and increased revenue for farmers by improving resource efficiency and reducing waste.
Or: Beyond linear.
Abstract: Equivariant neural networks are neural networks that incorporate symmetries. The nonlinear activation functions in these networks result in interesting nonlinear equivariant maps between simple representations, and motivate the key player of this talk: piecewise linear representation theory.
Disclaimer: No one is perfect, so please mind that there might be mistakes and typos.
dtubbenhauer@gmail.com
Corrected slides: dtubbenhauer.com/talks.html
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
Current Ms word generated power point presentation covers major details about the micronuclei test. It's significance and assays to conduct it. It is used to detect the micronuclei formation inside the cells of nearly every multicellular organism. It's formation takes place during chromosomal sepration at metaphase.
BREEDING METHODS FOR DISEASE RESISTANCE.pptxRASHMI M G
Plant breeding for disease resistance is a strategy to reduce crop losses caused by disease. Plants have an innate immune system that allows them to recognize pathogens and provide resistance. However, breeding for long-lasting resistance often involves combining multiple resistance genes
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
The binding of cosmological structures by massless topological defectsSérgio Sacani
Assuming spherical symmetry and weak field, it is shown that if one solves the Poisson equation or the Einstein field
equations sourced by a topological defect, i.e. a singularity of a very specific form, the result is a localized gravitational
field capable of driving flat rotation (i.e. Keplerian circular orbits at a constant speed for all radii) of test masses on a thin
spherical shell without any underlying mass. Moreover, a large-scale structure which exploits this solution by assembling
concentrically a number of such topological defects can establish a flat stellar or galactic rotation curve, and can also deflect
light in the same manner as an equipotential (isothermal) sphere. Thus, the need for dark matter or modified gravity theory is
mitigated, at least in part.
Cytokines and their role in immune regulation.pptx
Intelligent Systems for Process Mining
1. 09/09/2022 - KES 2022, Verona, Italy
Intelligent systems
for process mining
Marco Montali
Free University of Bozen-Bolzano, Italy
2. What do we do
We develop novel, foundational and applied
techniques
grounded in arti
fi
cial intelligence, logics, and
formal methods,
to create intelligent agents and information
systems that combine
processes and data.
3. How to attack these challenges?
Arti
fi
cial
Intelligence
Knowledge representation
Automated reasoning
Computational logics
Information
Systems
Business process
management
Data management
Decision management
Formal
Methods
In
fi
nite-state systems
Veri
fi
cation
Petri nets
Data
Science
Process mining
Data access
and integration
14. Work process
A set of logically related tasks performed to achieve a de
fi
ned business
outcome for a particular customer or market.
(Davenport, 1992)
A collection of activities that take one or more kinds of input and create an
output that is of value to the customer.
(Hammer & Champy, 1993)
A set of activities performed in coordination in an organizational and technical
environment. These activities jointly realize a business goal.
(Weske, 2011)
15. Why “work process” and not “business process”?
Steady transition from business process management to process
management in the large
• Production processes
• Normative processes
• Manufacturing processes
• Systems with humans in the loop (knowledge-intensive processes)
• Automated systems (autonomous processes with little human
intervention)
Process mining applicable on any dynamic systems with exogenous
events and endogenous activities (“work” to be carried out)
16. The two realities within organisations
managers &
analysts
knowledge
workers
• de
fi
nition of business
goals
• de
fi
nition of operational
processes
• strategic decision making
• resource planning
• …
• day-by-day realization of
goals
• execution of operational
processes
• runtime decision making
• resource allocation
• …
produce
models
generate
data
28. The process mining framework
Original picture by Wil van der Aalst
1.3 Process Mining 9
29. The process mining framework
Original picture by Wil van der Aalst
1.3 Process Mining 9
30. Process mining streams
Replay: Connecting events to mod
elements is essential for process m
event log process model
Play-In
event log
process model
Play-Out
Replay
• extended model
elements is essential for process m
event log process model
Play-In
event log
process model
Play-Out
Replay
• extended model
showing times,
frequencies, etc.
event log process model
Play-In
event log
process model
Play-Out
event log process model
Replay
• extended model
showing times,
frequencies, etc.
• diagnostics
• predictions
• recommendations
Play in
Play out
Replay
31. Event data
5.2 Event Logs 129
Table 5.1 A fragment of some event log: each line corresponds to an event
Case id Event id Properties
Timestamp Activity Resource Cost ...
1 35654423 30-12-2010:11.02 register request Pete 50 ...
35654424 31-12-2010:10.06 examine thoroughly Sue 400 ...
35654425 05-01-2011:15.12 check ticket Mike 100 ...
35654426 06-01-2011:11.18 decide Sara 200 ...
35654427 07-01-2011:14.24 reject request Pete 200 ...
2 35654483 30-12-2010:11.32 register request Mike 50 ...
35654485 30-12-2010:12.12 check ticket Mike 100 ...
35654487 30-12-2010:14.16 examine casually Pete 400 ...
35654488 05-01-2011:11.22 decide Sara 200 ...
35654489 08-01-2011:12.05 pay compensation Ellen 200 ...
3 35654521 30-12-2010:14.32 register request Pete 50 ...
35654522 30-12-2010:15.06 examine casually Mike 400 ...
35654524 30-12-2010:16.34 check ticket Ellen 100 ...
35654525 06-01-2011:09.18 decide Sara 200 ...
32. Event data
5.2 Event Logs 129
Table 5.1 A fragment of some event log: each line corresponds to an event
Case id Event id Properties
Timestamp Activity Resource Cost ...
1 35654423 30-12-2010:11.02 register request Pete 50 ...
35654424 31-12-2010:10.06 examine thoroughly Sue 400 ...
35654425 05-01-2011:15.12 check ticket Mike 100 ...
35654426 06-01-2011:11.18 decide Sara 200 ...
35654427 07-01-2011:14.24 reject request Pete 200 ...
2 35654483 30-12-2010:11.32 register request Mike 50 ...
35654485 30-12-2010:12.12 check ticket Mike 100 ...
35654487 30-12-2010:14.16 examine casually Pete 400 ...
35654488 05-01-2011:11.22 decide Sara 200 ...
35654489 08-01-2011:12.05 pay compensation Ellen 200 ...
3 35654521 30-12-2010:14.32 register request Pete 50 ...
35654522 30-12-2010:15.06 examine casually Mike 400 ...
35654524 30-12-2010:16.34 check ticket Ellen 100 ...
35654525 06-01-2011:09.18 decide Sara 200 ...
130 5 Getting the Data
33. Event data
5.2 Event Logs 129
Table 5.1 A fragment of some event log: each line corresponds to an event
Case id Event id Properties
Timestamp Activity Resource Cost ...
1 35654423 30-12-2010:11.02 register request Pete 50 ...
35654424 31-12-2010:10.06 examine thoroughly Sue 400 ...
35654425 05-01-2011:15.12 check ticket Mike 100 ...
35654426 06-01-2011:11.18 decide Sara 200 ...
35654427 07-01-2011:14.24 reject request Pete 200 ...
2 35654483 30-12-2010:11.32 register request Mike 50 ...
35654485 30-12-2010:12.12 check ticket Mike 100 ...
35654487 30-12-2010:14.16 examine casually Pete 400 ...
35654488 05-01-2011:11.22 decide Sara 200 ...
35654489 08-01-2011:12.05 pay compensation Ellen 200 ...
3 35654521 30-12-2010:14.32 register request Pete 50 ...
35654522 30-12-2010:15.06 examine casually Mike 400 ...
35654524 30-12-2010:16.34 check ticket Ellen 100 ...
35654525 06-01-2011:09.18 decide Sara 200 ...
130 5 Getting the Data
<log xes.version="1.0"
xes.features="nested-attributes"
openxes.version="1.0RC7">
<extension name="Time"
prefix="time"
uri="http://www.xes-standard.org/time.xesext"/>
<classifier name="Event Name" keys="concept:name"/>
<string key="concept:name" value="XES Event Log"/>
...
<trace>
<string key="concept:name" value="1"/>
<event>
<string key="User" value="Pete"/>
<string key="concept:name" value="create paper"/>
<int key="Event ID" value="35654424"/>
...
</event>
<event>
...
<string key="concept:name" value="submit paper"/>
...
</event>
attKey: String
attType: String
Attribute
extName: String
extPrefix: String
extUri: String
Extension
attValue: String
ElementaryAttribute CompositeAttribute
{disjoint}
ca-contains-a
*
*
logFeatures: String
logVersion: String
Log Trace Event
GlobalAttribute
GlobalEventAttribute
GlobalTraceAttribute
EventClassifier
TraceClassifier
name: String
Classifier
a-contains-a
*
*
*
* e-usedBy-a
e-usedBy-l
*
*
l-contains-t t-contains-e
* *
*
1..*
l-contains-e
*
*
* * *
*
*
*
l-has-a
t-has-a
e-has-a
l-has-gea *
1..*
l-contains-ec
1..*
*
1..*
l-contains-tc
*
ec-definedBy-gea
1..*
*
1..*
1..*
* tc-definedBy-gea
l-has-gta
*
{disjoint}
{disjoint}
Fig. 14: A more comprehensive event schema, capturing all main abstractions of the
XES standard
IEEE standard
34. Play in: discovery
register travel
request (a)
get detailed
motivation
letter (c)
get support
from local
manager (b)
check budget
by finance (d)
decide (e)
accept
request (g)
reject
request (h)
reinitiate
request (f)
start end
Case Activity Timestamp Resource
432 register travel request (a) 18-3-2014:9.15 John
432 get support from local manager (b) 18-3-2014:9.25 Mary
432 check budget by finance (d) 19-3-2014:8.55 John
432 decide (e) 19-3-2014:9.36 Sue
432 accept request (g) 19-3-2014:9.48 Mary
…
37. Extended framework information system(s)
current
data
“world”
people
machines
organizations
business
processes documents
historic
data
resources/
organization
data/rules
control-flow
de jure models
resources/
organization
data/rules
control-flow
de facto models
provenance
explore
predict
recommend
detect
check
compare
promote
discover
enhance
diagnose
cartography
navigation auditing
event logs
models
“pre
mortem”
“post
mortem”
From post-mortem
event data to runtime
operational support
38. Large industry uptake
• In the last 15 years: from few attempts to a
fl
ourishing market of 30+
process mining vendors
• Many success stories (e.g., from startups to large companies in 10 years,
large acquisitions of process mining solutions, …)
• Growing market:
see 2021 Gartner’s
Market guide for
process mining
40. Opportunity #1: solidity
In terms of correctness, formal guarantees, performance
Replay: Connecting events to model
elements is essential for process mining
event log process model
ay-In
event log
process model
y-Out
play
• extended model
showing times,
frequencies, etc.
• diagnostics
• predictions
Replay: Connecting events to model
elements is essential for process mining
PAGE 3
event log process model
lay-In
event log
process model
ay-Out
event log process model
eplay
• extended model
showing times,
frequencies, etc.
• diagnostics
• predictions
• recommendations
Replay: Connecting events to model
elements is essential for process mining
PAGE 3
event log process model
Play-In
event log
process model
lay-Out
event log process model
Replay
• extended model
showing times,
frequencies, etc.
• diagnostics
• predictions
• recommendations
SAT
solvers
SMT/OMT
solvers
Planners
Automated
reasoners
Automata
Neural
networks
Constraint
solvers
Neurosymbolic
reasoners
Model
checkers
Encoding
Extension
…
41. On control and flexibility
complexity ->
predictability <-
repetitiveness <-
42. On control and flexibility
complexity ->
predictability <-
repetitiveness <-
Control
degree to which a
central
orchestrator
decides how to
execute the
process
43. On control and flexibility
complexity ->
predictability <-
repetitiveness <-
Control
degree to which a
central
orchestrator
decides how to
execute the
process
Flexibility
degree to which
process
stakeholders
locally decide how
to execute the
process
44. On control and flexibility
complexity ->
predictability <-
repetitiveness <-
Control
degree to which a
central
orchestrator
decides how to
execute the
process
Flexibility
degree to which
process
stakeholders
locally decide how
to execute the
process
45. Which Italian food do you like best?
complexity ->
predictability <-
repetitiveness <-
Control
degree to which a
central
orchestrator
decides how to
execute the
process
Flexibility
degree to which
process
stakeholders
locally decide how
to execute the
process
Lasagna
processes Spaghetti
processes
47. Opportunity #2: reasoning support
Taming spaghetti processes calls for “intelligent”
fi
ltering,
querying, analysis, inference, abstraction, classi
fi
cation, …
• All in all: automated reasoning
Cognitive dimensions of notations (by Green): high
abstraction gradient and presence of hidden
dependencies call for underlying support
• Intelligent systems at the service of process mining
52. The issue of flexibility is widely known
Different ways to address
fl
exibility in processes
We are interested here in
fl
exibility by design
53. Two process modelling paradigms
Imperative models
• Repetitive, predictable processes
• Metaphor:
fl
owchart
speci
fi
cations
• Highly-variable processes
• Metaphor: rules/constraints
Dec t i
lar
a
e
v
55. Imperative modelling
Focus: how things must be done
• Explicit description of the process control-
fl
ow
Closed modelling
• All that is not explicitly modelled is forbidden
• Exceptions/uncommon behaviours: explicitly enumerated
at design time
56. The effect of imperative modelling
Receive
order
Check
availability
Article available?
Ship article
Financial
settlement
yes
Procurement
no
Payment
received
Inform
customer
Late delivery
Undeliverable
Customer
informed
Inform
customer
Article
removed
Remove
article from
catalogue
57. The effect of imperative modelling
Receive
order
Check
availability
Article available?
Ship article
Financial
settlement
yes
Procurement
no
Payment
received
Inform
customer
Late delivery
Undeliverable
Customer
informed
Inform
customer
Article
removed
Remove
article from
catalogue
Input Output
58. The effect of imperative modelling
Receive
order
Check
availability
Article available?
Ship article
Financial
settlement
yes
Procurement
no
Payment
received
Inform
customer
Late delivery
Undeliverable
Customer
informed
Inform
customer
Article
removed
Remove
article from
catalogue
Input Output
Receive
order
Check
availability
Article available?
Ship article
Financial
settlement
yes
Procurement
no
Payment
received
Inform
customer
Late delivery
Undeliverable
Customer
informed
Inform
customer
Article
removed
Remove
article from
catalogue
61. Example
Flexible shopping (without considering data)
• Tasks: = {pick item, close order, pay}
• A customer may pick items, close several orders, and pay one or
more orders at once
• She may even close an empty order and pay for it, or pay multiple
times (whole orders or order parts; a no-op if nothing is left to pay)
• Process constraint: whenever you close an order, you have to
pay later at least once
Σ
62. Quiz: which traces belong to the process?
“Whenever you close an order, you have to pay later at least once”
1.<> (empty trace)
2.<i,i,i>
3.<i,i,i,c,p>
4.<i,i,i,c,p,p>
5.<i,i,i,p,c>
6.<i,c,p,i,i,c,p>
63. Quiz: which traces belong to the process?
“Whenever you close an order, you have to pay later at least once”
1.<> (empty trace)
2.<i,i,i>
3.<i,i,i,c,p>
4.<i,i,i,c,p,p>
5.<i,i,i,p,c>
6.<i,c,p,i,i,c,p>
64. Quiz: How to model the process with BPMN?
“Whenever you close an order, you have to pay later at least once”
First attempt:
1. <> (empty trace)
2. <i,i,i>
3. <i,i,i,c,p>
4. <i,i,i,c,p,p>
5. <i,i,i,p,c>
6. <i,c,p,i,i,c,p>
pick
item
close
order
pay
quit
order paid
65. Quiz: How to model the process with BPMN?
“Whenever you close an order, you have to pay later at least once”
First attempt:
1. <> (empty trace)
2. <i,i,i>
3. <i,i,i,c,p>
4. <i,i,i,c,p,p>
5. <i,i,i,p,c>
6. <i,c,p,i,i,c,p>
pick
item
close
order
pay
quit
order paid
Simple, correct, but not general enough
66. Quiz: How to model the process with BPMN?
“Whenever you close an order, you have to pay later at least once”
Most liberal attempt:
1. <> (empty trace)
2. <i,i,i>
3. <i,i,i,c,p>
4. <i,i,i,c,p,p>
5. <i,i,i,p,c>
6. <i,c,p,i,i,c,p>
pick
item
close
order
pay
pick
item
pay
67. Quiz: How to model the process with BPMN?
“Whenever you close an order, you have to pay later at least once”
Most liberal attempt:
1. <> (empty trace)
2. <i,i,i>
3. <i,i,i,c,p>
4. <i,i,i,c,p,p>
5. <i,i,i,p,c>
6. <i,c,p,i,i,c,p>
Correct, general (captures the rule in a maximal way), but unreadable
pick
item
close
order
pay
pick
item
pay
72. Idea
Focus: what has to be accomplished
• Explicit description of the relevant process constraints
• Control-
fl
ow left implicit
Open modelling
• All behaviors are possible as long as all constraints are
satis
fi
ed
79. Constraint templates
Constraint types de
fi
ned on activity placeholders, each with a speci
fi
c
meaning
• … then instantiated on actual activities (by grounding)
Dimensions
• Activities: how many are involved
• Time: temporal orientation (past, future, either) and strength (when)
• Expectation: negative vs positive
Much richer than the precedence
fl
ow relation of imperative languages
80. Playing with the three dimensions…
Unary templates (over a single activity A)
• Existence: you should/cannot do A (for a certain number of times)
Binary templates (apply to two - or more - tasks)
• Choice: do A (x-)or B
• Relation: if A occurs, then B must also occur (when?)
• Negation: if A then B cannot occur (when?)
Once instantiated, binary constraints may branch to represent potential
alternatives
85. Declare specification
A set of constraints (templates grounded on the
activities of interest)
• Constraints have to be all satis
fi
ed globally over
each, complete trace
• Compositional approach by conjunction
86. Flexible shopper in Declare
“Whenever you close an order, you have to pay later at least once”
1. <> (empty trace)
2. <i,i,i>
3. <i,i,i,c,p>
4. <i,i,i,c,p,p>
5. <i,i,i,p,c>
6. <i,c,p,i,i,c,p>
Pick
item
Close
order
Pay
Accepts all {i, c, p}*
87. Flexible shopper in Declare
“Whenever you close an order, you have to pay later at least once”
1. <> (empty trace)
2. <i,i,i>
3. <i,i,i,c,p>
4. <i,i,i,c,p,p>
5. <i,i,i,p,c>
6. <i,c,p,i,i,c,p>
Pick
item
Close
order
Pay
response
88. Interaction among constraints
Aka hidden dependencies [____,TWEB2010]
Cancel
order
Close
order
Pay
If you cancel the order,
you cannot pay for it
If you close the order,
you must pay for it
Pick
item
To close an order, you
must
fi
rst pick an item
89. Interaction among constraints
Aka hidden dependencies [____,TWEB2010]
Cancel
order
Close
order
Pay
If you cancel the order,
you cannot pay for it
If you close the order,
you must pay for it
Pick
item
To close an order, you
must
fi
rst pick an item
Implied: cannot
cancel and
con
fi
rm!
90. Interaction among constraints
Aka hidden dependencies [____,TWEB2010]
Cancel
order
Close
order
Pay
If you cancel the order,
you cannot pay for it
If you close the order,
you must pay for it
Pick
item
To close an order, you
must
fi
rst pick an item
Implied: cannot
cancel and
con
fi
rm!
Key questions
1. How to
characterise the
language of a
Declare
speci
fi
cation?
2. How to
understand
whether a
speci
fi
cation is
correct?
92. Back to the roots
Patterns in Linear
Temporal Logic
(LTL)
93. LTL: a logic for infinite traces
Logic interpreted over in
fi
nite traces
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
Atomic propositions
Boolean connectives
At next step φ holds
At some point φ2 holds, and φ1 holds until φ2 does
φ eventually holds
φ always holds
φ1 holds forever or until φ2 does
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
::= A | ¬' | '1 ^ '2 | ' | '1U'2
'1 ^ '2 | ' | '1U'2
| ' | '1U'2
⌃' ⌘ true U'
⇤' ⌘ ¬⌃¬'
'1W'2 = '1U'2 _ ⇤'1
⇤' ⌘ ¬⌃¬'
…
Each state indicates which
propositions hold
94. LTL: a logic for infinite traces
Logic interpreted over in
fi
nite traces
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
Atomic propositions
Boolean connectives
At next step φ holds
At some point φ2 holds, and φ1 holds until φ2 does
φ eventually holds
φ always holds
φ1 holds forever or until φ2 does
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
::= A | ¬' | '1 ^ '2 | ' | '1U'2
'1 ^ '2 | ' | '1U'2
| ' | '1U'2
⌃' ⌘ true U'
⇤' ⌘ ¬⌃¬'
'1W'2 = '1U'2 _ ⇤'1
⇤' ⌘ ¬⌃¬'
…
Each state indicates which
propositions hold
Can be
seamlessly
extended
with past-
tense
operators
95. LTL: a logic for infinite traces
Logic interpreted over in
fi
nite traces
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
Atomic propositions
Boolean connectives
At next step φ holds
At some point φ2 holds, and φ1 holds until φ2 does
φ eventually holds
φ always holds
φ1 holds forever or until φ2 does
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
::= A | ¬' | '1 ^ '2 | ' | '1U'2
'1 ^ '2 | ' | '1U'2
| ' | '1U'2
⌃' ⌘ true U'
⇤' ⌘ ¬⌃¬'
'1W'2 = '1U'2 _ ⇤'1
⇤' ⌘ ¬⌃¬'
…
Each state indicates which
propositions hold
Can be
seamlessly
extended
with past-
tense
operators
Trace t satis
fi
es a
formula is now
formally de
fi
ned:
φ
t ⊧ φ
97. Semantics of Declare via LTL
Atomic propositions are activities
Each constraint is an LTL formula (built from template formulae)
…
Each state contains
a single activity
delete
order
close
order
pay
pick
item
98. delete
order
pick
item
Semantics of Declare via LTL
Atomic propositions are activities
Each constraint is an LTL formula (built from template formulae)
…
Each state contains
a single activity
close
order
pay
□ (
𝚌
𝚕
𝚘
𝚜
𝚎
→ ◊
𝚙
𝚊
𝚢
)
99. Semantics of Declare via LTL
Atomic propositions are activities
A Declare speci
fi
cation is the conjunction of its constraint formulae
…
Each state contains
a single activity
delete
order
pick
item
close
order
pay
□ (
𝚌
𝚕
𝚘
𝚜
𝚎
→ ◊
𝚙
𝚊
𝚢
)
∧ □ (
𝚌
𝚕
𝚘
𝚜
𝚎
→ ◊
𝚒
𝚝
𝚎
𝚖
)
∧ □ (
𝚌
𝚊
𝚗
𝚌
𝚎
𝚕
→ ¬ □
𝚙
𝚊
𝚢
)
100. An unconventional use of logics!
From …
Temporal logics for specifying (un)desired properties of
a dynamic system
… to …
Temporal logics for specifying the dynamic system itself
102. Wait a moment…
Close
order
Pay
□ (
𝚌
𝚕
𝚘
𝚜
𝚎
→ ◊
𝚙
𝚊
𝚢
)
Just what we needed!
But recall: each process
instance should complete
in a
fi
nite number of steps!
103. LTLf: LTL over finite traces
[DeGiacomoVardi,IJCAI2013]
LTL interpreted over
fi
nite traces
' ::= A | ¬' | '1 ^ '2 | ' | '1U'2
No successor!
Same syntax of LTL
In LTL, there is always a next moment… in LTLf, the contrary!
φ always holds from current to the last instant
The next step exists and at next step φ holds
(weak next) If the next step exists, then at next step φ holds
last instant in the trace
' | '1 ^ '2 | ' | '1U'2
⇤'
Last ⌘ ¬ true
' ⌘ ¬ ¬'
104. Look the same, but they are not the same
Many researchers: misled by moving from in
fi
nite to
fi
nite traces
In [____,AAAI14], we studied why!
• People typically focus on “patterns”, not on the entire logic
• Many of such patterns in BPM, reasoning about actions, planning, etc. are
“insensitive to in
fi
nity”
105. Quiz: does this specification accept traces?
b c
0..1
1..*
a d
106. Quiz: does this specification accept traces?
b c
0..1
1..*
a d
107. Quiz: does this specification accept traces?
b c
0..1
1..*
a d
108. Quiz: does this specification accept traces?
b c
0..1
1..*
a d
109. Quiz: does this specification accept traces?
b c
0..1
1..*
a d
110. Quiz: does this specification accept traces?
b c
0..1
1..*
a d
112. Quiz: does this specification accept traces?
a b
Only the empty trace <>,
due to
fi
nite-trace
semantics
113. How to do this, systematically?
Declare speci
fi
cation: encoded in LTLf
LTLf: the star-free fragment of regular expressions
Regular expressions: intimately connected to
fi
nite-state
automata
• No exotic automata models over in
fi
nite structures!
• Just the good-old deterministic
fi
nite-state automata
(DFAs) you know from a typical course in programming
languages and compilers
114. From Declare to automata
LTLf NFA
nondeterministic
DFA
deterministic
LTLf2aut determin.
'
nt
lf formulas can be translated into equivalent nfa:
t |= Ï iff t œ L(AÏ)
/ldlf to nfa (exponential)
to dfa (exponential)
often nfa/dfa corresponding to ltlf /ldlf are in fact small!
ompile reasoning into automata based procedures!
Process engine!
[DeGiacomoVardi,IJCAI2013]
[____,TOSEM2022]
115. Vision realised!
LTLf NFA
nondeterministic
DFA
deterministic
LTLf2aut determin.
'
nt
lf formulas can be translated into equivalent nfa:
t |= Ï iff t œ L(AÏ)
/ldlf to nfa (exponential)
to dfa (exponential)
often nfa/dfa corresponding to ltlf /ldlf are in fact small!
ompile reasoning into automata based procedures!
Process engine!
[DeGiacomoVardi,IJCAI2013]
[____,TOSEM2022]
118. Few lines of code
[____,TOSEM2022] 0:8 G. De Giacomo et al.
(tt, ⇧) = true
(↵ , ⇧) = false
( , ⇧) = (h itt, ⇧) ( prop.)
('1 ^ '2, ⇧) = ('1, ⇧) ^ ('2, ⇧)
('1 _ '2, ⇧) = ('1, ⇧) _ ('2, ⇧)
(h i', ⇧) =
⇢
E (') if ⇧ |= ( prop.)
false if ⇧ 6|=
(h ?i', ⇧) = ( , ⇧) ^ (', ⇧)
(h⇢1 + ⇢2i', ⇧) = (h⇢1i', ⇧) _ (h⇢2i', ⇧)
(h⇢1; ⇢2i', ⇧) = (h⇢1ih⇢2i', ⇧)
(h⇢⇤
i', ⇧) = (', ⇧) _ (h⇢iF h⇢⇤i', ⇧)
([ ]', ⇧) =
⇢
E (') if ⇧ |= ( prop.)
true if ⇧ 6|=
([ ?]', ⇧) = (nnf (¬ ), ⇧) _ (', ⇧)
([⇢1 + ⇢2]', ⇧) = ([⇢1]', ⇧) ^ ([⇢2]', ⇧)
([⇢1; ⇢2]', ⇧) = ([⇢1][⇢2]', ⇧)
([⇢⇤
]', ⇧) = (', ⇧) ^ ([⇢]T [⇢⇤]', ⇧)
(F , ⇧) = false
(T , ⇧) = true
Fig. 1: Definition of , where E (') recursively replaces in ' all occurrences of atoms of
the form T and F by .
1: algorithm LDLf 2NFA
2: input LDLf formula '
3: output NFA A(') = (2P
, S, s0, %, Sf )
4: s0 {'} . set the initial state
5: Sf {;} . set final states
6: if ( (', ✏) = true) then . check if initial state is also final
7: Sf Sf [ {s0}
8: S {s0, ;}, % ;
9: while (S or % change) do
10: for (s 2 S) do
11: if (s0
|=
V
( 2s) ( , ⇧) then . add new state and transition
12: S S [ {s0
}
13: % % [ {(s, ⇧, s0
)}
14: if (
V
( 2s0) ( , ✏) = true) then . check if new state is also final
15: Sf Sf [ {s0
}
Fig. 2: NFA construction.
' NFA
119. Few lines of code
[____,TOSEM2022] 0:8 G. De Giacomo et al.
(tt, ⇧) = true
(↵ , ⇧) = false
( , ⇧) = (h itt, ⇧) ( prop.)
('1 ^ '2, ⇧) = ('1, ⇧) ^ ('2, ⇧)
('1 _ '2, ⇧) = ('1, ⇧) _ ('2, ⇧)
(h i', ⇧) =
⇢
E (') if ⇧ |= ( prop.)
false if ⇧ 6|=
(h ?i', ⇧) = ( , ⇧) ^ (', ⇧)
(h⇢1 + ⇢2i', ⇧) = (h⇢1i', ⇧) _ (h⇢2i', ⇧)
(h⇢1; ⇢2i', ⇧) = (h⇢1ih⇢2i', ⇧)
(h⇢⇤
i', ⇧) = (', ⇧) _ (h⇢iF h⇢⇤i', ⇧)
([ ]', ⇧) =
⇢
E (') if ⇧ |= ( prop.)
true if ⇧ 6|=
([ ?]', ⇧) = (nnf (¬ ), ⇧) _ (', ⇧)
([⇢1 + ⇢2]', ⇧) = ([⇢1]', ⇧) ^ ([⇢2]', ⇧)
([⇢1; ⇢2]', ⇧) = ([⇢1][⇢2]', ⇧)
([⇢⇤
]', ⇧) = (', ⇧) ^ ([⇢]T [⇢⇤]', ⇧)
(F , ⇧) = false
(T , ⇧) = true
Fig. 1: Definition of , where E (') recursively replaces in ' all occurrences of atoms of
the form T and F by .
1: algorithm LDLf 2NFA
2: input LDLf formula '
3: output NFA A(') = (2P
, S, s0, %, Sf )
4: s0 {'} . set the initial state
5: Sf {;} . set final states
6: if ( (', ✏) = true) then . check if initial state is also final
7: Sf Sf [ {s0}
8: S {s0, ;}, % ;
9: while (S or % change) do
10: for (s 2 S) do
11: if (s0
|=
V
( 2s) ( , ⇧) then . add new state and transition
12: S S [ {s0
}
13: % % [ {(s, ⇧, s0
)}
14: if (
V
( 2s0) ( , ✏) = true) then . check if new state is also final
15: Sf Sf [ {s0
}
Fig. 2: NFA construction.
' NFA
Automata manipulations
much easier to handle
than in the in
fi
nite case,
with huge performance
improvements
[ZhuEtAl,
IJ
CAI17]
[Westergaard,BPM11]
120. Constraint automata
Template: pre-compiled into a DFA
Constraint: grounds the template DFA on speci
fi
c activities
close
order
delete
order
pay
pick
item
close
order
pay
0 1
2
i
Σ
{i,c}
Σ∖ Σ
c
1
1
{c}
Σ∖
c
{p}
Σ∖
p
0 1
2
p
Σ
{d}
Σ∖
d
{p}
Σ∖
126. From local automata to global automaton
Entire speci
fi
cation: product automaton of all local
automata
• Corresponds to the automaton of the conjunction of
all formulae
• Many optimisations available
Declare speci
fi
cation consistent if and only if its
global automaton is non-empty
127. (Anticipatory) monitoring
t
Evolving trace
Declare speci
fi
cation
Monitor
Continuous feedback
Track a running process execution to check conformance to a reference process
model
• Goal: Detect and report
fi
ne-grained feedback and deviations as early as possible
One of the main operational support tasks
• Complementary to predictive monitoring!
128. (Anticipatory) monitoring
Track a running process execution to check conformance to a reference process
model
• Goal: Detect and report
fi
ne-grained feedback and deviations as early as possible
One of the main operational support tasks
• Complementary to predictive monitoring!
…
…
t
…
…
…
Declare speci
fi
cation
Monitor
Continuous feedback
Evolving trace
129. Fine-grained feedback
Re
fi
ned analysis of the “truth value”
of a constraint, looking into (all)
possible futures
Consider a partial trace t, and a
constraint C...
…
…
…
t
C
satis
fi
ed?
…
…
C
satis
fi
ed?
130. RV-LTL truth values
[BauerEtAl,InfCom2010]
C is permanently satis
fi
ed if t
satis
fi
es C and no matter how t is
extended, C will stay satis
fi
ed
C is currently satis
fi
ed if t satis
fi
es C
but there is a continuation of t that
violates C
…
…
t
…
…
…
t
131. RV-LTL truth values
C is currently violated if t violates C
but there is a continuation that leads
to satisfy C
C is permanently violated if t violates
C and no matter how t is extended, C
will stay violated
t
…
…
…
…
…
[BauerEtAl,InfCom2010]
132. RV-LTL on finite traces
[____,BPM2011] [____,BPM2014] [____,TOSEM2022]
close
order
delete
order
pay
pick
item
close
order
pay
0 1
2
i
Σ
{i,c}
Σ∖ Σ
c
1
1
{c}
Σ∖
c
{p}
Σ∖
p
0 1
2
p
Σ
{d}
Σ∖
d
{p}
Σ∖
Su
ffi
xes of the current trace: each with unbounded,
fi
nite length
• Again standard DFAs -> all formulae of LTLf are monitorable
Each state of the DFA: colored with an RV-LTL value via simple reachability checks
133. RV-LTL on finite traces
[____,BPM2011] [____,BPM2014] [____,TOSEM2022]
Su
ffi
xes of the current trace: each with unbounded,
fi
nite length
• Again standard DFAs -> all formulae of LTLf are monitorable
Each state of the DFA: colored with an RV-LTL value via simple reachability checks
Σ
Σ
Σ
0
[cs]
1
[ps]
2
[pv]
1
[cv]
0
[cs]
0
[cs]
1
[cs]
2
[pv]
c
{i,c}
Σ∖
i
{c}
Σ∖
c
{p}
Σ∖
p
p
{d}
Σ∖
d
{p}
Σ∖
close
order
delete
order
pay
pick
item
close
order
pay
146. LDLf
linear dynamic logic
over
fi
nite traces
[DeGiacomoVardi,
IJ
CAI2013]
MSOL over
fi
nite traces
Regular
expressions
Can we do more?
LTLf
FOL
over
fi
nite traces
Star-free
regular
expressions
Finite-state
automata
147. LDLf
linear dynamic logic
over
fi
nite traces
[DeGiacomoVardi,
IJ
CAI2013]
MSOL over
fi
nite traces
Regular
expressions
Can we do more?
[____,BPM2014] [____,TOSEM2022]
LTLf
FOL
over
fi
nite traces
Star-free
regular
expressions
Finite-state
automata
148. LDLf
linear dynamic logic
over
fi
nite traces
[DeGiacomoVardi,
IJ
CAI2013]
MSOL over
fi
nite traces
Regular
expressions
Can we do more?
[____,BPM2014] [____,TOSEM2022]
LTLf
FOL
over
fi
nite traces
Star-free
regular
expressions
Finite-state
automata
149. LDLf
linear dynamic logic
over
fi
nite traces
[DeGiacomoVardi,
IJ
CAI2013]
MSOL over
fi
nite traces
Regular
expressions
Can we do more?
[____,BPM2014] [____,TOSEM2022]
LTLf
FOL
over
fi
nite traces
Star-free
regular
expressions
Finite-state
automata
150. From constraints to metaconstraints
[____,BPM2014] [____,TOSEM2022]
LDLf expresses RV-LTL monitoring states of LDLf constraints
• Support for metaconstraints predicating over the monitoring status of
other constraints
Example: a form of “contrary-to-duty” process constraint
• If constraint C1 gets permanently violated, eventually satisfy a
compensation constraint C2
Interesting open problem: relationship with normative frameworks and
defeasible reasoning
• [Governatori,EDOC2015] -> LTL cannot express normative notions
• [AlechinaEtAl,FLAP2018]-> not true!
151. Tooling
0:28 G. De Giacomo et al.
Fig. 10: Screenshot of one of the LDL MONITOR clients.
adopted by the IEEE task force on process mining. The response produced by the LDL
MONITOR provider is composed of two parts. The first part contains the temporal infor-
Fully implemented as part
of the RuM toolkit
(rulemining.org)
152. And the story continues…
Probabilistic constraints dealing with quanti
fi
ed
uncertainty for constraints which may or not hold.
Combination of probabilistic and logical reasoning.
Mixed-paradigm speci
fi
cations dealing with the interplay
of declarative and imperative processes in a loose or
tightly integrated fashion.
Data-aware speci
fi
cations dealing with case/event data
and conditisions, and with multiple co-evolving objects.
Undecidability issues, data-abstraction techniques
needed.
see lecture on
process mining
over multiple
behavioral
dimensions
161. Processes are not flat
Package
1
p1
p2 p3
Figure 1: Structure of order, item, and package data objects in an order-to-delivery sce-
nario whereuv items from di↵erent orders are carried in several packages
event log for orders
timestamp overall log order o1 order o2 order o3
2019-09-22 10:00:00 create order o1 create order
2019-09-22 10:01:00 add item i1,1 to order o1 add item
2019-09-23 09:20:00 create order o2 create order
2019-09-23 09:34:00 add item i2,1 to order o2 add item
2019-09-23 11:33:00 create order o3 create order
2019-09-23 11:40:00 add item i3,1 to order o3 add item
2019-09-23 12:27:00 pay order o3 pay order
2019-09-23 12:32:00 add item i1,2 to order o1 add item
2019-09-23 13:03:00 pay order o1 pay order
2019-09-23 14:34:00 load item i1,1 into package p1 load item
2019-09-23 14:45:00 add item i2,2 to order o2 add item
2019-09-23 14:51:00 load item i3,1 into package p1 load item
2019-09-23 15:12:00 add item i2,3 to order o2 add item
2019-09-23 15:41:00 pay order o2 pay order
2019-09-23 16:23:00 load item i2,1 into package p2 load item
2019-09-23 16:29:00 load item i1,2 into package p2 load item
2019-09-23 16:33:00 load item i2,2 into package p2 load item
2019-09-23 17:01:00 send package p1 send package send package
2019-09-24 06:38:00 send package p2 send package send package
2019-09-24 07:33:00 load item i2,3 into package p3 load item
2019-09-24 08:46:00 send package p3 send package
2019-09-24 16:21:00 deliver package p1 deliver package deliver package
2019-09-24 17:32:00 deliver package p2 deliver package deliver package
2019-09-24 18:52:00 deliver package p3 deliver package
2019-09-24 18:57:00 accept delivery p3 accept delivery
2019-09-25 08:30:00 deliver package p1 deliver package deliver package
2019-09-25 08:32:00 accept delivery p1 accept delivery accept delivery
2019-09-25 09:55:00 deliver package p2 deliver package deliver package
2019-09-25 17:11:00 deliver package p2 deliver package deliver package
2019-09-25 17:12:00 accept delivery p2 accept delivery accept delivery
Table 1: Overall log of of an order-to-delivery scenario whose structure is illustrated in
Figure 1, and its flattening using the viewpoint of orders.
162. Processes are not flat
Package
1
p1
p2 p3
Figure 1: Structure of order, item, and package data objects in an order-to-delivery sce-
nario whereuv items from di↵erent orders are carried in several packages
event log for orders
timestamp overall log order o1 order o2 order o3
2019-09-22 10:00:00 create order o1 create order
2019-09-22 10:01:00 add item i1,1 to order o1 add item
2019-09-23 09:20:00 create order o2 create order
2019-09-23 09:34:00 add item i2,1 to order o2 add item
2019-09-23 11:33:00 create order o3 create order
2019-09-23 11:40:00 add item i3,1 to order o3 add item
2019-09-23 12:27:00 pay order o3 pay order
2019-09-23 12:32:00 add item i1,2 to order o1 add item
2019-09-23 13:03:00 pay order o1 pay order
2019-09-23 14:34:00 load item i1,1 into package p1 load item
2019-09-23 14:45:00 add item i2,2 to order o2 add item
2019-09-23 14:51:00 load item i3,1 into package p1 load item
2019-09-23 15:12:00 add item i2,3 to order o2 add item
2019-09-23 15:41:00 pay order o2 pay order
2019-09-23 16:23:00 load item i2,1 into package p2 load item
2019-09-23 16:29:00 load item i1,2 into package p2 load item
2019-09-23 16:33:00 load item i2,2 into package p2 load item
2019-09-23 17:01:00 send package p1 send package send package
2019-09-24 06:38:00 send package p2 send package send package
2019-09-24 07:33:00 load item i2,3 into package p3 load item
2019-09-24 08:46:00 send package p3 send package
2019-09-24 16:21:00 deliver package p1 deliver package deliver package
2019-09-24 17:32:00 deliver package p2 deliver package deliver package
2019-09-24 18:52:00 deliver package p3 deliver package
2019-09-24 18:57:00 accept delivery p3 accept delivery
2019-09-25 08:30:00 deliver package p1 deliver package deliver package
2019-09-25 08:32:00 accept delivery p1 accept delivery accept delivery
2019-09-25 09:55:00 deliver package p2 deliver package deliver package
2019-09-25 17:11:00 deliver package p2 deliver package deliver package
2019-09-25 17:12:00 accept delivery p2 accept delivery accept delivery
Table 1: Overall log of of an order-to-delivery scenario whose structure is illustrated in
Figure 1, and its flattening using the viewpoint of orders.
Item
Package
contains
carried
in
1
*
*
1
i1,1 i1,2 i2,1 i2,2 i2,3 i3,1
p1 p2 p3
Order o1 o2 o3
163. Processes are not flat
Package
1
p1
p2 p3
Figure 1: Structure of order, item, and package data objects in an order-to-delivery sce-
nario whereuv items from di↵erent orders are carried in several packages
event log for orders
timestamp overall log order o1 order o2 order o3
2019-09-22 10:00:00 create order o1 create order
2019-09-22 10:01:00 add item i1,1 to order o1 add item
2019-09-23 09:20:00 create order o2 create order
2019-09-23 09:34:00 add item i2,1 to order o2 add item
2019-09-23 11:33:00 create order o3 create order
2019-09-23 11:40:00 add item i3,1 to order o3 add item
2019-09-23 12:27:00 pay order o3 pay order
2019-09-23 12:32:00 add item i1,2 to order o1 add item
2019-09-23 13:03:00 pay order o1 pay order
2019-09-23 14:34:00 load item i1,1 into package p1 load item
2019-09-23 14:45:00 add item i2,2 to order o2 add item
2019-09-23 14:51:00 load item i3,1 into package p1 load item
2019-09-23 15:12:00 add item i2,3 to order o2 add item
2019-09-23 15:41:00 pay order o2 pay order
2019-09-23 16:23:00 load item i2,1 into package p2 load item
2019-09-23 16:29:00 load item i1,2 into package p2 load item
2019-09-23 16:33:00 load item i2,2 into package p2 load item
2019-09-23 17:01:00 send package p1 send package send package
2019-09-24 06:38:00 send package p2 send package send package
2019-09-24 07:33:00 load item i2,3 into package p3 load item
2019-09-24 08:46:00 send package p3 send package
2019-09-24 16:21:00 deliver package p1 deliver package deliver package
2019-09-24 17:32:00 deliver package p2 deliver package deliver package
2019-09-24 18:52:00 deliver package p3 deliver package
2019-09-24 18:57:00 accept delivery p3 accept delivery
2019-09-25 08:30:00 deliver package p1 deliver package deliver package
2019-09-25 08:32:00 accept delivery p1 accept delivery accept delivery
2019-09-25 09:55:00 deliver package p2 deliver package deliver package
2019-09-25 17:11:00 deliver package p2 deliver package deliver package
2019-09-25 17:12:00 accept delivery p2 accept delivery accept delivery
Table 1: Overall log of of an order-to-delivery scenario whose structure is illustrated in
Figure 1, and its flattening using the viewpoint of orders.
Item
Package
contains
carried
in
1
*
*
1
i1,1 i1,2 i2,1 i2,2 i2,3 i3,1
p1 p2 p3
Order o1 o2 o3
164. Processes are not flat
Package
1
p1
p2 p3
Figure 1: Structure of order, item, and package data objects in an order-to-delivery sce-
nario whereuv items from di↵erent orders are carried in several packages
event log for orders
timestamp overall log order o1 order o2 order o3
2019-09-22 10:00:00 create order o1 create order
2019-09-22 10:01:00 add item i1,1 to order o1 add item
2019-09-23 09:20:00 create order o2 create order
2019-09-23 09:34:00 add item i2,1 to order o2 add item
2019-09-23 11:33:00 create order o3 create order
2019-09-23 11:40:00 add item i3,1 to order o3 add item
2019-09-23 12:27:00 pay order o3 pay order
2019-09-23 12:32:00 add item i1,2 to order o1 add item
2019-09-23 13:03:00 pay order o1 pay order
2019-09-23 14:34:00 load item i1,1 into package p1 load item
2019-09-23 14:45:00 add item i2,2 to order o2 add item
2019-09-23 14:51:00 load item i3,1 into package p1 load item
2019-09-23 15:12:00 add item i2,3 to order o2 add item
2019-09-23 15:41:00 pay order o2 pay order
2019-09-23 16:23:00 load item i2,1 into package p2 load item
2019-09-23 16:29:00 load item i1,2 into package p2 load item
2019-09-23 16:33:00 load item i2,2 into package p2 load item
2019-09-23 17:01:00 send package p1 send package send package
2019-09-24 06:38:00 send package p2 send package send package
2019-09-24 07:33:00 load item i2,3 into package p3 load item
2019-09-24 08:46:00 send package p3 send package
2019-09-24 16:21:00 deliver package p1 deliver package deliver package
2019-09-24 17:32:00 deliver package p2 deliver package deliver package
2019-09-24 18:52:00 deliver package p3 deliver package
2019-09-24 18:57:00 accept delivery p3 accept delivery
2019-09-25 08:30:00 deliver package p1 deliver package deliver package
2019-09-25 08:32:00 accept delivery p1 accept delivery accept delivery
2019-09-25 09:55:00 deliver package p2 deliver package deliver package
2019-09-25 17:11:00 deliver package p2 deliver package deliver package
2019-09-25 17:12:00 accept delivery p2 accept delivery accept delivery
Table 1: Overall log of of an order-to-delivery scenario whose structure is illustrated in
Figure 1, and its flattening using the viewpoint of orders.
Item
Package
contains
carried
in
1
*
*
1
i1,1 i1,2 i2,1 i2,2 i2,3 i3,1
p1 p2 p3
Order o1 o2 o3
create order
(3)
add item
(6)
pay order
(3)
load item
(6)
send package
(5)
accept delivery
(5)
deliver package
(11)
3
3
3
3
3
4 1
1
2
3
2
5
6
3
Intelligent systems
needed in the
whole process
mining pipelines
165. 3D processes require 3D models
Object-centric constraints [____,DL17] [____,BPM19]
Constraints co-referring via data model: co-evolution of multiple objects
• From propositional to
fi
rst-order temporal logics
• Interesting balance between expressiveness and decidability
• Discoverable from data (ongoing work)
Item
Order Package
create
order
add
item
get
package
0..1 *
* 0..1
includes is in
166. Conclusions
Process mining: a fascinating
fi
eld at the intersection of data
science and process management
Intelligent systems required for solid, e
ff
ective process
mining
•No ad-hoc algorithms, but careful usage and further
development of well-established formalisms and techniques
AI and formal methods to the rescue for anticipatory
monitoring
170. Subsumption and incompatibility
Chain response(a,b): whenever a is executed, the next executed activity is b
Alternate response(a,b): whenever a is executed, a cannot be executed again
until b is executed
Response(a,b): whenever a is executed, b must be later executed
Negation succession(a,b): whenever a is executed, b cannot be later executed
171. Subsumption and incompatibility
Chain response(a,b): whenever a is executed, the next executed activity is b
Alternate response(a,b): whenever a is executed, a cannot be executed again
until b is executed
Response(a,b): whenever a is executed, b must be later executed
Negation succession(a,b): whenever a is executed, b cannot be later executed
Subsumes (is stronger than)
Subsumes (is stronger than)
Incompatible
175. Declarative process discovery
Simply stated…
Process discovery aiming at extracting
a declarative speci
fi
cation from a log
In our case: Declare
176. Declarative process discovery
Two settings
Discriminative mining
Speci
fi
cation mining
Event
log
Event
log
Partition
Discover
Discover
Speci
fi
cation that
covers all green traces
and rejects all red ones
Speci
fi
cation that
covers well all traces in
the log
179. General idea
1.De
fi
ne suitable metrics to capture the meaning of covering
well -> interesting satisfaction
• Starting point: support and con
fi
dence from data mining
• Issues: not enough, not easy to import (see next slides)
2.Approach discovery by combining:
• Metrics for interestingness
• Temporal reasoning for logical correctness and for
computing the inputs of metrics
180. Naive support is not enough
Constraint support (naive) =
Issue: consider response(a,b)
• <a,c,b,a>
• <c,d,e,c,d,e,f>
• <b,c,d,e>
• <a,b>
• <a,a,b,a,b,a,b,a,a,a,b>
# traces that satisfy the constraint
total # traces in the log
181. Naive support is not enough
Constraint support (naive) =
Issue: consider response(a,b)
• <a,c,b,a>
• <c,d,e,c,d,e,f>
• <b,c,d,e>
• <a,b>
• <a,a,b,a,b,a,b,a,a,a,b>
Support: 4/5 (not informative)
# traces that satisfy the constraint
total # traces in the log
Need to:
1. Account for vacuous
satisfaction
2. Distinguish satisfying traces
based on interestingness
3. De
fi
ne event-based measures
182. Constraint activation and target
Activation: activity/moment that triggers an expectation in the constraint
Target: LTLf formula for the expectation
Systematically extracted from a normal form of constraints
AtLeastOne(a) Start Do a sooner or later
Response(a,b) a Do b afterwards
Precedence(a,b) b Did a before
NegationResponse(a,b) a Don’t do b afterwards
Template Activation Target
183. Trace-based support, refined
Constraint support (trace-based) =
Response(a,b)
• <a,c,b,a>
• <c,d,e,c,d,e,f>
• <b,c,d,e>
• <a,c,d,b>
• <a,a,c,b,a,b,a,d,b,a,a,c,a,b>
Support: from 4/5 to 2/5 (informative, but not re
fl
ecting what happens within a trace)
# traces that satisfy the constraint and activate it
total # traces in the log
184. Trace-based support, refined
Constraint support (trace-based) =
Response(a,b)
• <a,c,b,a>
• <c,d,e,c,d,e,f>
• <b,c,d,e>
• <a,c,d,b>
• <a,a,c,b,a,b,a,d,b,a,a,c,a,b>
Support: from 4/5 to 2/5 (informative, but not re
fl
ecting what happens within a trace)
# traces that satisfy the constraint and activate it
total # traces in the log
185. Event-based support
Constraint support (event-based) =
Response(a,b)
• <a,c,b,a>
• <c,d,e,c,d,e,f>
• <b,c,d,e>
• <a,c,d,b>
• <a,a,c,b,a,b,a,d,b,a,a,c,a,b>
Support: 9/33 (informative, but not re
fl
ecting trace satisfaction/violation)
# events that satisfy the constraint activation and its target
total # events in the log
186. Event-based support
Constraint support (event-based) =
Response(a,b)
• <a,c,b,a>
• <c,d,e,c,d,e,f>
• <b,c,d,e>
• <a,c,d,b>
• <a,a,c,b,a,b,a,d,b,a,a,c,a,b>
Support: 9/33 (informative, but not re
fl
ecting trace satisfaction/violation)
# events that satisfy the constraint activation and its target
total # events in the log
187. Trace- and event-based confidence
Activation and target: solid basis to import the notion of con
fi
dence from
association rule mining
Constraint con
fi
dence (trace-based) =
Constraint con
fi
dence (event-based) =
# traces that satisfy the constraint and activate it
# traces that activate the constraint
# events that satisfy the constraint activation and its target
# events that satisfy the constraint activation
188. Finally, discovery
1. Select templates of interest
2. Compute metrics for corresponding constraints (grounded on log activities)
3. Filter based on minimum thresholds
4. Redundant constraints?
• Keep the most liberal if metrics are better for it
• Keep the most restrictive in case of equal metrics
5. Incompatible constraints?
• Keep only the one with better metrics
6. Further processing to ensure consistency, minimality, …