This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Survey of Process-Oriented Data Science and Analytics for supporting Business Process Management

Asjad Khan University of WollongongNorthfields AveWollongongNSWAustralia Aditya Ghose University of WollongongNorthfields AveWollongongNSWAustralia Hoa Dam University of WollongongNorthfields AveWollongongNSWAustralia [email protected]  and  Arsal Syed University of Nevada1664 N Virginia St, RenoLas VegasNevadaU.S.A [email protected]
(2018)
Abstract.

Process analytics approaches allow organizations to support the practice of Business Process Management and continuous improvement by leveraging all process-related data to extract knowledge, improve process performance and support decision making across the organization. Process execution data once collected will contain hidden insights and actionable knowledge that are of considerable business value enabling firms to take a data-driven approach for identifying performance bottlenecks, reducing costs, extracting insights and optimizing the utilization of available resources. Understanding the properties of ‘current deployed process’ (whose execution trace is often available in these logs), is critical to understanding the variation across the process instances, root-causes of inefficiencies and determining the areas for investing improvement efforts. In this survey we discuss various methods that allows organizations to understand the behaviour of their processes, monitor currently running process instances, predict the future behavior of those instances and provide better support for operational decision-making across the organization.

Process mining, Business Process Monitoring, Process Predictive analytics, process automation, decision-support
copyright: acmcopyrightjournalyear: 2018doi: 10.1145/1122445.1122456

1. Introduction

Business processes are at the heart of every organization in the emerging knowledge economy. Process execution data contains hidden insights and actionable knowledge of considerable business value. Process analytic approaches allow organizations to support the practice of Business Process Management and continuous improvement by leveraging all process-related data to identify performance bottlenecks, reducing costs, extracting insights and optimizing the utilization of available resources. Over the past two decades, the process mining research community has primarily focused on investigating problems such as (automated) process discovery, process conformance checking and process enhancement. Within the broad umbrella of process mining, several methods have been proposed to provide support during the process redesign and process analysis/diagnosis phases of the BPM life-cycle. e.g constructing simulation models from process logs, performing organizational or social-network mining, case outcome predictions and so on (Van Der Aalst et al., 2011). These methods allow organizations to leverage process execution logs to answer questions like; Does the process behave as expected? or Are there any bottlenecks that negatively impact process performance? and Do the logged instances conform to applicable laws and regulations? (Dumas et al., 2013a). While process mining topics like discovery, conformance checking, and enhancement continue to be active areas of research, several newer methods have grown in prominence that leverage a much more diverse range of data sources and offer a much more sophisticated set of capabilities.

Process analytics allows organizations to support the practice of Business Process Management by leveraging all process-related data to extract knowledge, improve process performance and support managerial-decision making across the organization(Dumas, 2018). Process analytics involves a sophisticated layer of data analytics built over the traditional notion of process mining. While process mining addresses the problem of reverse engineering process designs from process execution data (process logs), process analytics extends the scope and addresses the more general problem of leveraging data generated by, or associated with, process execution to obtain actionable insights about business processes. Process analytics leverages a range of data, including, but not limited to process logs, event logs, provisioning logs, decision logs and process context. Process analytics can be used to obtain predictive insights (how will a given process work out?), diagnostic insights (why did a process generate a given outcome?) as well as prescriptive insights (what should be done next in a process?). Process data can for example be used to prescribe what resources should be allocated to process tasks, or predict whether a process will achieve its goals.

Process analytics enables firms to take a data-driven approach for identifying performance bottlenecks, reducing costs, extracting insights and optimizing the utilization of available resources.It encompasses methods, tools and techniques that allow firms to understand the behaviour of their processes, monitor currently running process instances, predict the future behaviour of those instances and provide better support for operational decision-making across the organization (Benatallah et al., 2016). Process analytics can also be viewed as an organizational capability that enables firms to stay competitive by better understanding their business processes and identifying areas of improvement. Some examples of these capabilities include:

  • Mining the logs of historical (completed) traces to better understand the process behaviour, retrospectively.

  • methods to analyse processes through data to gain insights about performance bottlenecks and identifying the root causes of undesired process behaviour/outcomes to support evidence-based process analysis and improvement

  • Understanding resource behaviour and optimizing the utilization of available resources for future instances.

  • Supporting operational decision-making across the organization during process execution by for example, making process behaviour forecasts for running cases.

  • Providing decision support to knowledge workers involved in the execution of knowledge-intensive, unstructured or semi-structured processes.

  • Supporting the practice of risk management by assisting in early identification and possible mitigation of undesired effects.

In this work, we aim to cover process analytic techniques that have received relatively less attention from the research community and therefore offer a fertile ground for exploration and contributions. We can broadly classify them into three major themes: Mining Process Behaviour, Predictive Process Monitoring, and Process Decision Support. In mining process behaviour, we focus on techniques that can help us analyze historical (completed) traces to better understand the process behaviour, retrospectively (Van Der Aalst et al., 2011). Process Mining methods are used for post-mortem analysis of completed business processes to assist with process improvement and redesign efforts. They can help in optimizing process performance by mining process behaviour, identifying bottlenecks (or various sources of inefficiencies and frequent defects), and understanding their root causes. Predictive Process Monitoring methods on the other hand aim at analyzing execution data (e.g. event logs) of a business process at run-time to forecast the future state of the executions of a business process (Teinemaa et al., 2019). These methods are an effective decision support tool for continuously monitoring performance of process instances and reducing the overall risk associated with negative outcomes. Lastly, Assisting organizations in operational decision support remains one of the major goals of process analytics and process management systems. We discuss the topic of decision support, specifically in the context of: (i) supporting knowledge-intensive processes by recommending the best alternative given a decision-making point during process execution. (ii) providing intelligent assistance for resource allocation and (iii) reducing risks associated with operational and strategic decisions for processes operating in dynamic environments. Overall, We see these capabilities and themes in combination represent an important area for further exploration and research.

The rest of the paper is structured as follows: Section 2 reviews basic concepts of BPM, and discusses the relationship between Process Analytics and Business Process Management. Section 3 describes the search protocol used for conducting the systematic literature review. Section 4 discusses methods for understanding process behaviour using logs of historical (completed) traces. Section 5 surveys the selected studies related to predictive process monitoring methods and provides a taxonomy to classify them. In Section 6, we investigate the theme of decision support in the context of structured, unstructured and semi-structured knowledge intensive processes.

2. PRELIMINARY CONCEPTS

Business Process Management:

Business Process Management (BPM) follows a paradigm of “process thinking” where we take a process-centered approach towards understanding and improving business processes (Vom Brocke and Rosemann, 2010). Business processes are key instruments in organizing activities whose outcome is a product or a service (Di Ciccio et al., 2015) Processes are described by ‘Events, activities and decisions involving multiple actors and resources, that collectively lead to an outcome that is of value to an organisation or its customers’ (Dumas et al., 2013a). BPM is defined as:”a body of principles, methods and tools to design, analyse, execute and monitor business processes, with the aim of improving their performance.” Effective management of enterprise-wide processes enables organizations to control the process outcomes and focus their efforts on identifying and improving areas of high impact. BPM is concerned with management, compliance, and the redesign of business processes. It aims at bringing predictability, transparency and consistency to business operations (Dumas et al., 2013a). BPM practice is supported by various modeling tools and analysis methods that have been proposed over the years. Such methods and tools enable continuous performance improvement by optimally allocating scarce resources and tracing performance bottlenecks as possible examples.

Process Management Systems: Formerly known as workflow management systems, Process Management Systems (PMS) are designed to facilitate the operations of an organization by supporting all the phases in the business process life cycle (e.g via managing the process routing and allocation of resources) (Di Ciccio et al., 2015). In doing so, a PMS also records event data representing the execution trace of various deployed processes. Currently, there exist various types of Business Process Management systems that can be distinguished based on degree of support’ they provide and ‘orientation on process data’ (Dumas et al., 2013a). Examples include groupware systems, ad-hoc workflow systems, production workflow systems and case handling systems (Ouyang et al., 2009).

Process Data

Processes deployed inside and across an enterprise, when executed can leave an operational data footprint in the form of an event logs, which if analyzed can be a valuable source of insights to support the management and improvement of business processes. This Process execution data can transformed into an event log describing the sequence of activities that were performed, along with the resources involved in the execution. Each entry of the process execution log(also known as event logs) represents an event, which records the occurrence of an activity at a particular point in time and belongs to precisely one case (case represents a unique process instance). Events can be characterised by multiple descriptors (attributes), including an event (or activity name), a unique case identifier (e.g case ID), a timestamp and optionally details of resource responsible for executing the task of the process instance. An event refers to an activity (or a step) in the process and belongs to a process instance or a case (Dumas et al., 2013a)(Weske, 2019). The sequence of ordered events within a case form a trace. We use the following definition to define an event:

Definition (Van der Aalst, 2016): An event ee is tuple e=(c,a,t,r)e=(c,a,t,r)\in 𝒰case ×𝒰act ×𝒰time ×𝒰res \mathcal{U}_{\text{case }}\times\mathcal{U}_{\text{act }}\times\mathcal{U}_{\text{time }}\times\mathcal{U}_{\text{res }} referring to case cc, activity a, timestamp tt, and resource rr of event e. An event logL\log L is a multiset of events, i.e., L(𝒰case ×𝒰act ×𝒰time ×𝒰res )L\in\mathcal{B}\left(\mathcal{U}_{\text{case }}\times\mathcal{U}_{\text{act }}\times\mathcal{U}_{\text{time }}\times\mathcal{U}_{\text{res }}\right).

Apart from event logs (which record sequential events representing activities in a case instance) organizations can leverage a diverse set of heterogeneous data sources to gain a holistic view of all process-related aspects. Various sources of structured and unstructured process-related execution data are available but often times are scattered across several systems. Examples include:

  • Enterprise Resource Planning Systems

  • Contextual data e.g from real-time IOT sensors

  • Web data. e.g Social Media Sentiment

  • Provisioning Logs and decision logs

  • Unstructured and Semi-Structured data

BPM lifecycle:

Heiskanen et al. (Heiskanen and Newman, 1997) have described the BPM lifecycle as ”The entry point to the cycle is the design and analysis phase, where the business processes are identified and provided with a formal representation. Newly created models and models from past iterations are verified and validated against current process requirements. In the configuration phase, the systems to use are selected, and the business processes identified before are implemented, tested, and deployed. During the enactment phase, the processes are operated, and the process execution is monitored and maintained. The resulting execution data is processed by the techniques of the evaluation phase, for example process mining. Using the knowledge gained from one iteration, the next iteration can be started by redesigning the business processes.” We briefly describe the steps below:

  • Process Identification/(Re)design: In process identification, after requirement analysis, process models are designed using a suitable modeling language. Existing models can be further refined/improved based on insights gathered in the previous cycle.

  • Process Analysis/diagnosis: In this step, process logs are typically analysed to diagnose problems and identify areas of potential improvement. Process Discovery methods are also useful here, allowing the analyst to reverse engineer the models from recorded process execution logs (Di Ciccio et al., 2012).

  • Process Implementation: In Process implementation (also known as process enactment), a process management engine is sometimes used to support the process enactment by instantiating a process instance where tasks are assigned to the relevant resources for execution. A PMS can also manage process routing(control-flow) by considering which tasks are enabled for execution (Benatallah et al., 2016).

  • Process Monitoring and Controlling: Organizational process are monitored by the various process support and management tools while executed tasks are tracked and recorded to generate execution traces for later analysis.

Business Process Paradigms:

It is useful to characterise various types of processes based on the degree of structuring and predictability they exhibit. Process paradigms describe work activities that a process management (or process-support) system can handle (Benatallah et al., 2016). Traditional process management systems have provided support for modeling, monitoring and management of structured processes. However, business processes often contain reusable patterns with elements of unpredictable nature, requiring a certain degree of flexibility. Ciccio et al. (Di Ciccio et al., 2015) have described the spectrum of process management as follows:

  • Structured processes: Structured processes represent predictable routine work (e.g administrative processes). They have a predefined schema (defined apriori in a process model) where process logic and ordering of activities, their dependencies and allocation of resources is known in advance.

  • Structured processes with ad-hoc exceptions: They allow a certain degree of flexibility as unanticipated exceptions can cause deviations during the execution of a process instance. Some of these deviations can also be anticipated in advance and incorporated into the process design via exception handlers.

  • Unstructured processes with predefined segments: Here structured process fragments can be pre-defined on a per-case basis(based on policies and regulations) but overall process logic cannot be specified in advance.

  • Loosely structured processes: Loosely structured processes require adaptation strategies, as the ordering of activities is hard to anticipate in advance, making the overall process structure less rigid. Here constraints (described by policies and business rules) are defined ahead of time to prohibit undesirable behaviour.

  • Unstructured processes: Unstructured processes are often knowledge-centric (dependent on human judgement and expertise) and represent complex, non-routine business process. They are collaborative in nature, driven by rules and dynamic events where the structure of a process evolves based on the operational context.

Readers can refer to (Kemsley, 2011) (Harrison-Broninski, 2018) (Rosenfeld, 2011) for a more detailed discussion on the process spectrum.

3. Research Methodology

We identified and classified the most relevant process analytics studies by conducting a systematic literature review (SLR) according to the scientifically rigorous guidelines described in (Kitchenham, 2004). First, we formulated a list of research questions representing our research goals. Next, guided by these questions, we describe the relevant search strings used for querying a database of academic papers. We then applied inclusion/exclusion criteria to the retrieved studies and filtered out the irrelevant ones. Finally we divided all relevant studies into primary and subsumed ones based on their contribution.

3.1. Research Questions

Our paper aims to conduct a systematic literature review by analysing research studies related to process analytics. Our goal is to understand the recent developments in the field of process analytics by examining the following research questions:

  1. RQ1 What is the body of recent and relevant academic publications within the field of business process analytics?

  2. RQ2 What family of methods exists for diagnostic analytics that lets us mine relevant process data in a retrospective (post-mortem) fashion?

  3. RQ3 What aspects of business processes can be predicted?

  4. RQ4 What are some of the state-of-art methods for predictive monitoring tasks?

  5. RQ5 What role can prescriptive process analytic methods play in automating or supporting decision making in the context of business process execution?

  6. RQ6 How do we characterise these findings in a taxonomy?

  7. RQ7 What should be the future research focus of BPM and process analytics?

RQ1 is the core question that identities existing methods to support the practice of business process management. It allows us to identify a set of classification criteria. Given the richness of the process analytics literature, this overarching question is then decomposed into subsequent research questions to have a well-delimited and manageable scope. RQ2 aims to identify methods that attempt to understand the process behaviour using offline data in order to diagnose possible performance problems and identify areas of potential improvements. Given the vast number of publications in the broader field of process mining, we focus on contemporary topics that have not been examined in great detail so far. RQ3 and RQ4 investigate aspects of business processes that can be predicted by means of machine learning techniques (also known as business process monitoring techniques). RQ5 explores prescriptive methods with the potential of providing decision support both at strategic and operational level.

We categorize our findings based on input data required, type of algorithm employed, validation method, and tool support. (Note to self: The strategy for taxonomy isnt finalized yet).

3.2. Study Retrieval

Keeping the research goals in mind, and following the guidance given in (Kitchenham, 2004) we define the search string. We drive a set of relevant keywords from our subject matter expertise:

  • “business process” — generic term for retrieving most process analytics papers.

  • “mining” - Here a relevant study must take as input a event dataset and proposes a technique for analysing, extracting actionable knowledge.

  • “diagnostic” — diagnostic methods perform post-mortem analysis of process data in an attempt to understand what happened in the past and to analyse the root causes of performance issues.

  • “monitoring“ — monitoring methods are concerned with run-time predictions of process outcomes.

  • “prediction” — a relevant study that estimates or predicts various aspects/properties of a business process.

  • “prescriptive” - studies that are concerned with action recommendations

  • “decision-support” - Methods that provide decision support for achieving process goals.

  • “recommendations” - Methods that provide recommendations for decision support.

Based on these selected keywords and criteria, we constructed the following search phrases:

  • “process mining AND performance improvement”

  • “process mining AND resource management”

  • “business process prediction”

  • “predictive business process monitoring”

  • “prescriptive process analytics AND decision Support”

  • “prescriptive process analytics”

  • “process recommendations AND decision-support”

  • “decision support AND process analytics”

  • “decision support AND knowledge-intensive processes”

3.3. Study Selection

We then picked the Google Scholar database (Gusenbauer, 2019) as our primary search engine for retrieving relevant studies based on our final search strings. We double-checked the retrieved studies with six other literature search sources, including SpringerLink, Scopus, IEEE Xplore, ScienceDirect and ACM Digital Library. These major database electronic literature databases cover the scientific publications within the field of computer science. After search was conducted(in August 2020), it returned more than 5000 papers.

Exclusion criteria: Non-English and Duplicate studies (appearing in multiple databases) were removed. Short Papers with page length 6 and workshop papers were also excluded. Studies which propose method whose input is not process data are also excluded. Lastly, studies where the main contribution of the paper is a case study or a tool implementation instead of novel model with proper evaluation are excluded.

Inclusion criteria: Based on the meta-data (titles, abstract) of remaining papers, we then filtered studies that appeared out of scope based on inclusion criteria(as suggested by (Okoli, 2015) (Fink, 2019)).

IN1: We picked the study only if: (i) it is concerned with analysis and mining process data. (ii) proposes techniques for predictive monitoring of processes. (iii) investigates methods related to prescriptive analytics and decision support in the context of structured, unstructured and knowledge-intensive business processes.

IN2: Study is published in 2011 and later were all included, even if they had fewer than 5 citations. However, for papers published before 2011 we used the snowballing technique.

IN3: The study clearly defined research context, goals and proposes a novel method that has been properly evaluated with sound experiments.

IN4: The study is peer-reviewed and published in a reputable venue.

For final inclusion, a given study must meet the above inclusion criteria. We applied the inclusion criteria IN2 by configuring the search engine’s filter settings. For applying the IN1 and IN3 criterion, was assessed by reading title, abstract and skimming the relevant sections of the paper. After filtering the search results by applying inclusion and exclusion criteria, we identified primary and subsumed studies. Primary studies constitute an original contribution to the field and subsumed are ones that do not substantially improvement with respect to the original contribution. The application of the exclusion criteria resulted in 734 relevant studies out of 1319 works selected in the previous step.

3.4. Taxonomy

We categorize the selected works using different dimensions specifying the typology of the existing methods and their characteristics.

Refer to caption
Figure 1. Publications dates of included literature

In particular, each study can be decomposed and organized along the following dimensions:
- Input data
- Outcome
- Process perspective (control flow, resources, data)
- Family of algorithms(the main algorithm used in the study)
- Evaluation data (real-life or artificial logs) and application domain (e.g., insurance, banking, healthcare)
- Implementation (standalone or plug-in, and tool accessibility)

Lastly, we also observed that currently, many of the same mining problems are being studied under different names, and proposed methods are being evaluated in an ad-hoc manner with no common experimental setups or evaluation measures.

Refer to caption
Figure 2. Proposed Taxonomy

4. Mining Process Behaviour

In a given enterprise, business processes, when executed often leave an operational data footprint in the form of logs, documentation, and various data artifacts containing insights and knowledge that are of considerable business value. Process mining techniques can leverage this execution data to mine actionable knowledge, discover insights about performance bottlenecks, frequent defects, their root causes and other sources of inefficiencies. This allows organizations to gain behaviour visibility in order to optimize process performance and support both operational and managerial decision-making (Benatallah et al., 2016) (Dumas, 2018). Enterprises adopt process mining tools that typically support business process improvement by offering techniques for process automation, auditing and compliance checking, and recently digital transformation (Kerremans, 2018).

Traditional process mining techniques cover various process-related perspectives. e.g control-flow perspective, organizational perspective, case perspective and time perspective. In behaviour analysis, we start by understanding the current ‘as is’ state of processes. Here process discovery techniques can for example by leveraged event data to for example reveal (in terms of an abstract process model) the common workflows used by an enterprise for executing various types of functionality. i.e. we describe how process activities currently are being performed by mining the process model from event log. The control-flow model can then be extended with other perspectives to obtain a holistic view covering all process-related aspects and gain a complete understanding of current process behaviour. The insights and knowledge gained during this process can be used for the redesigning processes and offers a means for improving process performance such that it leads to positive (value-adding) process outcomes (Zur Muehlen and Shapiro, 2015).

Behaviour analysis also aims to provide an analytical support layer that addresses the information needs of process field analysts (Klinkmüller et al., 2019). Such support involves combining multiple analytic techniques to: (i) become familiar with the dataset at hand in order to answer a particular question (ii) obtain performance statistical summaries(e.g wait times, historical cycle times) using aggregation, correlation and evaluation techniques (iii) uncovering insights, patterns and discovery of hidden relationships among various organizational elements/artifacts (Hamilton, 2015). During the ’exploratory data analysis’ phase, process analysts often engage in techniques like deductive (hypothesis-based) and inductive (pattern-based) reasoning to formulate and to answer one or more questions (Hamilton, 2015). Traditional process mining and discovery algorithms (Klinkmüller et al., 2019) for generating procedural models (like Petri nets, causal nets, BPMN models process trees) represent the control-flow of the process and can give a big picture view of existing deployed business process models. Visualizing the control-flow or process models through process discovery is a good starting point for process analysts to perform further analysis by asking questions in an iterative manner (Van Der Aalst et al., 2011). The initial analysis then guides the process mining activities as information needs emerge during familiarization and discovery phase (Van Der Aalst et al., 2011)(Klinkmüller et al., 2019). In (Klinkmüller et al., 2019) Muller et al. have investigated challenges faced by process analysts in practice. According to their findings, analysts spend significant time investigating the case perspective (which focuses on properties of cases) and organizational perspective(which focuses on organizational resources). In (Van der Aalst, 2009) Van der Aalst has also shared some lessons learned after conducting practical real world process mining projects.

Process mining algorithm’s performance is traditionally measured by how well it achieves pareto-optimality of the mined model in terms of various properties such as fitness and precision with respect to the available event log. The algorithm has an additional goal of achieving generalization on future process instances. In many practical settings, the target search space of models is quite large for an exhaustive search; therefore, process mining algorithms, enforce a specific representational bias to to make a trade-off (e.g. between higher fitness and lower precision). In many real-world settings, process behaviour is not completely captured or available for mining in the event logs (Augusto et al., 2018) as these logs are often noisy and incomplete. Process discovery algorithms when applied to real-world complex event logs often produce either noisy or incomprehensible models that either poorly fit the event log (low fitness) or over-generalize it (low precision or low generalization) (Augusto et al., 2018).

Process mining techniques discussed in this section along with classic methods like conformance checking and process enhancement, collectively help us construct an enriched process model through the offline analysis of all available process-related data. This enriched process model provides us with a holistic process view, covering all relevant aspects of the process at hand and is helpful for: (a) supporting managerial decision-making and generating insights about performance improvement (b) constructing a simulation model which captures the on-ground reality And; (c) Performing ”what-if” analysis using the simulated model. e.g., if I allocate a given resource set to the next task, what would the predicted completion time for the process instance be?. Next, we will focus on reviewing process mining techniques aimed at behaviour analysis of all the available structured and unstructured process-related data, in order to:

  • Assist business analysts in gaining visibility and understanding, of process behaviour captured in process execution logs and other process-related data by Answer a wide range of process-related questions.

  • Uncover patterns, extract insights and discovering hidden relationships among various organizational elements/artifacts.

  • Support managerial decision-making by mining actionable knowledge and insights related to performance bottle-necks, optimal resource allocation and uncovering root-causes that lead to undesired process outcomes.

4.1. Process Performance Mining

Prior to process design, organizations often engage in an exercise of strategic planning by using balanced scorecards where the mission and strategy of an organization (Goodspeed, 2004) are translated into a set of performance measures (also known as KPIs). Such performance measures, can help organization keep track of progress towards the specific defined organizational goals. Moreover, for organizations implementing business process management, organizational objectives are used to formulate process performance measures, characterized by metrics like per-instance cost, cycle time efficiency (CTE), resource utilization, quality of service and so on (Zur Muehlen and Shapiro, 2015). Several performance measurement systems exist, and appropriate ones can be picked depending on the organizational objectives or strategic success factors (Vom Brocke and Rosemann, 2010). Devil’s quadrangle can for example, measures dimensions such as process cost, quality (e.g. visiting frequencies, error rates) and cycle time (also known as throughput time)(Jansen-Vullers et al., 2007). Jansen-Vullers et. al (Jansen-Vullers et al., 2007) have proposed a framework for quantifying the impact of best practices and discuss different dimensions for performance measurement. Several domain-specific performance reference models also exist. e.g Supply Chain Operations reference model, IT infrastructure Library (ITIL) etc.

Performance analysis techniques allow us to extract, analyze and enhance existing process models by identifying performance improvement areas and mining KPI values using the available process logs. Early identification of large deviations between the planned and actual KPI values allows organisations to take corrective actions to achieve the desired goals. Performance analysis techniques include Bottleneck Analysis (where we identify activity, resource and waiting bottlenecks), workload and demand analysis (where we analyze resource usage to identify under-utilized or over-utilized resources), rework analysis (where we identify errors or defects) and over-processing analysis. Organizations can leverage these techniques to identify a range of issues related time, costs and quality based on event log data. These insights can later be used in the process redesign phase for improving or enhancing the apriori process model. We note that process performance analysis has several titles in the literature. e.g Business Process perspective, Performance perspective, Process Performance Management etc (Hornix, 2007) (Zur Muehlen and Shapiro, 2015).

Performance analysis techniques can be divided into Qualitative or Quantitative analysis (Dumas et al., 2013a). In Qualitative Process Analysis, we can perform Value-Added Analysis(which aims at identifying waste) and Root Cause Analysis(e.g cause-effect analysis and why–why analysis). Here we measure process performance using metrics like per-instance cost, cycle time efficiency (CTE), resource utilization, quality of service, and so on (Zur Muehlen and Shapiro, 2015). Quantitative process analysis is based on historical process execution data or simulation models and involves using techniques like flow analysis, queuing theory and process simulation (Dumas et al., 2013a). Process simulation is a popular technique for quantitative analysis of process models where synthetic data based on hypothetical instances is used to estimate cycle times, average waiting times and average resource utilization. Similarly, flow analysis provides a set of techniques that utilize activity performance data to estimate the process’s overall performance but ignores the resources utilization aspect from the analysis (Dumas et al., 2013a).

Business process monitoring deals with event analysis and support of real-time executions of a given process. It aims to understand the present process performance by presenting a real-time picture of key performance indicators(often using dashboards) such as mean execution time, resource utilization or error rate etc.(Dumas et al., 2013b). Process monitoring tools help determine how well business processes under consideration are aligned with the goals of an organizations. They enable continuous, real-time monitoring of processes based on performance-related KPI values of a given process instance to track various business objectives. Process monitoring tools like performance dashboards can monitor the actual KPI values(e.g the distribution of cycles times, visualize the activities and execution patterns, and total duration) of a business process(Tardío and Peral, 2015). Several software vendors offer various performance analysis features where we can find the causes of these deviations(when they happen) and project performance analysis results onto process models to show bottlenecks, service performance levels, throughput times, and frequencies (Hornix, 2007).

4.2. Causal Process Mining

To inform decision making under uncertainty, Businesses often draw on previous experience and recorded data to understand the potential downstream effects of certain specific decisions and actions. Historically, to determine the effect of a given intervention on a particular process would require either guesswork or launching numerous A/B tests, which can be costly and time consuming (Kohavi and Longbotham, 2017). Theory of causality provides a better alternative for potential downstream effects of decisions that are being considered (Pearl et al., 2009). Causal process mining seeks to use the process execution logs to discover and quantify cause-effect relations. Causal Process Mining can answer the fundamental question: What changes, if implemented, will cause an improvement to the process?. Existing process discovery techniques allow us to discover correlation but not causation. In causal analysis we try to develop an understanding that goes beyond the control-flow perspective. We are interested in understanding the outcomes(desired or undesired effects) associated with particular action(s) (interventions) taken during the course of process execution. There has been much recent interest in studying Causal analysis and Causal reasoning in the data science and machine learning research communities.

Leveraging process data would allows us to infer future process state based on what happened in the past and use that evidence to establish causality. However, pinning down causal effects rigorously is challenging. We briefly cover some of the recent contributions to process analytics in this section. Koorn et. al. (Koorn et al., 2020) have proposed a method that uses statistical tests to discover action-response-effect patterns. Their method can identify causal relations between responses (interventions) and effects for certain pre-defined sub-populations. This can support the decision making processes by giving insights into how certain actions lead to desired outcomes (e.g improved performance). In (Mahnaz Sadat Qafari, 2020) Qafari et. al. showed that root cause analysis using structural equation modeling is useful for testing if a predetermined causal relation (identified earlier by a process analyst) holds. Bozorgi et. al. (Dasht Bozorgi et al., 2020) proposed an approach that extracts recommendation rules from event logs. Causal effect can then be assessed and rules with highest incremental effect (uplift) can be used as recommendations in the form of interventions that can influence process outcomes positively. Their framework also computes a cost-benefit model of a particular intervention, identifying particular cases to which it is applicable (case-level recommendations). Agarwal et. al (Narendra et al., 2019) take a similar approach for structural causal model discovery and performing counterfactual reasoning. Discovering cause-effect relationships in event data is useful for root cause analayis of process performance issues but remains a challenging open problem. Hompes et. al. (Hompes et al., 2017) propose a method which uses Granger causality to generate a graph of causal factors explaining factors (or their combinations) that may affect process performance.

Causal Analysis techniques has also been used to understand process deviation, explain predictions and make recommendations. We will discuss these techniques in next sections.

4.3. Deviance Mining

Business process variant (also known as deviance analysis or drift detection) refers to process instances that maybe deviate from the desired course of execution, resulting in an unexpected or unplanned outcome. Variations may occur due to contextual/environmental factors, human factors or because of explicit decisions made by process participants. In a given process log, process variants are a subset of instances that violate the behavior prescribed by the model and can distinguished based on a certain characteristics they are correlated with. e.g. representing violation compliance rules or missing set performance targets. In deviance mining, we are interested in identifying various variants of the processes that may exist and diagnosing the root-causes of process variations by analyzing or comparing two or more event logs. i.e Given a set of event logs of two or more process variants, how can we identify and explain the differences among these variants? (Taymouri et al., 2021a). Understanding the factors leading to variation can help managers make better decisions and improve overall process performance.

Several methods have been proposed in the last decade to analyze the execution logs to identify and explain differences between two or more process variants. One class of techniques uses frequent pattern mining which typically takes as input two event logs (corresponding to two variants of a process) and produces as output a list of differences (with respect to established performance objectives). Machine Learning techniques on the other hand can classify instances as “normal” or “deviant” when trained with labelled examples. In (Nguyen et al., 2016) Nguyen et al. have surveyed sequence classification and mining techniques. They divide the techniques into those based on frequent pattern mining and rest based on discriminative pattern mining. They also provide benchmark comparisons of representative techniques using various real-life event logs. Similarly, Taymouri et al. (Taymouri et al., 2021a) have tried to present a unified view of variant analysis techniques by surveying and classifying existing methods based on data, type of algorithms and analysis performed. Several methods rely on identifying frequent patterns, while others use generative approaches to discover and compare models of process variants. They have categorized process variant analysis outcomes as Rule-based, model based or descriptive (representing discrepancies and behavior of the different process variants. Modeling and management of process variants is another important challenge that contemporary BPM tools do not adequately support (Reichert et al., 2015). Existing tools require that variants be specified in separate process models or ’expressed in terms of conditional branches within the same process model’.

4.4. Decision Mining

Operational Decision-making associated with frequently executed processes and cases can significantly impact process outcomes. Traditionally, a business process model specifies activities within which the decision-making occurs. Decision logic can define the specific logic used to make individual decisions, such as business rules or executable analytic models. Decision modelling can represent complex, multi-criteria business rules and thus provides another perspective, creating a bridge between business process models and decision logic models (Figl et al., 2018).

Decision Model and Notation (DMN) is a modeling notation for decisions, published by OMG and allows for decoupling between decisions and control flow logic (Model, 2016). It provides an understandable symbolic representation of operational decisions and supports supports decision management, and business rules specification (Kluza et al., 2019).

Decision mining aims at mining the Decision model along with along with associated decision logic. The Decision model will define the decisions to be made in tasks (defined by the process model), their interrelationships, and their requirements for decision logic. Process execution data, depicting underlying rules (governing the choices) can be mined for frequently made decisions within an organization. However, decision mining is challenging due to several factors (e.g incompleteness of available data). Decisions are often dependent on personal expertise and contextual factors, information that may not be present in event logs. Decision mining techniques can identify and derive decision rules from relationships between process context, path decisions, and process outcomes. Several techniques proposed in the literature has tried to tackle the problem of decision mining partially. In mannhardt et al. (Mannhardt et al., 2016) propose a decision-tree based learning method to discover overlapping decision rules from event data. Similar to decision mining, Batoulis et. al. (Batoulis et al., 2015) have proposed a method for extracting decision logic from process models. The paper explains a semi-automatic approach where execution logs are not required and decision points can be replaced by generating a dedicated decision model, allowing decision logic to modeled separately from process logic. Rozinat et. al. (Rozinat and van der Aalst, 2006) propose a method for decision point analysis that determines how data attribute values(or data dependencies) affect the routing of a case. Their technique can identify decision points in Petri net models and has been implemented as a ProM Plug-in. Leoni et. al (De Leoni and van der Aalst, 2013) extended the technique and proposed a general-purpose method for discovering Branching Conditions (where atoms are equalities or inequalities consisting of multiple variables and arithmetic operators). Similarly, Bazhenova et. al. (Bazhenova et al., 2016) use decision tree classification to derive of DMN based decision models. Their techniques are able to extract not only control flow decisions but data decisions and dependencies. Overall, a comprehensive set of methods and frameworks is required that assist orgnaizations with dynamic management of decisions via analyzing, modelling and improvement efforts. Such methods should be able to leverage available process data and mine Decision model along with along with decision logic and dependencies between decisions and data elements .

4.5. Resource Behaviour Analysis

Human resources often take the role of knowledge workers and play a major role in modern organizations where processes are increasingly complex have knowledge-intensive nature (Di Ciccio et al., 2015). In such knowledge-intensive scenarios, Resources (especially human resources) play a critical role in deciding the overall process outcomes. Event logs contain rich information about the task, resource and process outcome and can be leveraged to optimally realize the process goals (Rajan, 2018). Resource analysis (also known as organizational or resource mining/perspective) aims at analyzing event logs for extracting insights about organizational resources, evaluate resource performance and understand past resource allocation decisions in order to find areas of improvement (Pika et al., 2017). Over the past decade, various resource analysis techniques have been proposed to tackle problems like organizational structure discovery, classification of users in roles, discovery of organizational models, social network analysis (SNA), resource allocation, analysis of information flows between organizational entities and role mining (Zhao and Zhao, 2014) (Burattin et al., 2013) (Song and Van der Aalst, 2008)(Rajan, 2018).

Optimal allocation of resources to process tasks can significantly impact the overall process performance. Comparing and tracking the productivity of human resources is a challenging problem due to biases and lack of objectivity. Sindhgatta et al. (Sindhgatta et al., 2015) propose a method that uses process execution logs to identify resource allocations decisions which result in good outcomes(measured in terms of quality of service). In a similar work, Sindhgatta et al. (Sindhgatta et al., 2014) investigate the variation in resource efficiencies with varying case attributes and show that process outcomes are dependent on various contextual factors like complexity of work, task priority and capabilities(expertise) of the resources involved.

Accurately measuring resource behaviour allows us to effectively dispatching and suggest staffing policies that meet the contractual service levels (quality) of the service system and the business process. Huang et. al. (Huang et al., 2012) consider the problem of measuring resource behaviour from different perspectives such as preference, availability, competence and cooperation. Similarly, several methods for extracting useful knowledge about resource performance from event logs have been proposed in the literature. Pika et al. (Pika et al., 2017) propose a technique based on data envelopment analysis for analysing resource productivity using event logs. This technique is often used to measure the efficiency of companies. Linh ly et. al. (Ly et al., 2005) study the problem of mining staff assignment rules from event-based data using an organisational model. Similarly, Senderovich et. al. (Senderovich et al., 2014) explore mining of resource scheduling protocols from recorded event data. Related to the problem of resource assignment and allocation Cabanillas et. al. (Cabanillas et al., 2013) have studied the challenge of resource ranking and prioritization.

The success of process outcomes also depends on resources interaction. Social collaboration patterns between human resources can significantly impact process outcomes as optimal cooperative resource behaviour leads to enhanced service quality and process performance. Improving process performance by analyzing relationships and mining collaboration patterns between organizational entities is one of the key goals of resource analysis. It is based on the premise that process output not only depends on capability but compatibility as well. Social network analysis (Van der Aalst and Song, 2004) is one approach for understanding organizational structure and analyze relationships between originators involved in processes. Such an analysis is useful for many reasons such as improving handover relations, improving resource compatibility etc. In (Kumar et al., 2013) Kumar et al. propose a modeling technique that captures the compatibility between resources at the time of task assignment. Schonig et. al. (Schönig et al., 2015) proposed an approach to extract declarative process models. Their technique allows modelling of organizational relations using rule templates that can be represented in textual Declarative Process Intermediate Language (DPIL).

RBAC (Role based access control) is used to manage resources in workflow area and is a useful model for managing resources(Kuhlmann et al., 2003). Event logs can be used to to mine role based access control (RBAC) models, identifying the privileges or authorization of resources (Baumgrass, 2011) (Burattin et al., 2013). Roles refer to the assignment of organizational agents to job functions within an organization. Discovering an optimal set of roles remains a useful problem to investigate. Vaidya et. al. (Vaidya et al., 2007) define the perform as determining a role-based access control (RBAC) configuration and analyze its theoretical bounds. Similarly, frank et. al. (Frank et al., 2013) propose a probabilistic solution to learn the RBAC configuration and show how it generalizes well to new system users for a diverse range of data.

4.6. Distributed Privacy-Preserving Mining

Modern organizations routinely deploy process analytics, including process discovery techniques on their process data, both to gain insight into the reality of their operational processes and also to identify process improvement opportunities (Augusto et al., 2018). Process analytics techniques such as process discovery play an important role in mining event data and providing organizations with insights about the behaviour of their deployed processes. However, in many practical settings, process log data is often geographically dispersed, may contain information that may be deemed sensitive and may be subject to compliance obligations that prevent this data from being transmitted to sites distinct to the site where the data was generated. Traditional process mining techniques operate by assuming that all relevant available process data is available in a single repository. However, anonymising, giving control access and safely transferring sensitive data across organization/site boundaries while preserving priacy guarantees is non-trivial. However, in many practical settings, process log data is geographically dispersed and can contain information that may be deemed sensitive. A classical example of a privacy-preserving process mining problem of the first type is from the field of medical research involving impediments to data migration. Consider the case that a number of different hospitals wish to jointly mine their process logs for the purpose of medical research, but are faced with regulatory and legislative compliance hurdles that prevent clinical process histories being shared across health jurisdictions (hospitals, health districts, national boundaries etc.). Hospitals are therefore restricted from ever pooling their data or revealing it to each other leading to small dataset available for knowledge extraction. This negatively impacts the confidence with which clinicians might deploy the results thus obtained. Our inability to migrate clinical process data also implies that we miss out on the opportunities for extracting higher-impact insights that might have been possible if data from multiple health jurisdictions could have been analysed in juxtaposition (Lenz and Reichert, 2007).

Traditional process mining techniques operate by assuming that all relevant available process data has been curated into a central site for analysis. However, anonymising, giving control access and safely transferring sensitive data across organizations is non-trivial. Moreover, organizations face legal constraints, risk of data breaches (or hacks) along with data integration challenges, preventing them from building a centralised data warehouse (Dunkl et al., 2011). This leads to a scenario where event-log data is present in organizational silos and distributed among several custodians, none of whom are allowed to share/transfer their sensitive data directly with each other (Lang et al., 2008). Mining process data in such cross-silo settings can prove to be invaluable for providing relevant operational support to organizations if privacy guarantees can be offered (Jensen et al., 2012).

Differential Privacy provides us with a formal privacy notion for datasets that are released publicly or might come in contact with potentially malicious adversaries (McSherry and Talwar, 2007). It is considered as the de facto standard for ensuring privacy in a variety of domains. The definition proposed by Dwork et al. (Dwork et al., 2014) offers a mathematically rigorous gold standard for ensuring privacy protection when analyzing datasets like process logs(or results of a randomized algorithm) that might contain sensitive or private information. We modify the definition slightly for event logs:

Definition 1: Differential Privacy (adapted from (Dwork et al., 2014)) A randomized mechanism :𝒟\mathcal{M}:\mathcal{D}\rightarrow\mathcal{R} with a domain 𝒟(e.g.,\mathcal{D}(e.g.,, possible event logs) and range (e.g.\mathcal{R}(e.g., all possible trained models )) satisfies (ϵ,δ)(\epsilon,\delta)- differential privacy if for any two adjacent process logs l,l𝒟l,l^{\prime}\in\mathcal{D} and for any subset of outputs SS\subseteq\mathcal{R} it holds that Pr[(d)S]eϵPr[(d)S]+δ\operatorname{Pr}[\mathcal{M}(d)\in S]\leq e^{\epsilon}\operatorname{Pr}\left[\mathcal{M}\left(d^{\prime}\right)\in S\right]+\delta

Two process log ll and ll^{\prime} are defined to be adjacent if ll^{\prime} can be constructed by adding or removing a single instance(entry) from the log ll. By bounding the potential worst-case information loss, the above definition provides us with a strong formal privacy guarantee. Formally, under the (ε,δ)(\varepsilon,\delta)-differential privacy definition, we measure Differential Privacy properties of our method by epsilon and delta values. Epsilon(ϵ\epsilon) is the privacy loss parameter in differential privacy and is inversely proportional to the amount of noise added. i.e Lower values of ε\varepsilon imply stronger privacy guarantees. A Differentially private mechanism typically involves using a randomized mechanim that perturbs the input dataset, intermediate calculations, or the outputs of a function, using a calculated quantity of noise (usually at the cost of utility) (Dwork et al., 2006). Such a mechanism is considered private if it hides the isolated contribution of any single individual in the databases. i.e removing a single entry will not result in much difference in the output distribution (Acs and Castelluccia, 2012) (Dwork et al., 2014).

In Privacy-Preserving Distributed Process Discovery, our goal is to discover a global process model by privately mining multiple distributed process log independently and share only the resulting insights from each analysis. i.e mining a differentially private process model, without ever pooling the data to a central site, in a way that reveals nothing but the final discovery process model to the participating organizations.

Most typical methods presented in the literature rely on some form of data transformation in order preserve user privacy. These techniques are a trade-off between information loss and privacy(Bamiah et al., 2012). Techniques based on Secure Sum Protocol (Clifton et al., 2002) permits a network of nodes to transmit a numeric sum (from node to node) to which each node adds a node-specific number without having any node being able to compute what the individual contributions of the participating nodes were. The final node obtains the sum of the numbers contributed by all of the participating nodes, again without being able to compute what the individual nodespecific numbers were. This protocol can be used to create distributed versions of most process mining algorithms. Elkoumy et al. (Elkoumy et al., 2020) propose an architecture based on Sharemind, which uses multiparty compute to mine a directly follows graph. However they don’t provide differential privacy guarantees.

4.7. Knowledge-Centric Process Mining:

In process mining, while a lot of emphasis has been on analyzing and extracting process insights from the observed behaviour logged in event logs, the knowledge dimension associated with business processes has received very little attention (Di Ciccio et al., 2015). i.e Current process mining techniques are self-contained and have minimal capacity to leverage and reasoning using prior knowledge. In practice, this means process mining algorithms can’t perform inference that goes beyond the implicit knowledge which is recorded in the event logs. Traditional mining techniques focus on mining behaviour that inherently cannot represent all of the cascading hierarchical structure representing complex real-world processes (Van der Aalst, 2009). These algorithms can’t reason about abstract relationships between various objects involved in the process. For example, easily-drawn inferences that people can readily answer without direct training like smoke is seen so there must be fire happening cannot be inferred by the current process mining methods. This leads to an incomplete understanding of process behaviour where process analysts are left trying to abstract, simplify, and even leave out key relationships needed for complete understanding of process behaviour.

Process knowledge has many faces and it will differ from other forms of organizational knowledge as it will be highly contextual, sometimes tacit and relevant to a particular domain. Organizations have realised that knowledge and processes are interlinked and should explicitly be made a key component of business processes (Barclay and Murray, 1997). Knowledge is more complex than simple data or information with an additional characteristic of being subjective, as its often tacit in nature (e.g obtained with years of experience). Process knowledge has many faces and it will differ from other forms of organizational knowledge as it will be highly contextual, sometimes tacit and relevant to a particular domain. Sometimes this knowledge is documented but even then it exists in silos of unstructured documents, often halting the progress of processes to consult any knowledge bases that might exist.

For organizations, having access to the right knowledge at the right time is crucial to effective decision making as it allows companies to solve problems quickly, diffuse best practices amongst employees, cross fertilize ideas and enable them to stay competitive (Van Beveren, 2002). Managing Process knowledge requires a deliberate and systematic approach which should be reflected at all levels of organization. Organizations employ various knowledge managements tools to capture and create the knowledge that is either explicit or tacit by sometimes interviewing experts, telling war stores, and apprenticeship style training programs. Despite several attempts, modern organizations find capturing managing the available knowledge(in terms of capturing, codifying, and sharing) difficult and rely on knowledge workers for know-how. This leads to organizations routinely forgetting the lessons learned during the past when knowledge workers leave or switch projects. Overtime organizations have realised the importance of deliberate and systematic approach for creating a culture where company’s knowledge base in maintained (Von Rosing et al., 2014) and many expert systems, and Enterprise knowledge management systems have been proposed to capture knowledge effectively, categorize it and then make that knowledge available across an organization. Such tools can codify knowledge in the form of wikis, cognitive maps, decision trees, knowledge graphs etc. These movements were largely unsuccessful(Easterby-Smith and Lyles, 2011).

Knowledge management has largely been ignored in recent years. Solutions are needed that tackle knowledge management challenges and help organization extract actionable knowledge from all available process related data(e.g extracting meaningful knowledge from unstructured documents) and make it widely available inside the organisation. Similarly, Common-sense reasoning has been highlighted as one of the major challenges for the process analytics research community (Calvanese et al., 2021). It follows a broader trend in AI research where the need for solving complex tasks by incorporating knowledge and common-sense reasoning has been repeatedly highlighted (Davis and Marcus, 2015). Furthermore incorporating domain knowledge by means of constraints can support process analysts in their efforts to fully understand executed process behaviour recorded in real-world event logs and improve the outcome of process mining activities.

4.8. Discussion:

The topics covered in this section, along with classic topics like process discovery continue to provide opportunities for research contribution. We note that even with the availability of various state of the art process discovery methods, understanding business processes and resource behaviour from process logs alone remains a challenging problem(Benatallah et al., 2016). Furthermore systematic adoption of process mining in challenging domains like healthcare, also remains a challenge (Munoz-Gama et al., 2022). Process mining techniques can be of great value to answer process-related questions and is often a starting point for process analysts, however, more support is needed to address the challenges faced by process analysts. Carmona (Carmona, 2020) has highlighted some of the open problems and research challenges associated with process discovery and conformance. Their findings show that the existing mining methods struggle to deal with challenges like spaghetti models, concept drift and identifying events that occurred at different levels of abstraction. We also observe that traditional process mining and analysis techniques have not given significant attention to the analysis of resource behaviour and its affects on process outcomes. Resource analysis deserves more attention from the research community as knowledge about resource behaviour can be used for effective planning, strategizing and gaining insights which lead to better overall process performance. In (vom Brocke et al., 2021) Brocke et al. present an enterprise framework for analyzing the effects of process mining that emerge at various levels of an organization. Lastly, by understanding contextual factors and gaining detailed insights about how processes are being executed within a particular context also remains an interesting challenge.

Data Challenges: Process mining techniques are primarily reliant on process logs, which don’t always explicitly capture all the behaviour of past executed processes(Augusto et al., 2018). Such logs are susceptible to domain gaps, data bias (due to incompleteness) and quality issues (due to noise and erroneous data recordings). Many real world processes are unstructured in nature and for these processes most state-of-the-art process discovery algorithms produce, hard-to-interpret, spaghetti-like models which poorly fit the event log. This negatively influences the usefulness of the discovered process model. Often times discovered models are hard to interpret from a process analyst perspective while also prone to under-fitting or over-fitting the given event logs, offering only minuscule support for improving process outcomes. Furthermore, a particular challenge in process mining is the management of business-process variants and contemporary business process management tools do not provide adequate support for modeling and management of process variants (Taymouri et al., 2021b). For complex domains like healthcare, where improving clinical outcomes can directly impact the quality of life for patients, this implies that process analysts miss out on the opportunities for a complete understanding of the underlying process behaviour and subsequently extracting higher-impact insights.

Event data can come from various heterogeneous data sources and in several data formats. These logs are rarely in the desired format and form. It is common for process analysts to apply a series for pre-processing steps for consolidation, verification(checking for errors and inconsistencies) and transformation. To identify an event trace representative of a process instance, is challenging and is often done manually using techniques like extraction, correlation, and abstraction of the event data by the process analyst (Diba et al., 2020). There is also the challenge of ’Big data’, where data to be analyzed is of huge volume, velocity and variety, making it unfeasible to be analysed with traditional analysis tools. Other challenges include the diversity of formats and nonstandard data models (Sakr et al., 2018). To maximize the utility of available data, it is sometimes useful to build a data lake that provides a single consolidated view of organizations’ datasets. Data lake makes the preparation and analysis(by querying) of data more accessible. Such data can also include relevant context in which business processes were executed (this could be the context for an entire process or an individual task) and enable organizations to gain insights that help improve existing processes.

As observed in various other fields of AI and machine learning, an important driver of measuring progress is precisely defining Process Analytic tasks(using mathematically rigorous definitions) and building useful benchmark datasets that can be used for performance comparison (Blagec et al., 2020). The progress of the field depends on the cycle of identifying real-world problems, researchers proposing novel techniques and industry adaptation of such techniques. It is therefore important to make quantifiable progress by benchmarking process analytics tasks on standard datasets with well-defined performance metrics. We observe that many process mining published papers often lack strong reproducible experiments, resulting in poor comparability and relative merits of the proposed approach. So far, only in predictive process analytics, we have seen the adoption of rigorous evaluation criteria. We refer to work by (Augusto et al., 2018) as an example of such an effort where Augusto et al. review various state-of the art process discovery methods and benchmark the performance of various automated discovery algorithms. We hope future studies will follow the same approach where they critique and compare proposed methods against existing state-of-art process mining techniques via standard benchmarks.

5. Predictive Process Monitoring

Predictive business process monitoring is a family of techniques concerned with predicting the future state, outcomes and behaviour of ongoing cases of a business process(Teinemaa et al., 2019). For organizations, process monitoring is a useful capability that enables operational support for near real-time monitoring of processes and allows them to take preventive measures that help avoid conformance violations, undesired deviations and prevent delays.

Predictive analytics can be also be viewed as, computing a set of functions or a set of computer programs that carry out computation, over a (partially executed) process instance to perform continuous monitoring of process instances. Such methods can leverage historical(completed) executions logs, such as event logs, along with available contextual data, in order to: recommending appropriate actions at each stage, early detection of process variation or anomalies(for fraudulent behaviour), predict various properties of a case instance, help perform early risk assessments and assist with resource allocation decision(Weinzierl et al., 2020b).

The research goals of the field of Predictive Analytics can be framed as following research questions:

Given an an event log of a completed business process execution cases and the final outcome for each case, Can we:

  • Predict the next best activity to execute (in order to achieve optimal outcomes)? or Predict the outcome of a single activity

  • Predict the entire sequence of activities leading to the process end?

  • Predict if the running process instance will meet its performance targets

  • Predict the total remaining time to completion for a process instance?

  • Predict the performance outcome of an incomplete case, based on the given (partial) trace (redundant)

  • Predict the cycle time of a given process instance ?

  • Estimate the likely cost that will be incurred in executing the remainder of the process?

Machine learning allows us to build models from past experiences embedded in the process data and use it to provide real-time or near real-time decision support. Over the years many machine learning techniques, such as regressions, support vector machines etc. have been used to build predictive models from historical data. In Predictive Process Monitoring, we study methods for building predictive models that utilize historical execution traces to predict the likelihood of events and forecast outcomes such as next activity prediction, process outcome prediction or remaining time prediction (Maggi et al., 2014). Predictive analytic methods can determine factors that influence the process outcomes and violate performance targets (Wetzstein et al., 2009).

Deep Learning: Recent advances in neural network architectures and availability of large datasets has led to the popularization of using ’deep learning‘ methods for predictive analytics. Deep Learning methods have been proposed for predicting how the future of a given process instance will unfold and the likely occurrence of certain events. Deep Learning methods are particularly good at discovering intricate structure and robust representations from large quantities of raw data, thus significantly reducing the need to handcraft features which is typically required in traditional machine learning techniques (LeCun et al., 2015). DNN’s have a number of processing layers that can be used to learn representations of data with multiple tiers of abstraction. The different tiers are obtained by producing non-linear modules that are used to transform the representation at one tier into a representation at a more abstract tier. Deep Convolutional Neural Nets and Recurrent Neural Nets are two popular architectures of DNNs, that brought about breakthroughs in processing text, images, video, speech and audio (LeCun et al., 2015). Recurrent Neural Nets, especially the Long Short-Term Memory (LSTM) have brought about breakthroughs in solving complex sequence modelling tasks in various domains such video understanding, speech recognition and natural language processing (LeCun et al., 2015) (Schmidhuber, 2015). LSTMs work by maintaining a dynamic short-term memory vector, which stores the summarization of historical events, from which the next activity can be predicted. It has been shown that LSTM can consistently outperform classical techniques for a number of process analytics tasks such as predicting the next activity, time to the next activity and so on.(Navarin et al., 2017).

Machine Learning methods require vector representations of input data and manually encoding or crafting features is a tedious task in practice. Deep learning methods have an advantage over such classical methods as they can generalize well on various tasks without requiring explicit Feature engineering or configuration tuning (Arulkumaran et al., 2017). Further these methods exhibit ’robustness to noise’ and can show performance scaling as we input bigger and bigger datasets(Evermann et al., 2017). In Process analytics we can leverage the deep learning methods to automatically find highly compact low-dimensional representations (features) of high-dimensional data. In (Alexander Seeliger and Muhlhauser, 2021) Seeliger propose recurrent neural networks (RNNs) based architecture that automatically learns vector representations of cases. They train the network to predict the contextual factors of the corresponding case. This allows us to incorporate contextual factors of a case into a single compact vector representation, later to be used for process mining.

Predictive process analytics techniques support managers in operational decision-making processes by for example taking remedial actions as business processes unfold. Picking the best method depends on the domain, available datasets, choice of input features use train the models etc. For running process instances, making accurate and early predictions using limited computing resources still remains a challenge. Early accurate predictions of outcomes, explainability (reasoning behind why certain predictions were made) and prescribing actions to prevent undesired outcomes are some of the problems that are under active investigation by the research community. Various survey papers have tried to cover the literature on predictive analytics. Marquez-Chamorro et al. (Márquez-Chamorro et al., 2017) and Di Francescomarino et al.(2018) (Di Francescomarino et al., 2018) classify the literature based on input data, classification algorithm and prediction target. Similarly, (Teinemaa et al., 2019) (Verenich et al., 2019b) also survey the field by covering various datasets, propose task definitions and provide benchmark comparison of recently proposed algorithms. We briefly describe the relevant problems studied under predictive process monitoring:

5.1. Next Activity Prediction

Next Activity Prediction refers to the problem of predicting the next trace suffix likely to occur during the execution. In Next Activity Prediction, we attempt to derive a process-agnostic machinery for learning to generate recommendations and predict the future behaviour of a given partially executed process instance while assuming minimal domain knowledge. Models are trained using historic data and after training, the input is using the observed prefixes of running process cases (event stream). The problem has been tackled using various machine learning techniques where it is often treated as a multi-class classification problem. The next event(of a process instance) can be represented as one class that the ML classifier can predict.

Prediction techniques based on deep learning have a popular choice and have shown promising results(Evermann et al., 2017) (Márquez-Chamorro et al., 2017). Such techniques are often motivated by the applications of deep learning to Natural Language Processing tasks (e.g langauge modeling). Tama and Comuzi (Tama and Comuzzi, 2019) and Weinzierl et. al. (Weinzierl et al., 2020b) provide a comparison of architectures and encoding techniques for next activity prediction using real world logs. We can use several evaluation metrics(e.g Accuracy, F1-score etc.) to compare the methods. Brunk et al. (Brunk et al., 2020) consider the problem of context-sensitive process predictions(in business process monitoring) by employing evidence sensitivity analysis to determine if context is cause or effect of the next event during execution. This allow us to understand the impact that a context variable can have on a running instance and offers an explanation for understanding why a certain prediction was made. While LSTMs based techniques can theoretically deal with long event sequences, the long-term dependencies between distant events in a process get diffused into the memory vector. We therefore seek modeling methods that are more expressive and allow storing and retrieval of intermediate process states in a long-term memory. To tackle this, recently, Khan et al.(Khan et al., 2018) explore the application specific type of neural network known as the memory–augmented neural network (MANN) for the task of next event. In a typical setting, a MANN is a recurrent neural network (e.g., LSTM [8]) augmented with an external memory matrix. Differential Neural Computer architecture can be adapted the to account for a variety of tasks in predictive process analytics: (i) separating the encoding phase and decoding phase, resulting dual controllers, one for each phase; (ii) implementing a write-protected policy for the memory during the decoding phase.

5.2. Process Path Prediction

Predicting the evolution of running cases is an important problem and plays a key role in risk management, resource allocation and process improvement. Similar to Next Activity, Process Path Prediction methods can predict possible paths of a running instance up to its completion(Verenich, 2016). Tax et. al. (Tax et al., 2017) and Evermann et al. (Evermann et al., 2016) explore the use of Long Short-Term Memory (LSTM) to solve various process predictive problems. Tax. et al.(Tax et al., 2017) show the use of LSTM neural networks for predicting the sequence of the future activities. Similarly, Camargo et al. (Camargo et al., 2019) employ deep learning techniques to predict sequences of next events, their timestamp, and their associated resource pools. Considering context(as cause or effect) when making these predictions is also an interesting problem studied by Brunk et. al. (Brunk et al., 2020). Tama and Comuzzi provide a comprehensive empirical comparison and bechmarking for such techniques (Tama and Comuzzi, 2019).

5.3. Time-related Predictions

Time-related Predictions problems such as Predicting the remaining cycle time, ompletion time and case duration, predicting delayed process executions and deadline violations of running instances has been studied extensively. Indicators available in event logs can be exploited to make predictions about time-related risks(Van der Aalst et al., 2011). Verenich et. al. (Verenich et al., 2019b) have done an extensive survey of various methods used for predicting the remaining cycle time and present a cross-benchmark comparison. Van Dongen et. al. (van Dongen et al., 2008) apply non-parametric regression on event data to predict remaining cycle time. Similarly (Evermann et al., 2017) employ RNN to predict the duration of activities. Similarly, Tax. et al.(Tax et al., 2017) propose a prediction method based on Long Short-Term Memory (LSTM) neural networks.

5.4. Process Outcome Predictions

The problem of case outcome prediction aims at identifying process instances that will end up in an undesirable state (measured as likelihood and severity of fault occurrence or violation of compliance rules) (Verenich, 2016). Instances can be labelled as normal or deviant and then process related risks can be classified using any of the traditional modern machine learning techniques. The case outcomes can be assessed primarily by first checking if the process has met its ’hard goals’ and then soft goals (determined by KPIs such as time quality, cost etc.) Similarly, in Compliance monitoring techniques are aimed at preventing Compliance violations by monitoring ongoing executions of a process and checking if they comply with respect to certain business constraints(Verenich, 2016).

5.5. Discussion

Explainability Explainablility or Interpretability remains a key challenge in predictive business process monitoring. Interpretability is defined as ”the ability to explain or to present in understandable terms to a human” and often stems from incompleteness in problem formulation leading to unquantified bias (Doshi-Velez and Kim, 2017). Business process stakeholders and decision makers can only fully trust prescriptive support systems that offer an explanation for the decisions or recommendations made (Adadi and Berrada, 2018). Decision Support systems which leverage machine learning methods must therefore be able to explain the reasoning behind certain decisions, recommendations, predictions made or actions taken. Predictive process monitoring methods should explain why the predictive model was mistaken when the predictions are inaccurate. By making systems and models interpretable we want to ensure that decisions and data can be explained to end users in a transparent and easy to understand manner (Adadi and Berrada, 2018). Increasingly, Explainability and Interpretability is not just a desirable property but increasingly becoming a serious matter of public debate. In some countries compliance laws will require ’a right to explanation’ which means end-users can ask for an explanation of a certain decision that affects their lives (Goodman and Flaxman, 2017). This also means that use of black box methods (such as Deep Neural Networks) in predictive and prescriptive approaches will be infeasible as we can’t explain the decisions or predictions made. ’Explainablility’ and ’establishing trust’ in these models remains one of the key challenges, not only for sensitive domains like healthcare but now for everyday consumer-facing products as well. Therefore, developing systems that provide trustworthy explanations and the necessary chain of reasoning that led to particular decisions and outcomes remains a significant challenge for future research work in predictive and prescriptive monitoring. Several methods have been proposed to make systems more comprehensible. Rizzi et. al. (Rizzi et al., 2020) for example propose a method that uses post-hoc explainers and encoding for identifying the most common features that explains incorrect predictions. By reducing the impact of identified features, explanations can be used to improve model accuracy. Brunk et. al. (Brunk et al., 2020) take a different approach to the problem of making transparent context-sensitive predictions by proposing a next event prediction technique based on dynamic Bayesian network. Galanti et. al. (Galanti et al., 2020) propose a based framework based on game theory of Shapley Values for explaining the predictions. Their framework is based based on LSTM models, and can explain any generic KPI. Lastly, Verenich et. al. (Verenich et al., 2019a) employ flow analysis technique for predicting quantitative performance indicators of running process instances. Their technique can be used to estimate values of this performance indicator (e.g cycle time) by aggregating performance indicators of the activities composing the process. We refer the readers to the work by Doshi-Velez and Kim (Doshi-Velez and Kim, 2017) and Adadi and Berrada (Adadi and Berrada, 2018) for a more detailed discussion around the topics of ’Interpretability’ and ’Explainable AI’.

6. Decision Support or Goal-Directed Decision Making or Prescriptive Process Analytics

Business process management assists organizations in planning and executing activities that collectively deliver business value, usually in the form of a product or a service. Flexible execution of business process instances entails multiple critical decisions, involving various actors and objects, taken to achieve optimal process outcomes(Teinemaa et al., 2019).These decisions are of variable nature and context dependent as Business operate under uncertain real-world environments. Sub-optimal decisions during process execution, such as picking the wrong execution path, incorrect resource allocation, can affect business processes outcomes leading to cost overruns and missed deadlines (Ghattas et al., 2014). These decisions therefore require careful attention, therefore, ability to guide and automate decision making, therefore, is crucial to maintaining and improving business process performance(Gröger et al., 2014). Overall, Analytics-driven decision support for process users and knowledge workers remains a significant challenge for BPM research (Catalkaya et al., 2013).

Historically, Process support has largely focused on development of Workflow Management Systems and later evolving to Business Process Management Systems(BPMSs). Such systems have played a significant role in facilitating analysis, improvement, and enactment of business processes(Schonenberg et al., 2008). However, future systems will tackle the challenge of leveraging historic and current contextual data for providing intelligent assistance to process users, and in guiding process-related decisions. We foresee future process analytic decision-making systems or decision-support would take the role of:

  • Recommending the best suffix for a task sequence, which, if executed, will lead to a desired outcomes or desired performance characteristics. This means identify a process path that would yield best performance in a given context.

  • Providing support for Knowledge-intensive processes to assist knowledge workers. e.g. Generating optimal action recommendations for a process instance while accounting for various sources of uncertainty.

  • Providing support for risk-aware Process Management. e.g. by early identification of process associated risks and generating recommendations for corrective actions that can help avoid a predicted metric deviation.

  • Providing support for Resource Management. e.g. assess the suitability of a resource in executing a certain task and support resource allocation decisions by suggesting optimal work assignment policies

  • Providing strategic support for robust process execution and for developing robust strategic plans in adversarial settings while carefully balancing multiple objectives

We discuss some of these ideas in detail below:

6.1. Risk-Informed Decision Making

Modern organizations operate in dynamic environments and face the challenge of dealing with uncertainties(risks) associated with various process decisions due to environmental factors that are often not fully under their control. Managing process-related risks by making risk-informed decisions remains a key challenge for organizations. Risk reduction usually involves decreasing the likelihood and severity of faults during process execution and ensuring that desired performance goals are met.

Management of risks during process execution remains a challenging problem(Suriadi et al., 2014). The ability to detect potential metric deviations early on is valuable in giving process owners decision information for a timely intervention. One primary goal of process decision support during Business process execution is to guide risk-informed decisions at run-time by modeling, detecting and mitigating risks as early as possible. In practice, this means using predictive and prescriptive techniques to identify process instances that are likely to get delayed or to terminate abnormally. This involves using those predictions to intervene early, recommend actions or support resource allocations decisions at run-time that enable organizations to avoid or decrease the likelihood of metric overrun.

To tackle these challenges, Risk-based decision support (RDS)(Sometimes known as Risk-aware Business Process Management Systems) have been proposed to help with early detection of factors by predicting risks in terms of metrics deviations during process execution. They can make recommendations that can help minimize the predicted process risk and avoid negative outcomes. The performance criteria of business processes are typically described by Key Performance Indicators (KPIs), and their target values can be defined based on business goals. They can assist in risk management by monitoring KPIs, PPMs, and QoS metrics. Overall, reduce risks (in terms of metrics deviations during process execution) to provide decision support for a given process, e.g., recommending the next process activity, which minimizes process risks. When those deviations exceed a tolerance threshold, interventions can be taken to decrease the likelihood of unfavourable outcomes.

Risk reduction techniques allow us to identify process instances that are at risk of not meeting certain performance criteria and recommend preventive actions to process participants. Several techniques have been proposed in the literature for modeling and detection of risk. Suriadi et. al. (Suriadi et al., 2014). Provide a comprehensive review of techniques for managing process-related risks. Literature on prescriptive business process monitoring consists of techniques (Gröger et al., 2014) (Conforti et al., 2013a) (Schonenberg et al., 2008) that can be used to recommend preventive actions in order to support risk-informed decision making.

Future Business Process Management systems must, therefore, assist process users/stakeholders in risk-informed decision making by generating predictions and recommendations that help reduce such risks (Di Francescomarino et al., 2017). Risk Management and BPM were historically separate fields, and their integration is understudied, leaving room for future research contributions (Suriadi et al., 2014). Conforti et al. (Conforti et al., 2015) discuss the Risk-aware BPM lifecycle where each phase can be complemented with elements of risk management like Risk Identification, Risk-aware Execution and Risk monitoring etc. Lastly, problems like real-time risk detection, resource scheduling, Automated risk mitigation and Real-Time Risk Monitoring are also investigated under the umbrellas of risk-informed decision making(Conforti et al., 2011) (Conforti et al., 2012) (Conforti et al., 2013b). Overall, managing process-related risks by making risk-informed decisions remains a key challenge for organizations.

6.2. Supporting business process execution via Process-Aware Recommender Systems

In the previous section, we discussed how predictive monitoring techniques can provide operational support based on models learned using historic logs in order to predict what is likely going to happen next and use those prediction capabilities to influences the outcomes of running case instances. e.g. monitor processes and issue recommendations to workers and managers based on probability that a given case will violate the set performance targets. The output of Predictive business process monitoring techniques, is just predictions. Predictions can be used as early warnings for taking risk informed decisions but do not explicitly support answering of question like What action should we take next to achieve a particular goal? and Why should we do it?(Lepenioti et al., 2020). Compared to descriptive and predictive business analytics, prescriptive process analytics remains less mature (Eili et al., 2021). Marquez et. al. (Márquez-Chamorro et al., 2017) point out that ‘little attention has been given to providing recommendations’. Instead of providing specific action recommendations, literature on business process monitoring focuses on forecasting future process events(and outcomes) while leaving the action implementation part to the subjective judgment of process users and business decision makers(Dees et al., 2019).

Recommender Systems have found applications in information filtering system such as in video or music services as playlist generators or content recommendations for social media companies etc(Beheshti et al., 2020).

Process-aware Recommender Systems have been proposed to assist knowledge workers, in operational decision-making by recommending actions for executing a particular process/task, manage resource allocation policies and so on (Beheshti et al., 2020) (Schonenberg et al., 2008). Eili et al. (Eili et al., 2021) provide a systematic review of Recommender Systems in Process Mining and classify recommendation approaches as ‘pattern optimization’, ‘risk minimization’, or ‘metric-based’. For structured processes, Process-aware Recommender Systems have been proposed to assist knowledge workers, in a context-aware adaptable fashion by recommending actions for executing a particular process/task, manage resource allocation policies and so on (Beheshti et al., 2020) (Schonenberg et al., 2008). Such systems leverage technologies like machine learning to build recommender systems that monitor process instances, predict future process states and recommend appropriate actions(Dees et al., 2019).

We should note that most process-aware recommender techniques focus on supporting risk-informed decision making.Their major focus is on preventive measures early warning recommendations to for example avoid predicted metric deviation. Another major goal of Process-aware Recommender systems should be to support decisions that that maximize the likelihood of achieving business goals. e.g Weinzier et al. (Weinzierl et al., 2020a) consider problem of recommending next best actions that lead to optimal outcomes. Their technique relies on explicitly adding control-flow knowledge to their proposed technique via formal process model and uses process simulations to verify and filter the predictions of the trained predictive model. Groger et al. (Gröger et al., 2014) introduce the concept of recommendation-based business process optimization to support adaptive process execution. Their framework recommends actions for the next process step to take for a given process instance. For organizations data-driven process optimization enables decision support for real-time process optimization. e.g. shortening the reaction time of decision-makers to events that may affect changes in process performance.

6.3. Operational decision support for Knowledge Intensive Processes

Business Processes assist organizations in organizing activities that deliver business value, usually in the form of a product or a service. Over the last decade, automation has caused the landscape of work to change significantly and Knowledge workers are now regarded as the most valuable organizational assets. Knowledge work is characterized by unstructured processes which can be hard to specify at design-time Supporting knowledge workers involved in the execution of unstructured Knowledge-Intensive Processes by providing context-specific recommendations remains an interesting challenge.

Knowledge Intensive Processes (KIPs) are processes that require precise expert(tacit) knowledge, involvement of knowledge workers, and consisting of activities that do not have the same level of repeatability as structured processes(Di Ciccio et al., 2015). Instead of assuming a rigid process structure, knowledge-intensive processes (KiPs) are goal-oriented, often unstructured(with pre-defined fragments) and characterized by activities that cannot be anticipated(or modeled in advance). KIPs represent a shift from the traditional process management view (where process models are structured with repeatable tasks), to a model where task execution depends on knowledge workers as primary process participants (Di Ciccio et al., 2015).

Knowledge workers are highly trained and have specialized expertise in performing complex tasks autonomously and are considered a key asset for modern businesses. Supporting knowledge work when rigid definitions of process models are not available or cannot be designed apriori (with structured or unstructured data)remains a key challenge(Di Ciccio et al., 2012). They typically rely on their experience based intuition and domain expertise, for decision-making. Their work is less characterized by explicit procedures and more by creative thinking that usually cannot be planned a priori(Di Ciccio et al., 2015). An example of knowledge work is Clinical decision-making for patient treatment in the hospital emergency room, which is highly case-specific and requires a knowledge-driven approach. i.e Treatment Decisions are made based on highly specific medical domain knowledge, the context in the form of patient’s medical history, years of specialized experience and evidence that emerges from patient test results and real-time sensors.

Knowledge workers still lack the adequate decision support tools to assist them in executing knowledge-intensive processes(Di Ciccio et al., 2015). As we enter the knowledge economy, future process management systems will have to drive processes in modern enterprises that are highly knowledge-driven, are semi-structured or unstructured while leveraging a diverse range of process-related datasets. e.g recommend appropriate actions to knowledge workers while operating in dynamic environments. In (Di Ciccio et al., 2015) Ciccio et al. have provided a set of requirements for process-oriented systems to support knowledge-intensive processes. One of the inherent challenges that future process management systems must address is that of flexibility. Flexibility means instance-specific adaptations based on context and environment. Flexible execution of business process instances involves multiple critical decisions at each step. e.g. what task to perform next and what resources to allocate to a task and so on. Previously, to address the problem of supporting flexible knowledge-intensive process, many paradigms have been proposed. e.g Adaptive Process Management(ACM), Flexible Process Aware Information Systems, case handling systems and declarative processes. Adaptive Case Management gained the most popularity amongst the various paradigms.

Adaptive Case Management(ACM) is aimed at supporting knowledge workers involved in the execution of dynamic, unstructured knowledge intensive processes(KIPs) where course of action for the fulfillment of process goals is highly uncertain(Hauder et al., [n.d.]) (Motahari-Nezhad and Swenson, [n.d.]). ACM offers a way to manage the entire life-cycle of a “case” by following the ’planning-by-doing’ principle, where work is done by considering the context and is continually adapted based on the changing characteristics of the environment(Motahari-Nezhad and Swenson, [n.d.]). In the case management paradigm, the focus is on the case and its hard to pre-define the sequence of activities. For example, Case is the ‘Product’ being manufactured or a ‘patient’ being treated where primary driver of case progress is the case data and information that emerges as the case evolves. There however can exist template or patterns that represent the structured aspects of the process. Here process could be seen as a recipe for handling cases of specific type (van der Aalst et al., 2003). A Case template is created by the knowledge worker, allowing high degree of flexibility in executing a particular case. Templates can then be used to instantiate case instances and represents a middle ground between a completely specified structured process and an unstructured process. Case execution allows us to gather feedback and adapt the templates to be reused in a particular context(Marin et al., 2016). Adaptive Case Management (ACM) has been gaining significant interest for handling unpredictable situations in processes and still lacks ’common operational semantics’ and a ’proper theory’ (Hauder et al., [n.d.])(Hewelt and Weske, 2016).

In the context of Case Management, the problem of recommending ’next best steps’ in a case management system based on the knowledge of past similar cases(which is the focus of this work) has been addressed by Schonenberg et al. (Schonenberg et al., 2008) and Motahari-Nezhad et al. (Motahari-Nezhad and Bartolini, 2011). Schonenberg et al. attempt it by first finding similar cases based on abstraction, then using support and Trace Weights to consider the relative importance of a log trace. Similarly, Motahari-Nezhad et al. (Motahari-Nezhad and Bartolini, 2011) have looked at the problem of decision support for guiding case resolution based on how similar cases were resolved in the past. Such support can complement knowledge workers decisions, often made using personal experience and expert knowledge(Di Ciccio et al., 2015). We argue that the notion of recommendations under-pins decision support for not only structured processes but across the whole spectrum of process management. i.e we see recommendations as a general-purpose mechanism for providing operational decision-support for not only structured but semi-structured and unstructured processes as well. In knowledge-intensive processes, for example, recommendations can provide intelligent assistance to process users by offering concrete support in various process related decisions like resource allocation decisions or action recommendations etc. Such form of assistance allows knowledge workers to take preemptive actions that can avoid negative outcomes (Schonenberg et al., 2008)(Gröger et al., 2014). Data-centric AI approaches hold the promise of supporting knowledge intensive processes and case management practices, whereby enabling flexible process execution. Khan et al. (Khan et al., 2021) propose a data-driven reinforcement learning based recommender system for supporting knowledge workers that considers the past execution data in addition to characteristics of the objects involved (e.g product or user). The proposed system recommends the next best action (or sequence of actions) while taking into account asset characteristics and process context.

6.4. Decision Support for Resource Management

Resources are entities that are responsible for performing activities of a business process and must satisfy various (sometimes contradictory) business goals. Management of resources involves selecting the right resources, evaluating the efficiency of resources and optimally allocating tasks to relevant resources (Rajan, 2018). Decision support for resource management provides process users, intelligent assistance on the optimal allocation of resources based on their capabilities, past performance, current workload and process characteristics.

Past execution data containing resource allocation decisions and can be leveraged to provide real-time decision support regarding the allocation of resources at an appropriate time while considering the specific context. Taking this data-driven approach improve productivity and efficiency of business processes and helps avoid performance deviations. Selecting the most appropriate resource and assigning it to the right task is often a challenging decision as it is dependent on task complexity, task priority and expertise of the worker. Traditional resource allocation decisions are made based on profile matching(perceived capabilities), which often involves human judgment. Such approaches are not always optimal because of various challenges such as resource unavailability, overloading and uncertainty(of process execution and resource behaviors) (Sindhgatta et al., 2015). Traditional thinking also held that the performance of a process is determined by its design; thus, well-designed careflows would lead to better patient outcomes. More generally, though, process context also plays an important role. Process context can be defined as knowledge exogenous to a process, and not consumed as input to a process that nevertheless serves as a determinant of process performance and proposed a context-aware recommender system for identifying context and using it to support resource allocation and task allocation decisions.

A number of methods related to learning, reasoning, and planning resource allocation have been proposed in the literature. e.g. In (Rajan, 2018) Rajan et al. have explored addressed the question of context-aware process management. Such recommender systems can derive data-driven business process provisioning that supports effective dispatching and staffing policies and assist managers in meeting the desired quality of service (or performance) levels. Russel et al. (Russell et al., 2005) discuss workflow resource patterns and have identified three main allocation types, namely, capability-based allocation(where we match capabilities of available resources with and requirements of an activity), history-based allocation (using past execution data to make ) and finally Role-based allocation(which considers the organizational position and relation of the resource). Liu et al. (Liu et al., 2014) model the task allocation problem as a Markov Decision Processes (MDPs) and show their Q-Learning based method overcomes many of the shortcomings of traditional methods(e.g. load imbalance) to compute social relation between two resources. We can also consider various abstraction levels when allocating resources. e.g. Arias et al. (Arias et al., 2016) have proposed a recommendation system that dynamically allocates resources at a sub-process level based on multi-factor criteria (to assess resources). Their proposed tool considers metric scores in the various dimensions of the resource process cube(knowledge base) to present a ranked list of suitable resources. Similarly, machine learning approaches can be used to mine resource allocation rules. e.g. Huang et. al. (Huang et al., 2011) treat resource allocation as a sequential decision making optimization problem and propose a reinforcement learning based, resource allocation solution. Their Q-learning based framework allows adjustment of real-time allocation decisions by learning appropriate allocation policies based on available feedback.

6.5. Strategic Decision Modeling

Goal-orchestration for flexible process execution: Goal models holds the promise of delivering significant value Strategic Modelling, by providing a hierarchic representation of statements of stakeholder intent, with goals higher in the hierarchy (parent goals) related to goals lower in the hierarchy (sub-goals) via AND- or OR-refinement links. Goal models encode important knowledge about feasible, available alternatives for realizing stakeholder intent represented at varying levels of abstraction. A number of prominent frameworks leverage goal models, including KAOS, i* and Tropos (Wang and Han, 2004). Strategic Modelling for organizations selecting amongst alternative goal refinements (OR-refinements). Given a goal model that delimits that space of goals and sub-goals that an organization can seek to satisfy, this is a critical (and indeed, only) decision problem to be solved. An AND-refinement of a goal is a statement of know-how that tells the organization how to achieve a parent goal (although without sequencing information, and thus falling short of being a full procedure or process model). OR-refinements offer alternative specifications of know-how for a given parent goal.

Replacing tasks or activities with goals in-process models allows us to enact processes in, flexible, context-sensitive ways. In (Santipuri et al., 2017) Santipuri et al. introduce the concept of goal orchestration for modeling processes (or process behavior) to enable flexible process execution. Goal orchestrations offer abstract, strategy-level views on processes, which can aid human understanding and ease process redesign. Their proposed technique can mine goal orchestrations from enterprise event logs and compute alternative task-level realizations of a goal if the initial attempt at realizing the goal fails to achieve the desired results. Goal-oriented process mining is a promising sub-field. In (Ghasemi and Amyot, 2020) Ghasemi et al. present a survey various techniques for mining goals from event logs and argue that combining goals and process mining can potentially augment the precision, rationality and interpretability of mined models.

Achieving resilience in adversarial settings: The need to future-proof businesses is widely acknowledged as one of the hardest challenges facing business decision-makers. Businesses need to anticipate environmental changes and the likely behavior of competitors. Much of what happens in the business environment (the effects of moves by these actors) is adversarial in nature and adversarial moves prevent or impede the achievement of business goals. Decision-making in adversarial settings involves reasoning about chains of moves and counter-moves by the adversarial entities involved. Strategic resilience requires that businesses make decisions that are most resilient to adversarial moves by players in the business environment. e.g Red teaming allows businesses to informally reason about the adversary’s strategic planning process(Hoffman, 2017). Similarly, in (Gou et al., 2017) Gou et al. present a decision support framework for robust process enactment. They leverage adversarial game search (e.g Monte Carlo game tree search) to compute alternative flows in order to anticipate and account for possible ways in which the execution environment might impede a process from achieving its desired effects or outcomes. This notion can be extended for strategic decision-making as well, where decision-making is viewed through an abstraction of a two-player game. Here goal models (representing structured models of strategy) can be combined with game-tree search using augmented game trees to assist organizations in selecting amongst alternative goal refinements (OR-refinements). The final computational machinery, overall, provides business management with a resilient strategic decision-making framework that can reason the consequences of various decisions.

7. Discussion and Future Directions

In this section, we present an overarching picture of some of the key challenges and characteristics needed to develop future process support systems. We can also view these characteristics as feature requirements for building future business process management and decision support systems. We argue, that in addition to supporting the existing capabilities, the next generation of business process management systems will offer several additional analytics features such as process monitoring, resource allocation, risk management. Some of these have already been incorporated by existing vendor service and product offerings (e.g. Process Monitoring support in Apromore), while problems like agility support for unstructured knowledge-intensive processes still remain unaddressed.

In (Dumas et al., 2022) Dumas et. al present a vision of AI-Augmented Business Process Management Systems. In such systems execution flows are not pre-determined rather use AI technology to adapt and reason within a set of restrictions based on one or more performance indicators. This allows operates largely autonomously, within the boundaries set by the process frame. Keeping this in mind, We briefly discuss enabling AI technologies for supporting flexible execution of processes and attain the pre-defined goals of a given business process. One useful abstraction is to see Processes as a Sequential decision-making activity, where we must make a sequence of decisions in response to information about the outcomes of our actions as we proceed. Decisions can cover the management of given resources, interventions for achieving process goals, and supporting decisions or the kind discussed in this section. The problem gets further complicated when processes are deployed in stochastic environments, where the outcomes of our actions are uncertain. There exist many approaches for designing decision-making systems and the problem of “learning and decision making over time to achieve a goal” has been studied by multiple disciplines and they all provide interesting perspectives (Sutton, 2022). From an agent perspective, There are many methods for designing decision-making agents. They differ in the responsibilities of the designer and the tasks left to automation. We provide a brief overview of methods that can be applied to tackle sequential decision problems faced in process analytics.

Reinforcement Learning (Sutton and Barto, 2018) provides a framework for learning from interaction with the environment in order to achieve a goal (implicitly defined by the reward function). Reinforcement Learning has widely been used to model sequential decision problems and has shown great promise in solving large scale complex problems with long time horizons, partial observability, and high dimensionality of observation and action spaces(Berner et al., 2019).

Reinforcement Learning (RL) assumes that there is an agent operating in the real world. At each step tt the agent, Executes action AtA_{t}, Receives an observation OtO_{t} and Receives scalar reward RtR_{t}. The Problem can be formulated as a Markov Decision Process(Sutton and Barto, 2018) defined by (𝒮,𝒜,T,R)(\mathcal{S},\mathcal{A},T,R) tuples where 𝒮\mathcal{S} and 𝒜\mathcal{A} refer to the state and action spaces; T:𝒮×T:\mathcal{S}\times 𝒜𝒮\mathcal{A}\rightarrow\mathcal{S}, is the state transition function and R:𝒮×𝒜R:\mathcal{S}\times\mathcal{A}\rightarrow\mathbb{R} represents the reward function. The goal of the agent is to estimate an optimal policy π:𝒮𝒜\pi:\mathcal{S}\rightarrow\mathcal{A} or an optimal action value function qπ(s,a)=𝔼π[GtSt=s,At=a]q_{\pi}(s,a)=\mathbb{E}_{\pi}\left[G_{t}\mid S_{t}=s,A_{t}=a\right] which maximizes the expected return 𝔼[t=1LγtRtπ]\mathbb{E}\left[\sum_{t=1}^{L}\gamma^{t}R_{t}\mid\pi\right] over a given MDP(Sutton and Barto, 2018).

Reinforcement Learning can allow us to formulate process goals as the maximization of a cumulative reward which is a very powerful general-purpose idea. In process analytics, it can be applied to for example provide decision support in the form of interventions or action recommendations which are used to generate possibilities from which human workers can pick the best alternative given the additional context and experience they have access to. i.e RL allows us for general formulation of sequential decision problems under the assumption that the model is known and that the environment is fully observable.

Offline Reinforcement Learning in particular provides an excellent opportunity to build adaptive systems that learn from past experience and leverage feedback in the form of rewards. We can use offline Reinforcement Learning to learn the decision criteria representing optimal outcomes and use it to recommend optimal action based on the current state of the process. Offline Reinforcement Learning requires sufficiently diverse training data that is close to cases that system might encounter in the future. Further it requires that we define a reward function (manifested as the weighted soft goal score), which captures the goals and priorities of the specific process. A Reinforcement Learning based decision support system takes a state based view for supporting process related decisions. Such systems can consider context and recommend actions in each stage of the process(which assumes availability of effect log).

Building adaptive systems via BDI agents: optimal decision making in a sequential context requires reasoning about future sequences of actions and observations. In Artificial Intelligence, a rational agent will perceive its environment, use its internal knowledge base, along with reasoning and planning capabilities, to select actions that lead to the desired goal state(or close) according to some utility measure(Russell and Norvig, 2002). Agent-oriented programming deals with modelling and writing software systems with this concept of rational agent, in which each component, or agent, perceives the environment through sensors and acts on the environment with actuators(Shoham, 1993). The Belief-Desire-Intention (BDI) agent is a particularly popular and effective architecture for designing such agents. A typical BDI agent program consists of three components (Rao et al., 1995): First, a set of beliefs acquired potentially through sensor inputs. Secondly, a set of plans, where each plan has an associated triggering event, pre-defined context conditions and plan body consisting of a set of action sequences. Lastly, a set of goals that an agent wants to achieve. Goals are related to plans. i.e. Each goal can be achieved by executing plans in the plan library. There are many programming languages and platforms developed over the last few decades for implementing BDI-agent systems. Some of these languages include PRS (Procedural Reasoning System) (Ingrand et al., 1992), AgentSpeak(L) (Rao, 1996), Jack [25], dMARS (Distributed Multiagent Reasoning System) etc.

Mining agent programs allows organisation to quickly build agent programs that can potentially replace traditional software systems which are costly and hard to maintain. In (Xu et al., 2013) Xu et. al. propose a framework for learning BPI agent plans from process and effect logs. The authors propose a plan recognition framework for generating BDI style plans. The framework requires input in the form of behaviour logs generated by enterprise applications in order to mine a ‘draft’ version of agent code that can potentially replace some of the applications deployed inside the organisation. Specifically, a WF-Net is generated using ProM using a number of pre-defined transformation rules and then the generated WF-Net is transformed into a set of plans using the proposed algorithm. Lastly, effects logs are used to identify context which becomes the pre-condition for each of the extracted plan. The plans utilise both positive and negative example sets and uses norm learning mechanisms to infer normative plans.

Robotic Process Automation: Firms are interested in the identification of potential areas of automation to save costs and improve efficiency. Historically Business Process Management Systems (BPMSs) supported Business process automation (BPA) by executing process instances, supporting the distribution of work to process participants and delegating activities to various information systems deployed across the organization (e.g. checking the creditworthiness of an applicant) (Dumas, 2018).

Today we observe a rise of new technologies that can enable Business Process Automation by automating procedural work and supporting complex processes. Promising technologies like Robotic Process Automation(RPA) and Reinforcement Learning hold the promise of replacing human worker and automate repeated tasks. Task automation might mean replacing human workers entirely with intelligent agents, while decision automation means making decisions that humans previously made. In other scenarios, it facilitates organizations to automate human decision making. e.g. automated allocation of resources in complex knowledge-intensive scenarios might mean providing decision support(e.g. in the form of recommendations) for resources involved in process execution.

RPA represents an interesting shift, aiming to automate parts of business processes that consist of humans interacting with day-to-day software(e.g. transferring data from an Enterprise Resource Management system to a web application form). “Robotic Process Automation (RPA) is an emerging technology that allows organizations automating repetitive clerical tasks by executing scripts that encode sequences of fine-grained interactions with Web and desktop applications”. Alternatively “RPA is an umbrella term for tools that operate on the user interface of other computer systems in the way a human would do”

In the context of Process Analytics, RPA aims at automating business processes that consist of human interaction with software and provide decision support for resources involved in process execution. Implementing RPA is suitable in situations where processes that are too infrequent for traditional process automation to be profitable, but still repetitive enough to be formalized into an RPA process mode. We can identify opportunities of autommation from logs of interactions between workers and Web and desktop applications. Frameworks like Value-driven RPA (Kirchmer and Franz, 2019) are useful identify right sub-processes to automate, given the process context.

Automated Process Improvement: In traditional process mining, process optimization occurs as a result of post-mortem data analysis. In Automated Process Improvement, we strive for proactive improvement of business processes during process execution and attempt to automate as many aspects of a given process as possible. Automated Process Improvement is one of the most desired capability in process Analytics and can make processes more adaptive (Zur Muehlen and Shapiro, 2015). Here we are interested in an intelligent exploration of improvement strategies using domain knowledge and all available historical and current process-related data(generated by enterprise systems and sensory data).

Search-based optimization techniques allow us to identify and discover opportunities, for improving business processes, from event lots. We do so by considering various performance metrics. e.g the cycle time and process cost as key performance indicators(KPIs). Automated Process Improvement can identify various opportunities of improvement. It can also be used to streamline tasks execution where we identify control-flow related improvement opportunities. e.g re-ordering, merging and parallelization of tasks. It can also identify opportunities for task automation e.g Using RPA. Secondly, we can identify best Practices and give recommendations based on the analysis of best performing instances in the past. Similarly, Automated Process Improvement can also help with optimal resource allocation. As discussed earlier, we use historic data to come with up optimal resource allocation policies. For example, we can optimally design staff schedules by analyzing historical data and using process goals formulated as Key Performance Indicators and systematically evaluate the proposed changes. Lastly, we can optimize decision logic to improve the routing of the cases, which means adding or removing decision points or enhancing existing decision rules.

Gröger et. al. (Gröger et al., 2014) introduce the concept of recommendation-based business process optimization (rBPO). Such a system can generate action recommendations during process execution, enabling us to perform process optimization using pre-specified metrics. Apart from mining based approaches, predictive methods can generate action recommendations during process execution. Lastly, Bozorgi et. al. (Dasht Bozorgi et al., 2020) show that treatment recommendations (based on causal machine learning), when applied during the execution of a case can improve the overall outcome of a process.

8. Conclusion

Modern organizations routinely deploy process analytics, including process discovery and variant analysis techniques, both to gain insight into the reality of their operational processes and also to identify process improvement opportunities. Process analytics refers to the repertoire of techniques centred on process mining, predictive monitoring, decision and automation support. Process analytic approaches play a critical role in supporting the practice of Business Process Management and continuous process improvement by leveraging process-related data to identify performance bottlenecks, reducing costs, extracting insights, and optimizing the utilization of available resources. They also enable us to mine insights from process data (which encompasses process logs but include many other types of data as well), predict the behavior of process instances and provide operational and strategic decision support. In this work, we briefly surveyed the literature on process analytics and identified promising directions for future research.

References

  • (1)
  • Acs and Castelluccia (2012) Gergely Acs and Claude Castelluccia. 2012. Dream: Differentially private smart metering. arXiv preprint arXiv:1201.2531 (2012).
  • Adadi and Berrada (2018) Amina Adadi and Mohammed Berrada. 2018. Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI). IEEE Access 6 (2018), 52138–52160.
  • Alexander Seeliger and Muhlhauser (2021) Timo Nolle Alexander Seeliger, Stefan Luettgen and Max Muhlhauser. 2021. Learning of Process Representations Using Recurrent Neural Networks. In International Conference on Advanced Information Systems Engineering. Springer, 36–53.
  • Arias et al. (2016) Michael Arias, Eric Rojas, Jorge Munoz-Gama, and Marcos Sepúlveda. 2016. A framework for recommending resource allocation based on process mining. In International Conference on Business Process Management. Springer, 458–470.
  • Arulkumaran et al. (2017) Kai Arulkumaran, Marc Peter Deisenroth, Miles Brundage, and Anil Anthony Bharath. 2017. A brief survey of deep reinforcement learning. arXiv preprint arXiv:1708.05866 (2017).
  • Augusto et al. (2018) Adriano Augusto, Raffaele Conforti, Marlon Dumas, Marcello La Rosa, Fabrizio Maria Maggi, Andrea Marrella, Massimo Mecella, and Allar Soo. 2018. Automated discovery of process models from event logs: Review and benchmark. IEEE Transactions on Knowledge and Data Engineering 31, 4 (2018), 686–705.
  • Bamiah et al. (2012) Mervat Bamiah, Sarfraz Brohi, Suriayati Chuprat, et al. 2012. A study on significance of adopting cloud computing paradigm in healthcare sector. In 2012 International Conference on Cloud Computing Technologies, Applications and Management (ICCCTAM). IEEE, 65–68.
  • Barclay and Murray (1997) Rebecca O Barclay and Philip C Murray. 1997. What is knowledge management. Knowledge praxis 19, 1 (1997), 1–10.
  • Batoulis et al. (2015) Kimon Batoulis, Andreas Meyer, Ekaterina Bazhenova, Gero Decker, and Mathias Weske. 2015. Extracting decision logic from process models. In International conference on advanced information systems engineering. Springer, 349–366.
  • Baumgrass (2011) Anne Baumgrass. 2011. Deriving current state RBAC models from event logs. In 2011 Sixth International Conference on Availability, Reliability and Security. IEEE, 667–672.
  • Bazhenova et al. (2016) Ekaterina Bazhenova, Susanne Bülow, and Mathias Weske. 2016. Discovering decision models from event logs. In International Conference on Business Information Systems. Springer, 237–251.
  • Beheshti et al. (2020) Amin Beheshti, Shahpar Yakhchi, Salman Mousaeirad, Seyed Mohssen Ghafari, Srinivasa Reddy Goluguri, and Mohammad Amin Edrisi. 2020. Towards Cognitive Recommender Systems. Algorithms 13, 8 (2020), 176.
  • Benatallah et al. (2016) Boualem Benatallah, Sherif Sakr, Daniela Grigori, Hamid Reza Motahari-Nezhad, Moshe Chai Barukh, Ahmed Gater, Seung Hwan Ryu, et al. 2016. Process Analytics: concepts and techniques for querying and analyzing process data. Springer.
  • Berner et al. (2019) Christopher Berner, Greg Brockman, Brooke Chan, Vicki Cheung, Przemysław D\kebiak, Christy Dennison, David Farhi, Quirin Fischer, Shariq Hashme, Chris Hesse, et al. 2019. Dota 2 with large scale deep reinforcement learning. arXiv preprint arXiv:1912.06680 (2019).
  • Blagec et al. (2020) Kathrin Blagec, Georg Dorffner, Milad Moradi, and Matthias Samwald. 2020. A critical analysis of metrics used for measuring progress in artificial intelligence. arXiv preprint arXiv:2008.02577 (2020).
  • Brunk et al. (2020) Jens Brunk, Matthias Stierle, Leon Papke, Kate Revoredo, Martin Matzner, and Jörg Becker. 2020. Cause vs. effect in context-sensitive prediction of business process instances. Information Systems (2020), 101635.
  • Burattin et al. (2013) Andrea Burattin, Alessandro Sperduti, and Marco Veluscek. 2013. Business models enhancement through discovery of roles.. In CIDM. 103–110.
  • Cabanillas et al. (2013) Cristina Cabanillas, José María García, Manuel Resinas, David Ruiz, Jan Mendling, and Antonio Ruiz-Cortés. 2013. Priority-based human resource allocation in business processes. In International Conference on Service-Oriented Computing. Springer, 374–388.
  • Calvanese et al. (2021) Diego Calvanese, Sanja Lukumbuzya, Marco Montali, and Mantas Simkus. 2021. Process mining with common sense. In 2021 International Workshop on BPM Problems to Solve Before We Die, PROBLEMS 2021, Rome, September 6-10, 2021., Vol. 2938. CEUR-WS, 45–50.
  • Camargo et al. (2019) Manuel Camargo, Marlon Dumas, and Oscar González-Rojas. 2019. Learning accurate LSTM models of business processes. In International Conference on Business Process Management. Springer, 286–302.
  • Carmona (2020) Josep Carmona. 2020. Process Mining: Past, Present and (Likely) Future. XII Jornadas de Ciencia e Ingeniería de Servicios (JCIS2016) 220 (2020), 37.
  • Catalkaya et al. (2013) Semra Catalkaya, David Knuplesch, Carolina Chiao, and Manfred Reichert. 2013. Enriching business process models with decision rules. In International conference on business process management. Springer, 198–211.
  • Clifton et al. (2002) Chris Clifton, Murat Kantarcioglu, Jaideep Vaidya, Xiaodong Lin, and Michael Y Zhu. 2002. Tools for privacy preserving distributed data mining. ACM Sigkdd Explorations Newsletter 4, 2 (2002), 28–34.
  • Conforti et al. (2013a) Raffaele Conforti, Massimiliano De Leoni, Marcello La Rosa, and Wil MP Van Der Aalst. 2013a. Supporting risk-informed decisions during business process execution. In International Conference on Advanced Information Systems Engineering. Springer, 116–132.
  • Conforti et al. (2015) Raffaele Conforti, Massimiliano de Leoni, Marcello La Rosa, Wil MP van der Aalst, and Arthur HM ter Hofstede. 2015. A recommendation system for predicting risks across multiple business process instances. Decision Support Systems 69 (2015), 1–19.
  • Conforti et al. (2011) Raffaele Conforti, Giancarlo Fortino, Marcello La Rosa, and Arthur HM Ter Hofstede. 2011. History-aware, real-time risk detection in business processes. In OTM Confederated International Conferences” On the Move to Meaningful Internet Systems”. Springer, 100–118.
  • Conforti et al. (2013b) Raffaele Conforti, Marcello La Rosa, Giancarlo Fortino, Arthur HM Ter Hofstede, Jan Recker, and Michael Adams. 2013b. Real-time risk monitoring in business processes: A sensor-based approach. Journal of Systems and Software 86, 11 (2013), 2939–2965.
  • Conforti et al. (2012) Raffaele Conforti, Arthur HM ter Hofstede, Marcello La Rosa, and Michael Adams. 2012. Automated risk mitigation in business processes. In OTM Confederated International Conferences” On the Move to Meaningful Internet Systems”. Springer, 212–231.
  • Dasht Bozorgi et al. (2020) Zahra Dasht Bozorgi, Irene Teinemaa, Marlon Dumas, Marcello La Rosa, and Artem Polyvyanyy. 2020. Process Mining Meets Causal Machine Learning: Discovering Causal Rules from Event Logs. arXiv e-prints (2020), arXiv–2009.
  • Davis and Marcus (2015) Ernest Davis and Gary Marcus. 2015. Commonsense reasoning and commonsense knowledge in artificial intelligence. Commun. ACM 58, 9 (2015), 92–103.
  • De Leoni and van der Aalst (2013) Massimiliano De Leoni and Wil MP van der Aalst. 2013. Data-aware process mining: discovering decisions in processes using alignments. In Proceedings of the 28th annual ACM symposium on applied computing. 1454–1461.
  • Dees et al. (2019) Marcus Dees, Massimiliano de Leoni, Wil MP van der Aalst, and Hajo A Reijers. 2019. What if Process Predictions are not followed by Good Recommendations?(Technical Report). arXiv preprint arXiv:1905.10173 (2019).
  • Di Ciccio et al. (2012) Claudio Di Ciccio, Andrea Marrella, and Alessandro Russo. 2012. Knowledge-intensive Processes: An Overview of Contemporary Approaches.. In KiBP@ KR. 33–47.
  • Di Ciccio et al. (2015) Claudio Di Ciccio, Andrea Marrella, and Alessandro Russo. 2015. Knowledge-intensive processes: characteristics, requirements and analysis of contemporary approaches. Journal on Data Semantics 4, 1 (2015), 29–57.
  • Di Francescomarino et al. (2018) Chiara Di Francescomarino, Chiara Ghidini, Fabrizio Maria Maggi, and Fredrik Milani. 2018. Predictive process monitoring methods: Which one suits me best?. In International Conference on Business Process Management. Springer, 462–479.
  • Di Francescomarino et al. (2017) Chiara Di Francescomarino, Chiara Ghidini, Fabrizio Maria Maggi, Giulio Petrucci, and Anton Yeshchenko. 2017. An eye into the future: leveraging a-priori knowledge in predictive business process monitoring. In International Conference on Business Process Management. Springer, 252–268.
  • Diba et al. (2020) Kiarash Diba, Kimon Batoulis, Matthias Weidlich, and Mathias Weske. 2020. Extraction, correlation, and abstraction of event data for process mining. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 10, 3 (2020), e1346.
  • Doshi-Velez and Kim (2017) Finale Doshi-Velez and Been Kim. 2017. Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017).
  • Dumas (2018) Marlon Dumas. 2018. Business Process Analytics: From Insights to Predictions. In International Baltic Conference on Databases and Information Systems. Springer, 15–20.
  • Dumas et al. (2022) Marlon Dumas, Fabiana Fournier, Lior Limonad, Andrea Marrella, Marco Montali, Jana-Rebecca Rehse, Rafael Accorsi, Diego Calvanese, Giuseppe De Giacomo, Dirk Fahland, et al. 2022. Augmented business process management systems: a research manifesto. arXiv preprint arXiv:2201.12855 (2022).
  • Dumas et al. (2013a) Marlon Dumas, Marcello La Rosa, Jan Mendling, and Hajo A Reijers. 2013a. Business process management. Springer.
  • Dumas et al. (2013b) Marlon Dumas, Marcello La Rosa, Jan Mendling, Hajo A Reijers, et al. 2013b. Fundamentals of business process management. Vol. 1. Springer.
  • Dunkl et al. (2011) Reinhold Dunkl, Karl Anton Fröschl, Wilfried Grossmann, and Stefanie Rinderle-Ma. 2011. Assessing medical treatment compliance based on formal process modeling. In Symposium of the Austrian HCI and Usability Engineering Group. Springer, 533–546.
  • Dwork et al. (2006) Cynthia Dwork, Krishnaram Kenthapadi, Frank McSherry, Ilya Mironov, and Moni Naor. 2006. Our data, ourselves: Privacy via distributed noise generation. In Annual International Conference on the Theory and Applications of Cryptographic Techniques. Springer, 486–503.
  • Dwork et al. (2014) Cynthia Dwork, Aaron Roth, et al. 2014. The algorithmic foundations of differential privacy. Foundations and Trends in Theoretical Computer Science 9, 3-4 (2014), 211–407.
  • Easterby-Smith and Lyles (2011) Mark Easterby-Smith and Marjorie A Lyles. 2011. Handbook of organizational learning and knowledge management. Number 2nd ed. Wiley Online Library.
  • Eili et al. (2021) Mansoureh Yari Eili, Jalal Rezaeenour, and Mohammadreza Fani Sani. 2021. A Systematic Literature Review on Process-Aware Recommender Systems. arXiv preprint arXiv:2103.16654 (2021).
  • Elkoumy et al. (2020) Gamal Elkoumy, Stephan A. Fahrenkrog-Petersen, Marlon Dumas, Peeter Laud, Alisa Pankova, and Matthias Weidlich. 2020. Secure Multi-party Computation for Inter-organizational Process Mining. In Enterprise, Business-Process and Information Systems Modeling, Selmin Nurcan, Iris Reinhartz-Berger, Pnina Soffer, and Jelena Zdravkovic (Eds.). Springer International Publishing, Cham, 166–181.
  • Evermann et al. (2016) Joerg Evermann, Jana-Rebecca Rehse, and Peter Fettke. 2016. A deep learning approach for predicting process behaviour at runtime. In International Conference on Business Process Management. Springer, 327–338.
  • Evermann et al. (2017) Joerg Evermann, Jana-Rebecca Rehse, and Peter Fettke. 2017. Predicting process behaviour using deep learning. Decision Support Systems 100 (2017), 129–140.
  • Figl et al. (2018) Kathrin Figl, Jan Mendling, Gul Tokdemir, and Jan Vanthienen. 2018. What we know and what we do not know about DMN. Enterprise Modelling and Information Systems Architectures (EMISAJ) 13 (2018), 2–1.
  • Fink (2019) Arlene Fink. 2019. Conducting research literature reviews: From the internet to paper. Sage publications.
  • Frank et al. (2013) Mario Frank, Joachim M Buhman, and David Basin. 2013. Role mining with probabilistic models. ACM Transactions on Information and System Security (TISSEC) 15, 4 (2013), 1–28.
  • Galanti et al. (2020) Riccardo Galanti, Bernat Coma-Puig, Massimiliano de Leoni, Josep Carmona, and Nicolò Navarin. 2020. Explainable predictive process monitoring. In 2020 2nd International Conference on Process Mining (ICPM). IEEE, 1–8.
  • Ghasemi and Amyot (2020) Mahdi Ghasemi and Daniel Amyot. 2020. From event logs to goals: a systematic literature review of goal-oriented process mining. Requirements Engineering 25, 1 (2020), 67–93.
  • Ghattas et al. (2014) Johny Ghattas, Pnina Soffer, and Mor Peleg. 2014. Improving business process decision making based on past experience. Decision Support Systems 59 (2014), 93–107.
  • Goodman and Flaxman (2017) Bryce Goodman and Seth Flaxman. 2017. European Union regulations on algorithmic decision-making and a “right to explanation”. AI magazine 38, 3 (2017), 50–57.
  • Goodspeed (2004) Scott Winans Goodspeed. 2004. Translating strategy into action: The balanced scorecard. (2004).
  • Gou et al. (2017) Yingzhi Gou, Aditya Ghose, and Hoa Khanh Dam. 2017. Leveraging Game-tree search for robust process enactment. In International Conference on Advanced Information Systems Engineering. Springer, 461–476.
  • Gröger et al. (2014) Christoph Gröger, Holger Schwarz, and Bernhard Mitschang. 2014. Prescriptive analytics for recommendation-based business process optimization. In International Conference on Business Information Systems. Springer, 25–37.
  • Gusenbauer (2019) Michael Gusenbauer. 2019. Google Scholar to overshadow them all? Comparing the sizes of 12 academic search engines and bibliographic databases. Scientometrics 118, 1 (2019), 177–214.
  • Hamilton (2015) Booz Allen Hamilton. 2015. The field guide to data science.
  • Harrison-Broninski (2018) K Harrison-Broninski. 2018. Human Processes. BPTrends. (2018).
  • Hauder et al. ([n.d.]) Matheus Hauder, Simon Pigat, and Florian Matthes. [n.d.]. Research challenges in adaptive case management: a literature review. In 2014 IEEE 18th International Enterprise Distributed Object Computing Conference Workshops and Demonstrations. 98–107.
  • Heiskanen and Newman (1997) Ari Heiskanen and Michael Newman. 1997. Bridging the gap between information systems research and practice: the reflective practitioner as a researcher. ICIS 1997 Proceedings (1997), 8.
  • Hewelt and Weske (2016) Marcin Hewelt and Mathias Weske. 2016. A hybrid approach for flexible case modeling and execution. In International Conference on Business Process Management. Springer, 38–54.
  • Hoffman (2017) Bryce G Hoffman. 2017. Red Teaming: How Your Business Can Conquer the Competition by Challenging Everything. Crown business.
  • Hompes et al. (2017) Bart FA Hompes, Abderrahmane Maaradji, Marcello La Rosa, Marlon Dumas, Joos CAM Buijs, and Wil MP van der Aalst. 2017. Discovering causal factors explaining business process performance variation. In International Conference on Advanced Information Systems Engineering. Springer, 177–192.
  • Hornix (2007) Peter TG Hornix. 2007. Performance analysis of business processes through process mining. Master’s Thesis, Eindhoven University of Technology (2007).
  • Huang et al. (2012) Zhengxing Huang, Xudong Lu, and Huilong Duan. 2012. Resource behavior measure and application in business process management. Expert Systems with Applications 39, 7 (2012), 6458–6468.
  • Huang et al. (2011) Zhengxing Huang, Wil MP van der Aalst, Xudong Lu, and Huilong Duan. 2011. Reinforcement learning based resource allocation in business process management. Data & Knowledge Engineering 70, 1 (2011), 127–145.
  • Ingrand et al. (1992) François Felix Ingrand, Michael P Georgeff, and Anand S Rao. 1992. An architecture for real-time reasoning and system control. IEEE expert 7, 6 (1992), 34–44.
  • Jansen-Vullers et al. (2007) MH Jansen-Vullers, MWNC Loosschilder, PAM Kleingeld, and HA Reijers. 2007. Performance measures to evaluate the impact of best practices. In Proceedings of Workshops and Doctoral Consortium of the 19th International Conference on Advanced Information Systems Engineering (BPMDS workshop), Vol. 1. Tapir Academic Press Trondheim, 359–368.
  • Jensen et al. (2012) Peter B Jensen, Lars J Jensen, and Søren Brunak. 2012. Mining electronic health records: towards better research applications and clinical care. Nature Reviews Genetics 13, 6 (2012), 395–405.
  • Kemsley (2011) Sandy Kemsley. 2011. The changing nature of work: from structured to unstructured, from controlled to social. In International Conference on Business Process Management. Springer, 2–2.
  • Kerremans (2018) Marc Kerremans. 2018. Market guide for process mining. Gartner Inc (2018).
  • Khan et al. (2021) Asjad Khan, Aditya Ghose, and Hoa Dam. 2021. Decision Support for Knowledge Intensive Processes Using RL Based Recommendations. In Business Process Management Forum, Artem Polyvyanyy, Moe Thandar Wynn, Amy Van Looy, and Manfred Reichert (Eds.). Springer International Publishing, Cham, 246–262.
  • Khan et al. (2018) Asjad Khan, Hung Le, Kien Do, Truyen Tran, Aditya Ghose, Hoa Dam, and Renuka Sindhgatta. 2018. Memory-augmented neural networks for predictive process analytics. arXiv preprint arXiv:1802.00938 (2018).
  • Kirchmer and Franz (2019) Mathias Kirchmer and Peter Franz. 2019. Value-driven robotic process automation (RPA). In International Symposium on Business Modeling and Software Design. Springer, 31–46.
  • Kitchenham (2004) Barbara Kitchenham. 2004. Procedures for performing systematic reviews. Keele, UK, Keele University 33, 2004 (2004), 1–26.
  • Klinkmüller et al. (2019) Christopher Klinkmüller, Richard Müller, and Ingo Weber. 2019. Mining Process Mining Practices: An Exploratory Characterization of Information Needs in Process Analytics. In International Conference on Business Process Management. Springer, 322–337.
  • Kluza et al. (2019) Krzysztof Kluza, Weronika T Adrian, Piotr Wiśniewski, and Antoni Lig\keza. 2019. Understanding decision model and notation: DMN research directions and trends. In International Conference on Knowledge Science, Engineering and Management. Springer, 787–795.
  • Kohavi and Longbotham (2017) Ron Kohavi and Roger Longbotham. 2017. Online Controlled Experiments and A/B Testing. Encyclopedia of machine learning and data mining 7, 8 (2017), 922–929.
  • Koorn et al. (2020) Jelmer J Koorn, Xixi Lu, Henrik Leopold, and Hajo A Reijers. 2020. Looking for Meaning: Discovering Action-Response-Effect Patterns in Business Processes. In International Conference on Business Process Management. Springer, 167–183.
  • Kuhlmann et al. (2003) Martin Kuhlmann, Dalia Shohat, and Gerhard Schimpf. 2003. Role mining-revealing business roles for security administration using data mining technology. In Proceedings of the eighth ACM symposium on Access control models and technologies. 179–186.
  • Kumar et al. (2013) Akhil Kumar, Remco Dijkman, and Minseok Song. 2013. Optimal resource assignment in workflows for maximizing cooperation. In Business process management. Springer, 235–250.
  • Lang et al. (2008) Martin Lang, Thomas Bürkle, Susanne Laumann, and Hans-Ulrich Prokosch. 2008. Process mining for clinical workflows: challenges and current limitations. In MIE, Vol. 136. 229–234.
  • LeCun et al. (2015) Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. nature 521, 7553 (2015), 436–444.
  • Lenz and Reichert (2007) Richard Lenz and Manfred Reichert. 2007. IT support for healthcare processes–premises, challenges, perspectives. Data & Knowledge Engineering 61, 1 (2007), 39–58.
  • Lepenioti et al. (2020) Katerina Lepenioti, Alexandros Bousdekis, Dimitris Apostolou, and Gregoris Mentzas. 2020. Prescriptive analytics: Literature review and research challenges. International Journal of Information Management 50 (2020), 57–70. https://doi.org/10.1016/j.ijinfomgt.2019.04.003
  • Liu et al. (2014) Xingmei Liu, Jian Chen, Yu Ji, and Yang Yu. 2014. Q-learning algorithm for task allocation based on social relation. In International Workshop on Process-Aware Systems. Springer, 49–58.
  • Ly et al. (2005) Linh Thao Ly, Stefanie Rinderle, Peter Dadam, and Manfred Reichert. 2005. Mining staff assignment rules from event-based data. In International Conference on Business Process Management. Springer, 177–190.
  • Maggi et al. (2014) Fabrizio Maria Maggi, Chiara Di Francescomarino, Marlon Dumas, and Chiara Ghidini. 2014. Predictive monitoring of business processes. In International conference on advanced information systems engineering. Springer, 457–472.
  • Mahnaz Sadat Qafari (2020) Wil van der Aalst Mahnaz Sadat Qafari. 2020. Root Cause Analysis in Process Mining Using Structural Equation Models.
  • Mannhardt et al. (2016) Felix Mannhardt, Massimiliano De Leoni, Hajo A Reijers, and Wil MP Van Der Aalst. 2016. Decision mining revisited-discovering overlapping rules. In International Conference on Advanced Information Systems Engineering. Springer, 377–392.
  • Marin et al. (2016) Mike A. Marin, Matheus Hauder, and Florian Matthes. 2016. Case Management: An Evaluation of Existing Approaches for Knowledge-Intensive Processes. In Business Process Management Workshops, Manfred Reichert and Hajo A. Reijers (Eds.). Springer.
  • Márquez-Chamorro et al. (2017) Alfonso Eduardo Márquez-Chamorro, Manuel Resinas, and Antonio Ruiz-Cortes. 2017. Predictive monitoring of business processes: a survey. IEEE Transactions on Services Computing 11, 6 (2017), 962–977.
  • McSherry and Talwar (2007) Frank McSherry and Kunal Talwar. 2007. Mechanism design via differential privacy. In 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS’07). IEEE, 94–103.
  • Model (2016) Decision Model. 2016. Notation (DMN). Version 1.1. Object Management Group, Inc (2016).
  • Motahari-Nezhad and Bartolini (2011) Hamid Reza Motahari-Nezhad and Claudio Bartolini. 2011. Next best step and expert recommendation for collaborative processes in it service management. In International Conference on Business Process Management. Springer, 50–61.
  • Motahari-Nezhad and Swenson ([n.d.]) Hamid R Motahari-Nezhad and Keith D Swenson. [n.d.]. Adaptive case management: Overview and research challenges. In 2013 IEEE 15th Conf. on Business Informatics. 264–269.
  • Munoz-Gama et al. (2022) Jorge Munoz-Gama, Niels Martin, Carlos Fernandez-Llatas, Owen A Johnson, Marcos Sepúlveda, Emmanuel Helm, Victor Galvez-Yanjari, Eric Rojas, Antonio Martinez-Millana, Davide Aloini, et al. 2022. Process mining for healthcare: Characteristics and challenges. Journal of Biomedical Informatics 127 (2022), 103994.
  • Narendra et al. (2019) Tanmayee Narendra, Prerna Agarwal, Monika Gupta, and Sampath Dechu. 2019. Counterfactual Reasoning for Process Optimization Using Structural Causal Models. In International Conference on Business Process Management. Springer, 91–106.
  • Navarin et al. (2017) Nicolò Navarin, Beatrice Vincenzi, Mirko Polato, and Alessandro Sperduti. 2017. LSTM networks for data-aware remaining time prediction of business process instances. In 2017 IEEE Symposium Series on Computational Intelligence (SSCI). IEEE, 1–7.
  • Nguyen et al. (2016) Hoang Nguyen, Marlon Dumas, Marcello La Rosa, Fabrizio Maria Maggi, and Suriadi Suriadi. 2016. Business process deviance mining: review and evaluation. arXiv preprint arXiv:1608.08252 (2016).
  • Okoli (2015) Chitu Okoli. 2015. A guide to conducting a standalone systematic literature review. Communications of the Association for Information Systems 37, 1 (2015), 43.
  • Ouyang et al. (2009) Chun Ouyang, Marlon Dumas, Wil MP Van Der Aalst, Arthur HM Ter Hofstede, and Jan Mendling. 2009. From business process models to process-oriented software systems. ACM transactions on software engineering and methodology (TOSEM) 19, 1 (2009), 1–37.
  • Pearl et al. (2009) Judea Pearl et al. 2009. Causal inference in statistics: An overview. Statistics surveys 3 (2009), 96–146.
  • Pika et al. (2017) Anastasiia Pika, Michael Leyer, Moe T Wynn, Colin J Fidge, Arthur HM Ter Hofstede, and Wil MP Van Der Aalst. 2017. Mining resource profiles from event logs. ACM Transactions on Management Information Systems (TMIS) 8, 1 (2017), 1–30.
  • Rajan (2018) Renuka Sindhgatta Rajan. 2018. Data-Driven and Context-Aware Process Provisioning. Ph.D. Dissertation. University of Wollongong.
  • Rao (1996) Anand S Rao. 1996. AgentSpeak (L): BDI agents speak out in a logical computable language. In European Workshop on Modelling Autonomous Agents in a Multi-Agent World. Springer, 42–55.
  • Rao et al. (1995) Anand S Rao, Michael P Georgeff, et al. 1995. BDI agents: from theory to practice.. In ICMAS, Vol. 95. 312–319.
  • Reichert et al. (2015) Manfred Reichert, Alena Hallerbach, and Thomas Bauer. 2015. Lifecycle management of business process variants. In Handbook on Business Process Management 1. Springer, 251–278.
  • Rizzi et al. (2020) Williams Rizzi, Chiara Di Francescomarino, and Fabrizio Maria Maggi. 2020. Explainability in Predictive Process Monitoring: When Understanding Helps Improving. In International Conference on Business Process Management. Springer, 141–158.
  • Rosenfeld (2011) Austin Rosenfeld. 2011. BPM: structured vs. unstructured. Retrieved from www.bptrends.com (2011).
  • Rozinat and van der Aalst (2006) Anne Rozinat and Wil MP van der Aalst. 2006. Decision mining in ProM. In International Conference on Business Process Management. Springer, 420–425.
  • Russell et al. (2005) Nick Russell, Wil MP van der Aalst, Arthur HM Ter Hofstede, and David Edmond. 2005. Workflow resource patterns: Identification, representation and tool support. In International Conference on Advanced Information Systems Engineering. Springer, 216–232.
  • Russell and Norvig (2002) Stuart Russell and Peter Norvig. 2002. Artificial intelligence: a modern approach. (2002).
  • Sakr et al. (2018) Sherif Sakr, Zakaria Maamar, Ahmed Awad, Boualem Benatallah, and Wil MP Van Der Aalst. 2018. Business process analytics and big data systems: A roadmap to bridge the gap. IEEE Access 6 (2018), 77308–77320.
  • Santipuri et al. (2017) Metta Santipuri, Aditya Ghose, Hoa Khanh Dam, and Suman Roy. 2017. Goal orchestrations: Modelling and mining flexible business processes. In International Conference on Conceptual Modeling. Springer, 373–387.
  • Schmidhuber (2015) Jürgen Schmidhuber. 2015. Deep learning in neural networks: An overview. Neural networks 61 (2015), 85–117.
  • Schonenberg et al. (2008) Helen Schonenberg, Barbara Weber, Boudewijn Van Dongen, and Wil Van der Aalst. 2008. Supporting flexible processes through recommendations based on history. In International Conference on Business Process Management. Springer, 51–66.
  • Schönig et al. (2015) Stefan Schönig, Cristina Cabanillas, Stefan Jablonski, and Jan Mendling. 2015. Mining the organisational perspective in agile business processes. In Enterprise, Business-Process and Information Systems Modeling. Springer, 37–52.
  • Senderovich et al. (2014) Arik Senderovich, Matthias Weidlich, Avigdor Gal, and Avishai Mandelbaum. 2014. Mining resource scheduling protocols. In International Conference on Business Process Management. Springer, 200–216.
  • Shoham (1993) Yoav Shoham. 1993. Agent-oriented programming. Artificial intelligence 60, 1 (1993), 51–92.
  • Sindhgatta et al. (2014) Renuka Sindhgatta, Gaargi Banerjee Dasgupta, and Aditya Ghose. 2014. Analysis of operational data for expertise aware staffing. In International Conference on Business Process Management. Springer, 317–332.
  • Sindhgatta et al. (2015) Renuka Sindhgatta, Aditya Ghose, and Gaargi Banerjee Dasgupta. 2015. Learning ‘Good Quality’Resource Allocations from Historical Data. In Service-Oriented Computing-ICSOC 2014 Workshops. Springer, 84–95.
  • Song and Van der Aalst (2008) Minseok Song and Wil MP Van der Aalst. 2008. Towards comprehensive support for organizational mining. Decision Support Systems 46, 1 (2008), 300–317.
  • Suriadi et al. (2014) Suriadi Suriadi, Burkhard Weiß, Axel Winkelmann, Arthur HM ter Hofstede, Michael Adams, Raffaele Conforti, Colin Fidge, Marcello La Rosa, Chun Ouyang, Anastasiia Pika, et al. 2014. Current research in risk-aware business process management―overview, comparison, and gap analysis. Communications of the Association for Information Systems 34, 1 (2014), 52.
  • Sutton and Barto (2018) Richard Sutton and Andrew Barto. 2018. Reinforcement learning:An introduction. MIT-Press.
  • Sutton (2022) Richard S Sutton. 2022. The Quest for a Common Model of the Intelligent Decision Maker. arXiv preprint arXiv:2202.13252 (2022).
  • Tama and Comuzzi (2019) Bayu Adhi Tama and Marco Comuzzi. 2019. An empirical comparison of classification techniques for next event prediction using business process event logs. Expert Systems with Applications 129 (2019), 233–245.
  • Tardío and Peral (2015) Roberto Tardío and Jesús Peral. 2015. Obtaining key performance indicators by using data mining techniques. In International Conference on Conceptual Modeling. Springer, 144–153.
  • Tax et al. (2017) Niek Tax, Ilya Verenich, Marcello La Rosa, and Marlon Dumas. 2017. Predictive business process monitoring with LSTM neural networks. In International Conference on Advanced Information Systems Engineering. Springer, 477–492.
  • Taymouri et al. (2021a) Farbod Taymouri, Marcello La Rosa, Marlon Dumas, and Fabrizio Maria Maggi. 2021a. Business process variant analysis: Survey and classification. Knowledge-Based Systems 211 (2021), 106557.
  • Taymouri et al. (2021b) Farbod Taymouri, Marcello La Rosa, Marlon Dumas, and Fabrizio Maria Maggi. 2021b. Business process variant analysis: Survey and classification. Knowledge-Based Systems 211 (2021), 106557.
  • Teinemaa et al. (2019) Irene Teinemaa, Marlon Dumas, Marcello La Rosa, and Fabrizio Maria Maggi. 2019. Outcome-oriented predictive process monitoring: Review and benchmark. ACM Transactions on Knowledge Discovery from Data (TKDD) 13, 2 (2019), 1–57.
  • Vaidya et al. (2007) Jaideep Vaidya, Vijayalakshmi Atluri, and Qi Guo. 2007. The role mining problem: finding a minimal descriptive set of roles. In Proceedings of the 12th ACM symposium on Access control models and technologies. 175–184.
  • Van Beveren (2002) John Van Beveren. 2002. A model of knowledge acquisition that refocuses knowledge management. Journal of knowledge management (2002).
  • Van Der Aalst et al. (2011) Wil Van Der Aalst, Arya Adriansyah, Ana Karla Alves De Medeiros, Franco Arcieri, Thomas Baier, Tobias Blickle, Jagadeesh Chandra Bose, Peter Van Den Brand, Ronald Brandtjen, Joos Buijs, et al. 2011. Process mining manifesto. In International Conference on Business Process Management. Springer, 169–194.
  • Van der Aalst (2009) Wil MP Van der Aalst. 2009. Process-aware information systems: Lessons to be learned from process mining. In Transactions on petri nets and other models of concurrency II. Springer, 1–26.
  • Van der Aalst (2016) Wil MP Van der Aalst. 2016. Process mining: data science in action. Springer.
  • Van der Aalst et al. (2011) Wil MP Van der Aalst, M Helen Schonenberg, and Minseok Song. 2011. Time prediction based on process mining. Information systems 36, 2 (2011), 450–475.
  • Van der Aalst and Song (2004) Wil MP Van der Aalst and Minseok Song. 2004. Mining social networks: Uncovering interaction patterns in business processes. In International conference on business process management. Springer, 244–260.
  • van der Aalst et al. (2003) Wil MP van der Aalst, Moniek Stoffele, and JWF Wamelink. 2003. Case handling in construction. Automation in Construction 12, 3 (2003), 303–320.
  • van Dongen et al. (2008) Boudewijn F van Dongen, Ronald A Crooy, and Wil MP van der Aalst. 2008. Cycle time prediction: When will this case finally be finished?. In OTM Confederated International Conferences” On the Move to Meaningful Internet Systems”. Springer, 319–336.
  • Verenich (2016) Ilya Verenich. 2016. A general framework for predictive business process monitoring. Proceedings of CAiSE 2016 Doctoral Consortium: (2016), 1–9.
  • Verenich et al. (2019a) Ilya Verenich, Marlon Dumas, Marcello La Rosa, and Hoang Nguyen. 2019a. Predicting process performance: A white-box approach based on process models. Journal of Software: Evolution and Process 31, 6 (2019), e2170.
  • Verenich et al. (2019b) Ilya Verenich, Marlon Dumas, Marcello La Rosa, Fabrizio Maria Maggi, and Irene Teinemaa. 2019b. Survey and cross-benchmark comparison of remaining time prediction methods in business process monitoring. ACM Transactions on Intelligent Systems and Technology (TIST) 10, 4 (2019), 1–34.
  • vom Brocke et al. (2021) Jan vom Brocke, Mieke Jans, Jan Mendling, and Hajo A Reijers. 2021. A Five-Level Framework for Research on Process Mining.
  • Vom Brocke and Rosemann (2010) Jan Vom Brocke and Michael Rosemann. 2010. Handbook on business process management 2. Springer.
  • Von Rosing et al. (2014) Mark Von Rosing, Henrik Von Scheel, and August-Wilhelm Scheer. 2014. The Complete Business Process Handbook: Body of Knowledge from Process Modeling to BPM, Volume 1. Vol. 1. Morgan Kaufmann.
  • Wang and Han (2004) Jianyong Wang and Jiawei Han. 2004. BIDE: Efficient Mining of Frequent Closed Sequences. In Proceedings of the 20th International Conference on Data Engineering. 79–90. https://doi.org/10.1109/ICDE.2004.1319986
  • Weinzierl et al. (2020a) Sven Weinzierl, Sebastian Dunzer, Sandra Zilker, and Martin Matzner. 2020a. Prescriptive business process monitoring for recommending next best actions. In International Conference on Business Process Management. Springer, 193–209.
  • Weinzierl et al. (2020b) Sven Weinzierl, Sandra Zilker, Jens Brunk, Kate Revoredo, A Nguyen, Martin Matzner, Jörg Becker, and Björn Eskofier. 2020b. An empirical comparison of deep-neural-network architectures for next activity prediction using context-enriched process event logs. arXiv preprint arXiv:2005.01194 (2020).
  • Weske (2019) M Weske. 2019. Business Process Management–Concepts, Languages, Architectures, Verlag. Berlin (2019).
  • Wetzstein et al. (2009) Branimir Wetzstein, Philipp Leitner, Florian Rosenberg, Ivona Brandic, Schahram Dustdar, and Frank Leymann. 2009. Monitoring and analyzing influential factors of business process performance. In 2009 IEEE International Enterprise Distributed Object Computing Conference. IEEE, 141–150.
  • Xu et al. (2013) Hongyun Xu, Bastin Tony Roy Savarimuthu, Aditya Ghose, Evan Morrison, Qiying Cao, and Youqun Shi. 2013. Automatic BDI plan recognition from process execution logs and effect logs. In International Workshop on Engineering Multi-Agent Systems. Springer, 274–291.
  • Zhao and Zhao (2014) Weidong Zhao and Xudong Zhao. 2014. Process mining from the organizational perspective. In Foundations of intelligent systems. Springer, 701–708.
  • Zur Muehlen and Shapiro (2015) Michael Zur Muehlen and Robert Shapiro. 2015. Business process analytics. In Handbook on Business Process Management 2. Springer, 243–263.