What Is Process Intelligence (and why should you care)?
Process Intelligence is a term that you might have encountered a bit more often these days, and one big driver of this is the popularity of Process and Task Mining. However, there are also voices out there who think that there is more to Process Intelligence than mining, and some who are thinking it is just a subset of traditional Business Intelligence (BI) and wonder what all the buzz is about.
This article attempts to bring some clarity into this conversation, starting with an analyst definition and a spin on this, while then looking at the underlying problem, the new(-ish) approach to it, and the different methods that exist in some cases for decades and already had their “day in the sun.”
What is Process Intelligence?
Let’s start our conversation with looking at how Bernhard Schaffrik from Forrester defines Process Intelligence:
It is interesting that Forrester is focusing on the software more than the discipline, which might be a function of the context in which the analyst report was written, but I have seen a lot of instances where clients were seeing Process Intelligence as a pure software category spiel. However, nothing could be further from the truth and we will discuss the different aspects of Process Intelligence further below and in a separate series of articles in which we have a look at the six steps for a successful process mining project.
So, what is the specialty of Process Intelligence according to Forrester? It is the mining technology, the focus on the complete process that includes not only system tasks (which might be a challenge to capture if you just use process mining software and not also task mining software), and the focus on decision-making … but notice the word “enabling” – he didn’t say “provide”, which is a reflection of the fact that there is a lot of human work to be done to get to the point where one can make a decision about process or other improvements that then later will be implemented in projects.
What is the problem?
Let’s take a step back and have a look at the problem that exist in how organizations run their processes.
There is a huge lack of visibility in all processes that don’t create physical goods that you can watch, like a production process, or a retail sales process where you can see the capacity of the process with the naked eye, answering the question of “what is the output of my process?” For most processes, the output is somewhat hidden – it is the amount of loans that have been underwritten or the number of candidates being hired and ready for onboarding. We also cannot easily see how long a process or its instances takes and if Service Level Agreements are met.
But these are no new problems and there were attempts to bring in this visibility by using process analysis techniques and BI in the past, and a lot of people have made money by providing services to create visibility and analyzing processes.
The ultimate goal in all of these efforts was (and is) that you change the execution of the runtime processes and create results and changes. This can be changes in the process itself, the organization, systems, etc. And you need to be able to measure the performance of these changes then as well. The graphic below shows the problem in an overview.
The new characteristic of process analysis
The key aspect of the problem with process visibility is not that we cannot measure process performance – back in the day people were observing the process with stopwatches and gathered the results on sheets of papers – but that there is new focus on technology that allows an automated capturing of visibility data that has a much higher quality and objectivity than anything else that has been captured manually before.
Mining technologies allow to capture *all* activities – as Forrester has defined above – independent if they leave traces in systems (Process Mining) or are captured by observing unstructured activities on user machines (Task Mining). They also calculate metrics that can give context and insights out-of-the-box, like durations (minimum, maximum, average, median times) and so on, which would take some effort to calculate otherwise.
This allows analysts not only to get to results faster, but they also are based on facts that are in the data set and there is just a limited amount of conversation possible – for example when the context of a situation is needed when you see an interesting visualization in the process mining charts. This improves the quality of the decisions made significantly because the subjectivity – “is what I hear from that Subject Matter Expert the main information, or is he telling me what he wants me to hear?” – will be taken out of it.
The big challenge with mining (and other) tools is that they just support decision-making, but not make the decision for you. You could trust generative AI tools for this, and we will see a lot of them being built into any tool that we use soon, but at the core they are just guessing the next word of an answer by calculating probabilities.
Would you risk the future of your company by trusting a probability calculation that sometimes can lead to “hallucinations” of the AI software? I am skeptical of this and rather see AI support more traditional process analysis techniques in the form of “little helpers” that save time in doing the busywork, but have a human identify recommendations and make decisions on changes.
How do various analysis techniques fit into this continuum?
Process analysis is not a new thing, and some of the methods have been around for decades and had their heyday in some cases decades ago, but are still relevant. And we will see an integration of AI into the underlying tools, which might make them more powerful and accessible for users.
The figure below shows various process analysis methods mapped on the Venn diagram:
Let’s have a quick look at each method:
- Business Intelligence is the practice of creating visibility in how the business is doing by connecting to source systems and visualize metrics in dashboards. It is the grandfather of what is done now in modern mining tools, but the success of it depends on the creativity and skill of the creator of a dashboard to convey the right information to the right stakeholder.
One additional drawback of BI is also the fact that mostly lagging indicators are used for the visualization and predictive technology is not widely implemented in the BI tools. They are also not necessarily process-oriented, even though some vendors provide process mining plug-ins for BI tools like PowerBI or Qlic. - Execution Monitoring is typically built into automation / workflow tools and can be configured to show process metrics. Unfortunately, they typically are focussed on the currently running processes in the system (that run in the one system that has this functionality built in, but might have difficulties to include data from other runtime systems in the process) and are not focussing on creating a higher-level summary of all processes and help identifying patterns of process performance. Independent of the shortcomings they provide insights into the execution and have alerting/notification capabilities if something goes wrong. And some of the runtime tools also allow to export data that then can be used in traditional BI tools or dashboards.
- Process/Task Mining tools fill that gap. They are designed to be process-centric and been fed by multiple systems in the process. However, they might have different levels of analysis tools built in and cannot bring in non-process metrics for visualization purposes. Some of the Process Mining tools can export data to BI tools and have machine-learning/AI capabilities that can give predictions on how cases might end in the future. They can also trigger notifications/alerts and trigger actions in runtime tools. However, they are not necessarily complete regarding their analysis capabilities and Simulation or detailed statistical analysis might be in a different module of a vendor’s platform.
- Six Sigma is a method to fill this gap. It is meant to calculate deviations in metrics that typically are used to improve quality of processes (“How well is it executed?”) and reducing defects.
Sometimes this can be a bit “nerdy” for the regular analyst and go into statistical details that they might see as “overdoing things”. Independent of that the statistical and data-driven approach that Six Sigma brings to the table will improve the quality of the analysis and a solid understanding of the statistical approach will be helpful in mining projects as well. - Lean is another methodology that aims to find waste in processes, such as overproduction, defects, inventory, etc. Those shall be optimized and can be helpful to give your mining projects an initial focus if your client has trouble finding hypotheses for (bad) process performance. It also comes with various visualization and analysis approaches, like Fishbone Diagrams, that are also helpful in other analysis types – and it is often combined with Six Sigma. I highly recommend to gain a good understanding of the Lean method for the average Business Analyst without the need for a formal “belt” certification (Green Belt, Black Belt, and so on).
- For a long time Process Modeling was the analysis technique of choice – visualize the content in a diagram and find consensus how work is done or should be done. Add objectives and measurements to them (current state and future state) and run a couple of reports on the diagrams to find improvement ideas. In a more advanced approach to this, tools like Process Simulation was used to create a performance baseline and create scenarios to resolve bottlenecks and other unwanted behaviors of the process.
- Interviews are the easiest analysis technique – talk with as many people as possible (process performers, process owners, system owners, etc.) and see what “bubbles up” or if your initial understanding of the problems are validated.
Even though it sound as if Interviews are the least professional analysis technique, they are still needed in many other approaches as well – the interview partners who have done the process for many, many years will give you the context and the details of the process, and thereby deliver the context for your analysis and ultimately the improvement opportunities that you will identify. However, the risk with this technique is that you are dealing with people and they are highly subjective – they might just tell you what they want you to take away, they might have a biased perspective on the process and its problems, or just think of process exceptions while the majority of process instances might run smoothly within the defined service levels. - The last technique on the list is somewhat new and might be applied to all other methods: data storytelling and dashboarding. While dashboards are used in BI and other approaches, they might be seen as a technical thing that visualizes information. The risk is that there is no context to the data that is shown. The context that you get from interviews will allow you to tell a story with the underlying data and therefore transform it into information and knowledge … and sometimes wisdom when you can see what you might not want to do 😉 Data storytelling is a skill that I have not seen in a lot of Business Analysts or Dashboard Developers, which might approach the creation of them from a more technical view (“Which widget do you want for that data and where on the canvas should it be?”).
As you might have guessed by now, to do a proper process analysis you will have to take pieces of each methodology and combine them to your situation. If you are doing a process mining project for example you will have to do interviews with stakeholders but also extract and transform process data to show them in dashboard, that are aligned to the hypotheses of process performance that drive your analysis.
What does that mean for me implementing Process Intelligence?
Given the information above, it is clear that the Forrester definition falls a bit short – and it slaps the title “Process Intelligence” on Process/Task Mining tools. When you start with your Process Intelligence implementation, I recommend to focus on the three steps that we outlined in the beginning:
- Create visibility: this can be the creation of dashboards (as the BI folks do) or the creation of models and diagrams, which is what traditional Business Analysts have done and then add performance data to these artifacts.
In both cases the objective is clear – get a performance baseline and make it visible, because only what you see can be improved. - Analyze the process: combining data-driven and human-focused approaches is the right way going forward. What you want, though, is an analysis result that is aligned to objective goals and not being tinted by the perspective of -important- stakeholders. Your recommendations of opportunities should be based on facts and be reproducible and understandable by another person.
- Execution: the best ideas are worth nothing if you don’t change things and make them better. This means that your recommendations need to be realistic from an implementation perspective – this includes things like feasibility, cost, or organizational/technical readiness for your ideas.
When you implement your change, you should always think of the KPIs and metrics that you have defined and make it an integral part of your solution design to capture those and compare them with the baseline metrics that you identified in the previous steps and were used in your process analysis.
Roland Woldt is a well-rounded executive with 25+ years of Business Transformation consulting and software development/system implementation experience, in addition to leadership positions within the German Armed Forces (11 years).
He has worked as Team Lead, Engagement/Program Manager, and Enterprise/Solution Architect for many projects. Within these projects, he was responsible for the full project life cycle, from shaping a solution and selling it, to setting up a methodological approach through design, implementation, and testing, up to the rollout of solutions.
In addition to this, Roland has managed consulting offerings during their lifecycle from the definition, delivery to update, and had revenue responsibility for them.
Roland has had many roles: VP of Global Consulting at iGrafx, Head of Software AG’s Global Process Mining CoE, Director in KPMG’s Advisory (running the EA offering for the US firm), and other leadership positions at Software AG/IDS Scheer and Accenture. Before that, he served as an active-duty and reserve officer in the German Armed Forces.