Skip to content

GitHub repository of Explainable Predictive Process Monitoring: Evaluation Metrics and Guidelines for Process Outcome Prediction

Notifications You must be signed in to change notification settings

AlexanderPaulStevens/Explainability-in-Process-Outcome-Prediction

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Explainability in Process Outcome Prediction: Guidelines to Obtain Interpretable and Faithful Models

Complementary code to reproduce the work of "Explainable Predictive Process Monitoring: Evaluation Metrics and Guidelines for Process Outcome Prediction"

Guidelines for accurate and eXplainable Models for process Outcome Prediction

This figure contains the guideline to obtain eXplainable Models for Outcome Prediction (X-MOP) An overview of the files and folders:

The different folders

figures

The folder figures contain the high-resolution figures (PDF format) that have been used in the paper.

labeled_logs_csv_processed

This folder contains cleaned and preprocessed event logs that are made available by this GitHub repository: Benchmark for outcome-oriented predictive process monitoring. They provide 22 event logs, and we have selected 13 of them. The authors of this work an GitHub repository provide a Google drive link to download these event logs.

metrics

The metrics introduced in this paper are created in a seperate class (and .py) file. Hopefully, this will support reproducibility. The metrics are intended to work with all the benchmark models as mentioned in the paper. If there are any problems or questions, feel free to contact me (corresponding author and email related to the experiment of this study can be found on this site.

models

This folder contains only one file (LLM.py), where LLM model (originally written in R) is translated to Python. Note that there are some different design choices made compared to the original paper (link: here).

params_dir

The different hyperparameter files (in .pickle and .csv)

results_dir

The different result files (in .csv) for the ML models (results_dir_ML) and the DL models (results_dir_DL).

transformers

This folder contains the files to perform the aggregation sequence encoding (code stems from Teinemaa et. al. (2019)).

The different .py files

Preprocessing files

The preprocessing and hyperoptimalisation are derivative work based on the code provided by Outcome-Oriented Predictive Process Monitoring: Review and Benchmark. We would like to thank the authors for the high quality code that allowed to fastly reproduce the provided work.

  • dataset_confs.py
  • DatasetManager.py
  • EncoderFactory.py

Hyperoptimalisation of parameters

  • Hyperopt_ML.py
  • Hyperopt_DL (GC).ipynb

Training of the Machine Learning Models

Logistic Regression (LR), Logit Leaf Model (LLM), Generalized Logistic Rule Regression (GLRM), Random Forest (RF) and XGBoost (XGB)

  • Experiment_ML.py

Training of the Deep Learning Models (with Google Colab)

Long short-term memory neural networks (LSTM) and Convolutional Neural Network( CNN)

  • experiment_DL (GC).ipynb

We acknowledge the work provided by Building accurate and interpretable models for predictive process analytics for their attention-based bidirectional LSTM architecture to create the long short-term neural networks with attention layers visualisations.

About

GitHub repository of Explainable Predictive Process Monitoring: Evaluation Metrics and Guidelines for Process Outcome Prediction

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published