Fast and Accurate Declare Mining

This post has 274 words. Reading it will take approximately 1 minute.

Presentation I’ll give in two weeks.  How to be significantly faster at mining Declare than any existing tool, and at the same time get much better models.

Abstract

Declarative models make it possible to describe changing systems or systems with a lot of freedom by imposing global constraints instead of explicitly defining the flow of control. State-of-the-art process mining of declarative models either limits the types of constraints that can be mined, impose assumptions that are not always realistic, spend prohibitively long time on the mining process, or several of these. For example, contemporary Declare mining cannot cope with a process with two sub-processes, one rare and expensive, and one common and cheap. Contemporary approaches would be unable to discover anything meaningful for the arguably more important expensive sub-process.

In this presentation, I present a highly efficient means to mine Declare models. Our approach does away with all previous assumptions and limitations, and instead checks these. This means we can potentially construct much better models as we have much more information available. This allows us to also provide sensible data in the case with the two sub-processes. We use a number of existing and new techniques to achieve this, including symmetry reduction, parallelization, superscalar mining, and prefix sharing.

Significantly improved speed (from days or weeks to seconds or minutes) also makes it possible to use Declare mining as a sub-procedure for more general mining algorithms. For example, we can implement the Alpha algorithm from scratch in minutes (and extend it to cope with error for free), or we can discover block-structured models with ease.

Everything presented is implemented in ProM and available now.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.