Table 1: Errors (avg±stdDev) made by
AA-TP
and its com-
petitors, when using the bag abstraction mode.
AA-TP
(IB-k)
AA-TP
(RepTree)
CA-TP FSM
rmse
0.205±0.125 0.203±0.082 0.291±0.121 0.505±0.059
mae
0.064±0.058 0.073±0.033 0.142±0.071 0.259±0.008
mape
0.119±0.142 0.189±0.136 0.704±0.302 0.961±0.040
Table 2: Errors (avg±stdDev) made by
AA-TP
and its com-
petitors, when using the set abstraction mode.
AA-TP
(IB-k)
AA-TP
(RepTree)
CA-TP FSM
rmse
0.287±0.123 0.286±0.084 0.750±0.120 0.752±0.037
mae
0.105±0.061 0.112±0.035 0.447±0.077 0.475±0.009
mape
0.227±0.131 0.267±0.060 2.816±0.303 2.892±0.206
target performance measure, we will measure predic-
tion effectiveness by way of three classic error metrics
(computed via 10-fold cross validation): root mean
squared error (rmse), mean absolute error (mae), and
mean absolute percentage error (mape). For an easier
interpretation of results, the former two metrics will
be normalized w.r.t. the average dwell-time (ADT),
i.e., the average length of stay over all the containers
that passed through the terminal. In this way, all the
quality metrics will be dimensionless (and hopefully
ranging over [0,1]). Moreover, for the sake of sta-
tistical significance, all the error results shown in the
following have been averaged over 10 trials.
5.2 Test Results: Tuning of Parameters
We tried our approach (referred to as
AA-TP
here-
inafter) with different settings of its parameters, in-
cluding the base regression method (REGR) for in-
ducing the PPM of each discovered cluster. For the
sake of simplicity, we here only focus on the usage of
two basic regression methods: classic Linear regres-
sion (Draper and Smith, 1998), and the tree-based re-
gression algorithm RepTree (Witten and Frank, 2005).
In addition, we consider the case where each PPM
model simply encodes a k-NN regression procedure
(denoted by IB-k hereinafter), as a rough term of com-
parison with the family of instance-based regression
methods (including, in particular, the approach in (van
Dongen et al., 2008)). For all of the above regres-
sion methods, we reused the implementations avail-
able in the popular data-mining library Weka (Frank
et al., 2005). We remind that the other parameters
are: minSupp, (i.e., the minimum support for frequent
patterns), kTop (i.e., the maximal number of patterns
to keep, and then use in the clustering), and maxGap
(i.e., the maximal number of spurious events allowed
to occur among those of a given pattern, in a trace that
supports this latter). Figure 4 allows for analyzing the
three kinds of errors varies, respectively, when using
different regression methods (distinct curves are de-
picted for them), and different values of the parame-
ters (namely, maxGap = 0,4,8, ∞, kTop = 4,∞, and
minSupp = 0.1,... ,0.4).
Clearly, the underlying regression method is the
factor exhibiting a stronger impact on precision re-
sults. In particular, the disadvantage of using linear
regression is neat (no matter of the error metrics),
whereas both IB-k and RepTree methods performs
quite well, and very similarly. This is good news, es-
pecially for the RepTree method, which is to be pre-
ferred to IB-k for scalability reasons. Indeed, this lat-
ter may end up being too time-consuming at run-time,
when a large set of example traces must be kept – even
though, differently from pure instance-based methods
(like (van Dongen et al., 2008)), we do not need to
search across the whole log, but only within a single
cluster (previously selected, based on context data).
As the to remaining parameters, it is easily seen
that poorer results are obtained when minSupp = 0.1
and kTop = 4, as well as when minSupp = 0.4. As
a matter of fact, the former case epitomizes the cases
where we cut little (according to frequency) during
the generation of patterns, while trying to reduce their
number in the filtering phase; the latter, instead, is an
opposite situation where a rather high threshold sup-
port threshold is employed, at a higher risk of loos-
ing important pieces of information on process be-
haviour. In more detail, in the former case, the nega-
tive outcome is alleviated when setting maxGap = 0,
i.e., the patterns are required to exactly match a seg-
ment (i.e., subsequence of contiguous elements) in
their supporting traces. It is worth noticing that, apart
from these extreme cases,
AA-TP
exhibits good stabil-
ity and robustness, over a wide range of parameter set-
tings. Remarkably, when minSupp gets a value from
[0.2,... ,0.3], the remaining two parameters, namely
kTop and maxGap, do not seem to affect the quality
of predictions at all. In practice, it suffices to choose
carefully the regression method (and a middling value
of minSupp) to ensure good and stable prediction out-
comes, no matter of the other parameters – which
would be, indeed, quite harder to tune in general.
5.3 Comparison with Competitors
Let us finally compare our approach with two other
ones, defined in the literature for the discovery of a
PPM:
CA-TP
(Folino et al., 2012) and
FSM
(van der
Aalst et al., 2011). Tables 1 and 2 reports the average
errors (and associated standard deviations) made by
system
AA-TP
, while varying and the base regression
method (namely, Linear, RepTree and IB-k). In par-
ICEIS2013-15thInternationalConferenceonEnterpriseInformationSystems
64