5 CONCLUSIONS
AND OUTLOOK
In this article, we first presented a probabilistic clas-
sifier based on mixture models (CMM) that can be
used in the field of data mining to extract classifica-
tion rules from labeled sample data. Then, we defined
some objective interestingness measures that are tai-
lored to measure various aspects of the rules of which
this classifier consists. These measures are also based
on probabilistic methods. A data miner may use these
measures to investigate the knowledge extracted from
sample data in more detail. In three case studies using
well-known data sets we demonstrated the application
of our approach.
In our future work we will investigate the pos-
sibility to apply our objective measures to each of
the C class-specific mixture models to obtain an even
more detailed class-specific assessment of the compo-
nents. In this work we used the measures as a post-
processing step to prune a trained model. However,
it is also possible to use them as side conditions in
the objective functions that are used for the training
of CMM in order to support certain properties of a
classifier already during training. Additionally, we
will investigate how the measures can be combined
to perform a ranking of rules based on their interest-
ingness. There is a close relation of CMM to certain
kinds of fuzzy classifiers concerning the functional
form as outlined in (Fisch et al., 2010). Thus, it would
also be interesting to transfer the proposed measures
to that kind of classifiers and compare them to other
measures.
REFERENCES
Atzmueller, M., Baumeister, J., and Puppe, F. (2004).
Rough-fuzzy MLP: modular evolution, rule genera-
tion, and evaluation. In 15th International Conference
of Declarative Programming and Knowledge Man-
agement (INAP-2004), pages 203–213, Potsdam, Ger-
many.
Basu, S., Mooney, R. J., Pasupuleti, K. V., and Ghosh, J.
(2001). Evaluating the novelty of text-mined rules us-
ing lexical knowledge. In Proceedings of the seventh
ACM SIGKDD International Conference on Knowl-
edge Discovery and Data Mining (KDD 2001), pages
233–238, San Francisco, CA.
Bishop, C. (1994). Novelty detection and neural network
validation. IEE Proceedings – Vision, Image, and Sig-
nal Processing, 141(4):217–222.
Bishop, C. M. (2006). Pattern Recognition and Machine
Learning. Springer, New York, NY.
Di Fiore, F. (2002). Visualizing interestingness. In Zanasi,
A., Brebbia, C., Ebecken, N., and Melli, P., editors,
Data Mining III. WIT Press, Southampton, U.K.
Duda, R. O., Hart, P. E., and Stork, D. G. (2001). Pattern
Classification. John Wiley & Sons, Chichester, New
York, NY.
Fayyad, U. M., Piatetsky-Shapiro, G., and Smyth, P. (1996).
Knowledge discovery and data mining: Towards a
unifying framework. In Proceedings of the Second In-
ternational Conference on Knowledge Discovery and
Data Mining (KDD 1996), pages 82–88, Portland,
OR.
Fisch, D., K
¨
uhbeck, B., Ovaska, S. J., and Sick, B. (2010).
So near and yet so far: New insight into properties of
some well-known classifier paradigms. Information
Sciences, 180:3381–3401.
Fisch, D. and Sick, B. (2009). Training of radial ba-
sis function classifiers with resilient propagation and
variational Bayesian inference. In Proceedings of the
International Joint Conference on Neural Networks
(IJCNN ’09), pages 838–847, Atlanta, GA.
Frank, A. and Asuncion, A. (2010). UCI machine learning
repository.
Hebert, C. and Cremilleux, B. (2007). A unified view of
objective interestingness measures. In Perner, P., ed-
itor, Machine Learning and Data Mining in Pattern
Recognition, number 4571 in LNAI, pages 533–547.
Springer, Berlin, Heidelberg, Germany.
Hilderman, R. J. and Hamilton, H. J. (2001). Knowledge
Discovery and Measures of Interest. Kluwer Aca-
demic Publishers, Norwell, MA.
Liu, B., Hsu, W., Chen, S., and Ma, Y. (2000). Analyz-
ing the subjective interestingness of association rules.
IEEE Intelligent Systems, 15(5):47–55.
McGarry, K. (2005). A survey of interestingness measures
for knowledge discovery. The Knowledge Engineer-
ing Review, 20(1):39–61.
Nauck, D. D. (2003). Measuring interpretability in rule-
based classification systems. In Proceedings of the
12th IEEE International Conference on Fuzzy Sys-
tems (FUZZ-IEEE’03), volume 1, pages 196–201, St.
Louis, MO.
Padmanabhan, B. and Tuzhilin, A. (1999). Unexpectedness
as a measure of interestingness in knowledge discov-
ery. Decision Support Systems, 27(3):303–318.
Piatetsky-Shapiro, G. and Matheus, C. (1994). The inter-
estingness of deviations. In Proceedings of the AAAI-
94 Workshop on Knowledge Discovery in Databases
(KDD 1994), pages 25–36, Seattle, WA.
Silberschatz, A. and Tuzhilin, A. (1996). What makes
patterns interesting in knowledge discovery systems.
IEEE Transactions on Knowledge And Data Engi-
neering, 8:970–974.
Taha, I. and Ghosh, J. (1997). Evaluation and ordering of
rules extracted from feedforward networks. In Inter-
national Conference on Neural Networks, volume 1,
pages 408–413, Houston, TX.
Tan, P.-N., Kumar, V., and Srivastava, J. (2004). Selecting
the right objective measure for association analysis.
Information Systems, 29(4):293–313.
UCL (2007). Elena Database. http://www.ucl.ac.be/mlg/
index.php?page=Elena.
IN YOUR INTEREST - Objective Interestingness Measures for a Generative Classifier
423