tion only. This reduction was necessary in order to
apply a classical algorithm of decision tree learning
that was designed for a single level of abstraction.
The use of this cut in AVT-NBL is just a tradeoff be-
tween the complexity and accuracy of the classifier
(Zhang et al., 2006). We show that na¨ıve Bayesian
classifier can be extended to AVO without any reduc-
tion of a given ontology. Secondly, the semantics of
AVO allows to precise descriptions explicitly (posi-
tive observations) and implicitly (negative observa-
tions). Thirdly, the semantics of AVO allows to avoid
overfitting in a very effective way by analyzing results
for less precise values. The overfitting avoidance is a
very important issue in Data Mining and is very im-
portant from the practical point of view.
5 CONCLUSIONS
In this paper we proposed an extension of the na¨ıve
Bayesian classifier by an attribute value ontology
(AVO) aiming at the improvement of the analysis of
imperfect data. In the proposed approach, every at-
tribute is a hierarchy of concepts from the domain
knowledge base (ISA hierarchy). This semantic ap-
proach to Data Mining allows to describe examples
either very precisely or, when it is not possible, in a
more general way (using a concept from higher levels
of the hierarchy). As a result, users that are unable to
precisely describe the observation by a specific value
of an attribute, are allowed to use (less or more) im-
precise values.
Let us notice, that each imprecise value of AVO,
except the most abstract concept, is more precise than
the missing value, represented by this most abstract
concept. Therefore, introducing these abstract con-
cepts we improve the analysis of imperfect data. This
improvement is increased by each upgrade of the pre-
cision of information. We showed that even negative
observations improve this precision: ”knowing what
we do not know” is already information.
We could ask a question: how far should we pre-
cise the description? There is no single answer for
this question. Each cut through a hierarchy seems to
be a tradeoff between the complexity and accuracy.
Therefore, maintaning all the levels of abstraction is
an alternative approach to this problem. It allows to
compare results for many levels of abstraction. That
can be an efficient way to avoid the effect of overfit-
ting. Moreover, this comparison can be utilized by
a cost sensitive computing. High precision carries a
high cost. The challenge is to exploit the tolerance
for imprecision. Further research aims at experimen-
tal evaluation of the proposed approach.
ACKNOWLEDGEMENTS
Research supported by the Polish Ministry of Science
and Higher Education grant No. N N516 186437.
REFERENCES
Almuallim, H., Akiba, Y., and Kaneda, S. (1996). An
efficient algorithm for finding optimal gain-ratio
multiple-split tests on hierarchical attributes in deci-
sion tree learning. In Proceedings of the Thirteenth
National Conference on Artificial Intelligence. AAAI
Press.
Breiman, L., Friedman, J. H., Olshen, R. A., and Stone,
C. J. (1984). Classification and Regression Trees.
Wadsworth, Belmont, California, 3rd edition.
Han, J., Cai, Y., and Cercone, N. (1992). Knowledge dis-
covery in databases: An attribute-oriented approach.
In Proceedings of the 18th International Conference
on Very Large Data Bases. Morgan Kaufmann.
Haussler, D. (1988). Quantifying inductive bias: Ai learn-
ing algorithms and valiant’s learning framework. In
Artif. Intell., Vol. 36(2). Elsevier.
Kudoh, Y., Haraguchi, M., and Okubo, Y. (2003). Data
abstractions for decision tree induction. In Theoretical
Computer Science, Vol. 292(1). Elsevier.
N´u˜nez, M. (1991). The use of background knowledge in
decision tree induction. In Machine Learning, Vol.
6(3). Springer.
Tanaka, H. (1996). Decision tree learning algorithm with
structured attributes: Application to verbal case frame
acquisition. In Proceedings of the 16th International
Conference on Computational Linguistics. Center for
Sprogteknologi, Danmark.
Taylor, M. G., Stoffel, K., and Hendler, J. A. (1997).
Ontology-based induction of high level classification
rules. In Proceedings of the Workshop on Research
Issues on Data Mining and Knowledge Discovery.
ACM.
Walker, A. (1980). On retrieval from a small version of a
large data base. In Proceedings of the Sixth Interna-
tional Conference on Very Large Data Bases. IEEE
Computer Society.
Witten, I., Frank, E., and Hall, M. (2011). Data Mining.
Practical Machine Learning Tools and Techniques.
Morgan Kaufmann, Burlington, 3rd edition.
Zhang, J., Kang, D., Silvescu, A., and Honavar, V. (2006).
Learning accurate and concise naive bayes classifiers
from attribute value taxonomies and data. In Knowl.
Inf. Syst., Vol. 9(2). Springer.
Zhang, J., Silvescu, A., and Honavar, V. (2002). Ontology-
driven induction of decision trees at multiple levels of
abstraction. In LNCS Vol. 2371. Springer.
ICEIS2012-14thInternationalConferenceonEnterpriseInformationSystems
334