of the examples is stationary, the online error of naive-Bayes will decrease. When the
distribution changes, the naive-Bayes online error will increase. In that case the test
installed at this nodeis not appropriate for the actual distribution of the examples.When
this occurs all the subtree rooted at this node will be pruned. naive-Bayes classifier
central idea on this work is the control of the training dataset that goes through to the
learning algorithm. The algorithm forget the old examples and learns the new concept
with only the examples in the new concept. This methodology was tested with two
artificial data sets and one real world data set. The experimental results show a good
performance at the change of concept detection and also with learning the new concept.
Acknowledgments: The authors reveal its gratitude to the financial support given by the FEDER,
and the Plurianual support attributed to LIACC. This work was developed in the context of the
project ALES (POSI/SRI/39770/2001).
References
1. L. Breiman. Bagging predictors. Machine Learning, 24, 1996.
2. Leo Breiman. Random forests. Technical report, University of Berkeley, 2002.
3. P. Domingos and G. Hulten. Mining high-speed data streams. In Knowledge Discovery and
Data Mining, pages 71–80, 2000.
4. J. F¨urnkranz. Round robin classification. Journal of Machine Learning Research, 2:721–747,
2002.
5. J. Gama, P. Medas, and R. Rocha. Forest trees for on-line data. In ACM Symposium on
Applied Computing - SAC04, 2004.
6. Michael Harries. Splice-2 comparative evaluation: Electricity pricing. Technical report, The
University of South Wales, 1999.
7. Geoff Hulten, Laurie Spencer, and Pedro Domingos. Mining time-changing data streams. In
Knowledge Discovery and Data Mining, 2001.
8. J.Gama, R. Rocha, and P.Medas. Accurate decision trees for mining high-speed data streams.
In Procs. of the 9th ACM SigKDD Int. Conference in Knowledge Discovery and Data Mining.
ACM Press, 2003.
9. J. Kittler. Combining classifiers: A theoretical framework. Pattern analysis and Applications,
Vol. 1, No. 1, 1998.
10. R. Klinkenberg. Learning drifting concepts: Example selection vs. example weighting. In-
telligent Data Analysis, 2004.
11. R. Klinkenberg and I. Renz. Adaptive information filtering: Learning in the presence of
concept drifts. In Learning for Text Categorization, pages 33–40. AAAI Press., 1998.
12. Ralf Klinkenberg and Thorsten Joachims. Detecting concept drift with support vector ma-
chines. In Pat Langley, editor, Proceedings of ICML-00, 17th International Conference on
Machine Learning, pages 487–494, Stanford, US, 2000. Morgan Kaufmann Publishers, San
Francisco, US.
13. M. Kubat and G. Widmer. Adapting to drift in continuous domain. In Proceedings of the 8th
European Conference on Machine Learning, pages 307–310. Spinger, 1995.
14. W.-Y. Loh and Y.-S. Shih. Split selection methods for classification trees. Statistica Sinica,
1997.
15. Tom Mitchell. Machine Learning. McGraw Hill, 1997.
16. R. Motwani and P. Raghavan. Randomized Algorithms. Cambridge University Press, 1997.
17. R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers, Inc.,
1993.
18. Gerhard Widmer and Miroslav Kubat. Learning in the presence of concept drift and hidden
contexts. Machine Learning, 23:69–101, 1996.
158