6 CONCLUSIONS AND FUTURE
WORK
In this paper we introduced the Diffusion Maps di-
mensionality reduction algorithm as a framework for
the construction of ensemble classifiers which use a
single induction algorithm. The DM algorithm was
applied to the training set using different values for
its input parameter. This produced different versions
of the training set and the ensemble members were
constructed based on these training set versions. In
order to classify a new sample, it was first embed-
ded into the dimension-reducedspace of each training
set using the Nystr¨om out-of-sample extension algo-
rithm. The results in this paper show that the pro-
posed approach is effective. The results were supe-
rior in most of the datasets compared to the plain al-
gorithm. The authors are currently extending this ap-
proach to other dimensionality reduction techniques.
Additionally, other out-of-sample extension schemes
should also be explored e.g. the Geometric Harmon-
ics (Coifman and Lafon, 2006b). Lastly, a heteroge-
neous model which combines several dimensionality
reduction techniques should be investigated.
REFERENCES
Asuncion, A. and Newman, D. J. (2007). UCI machine
learning repository. http://archive.ics.uci.edu/ml/.
Belkin, M. and Niyogi, P. (2003). Laplacian eigenmaps
for dimensionality reduction and data representation.
Neural Computation, 15(6):1373–1396.
Bengio, Y., Delalleau, O., Roux, N. L., Paiement, J. F., Vin-
cent, P., and Ouimet, M. (2004). Learning eigenfunc-
tions links spectral embedding and kernel pca. Neural
Computation, 16(10):2197–2219.
Breiman, L. (1996). Bagging predictors. Machine Learn-
ing, 24(2):123–140.
Breiman, L., Friedman, J. H., Olshen, R. A., and Stone, C. J.
(1993). Classification and Regression Trees. Chap-
man & Hall, Inc., New York.
Chung, F. R. K. (1997). Spectral Graph Theory. AMS
Regional Conference Series in Mathematics, 92.
Coifman, R. R. and Lafon, S. (2006a). Diffusion maps. Ap-
plied and Computational Harmonic Analysis: special
issue on Diffusion Maps and Wavelets, 21:5–30.
Coifman, R. R. and Lafon, S. (2006b). Geometric harmon-
ics: a novel tool for multiscale out-of-sample exten-
sion of empirical functions. Applied and Computa-
tional Harmonic Analysis: special issue on Diffusion
Maps and Wavelets, 21:31–52.
Cox, T. and Cox, M. (1994). Multidimensional scaling.
Chapman & Hall, London, UK.
Donoho, D. L. and Grimes, C. (2003). Hessian eigen-
maps: new locally linear embedding techniques for
high-dimensional data. In Proceedings of the National
Academy of Sciences, volume 100(10), pages 5591–
5596.
Drucker, H. (1997). Improving regressor using boosting. In
Jr., D. H. F., editor, Proceedings of the 14th Interna-
tional Conference on Machine Learning, pages 107–
115. Morgan Kaufmann.
Feher, C., Elovici, Y., Moskovitch, R., Rokach, L., and
Schclar, A. (2012). User identity verification via
mouse dynamics. Information Sciences, 201:19–36.
Freund, Y. and Schapire, R. (1996). Experiments with a
new boosting algorithm. machine learning. In Pro-
ceedings for the Thirteenth International Conference,
pages 148–156, San Francisco. Morgan Kaufmann.
Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann,
P., and Witten, I. H. (2009). The weka data mining
software: An update. SIGKDD Explorations, 11:1.
Ham, J., Lee, D., Mika, S., and Scholk¨opf, B. (2004). A
kernel view of the dimensionality reduction of mani-
folds. In Proceedings of the 21st International Con-
ference on Machine Learning (ICML’04), pages 369–
376, New York, NY, USA. ACM Press.
Hegde, C., Wakin, M., and Baraniuk, R. G. (2007). Random
projections for manifold learning. In Neural Informa-
tion Processing Systems (NIPS).
Hein, M. and Audibert, Y. (2005). Intrinsic dimensional-
ity estimation of submanifolds in Euclidean space. In
Proceedings of the 22nd International Conference on
Machine Learning, pages 289–296.
Hotelling, H. (1933). Analysis of a complex of statistical
variables into principal components. Journal of Edu-
cational Psychology, 24:417–441.
Jimenez, L. O. and Landgrebe, D. A. (1998). Supervised
classification in high-dimensional space: geometrical,
statistical and asymptotical properties of multivariate
data. IEEE Transactions on Systems, Man and Cyber-
netics, Part C: Applications and Reviews,, 28(1):39–
54.
Kruskal, J. B. (1964). Multidimensional scaling by opti-
mizing goodness of fit to a nonmetric hypothesis. Psy-
chometrika, 29:1–27.
Kuncheva, L. I. (2004). Diversity in multiple classifier sys-
tems (editorial). Information Fusion, 6(1):3–4.
Lafon, S., Keller, Y., and Coifman, R. R. (2006). Data fu-
sion and multicue data matching by diffusion maps.
IEEE Transactions on Pattern Analysis and Machine
Intelligence, 28:1784–1797.
Leigh, W., Purvis, R., and Ragusa, J. M. (2002). Forecast-
ing the nyse composite index with technical analysis,
pattern recognizer, neural networks, and genetic algo-
rithm: a case study in romantic decision support. De-
cision Support Systems, 32(4):361–377.
Mangiameli, P., West, D., and Rampal, R. (2004). Model
selection for medical diagnosis decision support sys-
tems. Decision Support Systems, 36(3):247–259.
Nystr¨om, E. J. (1928).
¨
Uber die praktische au߬osung
von linearen integralgleichungen mit anwendungen
auf randwertaufgaben der potentialtheorie. Commen-
tationes Physico-Mathematicae, 4(15):1–52.
DiffusionEnsembleClassifiers
449