Table 1: Experimental analysis of EvoSK and control methods on benchmark problem set mapping to a q = 2-dimensional
space in terms of E
sk
. The best results in each line are shown in bold numbers.
EvoSK control
problem best mean dev worst MDS ISOMAP t-SNE
Digits 44.17 45.25 0.89 46.52 48.50 46.91 100.01
Housing 0.27 0.63 0.26 1.01 0.08 0.36 33.52
Image 4.09 4.54 0.34 5.10 4.24 4.30 32.89
Friedman 82.14 83.45 1.63 86.66 92.07 91.65 140.52
Wind 22.32 26.52 3.23 30.17 16.11 14.41 84.98
where D
X,−i
and D
Z,−i
are the distance matrices with
the i-th columns and rows missing.
5 CONCLUSIONS
The iterative evolutionary variant of MDS approx-
imates a low-dimensional representation that mini-
mizes the Shepard-Kruskal measure, i.e., optimizes
the maintenance of distances of the high-dimensional
space in its low-dimensional counterpart. This op-
timization objective is inspired by MDS. EvoSK is
an approximation heuristic with random elements, in
particular based on Gaussian sampling. The outcome
of the DR result depends on the order the patterns are
embedded. But the randomness is the key property to
achieve a linear runtime.
In the line of research of iterative constructive em-
beddings, EvoSK is the first variant that considers
only the last n embedded patterns resulting in linear
runtime and turning out to be sufficient for minimiza-
tion of the Shephard-Kruskal measure. Experimen-
tal results have shown that the introduced variants can
compete with the related method MDS, ISOMAP, and
with t-SNE.
The runtime of EvoSK is linear when employing
a window size k. For a choice of k = 50, the embed-
dings turn out to show good characteristics. But the
results show that EvoSK significantly depends on the
effort µ invested into the sampling process.
REFERENCES
Borg, I. and Groenen, P. (2005). Modern Multidimensional
Scaling: Theory and Applications. Springer.
Hastie, T., Tibshirani, R., and Friedman, J. (2009). The
Elements of Statistical Learning. Springer, Berlin.
Jolliffe, I. (1986). Principal component analysis. Springer
series in statistics. Springer, New York.
Kramer, O. (2015a). Supervised manifold learning with in-
cremental stochastic embeddings. In European Sym-
posium on Artificial Neural Networks (ESANN), pages
243–248.
Kramer, O. (2015b). Unsupervised nearest neighbor re-
gression for dimensionality reduction. Soft Comput.,
19(6):1647–1661.
Lee, J. A. and Verleysen, M. (2007). Nonlinear Dimension-
ality Reduction. Springer.
Meinicke, P., Klanke, S., Memisevic, R., and Ritter, H.
(2005). Principal surfaces from unsupervised kernel
regression. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 27(9):1379–1391.
Mitliagkas, I., Caramanis, C., and Jain, P. (2013). Memory
limited, streaming PCA. In Advances in Neural Infor-
mation Processing Systems (NIPS), pages 2886–2894.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,
Thirion, B., Grisel, O., Blondel, M., Prettenhofer,
P., Weiss, R., Dubourg, V., Vanderplas, J., Passos,
A., Cournapeau, D., Brucher, M., Perrot, M., and
Duchesnay, E. (2011). Scikit-learn: Machine learning
in Python. Journal of Machine Learning Research,
12:2825–2830.
Sch
¨
olkopf, B., Smola, A., and Mller, K.-R. (1998). Nonlin-
ear component analysis as a kernel eigenvalue prob-
lem. Neural Computation, 10(5):1299–1319.
Tenenbaum, J. B., Silva, V. D., and Langford, J. C. (2000).
A global geometric framework for nonlinear dimen-
sionality reduction. Science, 290:2319–2323.
van der Maaten, L. and Hinton, G. E. (2008). Visualiz-
ing high-dimensional data using t-sne. Journal of Ma-
chine Learning Research, 9:2579–2605.
Dimensionality Reduction with Evolutionary Shephard-Kruskal Embeddings
481