MWCD-LDA
∗∗
are only slightly improved compared
to MWCD-LDA
∗
.
Open problems concerning the newly proposed
methods as well as more general ideas for a fu-
ture research in the area of robust analysis of high-
dimensional data contain the following tasks.
• Finding more efficient algorithms for specific
choices of the target matrix T.
• Comparing various approaches to regularizing the
means, mainly comparing the effect of L
2
and
L
1
norm. In addition, comparing the effect of
shrinking the means towards the common mean
vs. towards zero.
• Comparing the performance and robustness of the
new methods with approaches based on robust
PCA.
• Investigating the non-robustness of other standard
regularized classification methods.
• Applying the regularized robust Mahalanobis dis-
tance to modify other methods based on the Ma-
halanobis distance, such as classification trees, en-
tropy estimators, k-means clustering, or dimen-
sionality reduction.
• Combining regularization and robustness to other
methods, including neural networks or SVM or
even linear regression (Jurczyk, 2012).
ACKNOWLEDGEMENTS
This publication was supported by the project ”Na-
tional Institute of Mental Health (NIMH-CZ)”, grant
number CZ.1.05/2.1.00/03.0078 of the European Re-
gional Development Fund. The work of J. Kalina was
financially supported by the Neuron Fund for Support
of Science. The work of J. Hlinka was supported
by the Czech Science Foundation project No. 13-
23940S. The authors are thankful to Anna Schlenker
for the data analyzed in Section 3.4.
REFERENCES
Cai, T. and Shen, X. (2010). High-dimensional data analy-
sis. World Scientific, Singapore.
Chen, Y., Wiesel, A., and Hero, A. O. (2011). Robust
shrinkage estimation of high dimensional covariance
matrices. IEEE Transactions on Signal Processing,
59:4097–4107.
Christmann, A. and van Messem, A. (2008). Bouligand
derivatives and robustness of support vector machines
for regression. Journal of Machine Learning Re-
search, 9:915–936.
Croux, C. and Dehon, C. (2001). Robust linear discrimi-
nant analysis using S-estimators. Canadian Journal
of Statistics, 29:473–493.
Davies, P. (2014). Data analysis and approximate mod-
els: Model choice, location-scale, analysis of vari-
ance, nonparametric regression and image analysis.
Chapman & Hall/CRC, Boca Raton.
Davies, P. L. and Gather, U. (2005). Breakdown and groups.
Annals of Statistics, 33:977–1035.
Duffau, H. (2011). Brain mapping. From neural basis of
cognition to surgical applications. Springer, Vienna.
Filzmoser, P. and Todorov, V. (2011). Review of robust mul-
tivariate statistical methods in high dimension. Ana-
lytica Chinica Acta, 705:2–14.
Friedman, J., Hastie, T., Simon, N., and Tibshirani,
R. (2015). glmnet: Lasso and elastic-net reg-
ularized generalized linear models. http://cran.r-
project.org/web/packages/glmnet/index.html.
Guo, Y., Hastie, T., and Tibshirani, R. (2007). Regularized
discriminant analysis and its application in microar-
rays. Biostatistics, 8:86–100.
Han, H. and Jiang, X. (2014). Overcome support vector
machine diagnosis overfitting. Cancer Informatics,
13:145–148.
Hastie, T., Tibshirani, R., and Friedman, J. (2008). The
elements of statistical learning. Springer, New York,
2nd edition.
Hausser, J. and Strimmer, K. (2009). Entropy inference and
the James-Stein estimator, with application to nonlin-
ear gene association networks. Journal of Machine
Learning Research, 10:1469–1484.
Hlinka, J., Palu
ˇ
s, M., Vejmelka, M., Mantini, D., and Cor-
betta, M. (2011). Functional connectivity in resting-
state fMRI: Is linear correlation sufficient? NeuroIm-
age, 54:2218–2225.
Huber, P. J. and Ronchetti, E. M. (2009). Robust statistics.
Wiley, New York, 2nd edition.
Hubert, M., Rousseeuw, P. J., and van Aelst, S. (2008).
High-breakdown robust multivariate methods. Statis-
tical Science, 23:92–119.
Jurczyk, T. (2012). Outlier detection under multicollinear-
ity. Journal of Statistical Computation and Simula-
tion, 82:261–278.
K
˚
urkov
´
a, V. and Sanguineti, M. (2005). Learning with gen-
eralization capability by kernel methods of bounded
complexity. Journal of Complexity, 21:350–367.
Kalina, J. (2012). Implicitly weighted methods in robust
image analysis. Journal of Mathematical Imaging and
Vision, 44:449–462.
Kalina, J. (2014). Classification analysis methods for
high-dimensional genetic data. Biocybernetics and
Biomedical Engineering, 34:10–18.
Kalina, J. and Schlenker, A. (2015). A robust and regular-
ized supervised variable selection. BioMed Research
International, 2015(320385).
Kindermans, P.-J., Schreuder, M., Schrauwen, B., Mller,
K.-R., and Tangermann, M. (2014). True zero-training
brain-computer interfacing–An online study. PLoS
One, 9(102504).
Highly Robust Classification: A Regularized Approach for Omics Data
25