KLIM_L estimator achieves higher classification
accuracy than LDA, RDA, LOOC and KLIM
estimators on COIL-20 data set. In the future work,
the kernel method combined with these
regularization discriminant methods will be studied
for small sample problem with high dimension and
the selection of kernel parameters will be
investigated under some criterion.
ACKNOWLEDGEMENTS
The research work described in this paper was fully
supported by the grants from the National Natural
Science Foundation of China (Project No.
90820010, 60911130513). Prof. Guo is the author to
whom the correspondence should be addressed, his
e-mail address is pguo@ieee.org
REFERENCES
Bishop, C. M., 2007. Pattern recognition and machine
learning, Springer-Verlag New York, Inc. Secaucus,
NJ, USA.
Everitt, B. S., Hand, D., 1981. Finite Mixture
Distributions, Chapman and Hall, London.
Friedman, J. H., 1989. Regularized discriminant analysis,
Journal of the American Statistical Association, vol.
84, no. 405, 165–175.
Hoffbeck, J. P. and Landgrebe, D. A., 1996. Covariance
matrix estimation and classification with limited
training data, IEEE Transactions on Pattern Analysis
and Machine Intelligence, vol. 18, no. 7, 763–767.
Schafer, J. and Strimmer, K., 2005. A shrinkage approach
to large-scale covariance matrix estimation and
implications for functional genomics, Statistical
Applications in Genetics and Molecular Biology, vol.
4, no. 1.
Srivastava, S., Gupta, M. R., Frigyik, B. A., 2007.
Bayesian quadratic discriminant analysis, J. Mach.
Learning Res. 8, 1277
–1305.
Bickel, P. J. and Levina, E., 2008. Regularized estimation
of large covariance matrices, Annals of Statistics, vol.
36, no. 1, 199–227.
Friedman, J., Hastie, T., and Tibshirani, R., 2008. Sparse
inverse covariance estimation with the graphical lasso,
Biostatistics, vol. 9, no. 3, 432–441.
Cao, G., Bachega, L. R., Bouman, C. A., 2011. The Sparse
Matrix Transform for Covariance Estimation and
Analysis of High Dimensional Signals. IEEE
Transactions on Image Processing, Volume 20, Issue
3, 625 – 640.
Rivals, I., Personnaz, L., 1999. On cross validation for
model selection, Neural Comput. 11,863
–870.
Guo, P., Jia, Y., and Lyu, M. R., 2008. A study of
regularized Gaussian classifier in high-dimension
small sample set case based on MDL principle with
application to spectrum recognition, Pattern
Recognition, Vol. 41, 2842~2854.
Redner, R. A., Walker, H. F., 1984. Mixture densities,
maximum likelihood and the EM algorithm, SIAM
Rev. 26, 195
–239.
Aeberhard, S., Coomans, de Vel, D., O., 1994.
Comparative analysis of statistical pattern recognition
methods in high dimensional settings, Pattern
Recognition 27 (8), 1065
–1077.
Rissanen, J., 1978. Modeling by shortest data description,
Automatica 14, 465
–471.
Barron, A., Rissanen, J., Yu, B., 1998. The minimum
description length principle in coding and modeling,
IEEE Trans. Inform. Theory 44 (6), 2743
–2760.
Kullback, S., 1959. Information Theory and Statistics,
Wiley, New York.
Nene, S. A., Nayar, S. K. and Murase, H., 1996. Columbia
Object Image library(COIL-20). Technical report
CUCS-005-96.
MULTI-REGULARIZATION PARAMETERS ESTIMATION FOR GAUSSIAN MIXTURE CLASSIFIER BASED ON
MDL PRINCIPLE
117