equates to a uniform distribution along the discarded
dimensions and a full Gaussian model along the re-
tained dimensions.
It is important to contrast this model properly to
the several existing methods for linear dimensionality
reduction in mixture models such as mixtures of PCA,
PPCA, and FA. One of the main points of difference is
that the linear transformation is not restricted to be or-
thogonal. Further, the linear model adopted, x = Ωy,
does not assume additive noise models and makes x
observable. On the ground of that, we can evaluate
density N (Ωy|µ,Σ) = N (x|µ,Σ) directly in x-space.
For learning the model, we have presented two differ-
ent methods; the first method learns the model’s pa-
rameters in a maximum likelihood framework (MLiT
(N)). Normalization is proposed as a way to regular-
ize this solution. Thus, a common scale is imposed
to all the transformations and a singularity problem
is avoided. Another simple yet powerful method for
learning the model’s parameters can be based on ran-
dom matrices (MLiT (R)). This method has offered
promising and computationally feasible results. How-
ever, the maximum likelihood solution delivered bet-
ter accuracy results in majority of the data sets sug-
gesting that it can be a better way for learning the
model’s parameters.
The experimental performance of MLiT has
proved to outperform that of MPPCA and GMM in al-
most all cases with improvements ranging from 0.2%
to 5.2% compared to the runner-up. The only case
where MLiT did not deliver the best accuracy is on
the OpticDigit data set where it was slightly outper-
formed by MPPCA by 0.2%. In addition to visual ob-
ject classification, the proposed method permits gen-
eral application for density modeling and classifica-
tion of other continuous numerical data requiring di-
mensionality reduction. Moreover, its re-estimation
formulas can be easily extended to suit boosting and
other weighted maximum likelihood targets and adapt
to a variety of pattern recognition frameworks.
ACKNOWLEDGEMENTS
The authors wish to thank the Australian Research
Council and iOmniscient Pty Ltd that have partially
supported this work under the Linkage Project fund-
ing scheme - grant LP0668325.
REFERENCES
Asuncion, A. and Newman, D. (2007). UCI machine learn-
ing repository.
Bartholomew, D. J. (1987). Latent Variable Models and
Factor Analysis. Charles Griffin & Co. Ltd., London.
Bellman, R. (1961). Adaptive control processes - A guided
tour. Princeton University Press, Princeton, New Jer-
sey.
Bingham, E. and Mannila, H. (2001). Random projection in
dimensionality reduction: applications to image and
text data. In Proceedings of the 7th ACM SIGKDD In-
ternational Conference on Knowledge Discovery and
Data Mining (KDD-2001), pages 245–250.
Bishop, C. M. (2006). Pattern Recognition and Machine
Learning. Springer.
Bolton, R. J. and Krzanowski, W. J. (1999). A characteri-
zation of principal components for projection pursuit.
The American Statistician, 53(2):108–109.
Breiman, L. and Spector, P. (1992). Submodel selection and
evaluation in regression: The x-random case. Interna-
tional Statistical Review, 60(3):291–319.
Fodor, I. (2002). A survey of dimension reduction
techniques. Technical Report UCRL-ID-148494,
Lawrence Livermore National Laboratory.
Ghahramani, Z. and Hinton, G. (1997). The EM algorithm
for mixtures of factor analyzers. Technical Report
CRG-TR-96-1, University of Toronto.
Hinton, G. E., Dayan, P., and Revow, M. (1997). Modeling
the manifolds of images of handwritten digits. IEEE
Transactions on Neural Networks, 8(1):65–74.
Johnson, W. B. and Lindenstrauss, J. (1984). Extensions
of lipschitz mappings into a hilbert space. In Confer-
ence in modern analysis and probability, Contempo-
rary Math, volume 26, pages 189–206.
Kaski, S. (1998). Dimensionality reduction by random
mapping: Fast similarity computation for clustering.
In Proceedings of IJCNN’98, International Joint Con-
ference on Neural Networks, volume 1, pages 413–
418. IEEE Service Center.
Kittler, J. (1998). Combining classifiers: A theoreti-
cal framework. Pattern Analysis and Applications,
1(1):18–27.
Otoom, A. e. a. (2007). Towards automatic abandoned ob-
ject classification in visual surveillance systems. In
Asia-Pacific Workshop on Visual Information Process-
ing, pages 143–149, Tanian, Taiwan.
Tipping, M. E. and Bishop, C. M. (1999a). Mixtures of
probabilistic principal component analyzers. Neural
Computation, 11(2):443–482.
Tipping, M. E. and Bishop, C. M. (1999b). Probabilistic
principal component analysis. Journal of the Royal
Statistical Society: Series B (Statistical Methodol-
ogy),, 61(3):611–622.
VISAPP 2010 - International Conference on Computer Vision Theory and Applications
518