Figure 2: The results for the three datasets are given in column 1 for Art1, column 2 for Art2 and column 3 for Iris. The 1-th
row is for the map from the faGTM model. The 2-nd row is for the graphs of the g sets of curves T
k
. The mesh resulting of
the first EM step with coordinates s
(1)
k
is in red dot line.
uniform distribution on [0;0.15]. This resulting
dataset counts n = 1000 vectors with d = 15 fea-
tures.
- Art2. This dataset is a random sample from one
half of a sphere centered at origin in R
3
with ra-
dius 1, plus a circular band surrounding the 2-
th hemisphere near the great circle. This dataset
counts n = 1479 vectors of d = 3 features. The
sample from the hemisphere is clustered artifi-
cially into 10 non-overlapping classes.
- Iris. The dataset of the Iris is compound of 150
vectors in a 4-dimensional space and 3 classes.
The trajectory plot is less relevant in this situation
to reveal the 3 clusters which are less separated.
The projections for the three datasets are shown in
Figure 2. The points for the different classes have
different colors on the graphics. The results are very
encouraging, the method adds flexibility to the vec-
tors of basis function, and leads to a novel graphical
representation for the GTM.
6 CONCLUSIONS AND
PERSPECTIVE
We have proposed a hierarchical factor prior with pa-
rameters C, ρ and λ for generalizing MPPCA and
GTM. The faGTM and its prior offer several per-
spectives. For instance, the trajectory map as a com-
plement to the magnification factors (Bishop et al.,
1997; Maniyar and Nabney, 2006; Tiˇno and Giannio-
tis, 2007) can be studied further.
REFERENCES
Baek, J., McLachlan, G., and Flack, L. (2009). Mixtures
of factor analyzers with common factor loadings: ap-
plications to the clustering and visualisation of high-
dimensional data. IEEE Transactions on Pattern Anal-
ysis and Machine Intelligence.
Bishop, C., Svensen, M., and Williams, C. (1997). Mag-
nification factors for the gtm algorithm. In Fifth In-
ternational Conference on Artificial Neural Networks,
pages 64 –69.
Bishop, C. M. (1995). Neural Networks for Pattern Recog-
nition. Clarendon Press.
Bishop, C. M., Svens´en, M., and Williams, C. K. I. (1998).
Developpements of generative topographic mapping.
Neurocomputing, 21:203–224.
Dempster, A., Laird, N., and Rubin, D. (1977). Maximum-
likelihood from incomplete data via the EM algo-
rithm. J. Royal Statist. Soc. Ser. B., 39, pages 1–38.
Ghahramani, Z. and Hinton, G. E. (1996). The EM algo-
rithm for mixtures of factor analyzers. Technical Re-
port CRG-TR-96-1.
Kohonen, T. (1997). Self-organizing maps. Springer.
Maniyar, D. M. and Nabney, I. T. (2006). Visual data
mining using principled projection algorithms and in-
formation visualization techniques. In Proceedings
of the 12th ACM SIGKDD international conference
on Knowledge discovery and data mining, KDD ’06,
pages 643–648. ACM.
McLachlan, G. J. and Peel, D. (2000). Finite Mixture Mod-
els. John Wiley and Sons, New York.
Tipping, M. E. and Bishop, C. M. (1999a). Mixtures of
probabilistic principal component analyzers. Neural
Computation, 11(2):443–482.
Tipping, M. E. and Bishop, C. M. (1999b). Probabilistic
principal component analysis. Journal of the Royal
Statistical Society. Series B (Statistical Methodology),
61(3):pp. 611–622.
Tiˇno, P. and Gianniotis, N. (2007). Metric properties
of structured data visualizations through generative
probabilistic modeling. In Proceedings of the 20th in-
ternational joint conference on Artifical intelligence,
pages 1083–1088.
GENERATIVE TOPOGRAPHIC MAPPING AND FACTOR ANALYZERS
287