tom row). As can be seen in the figure, simple user
added (dis)similarity constraints can significantly im-
prove the segmentations. Consider for instance the
flower image, there are three well separated color
components in the image: the green background, the
red leaves, and the yellow flower center. There are
thus three reasonable segmentations –separating each
one of the components from the other two– and it is
a-priori not clear which is desired by a user. How-
ever once a small set of (dis)similarity constraints are
added, the segmentation desired by the user is easily
identified.
4 CONCLUSIONS
In this paper we developed a semi-supervised di-
mensionality reduction method which uses pairwise
equivalence constraints to discover the groups in
high-dimensional data. To this end, we modified
LPP scheme such that its objective function takes into
account the equivalence constraints. Like LPP, our
algorithm first finds neighboring points to create a
weighted neighborhood graph. Then, the constraints
are used to modify the neighborhood relations and
weight matrix to reflect this weak form of supervision.
The optimal projection matrix according to our cost
function is then identified by solving for the smallest
eigenvalue solutions of an n × n eigenvector problem,
where n is the number of data points. Experimental
results show that our semi-supervised dimensional-
ity reduction method increases performance of subse-
quent clustering and classification algorithms. More-
over, it yields better results than methods applying un-
supervised dimensionality reduction followed by full-
rank metric learning.
In some applications, small subsets of data points
with same class labels, so-called ‘chunklets’, occur
naturally, e.g., for face recognition in video. In fu-
ture work, we will explore distance metrics between
chunklets as well as chunklets and points, rather than
between individual data points. Since these metrics
operate on richer data structures, we expect them to
significantly improve clustering and classification re-
sults.
REFERENCES
Basu, S., Banerjee, A., and Mooney, R. J. (2004). Active
semi-supervision for pairwise constrained clustering.
In the SIAM International Conference on Data Min-
ing.
Belkin, M. and Niyogi, P. (2001). Laplacian eigenmaps and
spectral techniques for embedding and clustering. In
Advances in Neural Information Processing Systems.
Bilenko, M., Basu, S., and Mooney, R. J. (2004). Integrat-
ing constraints and metric learning in semi-supervised
clustering. In the 21st International Conference on
Machine Learning.
Cevikalp, H., Larlus, D., Neamtu, M., Triggs, B., and Ju-
rie, F. (2007). Manifold based local classifiers: Linear
and nonlinear approaches. In Pattern Recognition in
review.
He, X. and Niyogi, P. (2003). Locality preserving direc-
tions. In Advances in Neural Information Processing
Systems.
Hertz, T., Shental, N., Bar-Hillel, A., and Weinshall, D.
(2003). Enhancing image and video retrieval: Learn-
ing via equivalence constraints. In the 2003 IEEE
Computer Society Conference on Computer Vision
and Pattern Recognition (CVPR’03).
Lazebnik, S., Schmid, C., and Ponce, J. (2005). A max-
imum entropy framework for part-based texture and
objcect recognition. In International Conference on
Computer Vision (ICCV).
Leibe, B. and Schiele, B. (2003). Interleaved object catego-
rization and segmentation. In British Machine Vision
Conference (BMVC).
Lowe, D. (2004). Distinctive image features from scale -
invariant keypoints. In International Journal of Com-
puter Vision, volume 60, pages 91–110.
Melia, M. and Shi, J. (2001). A random walks view of spec-
tral segmentation. In the 8th International Workshop
on Artificial Intelligence and Statistics.
Shental, N., Bar-Hillel, A., Hertz, T., and Weisnhall, D.
(2003). Computing gaussian mixture models with em
using equivalance constraints. In Advances in Neural
Information Processing Systems (NIPS).
Tsang, I. W. and Kwok, J. T. (2003). Distance metric learn-
ing with kernels. In the International Conference on
Artificial Neural Networks.
van de Weijer, J. and Schmid, C. (2006). Coloring local fea-
ture extraction. In European Conference on Computer
Vision (ECCV).
Wagstaff, K. and Rogers, S. (2001). Constrained k-means
clustering with background knowledge. In the 18th
International Conference on Machine Learning.
Xing, E. P., Ng, A. Y., Jordan, M. I., and Russell, S. (2003).
Distance metric learning with application to clustering
with side-information. In Advances in Neural Infor-
mation Processing Systems.
Yan, B. and Domeniconi, C. (2006). Subspace metric en-
sembles for semi-supervised clustering of high dimen-
sional data. In the 17th European Conference on Ma-
chine Learning.
VISAPP 2008 - International Conference on Computer Vision Theory and Applications
496