Figure 12: The left and the right side of a half ellipse.
value
˜
ε > 0 is used to compare the color information
of two representation vectors with respect to maxi-
mum norm. An upcoming article explicitly describes
the extraction of half ellipses and its color.
6.5 Learning and Recognition Scheme
used for COIL-100
As mentioned above the system uses e.g. 8 images
to learn an object. For one image it constructs e.g.
10 combinations of half ellipses. Each combination is
represented with e.g 6 feature vectors. Each vector is
labeled with the number N ∈ {1, ..., 100} of the object
it refers to.
Analyzing an image the system at first determines
the maximal length m ∈ N of the matched subse-
quences for each learned feature vector. Let the set
of such lengths be denoted as M. For ˜m = max M
the system depicts all feature vectors for which sub-
sequences of the length ˜m were matched. The ob-
ject with the greatest number of such feature vectors
will be returned as the recognized one. Having sev-
eral such objects the system chooses one of them ran-
domly.
7 SUMMARY AND FUTURE
WORK
The object recognition system presented in this paper
combines several important characteristics. It’s capa-
ble of handling 3D objects. The half ellipse extraction
is at least stable enough to wield COIL-100 images.
The trivial color representation used now has yet
to become illumination invariant. The optimization of
the running time doesn’t appear to be a great problem
as the central retrieval algorithm is highly paralleliz-
able. The greatest challenge seems to be the reduction
of the storage consumption without lost of perspective
robustness.
At the present the authors develop a flow estima-
tor based on the comparison of half ellipse combina-
tions. The flow estimator learns thousands of half
ellipse combinations on the first frame and tries to
match them on the second one. So in a near future
the system could gain an universal character being si-
multaneously capable of object recognition as well as
flow estimation.
REFERENCES
Arbter, K., Snyder, W. E., Burkhardt, H., and Hirzinger,
G. (1990). Application of affine-invariant fourier de-
scriptors to recognition of 3d objects. In IEEE Trans.
Pattern Analysis and Machine Learning.
Bishop, C. (2007). Neural Networks for Patternrecognition.
Oxford University Press.
Canny, J. (1986). A computational approach to edge detec-
tion. In IEEE Transactions on Pattern Analysis and
Machine Intelligence.
Caputo, B., Hornegger, J., Paulus, D., and Niemann, H.
(2000). A spin-glass markov random field for 3d ob-
ject recognition. NIPS 2000.
Dalal, N. and Triggs, B. (2005). Histograms of oriented
gradients for human detection. In IEEE Conference
Computer Vision and Pattern Recognition , San Diego.
Gyofri, L., Kohler, M., Krzyzak, A., and Walk, H. (2002).
A Distribution-Free Theory of Nonparametric Regres-
sion. Springer.
Hough, P. V. C. (1962). Method and Means of Recognising
Complex Patterns. US Patent 3069654.
Hu, M. K. (1962). Visual pattern recognition by moment
invariants. In IRE Transactions on Information The-
ory.
Jaehne, B. (2005). Digital Image Processing. Springer-
Verlag Berlin.
Nene, S. A., Nayar, S. K., and Murase, H. (1996). Columbia
Object Image Library (COIL-100).
Reiss, T. H. (1993). Recognizing Planar Objects Using In-
variant Image Features. Springer-Verlag Berlin Hei-
delberg.
Rosenblatt, F. (1962). Principles of Neurodynamics. Spar-
tan, New York.
Tsuji, S. and Matsumoto, F. (1978). Detection of ellipses by
a modified hough transform. In IEEE Trans. Comput.
Vapnik, V. N. (1998). Statistical Learning Theory. Wiley,
New York.
Yang, M. H., Roth, D., and Ahuja, N. (2000). Learning to
recognize 3d objects with snow. In ECCV 2000.
VISAPP 2011 - International Conference on Computer Vision Theory and Applications
400