Table 3: Clopper-Pearson intervals for the experiment re-
sults.
Dataset Feature CP Interval
UIUC Canny [92.48%, 98.69%]
SSE [97.85%, 100.0%]
IAIR Canny [78.83%, 90.48%]
SSE [87.81%, 96.45%]
mally distributed (p-values of all experiments in the
Shapiro-Wilk tests (Shapiro and Wilk, 1965) are <
10e-3). Comparing the independent groups Canny
and SSE for both datasets in the Wilcoxon-Mann-
Whitney test, we obtain p-values of 9.86e − 6 and
0.001872 for UIUC and IAIR, respectively. Thus,
the mean overlap value a
0
for the SSE edge detec-
tion tests is larger at the 95% confidence level than
the mean overlap for the Canny Edge Detection tests.
Additionally, we statistically evaluated the result-
ing localization errors for Canny and SSE edge fea-
tures in the same way as described above. We obtain
p-values of 0.005118 and 0.001696 for UIUC and
IAIR, respectively. Therefore, the mean localization
error for the SSE edge detection tests is lower at the
95% confidence level than the mean localization error
for the Canny Edge Detection tests.
5 CONCLUSIONS
We have shown that the object localization perfor-
mance obtained by the voting-based DGHT approach
in real-world tasks with variable background and clut-
ter can be significantly improved by a sophisticated
edge detection algorithm, namely the Structured Edge
Detector. This applies to general structured edge
features without additional training effort as well as
category-specific Structured Edge Detectors in par-
ticular. More precisely, we obtained absolute im-
provements in localization accuracy of 3.53% and
7.64% on a car and pedestrian localization task, re-
spectively. We conclude that the DGHT framework
can be successfully used for object localization also
in real-world images with larger and more variable
background.
In future work, we aim to integrate an intelligent
edge detection mechanism into the voting framework
and to explore strategies to handle object variability
(e.g. object size, rotation) as well as multi-object and
multi-class localization.
Figure 8: Error case: Detected pedestrian is not within the
annotated height range of 130-170 px. Bounding box col-
ors: yellow: prediction; green: ground truth annotation;
red: not in ground truth, because height 6∈ [130,170] px.
(Best viewed in color).
REFERENCES
Agarwal, S., Awan, A., and Roth, D. (2004). Learning to
detect objects in images via a sparse, part-based rep-
resentation. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 26(11):1475–1490.
Agarwal, S. and Roth, D. (2002). Learning a sparse repre-
sentation for object detection. In Computer Vision –
ECCV 2002, pages 113–127. Springer.
Andriluka, M., Roth, S., and Schiele, B. (2008).
People-tracking-by-detection and people-detection-
by-tracking. In IEEE Conference on Computer Vision
and Pattern Recognition, pages 1–8. IEEE.
Arbelaez, P., Maire, M., Fowlkes, C., and Malik, J. (2011).
Contour detection and hierarchical image segmenta-
tion. IEEE Transactions on Pattern Analysis and Ma-
chine Intelligence, 33(5):898–916.
Ballard, D. H. (1981). Generalizing the hough trans-
form to detect arbitrary shapes. Pattern recognition,
13(2):111–122.
Breiman, L. (2001). Random forests. Machine learning,
45(1):5–32.
Canny, J. (1986). A computational approach to edge de-
tection. IEEE Transactions on Pattern Analysis and
Machine Intelligence, (6):679–698.
Chaohui, Z., Xiaohui, D., Shuoyu, X., Zheng, S., and Min,
L. (2007). An improved moving object detection al-
gorithm based on frame difference and edge detec-
tion. In Fourth International Conference on Image
and Graphics, pages 519–523. IEEE.
Doll
´
ar, P. and Zitnick, C. L. (2013). Structured forests for
fast edge detection. In IEEE International Conference
on Computer Vision, pages 1841–1848. IEEE.
Doll
´
ar, P. and Zitnick, C. L. (2014). Fast edge detection
using structured forests.
Ecabert, O., Peters, J., Schramm, H., Lorenz, C., Von Berg,
J., Walker, M. J., Vembar, M., Olszewski, M. E., Sub-
ramanyan, K., Lavi, G., et al. (2008). Automatic
Structured Edge Detection for Improved Object Localization using the Discriminative Generalized Hough Transform
401