BBBD: Bounding Box Based Detector for Occlusion Detection and Order
Recovery
Kaziwa Saleh
1 a
and Zolt
´
an V
´
amossy
2 b
1
Doctoral School of Applied Informatics and Applied Mathematics,
´
Obuda University, Budapest, Hungary
2
John von Neumann Faculty of Informatics,
´
Obuda University, Budapest, Hungary
Keywords:
Occlusion Handling, Object Detection, Amodal Segmentation, Depth Ordering, Occlusion Ordering, Partial
Occlusion.
Abstract:
Occlusion handling is one of the challenges of object detection and segmentation, and scene understanding.
Because objects appear differently when they are occluded in varying degree, angle, and locations. Therefore,
determining the existence of occlusion between objects and their order in a scene is a fundamental requirement
for semantic understanding. Existing works mostly use deep learning based models to retrieve the order of
the instances in an image or for occlusion detection. This requires labelled occluded data and it is time-
consuming. In this paper, we propose a simpler and faster method that can perform both operations without
any training and only requires the modal segmentation masks. For occlusion detection, instead of scanning the
two objects entirely, we only focus on the intersected area between their bounding boxes. Similarly, we use the
segmentation mask inside the same area to recover the depth-ordering. When tested on COCOA dataset, our
method achieves +8% and +5% more accuracy than the baselines in order recovery and occlusion detection
respectively.
1 INTRODUCTION
Real-world scenes are complex and cluttered, as hu-
mans we observe and fathom them effortlessly even
when objects are not fully visible. We can easily de-
duce that an object is partially hidden by other ob-
jects. For machines, this is a challenging task partic-
ularly if the object(s) are occluded by more than one
object. Nevertheless, for a machine to comprehend its
surrounding, it has to be capable of inferring the order
of objects in the scene, i.e. to determine if the object
is occluded and by which object(s).
Working with occlusion is difficult because an ob-
ject can be hidden by other object(s) in varying ra-
tio, location, and angle. Yet, handling it plays a key
role in the machine perception. Many works in the
recent literature address occlusion in various appli-
cations (Saleh et al., 2021; T
´
oth and Sz
´
en
´
asi, 2020).
The focus is either on detecting and segmenting the
occluded object (Ke et al., 2021) (Zheng et al., 2021),
completing the invisible region (Wang et al., 2021)
(Ling et al., 2020), or depth-ordering (Zhan et al.,
a
https://orcid.org/0000-0003-3902-1063
b
https://orcid.org/0000-0002-6040-9954
2020)(Ehsani et al., 2018).
However, almost all the works in the literature
rely on deep learning based network to retrieve the
amodal mask (the mask for occluded region) of the
object and utilize it for occlusion presence detection
and depth ordering. Although this produces good re-
sults, it is time-consuming and it requires a labelled
occluded dataset which may not be available (Finta
and Sz
´
en
´
asi, 2019).
In this work, we propose a simpler approach.
The method only requires the modal masks and their
bounding boxes which can easily be obtained. In con-
trast to scanning the entire mask of objects to deter-
mine if they occlude each other or not, we only focus
on the intersected area (IA) of the bounding boxes.
We utilize the portion of the mask that falls into this
area. Since we concentrate on a smaller region, our
method is faster and can produce instant results.
Our method called Bounding Box Based Detector
(BBBD) takes the bounding boxes and modal masks
of the instances detected in the image as input, and
outputs a matrix that contains the order of the ob-
jects. By calculating the IA of the bounding boxes
and checking the mask in that area, we infer the exis-
tence of occlusion. The method is tested on COCOA
78
Saleh, K. and Vámossy, Z.
BBBD: Bounding Box Based Detector for Occlusion Detection and Order Recovery.
DOI: 10.5220/0011146600003209
In Proceedings of the 2nd International Conference on Image Processing and Vision Engineering (IMPROVE 2022), pages 78-84
ISBN: 978-989-758-563-0; ISSN: 2795-4943
Copyright
c
2022 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved