
tools for comparing distributions without requiring la-
belled data, making them ideal for unsupervised drift
detection in industrial applications. Similarly, adver-
sarial approaches (Rabanser et al., 2019), where a do-
main classifier discriminates between source and tar-
get data, offer a flexible method for identifying and
addressing drift in high-dimensional feature spaces.
Additionally, recent methods such as (Greco et al.,
2024) contribute to improving drift detection, while
the survey by (Hinder et al., 2023) provides a com-
prehensive review of the state-of-the-art approaches
in this field.
2.3 Unsupervised Domain Adaptation
for Object Detection
UDA relies on fully labeled instances in the source
domain while having no labels for the target domain.
This approach is particularly relevant in real-world
scenarios where new data often lacks annotations.
UDA has been widely researched for tasks like clas-
sification (Saito et al., 2018) and semantic segmenta-
tion (Toldo et al., 2020). Unlike image classification,
which only requires assigning a label to an entire im-
age, object detection involves both classification and
localization, making the task more complex. Domain
adaptation for object detection must ensure that both
the feature extraction and the bounding box predic-
tion generalize well across domains, adding another
layer of difficulty compared to tasks like image clas-
sification or segmentation.
The survey (Oza et al., 2023) categorizes exist-
ing UDA for object detection methods into different
types. Adversarial feature learning aligns learned fea-
tures across domains by training two competing mod-
els: a generator (feature extractor) and a discrimina-
tor (domain classifier). The generator minimizes the
task loss (e.g., object detection) while trying to con-
fuse the discriminator, which is trained to differentiate
domains. By doing so, the generator learns domain-
invariant features. DA-Faster (Chen et al., 2018) was
one of the first to apply this adversarial approach
proposed in (Ganin and Lempitsky, 2015) within the
Faster R-CNN framework, pioneering UDA for ob-
ject detection and influencing many subsequent works
(Chen et al., 2021).
On the other hand, mean-teacher methods (Cai
et al., 2019) use a teacher-student model to leverage
labeled source data and unlabeled target data, with the
teacher providing pseudo-labels and the student im-
proving performance by aligning with the teacher’s
predictions. Image-to-Image Translation methods
have also been explored aiming to translate images
from one domain to another and create intermediate
images between domains to reduce the gap. For ex-
ample, (Arruda et al., 2019) employs a strategy based
on this to adapt the model from detecting daytime
scenes to nighttime scenes. Pseudo-label based self-
training methods are also popular as they generate
pseudo-labels for unlabeled target data using model
predictions and then retrain the model with both la-
beled source data and these pseudo-labels to boost
performance (Kim et al., 2019).
3 MATERIALS AND METHODS
In this section, the materials and methods used in this
study are detailed. First, the dataset employed for the
experimentation is presented. Next, the process fol-
lowed for drift detection is described. Finally, the pro-
posed domain adaptation architecture is explained in
detail.
3.1 Datasets
In the manufacturing industry, obtaining high-quality
data is a persistent challenge due to strict privacy
and confidentiality around production processes. This
limitation restricts access to diverse datasets and com-
plicates the capture of variations or drift, which are
essential for developing robust models.
Collecting drift data in highly optimized manufac-
turing environments is particularly difficult, as pro-
duction plants aim to minimize variations and main-
tain strict control over operational variables. Conse-
quently, while some types of drift may occur, they are
infrequent and inadequately documented for creating
comprehensive datasets.
Variations in manufacturing can arise from several
factors. Changes in lighting due to fluctuations in nat-
ural or artificial sources can alter product appearance.
Sensor failures or machine vibrations may introduce
noise into images, while the movement of parts or
cameras can lead to blurriness. Dust and dirt particles
in the environment can obstruct visibility and distort
image characteristics.
Given the challenges in obtaining real data that
captures these variations, synthetic drift conditions
have been simulated using the original data. To sim-
ulate drift conditions, we employed traditional data
augmentation techniques, introducing variations such
as changes in brightness, particles, noise, and vibra-
tion. These augmentations were carefully chosen to
reflect real-world industrial scenarios. Ensuring that
the synthetic data mimics true industrial drift con-
ditions is essential for the robustness of the domain
adaptation method. Figure 1 shows a sample of the
VISAPP 2025 - 20th International Conference on Computer Vision Theory and Applications
422