2 PREVIOUS WORK
The field of active learning is a growing branch of the
very present machine learning domain. It is also re-
ferred to as optimal experimental design (Cohn, 1996).
(Settles, 2009) shows a broad overview of the cur-
rent state of the art in this discipline and gives an
outlook to multiple possible future work fields. Recent
methodological advances in the scientific community
mainly focused on classification problems. The main
application domains are speech recognition and text
information extraction (Settles, 2009).
While regression tasks in the context of active
learning have not been as popular, the methodolog-
ical development is relevant as well. (Sugiyama and
Rubens, 2008) propose an approach which actively
learns multiple models for the same task and picks
the best one to query new points. (Cai et al., 2013)
introduced an approach which uses expected model
change maximization (EMCM) to improve the active
learning progress for gradient boosted decision trees,
which was later extended to choose a set of informa-
tive queries and to gaussian process regression models
(GPs) by (Cai et al., 2017). (Park and Kim, 2020) pro-
pose a learning algorithm based on the EMCM, which
handles outliers more robustly than before. Those
publications focus on new criteria for single output
regression models to improve the active learning pro-
cess. (Zhang et al., 2016) present a learning algorithm
for multiple-output gaussian processes (MOGP) which
outperforms multiple single-output gaussian processes
(SOGP). However, this publication focuses on improv-
ing the prediction accuracy of one target output with
the help of several correlated auxiliary outputs. The
experiments indicate that a global consideration is ben-
eficial.
There were also advances in active learning for au-
tomotive calibration tasks for which the identification
of multiple process outputs in the same experiment is
more relevant to the application. (Klein et al., 2013)
applied a design of experiments for hierarchical local
model trees (HiLoMoT-DoE), which was presented by
(Hartmann and Nelles, 2013), successfully to an en-
gine calibration task. They presented two application
examples with two outputs each and five respectively
seven inputs. The two outputs were modeled with a
sequential strategy, which identifies an output model
completely before moving to the next one (Klein et al.,
2013).
(Dursun et al., 2015) applied the HiLoMoT-DoE
active learning algorithm to a drivability calibration ex-
ample characterized by multiple static regression tasks
with identical input spaces. They analyzed the sequen-
tial strategy already shown by (Klein et al., 2013) and
compared it to a round-robin strategy, which switches
the leading model after each iteration/measurement
(Dursun et al., 2015). The authors show that the round-
robin strategy outperforms offline methods and the
online sequential strategy in this experiment. It might
indicate, that round-robin is preferably used in gen-
eral, but further experiments are necessary. Since then,
no efforts have been made to analyze active learning
strategies for multiple outputs.
3 PROBLEM DEFINITION
The analyses of this paper are motivated by the field
of model-based drivability calibration. For this ap-
plication, an active learning algorithm learns a num-
ber of
M
different outputs, which are possibly non-
correlated. Their models are equally important for
succeeding optimizations, so the goal is to identify
adequate models for all outputs. The input dimensions
of all models are the same. Querying a new instance
corresponds to conducting a measurement on power-
train test benches. Therefore, a measurement point
is cost-sensitive, which is inherent to active learning
problems. Contrary to other applications, every single
measurement provides values for all
M
outputs
1
. Tasks
of simultaneously learning
M > 1
process outputs with
equal priority and multi-output measurements are not
known in the scientific community. In the following,
they are referred to as active output selection (AOS).
All measured outputs contain to some extent noise.
The signal-to-noise-ratio
SNR
m
of model
m
is the ra-
tio between the range of all measurements
y
m
and the
standard deviation
σ
N
of normally distributed noise:
SNR
m
=
max(y
m
)−min(y
m
)
σ
N
. The
SNR
for drivability cri-
teria lies approximately in a range of
(7 . . 100)
and
can be different for each criterion.
For applications on a test bench, conducting a mea-
surement is timely more expensive than the evaluation
of code. This is why the performance of code is not
crucial in this context and is only discussed openly in
this paper instead of analyzing it systematically.
4 ACTIVE OUTPUT SELECTION
STRATEGIES
This paper analyzes strategies for AOS tasks with
M > 1
regression models. In this paper, each of those
M
process outputs is modeled with a GP since they
1
This is in contrast to e. g. geostatistics, where measuring
any individual output, even at the same place (i. e. model
inputs), has its own costs (Zhang et al., 2016).
Active Output Selection Strategies for Multiple Learning Regression Models
151