2.3 Training and Classification
The hierarchy is trained by separately training the in-
dividual classifiers with the data {x
µ
∈ X
i
|t
µ
∈ C
i
}
that belong to the subsets of classes assigned to each
classifier. For the training the respective feature type
X
i
identified during the hierarchy generation phase
is used. The data will be relabelled so that all data
points of the classes belonging to one subset C
i,j
have
the same label j, i.e.
˜
t
µ
= j, x
µ
∈ X
i
, t
µ
∈ C
i,j
.
The number of input nodes of the single classifiers is
defined by the dimension d
i
of the respective feature
type X
i
assigned to the corresponding node i. The
number of output nodes equals the number of succes-
sor nodes s
i
. The classifiers are trained using super-
vised learning algorithms. The classifiers within the
hierarchy can be trained independently, i.e. all classi-
fiers can be trained in parallel.
Within the hierarchy different types of classifiers
can be used. Examples of classifiers would be radial
basis function (RBF) networks, linear vector quanti-
sation classifiers (Simon et al., 2002) or support vec-
tor machines (Schwenker, 2001). We chose RBF net-
works as classifiers. They were trained with a three
phase learning algorithm (Schwenker et al., 2001).
The classification result is obtained similar to the
retrieval process in a decision tree (Duda et al., 2001).
Starting with the root node the respective feature vec-
tor of the object to be classified is presented to the
trained classifier. By means of the classification out-
put the next classifier to categorise the data point is
determined, i.e. the classifier j
∗
corresponding to
the highest output value o(j
∗
) is chosen such that
j
∗
= argmax
j=1..s
i
(o(j)). Thus a path through the
hierarchy from the root node to an end node is ob-
tained which not only represents the class of the ob-
ject but also the subsets of classes to which the ob-
ject belongs. Hence the data point is not presented to
all classifiers within the hierarchy and the hierarchi-
cal decomposition of the classification problem yields
additional intermediate information.
If only intermediate results are of interest it might
not be necessary to evaluate the complete path. In
order to solve a task it might be sufficient to know
whether the object to be recognised belongs to a set
of classes and the knowledge of the specific category
of the object might not add any value. If the task for
example is to grasp a cup, it is not necessary to dis-
tinguish between red and green cups. Moreover, when
looking for a specific object it might in some cases not
be necessary to retrieve the final classification result
if a decision at a higher level of the hierarchy already
excludes this object.
2.4 Incremental Learning of New
Classes
While performing tasks a robot is likely to encounter
unfamiliar objects. Hence the ability to learn new
objects plays an important role. Given the situa-
tion that an object is in the robot’s visual field and
it is told the name of the new object by an instruc-
tor, the robot might look at the object from different
points of view and thereby generate a few data sam-
ples
˜
S := {(x
ν
, ˜c), ν = 1, ..., N } of the new class
˜c /∈ C. So compared to the already known classes
C the number of samples N of the new class ˜c is ap-
preciably lower. This scenario forms the basis for the
incremental learning approach developed here. In the
overall system the online learning is triggered by in-
struction sentences starting with ”this is” followed by
an object name.
In order to quickly achieve results the learning is
performed in two stages. In the first stage fast but
less sophisticated methods are used to obtain initial
results, i.e. the novel objects are learnt but the recog-
nition rate might be weak. In this first stage the recog-
nition rate can be improved by using a similar method
to retrain the new object with additional data gained
by taking different views of the object. In a second
stage more complex algorithms are used to adapt the
system and to further improve the classification re-
sults. This retraining is more advanced but also more
time consuming.
At first it is necessary to identify whether the pre-
sented object is already known or not. This is accom-
plished by presenting the new data to the trained clas-
sifier and taking the strength of the classifier response
into account. Thereby a strong response is consid-
ered as an unambiguous decision and weak responses
indicate a dubious decision which could be evoked
by unknown classes or if the object to classify bears
resemblance to more than one class. The thresholds
for this are derived from the classifier responses when
testing the known data. If the new data is unambigu-
ously classified as one class c without exception it is
assumed that the object is already known and belongs
to class c. Otherwise the object is regarded as a hith-
erto unidentified object.
If an object is identified as unfamiliar it is learnt by
fitting it into the hierarchy and if necessary retraining
the affected nodes. The new class ˜c /∈ C associated
with the unknown object is inserted as a new leaf. The
position of the new leaf is determined by classifying
all new data and evaluating the corresponding paths
through the hierarchy. The leaf will be inserted where
the paths start to diverge. As complete identicalness
for all data cannot be presumed even at the root node
since the network has not been trained with this data
a certain variance needs to be considered. Otherwise
ICINCO 2005 - ROBOTICS AND AUTOMATION
300