all cases, the suggested scheme is compared with the
supervised state of the art classification, resulting in
outperforming previous works.
A review of the sample selection scheme with its
spatial improvement is presented in Section 2. Sev-
eral classification alternatives are presented in Section
3. Results will be shown and analyzed in Section 5.
Finally, Section 6 presents some conclusions.
2 PRELIMINARIES
Nowadays, due to the improvement in the sensors,
databases used for segmentation and classification of
hyper-spectral satellite images are highly reliable in
terms of spectral and spatial resolution. Therefore,
we can consider that our feature space representation
of the data is also highly reliable. On the other hand,
in segmentation and classification of this kind of im-
ages the training data used has not been a concerned
so far, without worrying about providing the most re-
liable information (Comaniciu and Meer, 2002). The
scheme suggested in (Rajadell et al., 2011) was a first
attempt in this sense. It was proposed an unsupervised
selection of the training samples based on the analysis
of the feature space to provide a representative set of
labeled data. It proceeds as follows:
1. In order to reduce the dimensionality of the prob-
lem, a set of spectral bands, given a desired num-
ber, is selected by using a band selection method.
The WaLuMi band selection method (Mart´ınez-
Us´o et al., 2007) was used in this case, although
any other similar method could be used.
2. A clustering process is used to select the most rep-
resentative samples in the image. In this case,
we have used the Mode Seek clustering procedure
which is applied over the reduced feature space.
An improvement in the clustering process is in-
cluded by adding the spatial coordinates of each
pixel in the image as additional features. Since
the clustering is based on distances, spatial coor-
dinates should also be taken into account assum-
ing the class connection principle.
3. The modes (centers of the clusters) resulting of
the previous step define the training set for the
next step. The expert is involvedat this point, only
once, by providing the corresponding labels of the
selected samples.
4. The classification of the rest of non-selected sam-
ples is performed, using the training set defined
above to build the classifier. Three different clas-
sification experiments have been performed here:
a KNN classifier with k = 1, a direct classification
with the results of the clustering process, and an
extension will be presented for the use of SVM.
2.1 Mode Seek Clustering
Given a hyper-spectral image, all pixels can be con-
sidered as samples which are characterized by their
corresponding feature vectors (spectral curve). The
set of features defined is called the feature space
and samples (pixels) are represented as points in that
multi-dimensional space. A clustering method groups
similar objects (samples) in sets that are called clus-
ters. The similarity measure between samples is de-
fined by the cluster algorithm used. A crucial problem
lies in finding a good distance measure between the
objects represented by these feature vectors. Many
clustering algorithms are well known. A KNN mode
seeking method will be used in this paper (Cheng,
1995). It selects a number of modes which is con-
trolled by the neighborhood parameter (s). For each
class object x
j
, the method seeks the dissimilarity to
its s
th
neighbors. Then, for the s neighbors of x
j
,
the dissimilarities to their s
th
neighbors are also com-
puted. If the dissimilarity of x
j
to its s
th
neighbor is
minimum compared to those of its s neighbors, it is
selected as prototype. Note that the parameter s only
influences the scheme in a way that the bigger it is the
less clusters the method will get since more samples
will be grouped in the same cluster, that is, less modes
will be selected as a result. For further information
about the mode seek clustering method see (Cheng,
1995) and (Comaniciu and Meer, 2002)
2.2 Spatial Improvement
The clustering algorithm searches for local density
maxima where the density function has been calcu-
lated using the distances for each sample in its s
neighborhood using a dissimilarity measure as the
distance between pairs of samples. In that difference,
all features (dimensions) are considered. When fea-
tures do not include any spatial information the class
connection principle is missed (pixels that lie near
in the image are likely to belong to the same class).
Therefore, we suggest to include the spatial coordi-
nates among the feature of the samples. See Fig 1.(a)
where all samples have been represented in the three
first features space and in different color per class.
Notice that, when no spatial data is considered and
all classes are located in the same space and when
no prior knowledge is available for the clustering pro-
cess, finding representatives for each class would be
difficult since the classes themselves may lie together.
Moreover, different areas of the same class may be
AUTOMATIC SELECTION OF THE TRAINING SET FOR SEMI-SUPERVISED LAND CLASSIFICATION AND
SEGMENTATION OF SATELLITE IMAGES
413