categorized into two main approches: SUM and AV-
ERAGE. We denoted by SUM methods which group
observations into overlapping clusters while minimiz-
ing the sum of distances between each observation
and the sum of clusters’ representatives (prototypes
or centroids) to which the observation belongs to. Ex-
amples of these methods are Principal Cluster Anal-
ysis (PCL) (Mirkin, 1987b) with its variants (Mirkin,
1987a; Mirkin, 1990), the Alternating Least Square
algorithms (ALS) (Depril et al., 2008; Wilderjans
et al., 2012) and the Lowdimensional Additive Over-
lapping Clustering (Depril et al., 2012).
Conversely, methods based on AVERAGE ap-
proach group observations into overlapping clusters
while minimizing the sum of distances between each
observation and the average, instead of the sum, of
clusters’ representatives to which the observation be-
longs to. Examples of these methods are the Over-
lapping k-means (OKM) (Cleuziou, 2008), Kernel
Overlapping k-means (KOKM) (N’cir et al., 2010),
Overlapping k-Medoid (OKMED) (Cleuziou, 2009),
the Evidential c-means (ECM) (Masson and Denux,
2008) and Overlapping Clustering with Sparseness
Constraint (Lu et al., 2012).
All these methods extend k-means to take into ac-
count that an observation belongs to several clusters.
Despite different approaches are used by these meth-
ods, they are considered as generalization of k-means
to overlapping clustering (Cleuziou, 2008; Mirkin,
1990; Depril et al., 2008). If each observation is as-
signed to only one cluster, objectivecriteria optimized
by these methods exactly match with the objectivecri-
terion of k-means. The aim of this paper is to study
patterns induced by AVERAGE and SUM approaches
used to model overlapping clustering. We compare
effectiveness of OKM (AVERAGE based method) and
ALS (SUM based method) to identify overlapping
groups. We discuss cases in which these models can
be applied in real life applications.
This paper is organized as follows: Section 2 and
Section 3 describe respectively OKM and ALS meth-
ods. Then, Section 4 presents discussions on pat-
terns induced by OKM and ALS and describes clus-
tering applications in which these methods were ap-
plied. Section 5 presents experiments performed on
real overlapping data sets to check effectiveness of
OKM and ALS in detecting overlapping clusters. Fi-
nally Section 6 presents conclusion and future works.
2 OVERLAPPING k-Means (OKM)
OKM introduces the overlapping constraint (an ob-
servation can belong to more than one cluster) in the
usual squared error objective function. The function
models a local error on each observation x defined
by the squared Euclidean distance between x and it’s
representative in the clustering, denoted as “image”
(im(x)). Given a dataset X with N data over R
P
and a
number K of expected clusters, the aim of OKM is to
find the binary assignment matrix Π(N × K) and the
cluster representatives (prototypes) C = {c
1
,...,c
K
}
such that the following objective function is mini-
mized:
J
OKM
(Π,C) =
∑
x
i
∈X
k x
i
− im
Π,C
(x
i
) k
2
, (1)
where im
Π,C
(x
i
) is the average combination of cluster
representatives. Let Π
i
the set of clusters to which
x
i
belongs and |Π
i
| the number of clusters for x
i
, the
im
Π,C
(x
i
) is described by:
im
Π,C
(x
i
) =
∑
k∈Π
i
c
k
|Π
i
|
. (2)
The minimization of the objective function is per-
formed by iterating two principal steps:
1. computation of cluster representatives (C).
2. multi assignment of observationsto one or several
clusters (Π).
The update of representatives is performed locally for
each cluster. For the multiple assignment step, the
OKM method uses an heuristic to explore part of
the combinatorial set of possible assignments. The
heuristic consists, for each observation, in sorting
clusters from closest to the farthest, then assigning
the observation in the order defined while assignment
minimizes the distance between the observation and
its image. The stopping rule of algorithm is charac-
terized by two criteria: the maximum number of iter-
ations or the minimum improvement of the objective
function between two iterations.
3 ALTERNATING LEAST
SQUARE (ALS)
ALS is based on the Additive Overlapping Clustering
model (Mirkin, 1990). This model introduces the pos-
sibility that an observation belongs to more than one
cluster by considering variable values of an observa-
tion equals to the sum of the clusters’ profiles (pro-
totypes) to which the observation belongs to. Given
a dataset X with N data over R
P
and a number K of
expected clusters, a model matrix M = ΠC is looked
for to optimally approximate X. The matrix M can be
estimated by minimizing the least squares loss func-
tion:
OntheExtensionofk-MeansforOverlappingClustering-AverageorSumofClusters'Representatives?
209