posed dictionary-based hyperspectral CBIR system.
The paper is divided as follows: Sections 2 and 3
briefly review the NCD and the dictionary distances,
FDD and NDD, respectively. Section 4 introduces the
proposed Dictionary-based hyperspectral CBIR sys-
tem. Section 5 presents the experimental methodol-
ogy. Section 6 gives the results. Finally, we present
some conclusions and further work in Section 7.
2 NORMALIZED COMPRESSION
DISTANCE
The conditional Kolmogorov complexity of a signal x
given a signal y, K (x|y), is the length of the shortest
program running in an universal Turing machine, that
outputs x when fed with input y. The Kolmogorov
complexity of x, K(x), is the length of the shortest
program that outputs x when fed with the empty sig-
nal λ, that is, K (x) = K (x|λ). The information dis-
tance, E (x, y), is an universal metric distance defined
as the length of the shortest binary program in a Tur-
ing sense that, from input x outputs y, and from input
y outputs x. It is formulated as:
E (x, y) = max{K (x|y) , K(y|x)}. (1)
The normalized information distance, NID(x, y), is
defined as:
NID(x, y) =
E (x, y)
max{K(x), K (y)}
. (2)
The NID is sometimes known as the similarity
metric due to its universality property. Here, uni-
versality means that for every admissible distance
D(x, y), the NID is minimal, E (x, y) ≤ D(x, y), up to
an additive constant depending on D but not on x and
y. However, NID(x, y) relies on the notion of Kol-
mogorov complexity which is non-computable in the
Turing sense.
The normalized compression distance,
NCD(x, y), is a computable version of (2) based
on a given compressor, C. It is defined as:
NCD(x, y) =
C(xy) − min{C(x), C(y)}
max{C(x), C(y)}
(3)
where C(·) is the length of a compressed signal by
using compressor C, and xy is the signal resulting of
the concatenation of signals x and y. If the compres-
sor C is normal, then the NCD is a quasi-universal
similarity metric. In the limit case when C(·) = K (·),
the NCD(x, y) becomes “universal”. The NCD(x, y)
differs from the ideal NID(x, y)-based theory in three
aspects (Cilibrasi and Vitanyi, 2005): (a) The univer-
sality of NID(x, y) holds only for indefinitely long
sequences x,y. When dealing with sequences of fi-
nite length n, universality holds only for normalized
admissible distances computable by programs whose
length is logarithmic in n. (b) The Kolmogorov com-
plexity is not computable, and it is impossible to know
the degree of approximation of NCD(x, y) with re-
spect to NID(x, y). (c) To calculate the NCD(x, y)
an standard lossless compressor C is used. Although
better compression implies a better approximation to
Kolmogorov complexity, this may not be true for
NCD(x, y). A better compressor may not improve
compression for all items in the same proportion. Ex-
periments show that differences are not significant if
the inner requirements of the underlying compressor
C are not violated.
3 DICTIONARY DISTANCES
The use of NCD (3) for CBIR entails an unafford-
ably cost due to the requirement of compressing the
concatenated signals, C(xy). To deal with this prob-
lem, we propose the use of distances based on the
codewords of the dictionaries extracted by means of
dictionary-based compressors, such as the LZW for
text strings. This dictionary approach only requires
set operations to calculate the distance between two
signals given that the dictionaries have been previ-
ously extracted. Thus, dictionary distances are suit-
able for mining large image databases where the dic-
tionaries of the images in the database can be ex-
tracted off-line.
Given a signal x, a dictionary-based compression
algorithm looks for patterns in the input sequence
from signal x. These patterns, called words, are subse-
quences of the incoming sequence. The compression
algorithm result is a set of unique words called dictio-
nary. The dictionary extracted from a signal x is here-
after denoted as D(x), with D(λ) =
/
0 only if λ is the
empty signal. The union and intersection of the dic-
tionaries extracted from signals x and y are denoted as
D(x∪ y) and D(x ∩ y) respectively. The dictionaries
satisfy the following properties (correspondent proofs
can be found in (Macedonas et al., 2008)):
1. Idempotency: D(x ∪ x) = D(x).
2. Monotonicity: D(x∪ y) ≥ D(x).
3. Symmetry: D(x∪ y) = D(y∪x).
4. Distributivity: D(x ∪ y) + D(z) ≤ D(x∪ z) +
D(y ∪ z).
We have found two dictionary distance functions on
the literature, the Normalized Dictionary Distance
DICTIONARY BASED HYPERSPECTRAL IMAGE RETRIEVAL
427