The Mutual Information of x,y is defined as
MI(x,y) = H(x) − H(x|y) and can be explicitly ex-
pressed as
MI(x,y) =
∑
¯x∈X,¯y∈Y
p( ¯x, ¯y)log
p( ¯x, ¯y)
p( ¯x)p(¯y)
≥ 0 (1)
When the two variables are independent, the joint
probability distribution factorizes and the MI van-
ishes:
p( ¯x, ¯y) = p( ¯x)p(¯y) ⇒ MI(x, y) = 0.
Mutual Information is therefore a measure of depen-
dence between two discrete random variables and is
used by the REVEAL algorithm (Liang et al., 1998)
to infer causal relations between genes: for each gene
in the genome, a time series of its rate of expression
(called gene profile) is gathered from multiple DNA-
microarray experiments; an example is depicted in
Figure 1, with time samples on the x-axis and inten-
sity of gene expression on the y-axis.
Figure 1: example of time series representing the expression
of a gene.
To apply REVEAL algorithm, gene profiles are
then quantized in two levels, 0 (underexpressed) and 1
(overexpressed), and Mutual Information is computed
between all possible pairs of genes. In the specific
case probabilities are computed as the frequencies of
the symbols 0 or 1 within a given sequence; since the
sum of the probabilities being 0 or 1 must be equal to
unity, p(1) = 1 − p(0) and the formula for the entropy
becomes:
H(x) = −p(0)· log(p(0))− (1 − p(0)) ·log(1− p(0))
The joint probability is computed as a the proba-
bility of co-occurrence of two symbols.
Example 1. Consider two random variables x and y,
representing the quantization of two time series in two
levels, 0 and 1; for each variable, consider two se-
quences of 10 symbols: x
0
= {0,1,1,1,1,1,1, 0,0, 0}
and y
0
= {0,0,0,1, 1,0, 0,1, 1,1}. Then for variable x
we obtain
p(0) = 0.4 and p(1) = 0.6 = 1 − p(0)
that means 40 % of zeros and 60% of ones respec-
tively. As for joint probabilities, in one case out of 10
∃i : x
0
i
= 0 ∧ y
0
i
= 0, therefore
p(0,0) = 0.1
The remaining combinations of symbols are
p(0,1) = 0.3, p(1, 0) = 0.4 and p(1,1) = 0.2.
The algorithm infers causal relations between
pairs whose MI is above a given threshold.
3 FUZZY EXTENSION OF THE
REVEAL ALGORITHM
The classical REVEAL Algorithm is based on a
Boolean model, therefore it has to approximate a real
signal with just two symbols 0 and 1; it is clear that in
this way much information is lost.
Using the Fuzzy Sets framework it is possible to
obtain a more flexible and expressive model.
3.1 Membership Functions and
Conditional Probability
In this paper the point of view of Coletti and Scoz-
zafava (Coletti and Scozzafava, 2002; Coletti and
Scozzafava, 2006) has been adopted.
Let x be a random quantity with range X, the fam-
ily {x = ¯x, ¯x ∈ X} is obviously a partition of the sam-
ple space Ω (de Finetti, 1970); let then ϕ be any prop-
erty related to the random quantity x: notice that a
property, even if expressed by a proposition, does not
single out an event, since the latter needs to be ex-
pressed by a non-ambiguous statement that can be ei-
ther true or false. For this reason the event referred
by a property will be indicated with E
ϕ
, meaning
“You claim E
ϕ
” (in the sense of De Finetti (de Finetti,
1970)).
Coletti and Scozzafava state that a membership
function can be defined as a Conditional Subjective
Probability between two events E
ϕ
and x = ¯x, mean-
ing that “You believe that E
ϕ
holds given x = ¯x”.
µ
E
ϕ
( ¯x) = P(E
ϕ
|x = ¯x)
The membership degree µ
E
ϕ
( ¯x) is just the opin-
ion of a real (or fictitious) person, for instance, a
“randomly” chosen one, which is uncertain about it,
whereas the truth-value of that event x = ¯x is well de-
termined in itself. Notice that conditional probabil-
ity between events E
ϕ
and x = ¯x can be directly in-
troduced rather than being defined as the ratio of the
unconditional probabilities P(E
ϕ
∧ ¯x) and P(x = ¯x).
From the same paper we report also the following ex-
ample.
IJCCI 2009 - International Joint Conference on Computational Intelligence
26