timed data. This representation facilitates also the
building for the CPT tables.
The next section presents a short description of
the state of the art techniques concerning the
learning of Dynamic Bayesian Networks (DBN).
Section 3 introduces the basis of the Stochastic
Representation of TOM4L and the BJ-Measure.
Section 4 describes the learning principles we define
from the properties of the BJ-Measure, the DBN
learning algorithm that we proposes is proposed in
section 5 and an application to a theoretical example
is given in section 6 before showing a real life
application of the algorithm in section 7. Our
conclusions are presented in section 8.
2 RELATED WORKS
A BN is a couple <G, > where G denotes a Direct
Acyclic Graph in which the nodes represent the
variables and the edges represent the dependencies
between the variables (Pearl, 1988), and is the
Conditional Probabilities Tables (CP Tables)
defining the conditional probability between the
values of a variable given the values of the upstream
variables of G. BN learning algorithms aims at
discovering the couple <G, > from a given data
base.
BN learning algorithms fall into two main
categories: “search and scoring” and “dependency
analysis” algorithms. The “search and scoring”
learning algorithms can be used when the knowledge
of the edge orientation between the variables of the
system is given (Cooper, 1992), (Heckerman, 1997).
To avoid this problem, dependency analysis
algorithms uses conditional independence tests
(Cheng, 1997), (Cheesseman, 1995), (Friedman
1998), (Meyrs et al, 1999). But the number of test
exponentially increases the computation time
(Chickering, 1994).
(1)
For example, Cheng’s algorithm (Cheng, 1997) for
learning a BN from data falls in the dependency
analysis category and is representative of most of the
proposed algorithms. It is based on the d-separation
concept of (Pearl, 1988) to infer the structure G of
the Bayesian Network, and the mutual information
to detect conditional independency relations. The
idea is that the mutual information I(X, Y) (eq. 1)
tells when two variables are (1) dependent and (2)
how close their relationship is. The algorithm
computes the mutual information I(X, Y) between all
the pairs of variables (X, Y) producing a list L sorted
in descending order: pairs of higher mutual
information are supposed to be more related than
those having low mutual information values. The
List L is then pruned given an arbitrary value of the
parameter
ε
: each pair (X, Y) so that I(X, Y)< is
eliminated of L. In real world applications, list L
should be as small as possible using the parameter.
This first step (Drafting) creates a structure to start
with but it might miss some edges or it might add
some incorrect edges.
(2)
The second step (Thickening) phase tries to separate
each pair (X, Y) in L using the conditional mutual
information I(X, Y | E) (eq. 2) where E is a set of
nodes that forms a path between the current tested
nodes X and Y from L. When I(X, Y | E)>, then the
edges of the path E should be added between the
current nodes X and Y. This phase continues until the
end of list L is reached. The last step of the
algorithm (Thinning) searchs, for each edge in the
graph, if there are other paths besides this edge
between these two nodes. In that case, the algorithm
removes this edge temporarily and tries to separate
these two nodes using equation (2). If the two nodes
cannot be separated, then the temporarily removed
edge will be returned. After building the DBN
structure, the orientation of the edges and the CP
Tables’ computation is to be done. The procedure
used by (Cheng, 1997) is based on the idea of
searching for the nodes forming a V-Structure
X→Y←Z using the conditional mutual information,
and then trying to deduce the other edges from the
discovered one. This procedure have a very big
limitation which is that if a network does not contain
a V-Structure, no edge can be oriented.
The two main limitations of the methods of the
dependency analysis category are so the need of
defining the parameter and the exponential amount
of Conditional Independence tests to orient the edges
of the graph.
3 TOM4L FRAMEWORK
The Timed Observation Mining for Learning
process (TOM4L) proposes a solution to escape
from this problem (Le Goc, 2006).
ICSOFT 2010 - 5th International Conference on Software and Data Technologies
354