tion of the conjectured conclusion. Induction serves
little strategical function because the range of possi-
ble conclusions is restricted by the methods of gener-
alizing prior observed cases.
In this paper, we present a novel abductive search
method capable of handling the examples previously
described. Our model first uses abductive search for
hypothesis identification. To limit redundancy in the
abductive search results, we introduce two distinct
similarity metrics that compare causal structures of
variables. Additionally, to account for possible un-
familiar causes, we implement hypothesis generation
in our model as a method of generating novel ex-
planations—hypothesized causes that are not neces-
sarily observed in the background information. Fi-
nally, while abductive “confirmation” does not in-
dicate whether an abduced hypothesis logically pre-
cedes the observed effects, our model utilizes a hy-
pothesis comparison method to compare hypotheses
based on the likelihood of the explanation.
Cox et al. used abduction with surface deduction
to generate novel hypotheses from Horn-clauses, and
suggested extending this method’s application to ab-
duction from directed graphs (Cox et al., 1992). Our
abductive search model relies on Reichenbach’s Com-
mon Cause Principle rather than surface deduction
for hypothesis generation, and uses edit-distance and
Jaccard-based reasoning to distinguish redundant hy-
potheses. This combination of creative and proba-
bilistic abduction with similarity-based reasoning for
abductive search is distinct from the approach of Cox
et al. (Cox et al., 1992). The use of Reichenbach’s
Common Cause Principle is inspired by Schurz’s
theory on common cause abduction (Schurz, 2008).
While Schurz seems to deny the potential usefulness
of integrating common cause and Bayesian reasoning,
we introduce a form of Bayesian confirmation that
provides probabilistic explanations for the hypotheses
discovered through common cause abduction. Like-
wise, while our abductive model checks consistency
and simplicity similarly to Reiter’s heuristic diagno-
sis model (Reiter, 1987), we rely on Bayesian condi-
tioning during hypothesis generation and comparison,
which strengthens the plausibility of our model’s con-
jectured hypotheses.
2 GRAPHICALLY MODELING
ABDUCTION
Graphical models are tools for integrating logical and
probabilistic reasoning in order to represent rational
processes and causal relationships. Developed by
Pearl, they comprehensively account for complexity
and uncertainty within a dataset (Pearl, 1998). A
probabilistic graphical model is composed of nodes
representing random variables, and edges connecting
the nodes to indicate conditional independence or de-
pendence.
An abductive search problem can be represented
in a directed acyclic graph (DAG), in which a di-
rected edge from one node (the “parent”) to another
(the “child”) represents a causal relationship between
them. For edges of a DAG that are weighted with the
conditional probability P(child | parent) of the child
variable given the parent, the weight speaks to the
causal relationship’s influential strength.
2.1 Bayesian Networks
We adapt the definition of Bayesian network from
(Feldbacher-Escamilla and Gebharter, 2019), and
make use of conventional notation: Sets of objects, in-
cluding sets of sets, are represented by boldfaced up-
percase letters (e.g., S). Variables are represented by
upper-case letters (e.g., X ), and their respective real-
izations are represented by corresponding lower-case
letters (e.g., x). Additionally, a directed edge between
two variables is represented by an arrow, →, where
the parent node is at the arrow’s tail and the child node
is at the tip (e.g., X
i
→ X
j
).
Following the definitions in (Feldbacher-
Escamilla and Gebharter, 2019), BhV
V
V ,E
E
E,Pi is a
Bayesian network such that V
V
V is a set of random
variables, E
E
E is a set of directed edges, and P is a
probability distribution over V
V
V .
For all X
i
∈ V
V
V , P
P
Pa
a
ar
r
r(X
i
) is the set of X
i
’s parents:
P
P
Pa
a
ar
r
r(X
i
) = {X
j
∈ V
V
V | X
j
→ X
i
}. (1)
The set of X
i
’s children is defined as
C
C
Ch
h
h(X
i
) = {X
j
∈ V
V
V | X
i
→ X
j
}. (2)
We define the set of X
i
’s descendants to be
D
D
De
e
es
s
s(X
i
) = {X
j
∈ V
V
V | X
i
→ ... → X
j
}, (3)
and the set of X
i
’s ancestors to be
A
A
An
n
nc
c
c(X
i
) = {X
j
∈ V
V
V | X
j
→ ... → X
i
}. (4)
Within the context of this paper, all variables in V
V
V
are discrete. To properly incorporate continuous vari-
ables into the model, the discretization approach pre-
sented in (Chen et al., 2017) can be used with a dis-
cretization runtime of O(r ·n
2
), where r is the number
of class variable instantiations. Furthermore, Freid-
man et al. present a method of discretizing continuous
variables while learning the structure of the Bayesian
network using background information, that is, data
denoting the values of previous instantiations of vari-
ables in V
V
V (Friedman et al., 1996).
A Probabilistic Theory of Abductive Reasoning
563