and then learn the conditional probability p(MS
pi
│Sp
i
) representing the relationship between the
superpixel features and their corresponding labels
using a mixture of Gaussians model. In addition, we
empirically define the conditional probability
modeling the relationships between each superpixel
label and each edge state (i.e., true or false edge)
p(E
j
│Pa (E
j
) ), where Pa (E
j
) denotes the parent
superpixel nodes of E
j
.
During the testing stage, we learn the BN
structure encoding the contextual relationships
between superpixels and edge segments.
Specifically, each edge node has for parent nodes the
two superpixel nodes that are separated by this edge.
In other words, each superpixel provides contextual
information to judge whether the edge is on the
object boundary or not. If two superpixels have
different labels, it is more likely that there is a true
object boundary between them, i.e. E
j
=1, otherwise
E
j
=0. Although automatic segmentation methods
based on BN have shown great results in the state-
of-the-art, they may perform poorly in segmenting
low-contrast image regions and different regions
with similar features (Zhang and Ji, 2011). To
further improve the segmentation accuracy of BN,
we propose to include additional information
through cooperative learning using SRF.
2.3 SRF-BN Cooperative Learning
(One Layer)
To take advantage of the strengths of both classifiers
and overcome their limitations, we first propose a
one-layer cooperative learning strategy, where BN
benefits from the learned patch-to-patch mapping by
SRF. First, the trained SRF generates the
segmentation result, using the feature patches
extracted from the different MRI modalities of the
testing subject. Then, BN uses the SRF segmentation
result to define the prior probabilities p(Sp
i
) for each
superpixel region Sp
i
. Hence, with this cooperative
learning, the BN prior probabilities are estimated
based on the input SRF segmentation probability
maps. Such one-layer cooperative learning strategy
only boosts the BN performance since it is
performed in one way (from SRF to BN), while SRF
does not benefit from BN learning.
2.4 Deep SRF-BN Cooperative
Learning (Two Layers)
To address the aforementioned limitation of the one-
layer SRF-BN architecture, we further propose to
deepen the cooperative learning between SRF and
BN in the spirit of auto-context model (Tu and Bai,
2010). Basically, in the proposed deep auto-context
SRF-BN cooperative learning architecture, each
layer inputs the segmentation result of the previous
layer to boost the per formance of both SRF and BN
classifiers. In each layer, excluding the first one,
SRF classifier inputs the segmentation result of the
previous layer along with the original input
multimodal feature patches (Fig. 1). This allows the
integration of contextual features learned at both the
patch level (from SRF in the previous layer) and
superpixel level (from BN in the previous layer).
Similarly, BN inputs the segmentation result of the
previous layer along with the original input
multimodal superpixel features, while adding the
probability segmentation map output of the SRF in
the same layer. In this way, BN prior probabilities
are updated in each layer based on the posterior
probability of the previous layer and the SRF
probability map in the same layer.
2.5 Preprocessing and Features
To improve the performance of our segmentation
framework, we perform a few preprocessing steps.
Hence, we apply the N4 filter for inhomogeneity
correction, and use the histogram linear
transformation for intensity normalization. To train
the previous models, we use conventional features
(e.g., patch intensity) and we also propose a rich
feature set as follows:
▪ Statistical Features: First order operators
(mean, standard deviation, max, min, median,
Sobel, gradient); higher order operators
(laplacian, difference of gaussian, entropy,
curvatures, kurtosis, skewness); texture features
(Gabor filter); spatial context features
(symmetry, projection, neighborhoods)
(Prastawa et al., 2004).
▪ Symmetric Features: This is originally used to
describe and to exploit the symmetrical
properties of the brain structure. Thus, we define
the symmetry descriptor characterizing the
differences between symmetric pixels with
respect to the mid-sagittal plane. The adopted
symmetry measure is the intensity variance.
3 RESULTS AND DISCUSSION
In this section, we display the evaluation results of
our proposed brain tumor segmentation framework
on the Brain Tumor Image Segmentation Challenge
(BRATS, 2015) dataset. It contains brain MRI scans
ICAART 2018 - 10th International Conference on Agents and Artificial Intelligence
186