The F1-score is the harmonic mean between pre-
cision and recall, which an be formulated as follows:
F1 =
2.T P
2.T P + FP + FN
(10)
where TP, FP, FN are the true positive, false pos-
itive, and false negative rates, respectively. The true
negative (TP) rate is defined as T N = GT ∩NP, which
is the area not belonging to any of the two masks GT
and NP.
3.2 Experimental Results and
Discussion
The dataset used in this study has been obtained from
(Kumar et al., 2017). This dataset includes 30 WSI
images with annotations from 7 organs (breast, kid-
ney, colon, stomach, prostate, liver, and bladder) col-
lected at different medical centers. The size of each
image is 1000 × 1000.
The test data includes one image from every organ
that was not exposed to the network. The rest of im-
ages is used for training. Every image in the training
and testing data was scaled to 1024x1024 and divided
into four non-overlapping patches of size 512x512.
Further random cropping of 512x512 from every im-
age was applied, as well. The overall training data
had 4906 of 512x512 patches. A batch size of one im-
age was used due the limitation in resources, namely
the GPU memory. Proposed model was trained for
50 epoch. The stochastic gradient descent (SGD) was
used as an optimizer with an initial learning rate as
1e−1, momentum as 0.99 and weight decay as1e−8.
A Titan X GPU was used to run the experiments Table
1 shows the results of the proposed method and ones
of five state-of-the-art semantic segmentation based
on deep learning models nuclei segmentation. The
five models are Fully Convolutional Network (FCN),
U-Net, Mask R-CNN, and conditional GAN (cGAN).
As shown the proposed method achieves an F1-score
of 0.876 and AJI score of 0.735. These results are bet-
ter than the ones of the previous approach (Mahmood
et al., 2019). We also show that the addition of the
channel wise aggregation improves the performance
of the baseline framework (self attention mechanism
with CE2P).
Figure 3 shows the segmentation results of the
proposed method with different organs: breast, kid-
ney, liver, prostate, bladder, colon, and stomach. As
shown, our method produces good segmentation re-
sults with bladder and stomach histopathological im-
ages with AJI scores of 0.85 and 0.83, respectively.
The proposed method gives a segmentation results
lower than 0.67 with the liver image because of the
apparent overlap between several cell nuclei.
Table 1: Comparison between the proposed model and the
related methods: FCN, U-Net, Mask R-CNN, and cGAN.
Method F1−score AJI
FCN (Long et al., 2015) 0.35 0.35
U-Net (Ronneberger et al., 2015) 0.41 0.41
Mask R-CNN (He et al., 2017) 0.50 0.50
cGAN (Mahmood et al., 2019) 0.87 0.72
Baseline 0.88 0.73
Proposed 0.89 0.74
Figure 4 shows a comparison between the number
of cell nuclei in the predicted masks and the corre-
sponding ground truth. As shown, the number of cell
nuclei obtained by the proposed method is a bit higher
than the ones of the ground-truth.
The proposed model gives promising segmenta-
tion results when we have noisy ground truth masks
because of the ambiguous boundary between differ-
ent cell nuclei and the other objects that have a simi-
lar appearance beside the overlapping and clumped.
If the cell nuclei ground-truth is almost clean, the
channel-wise aggregation, label refinement, and the
self-correction training mechanisms can be seen as an
ensembling of clones of the basic segmentation model
(i.e, CE2P), which would improve the cell nuclei seg-
mentation results and produce a generalized model
that can be used with images of different organs ac-
quired at different medical centers.
4 CONCLUSION
In this paper, we propose a new technique for ag-
gregating the channel maps of semantic segmenta-
tion models in order to improve the segmentation re-
sults of cell nuclei in histopathological images. This
technique is integrated with a self-correction learning
mechanism that can handle noisy ground truth. We
show that the proposed nuclei segmentation method
gives promising results with images of different or-
gans (e.g., breast, bladder, and colon)collected from
medical centres that use devices of different manu-
facturers and stains. Our method achieves the new
state-of-the-art. Particularly, we achieve the AJI
score of 0.735 on the Multi-Organ Nuclei Segmen-
tation benchmark, which outperforms the previous
closest approaches. In the future work, we will ex-
plore the use of different aggregation functions to
improve the segmenting cell nuclei results. We will
also use the proposed segmentation model to segment
breast masses in other modalities, such as thermogra-
phy (Abdel-Nasser et al., 2016a; Abdel-Nasser et al.,
2016b).
VISAPP 2020 - 15th International Conference on Computer Vision Theory and Applications
470