The mAP metric used in evaluating the Mask R-
CNN model customized for this problem had results
between 0.2333 and 0.2585 for an IoU of 0.5. In the
individual evaluation of the AP metric of each image,
we verified that the AP values were lower for images
that contained many particles present in the sample
and higher AP values for images that contained few
particles present.
This factor may be associated with a small number
of labels in the validation set (only 10 per image), in-
creasing the probability for images with few particles
that the predicted value was associated with labeled
ground truth. For better AP results for these classes,
we suggest in future work that a greater proportion of
particles be labeled in the validation set and that the
training and validation sets have particles from differ-
ent classes labeled in the same image.
We emphasize that our implementation has the
challenge of working with images derived from an
industrial environment. These images are complex,
as they present homogeneity in color, texture, com-
plex background, overlapping, and occlusion. Fur-
thermore, we did not find any database available for
the implementation, and we designed our database.
From the results obtained in this step, it was pos-
sible to raise new hypotheses of approaches to im-
prove the algorithm to obtain the particle size distri-
bution of the quasi-particles present in a sample in fu-
ture work. The development of applied solutions with
deep learning can bring significant benefits, both in
the improvement of processes and in the insertion of
steelmaking processes in Industry 4.0.
REFERENCES
Bello, R.-W., Mohamed, A. S. A., and Talib, A. Z. (2021).
Contour extraction of individual cattle from an im-
age using enhanced mask r-cnn instance segmentation
method. IEEE Access, 9:56984–57000.
Chen, D., Guo, D., Liu, S., and Liu, F. (2020). Microstruc-
ture instance segmentation from aluminum alloy met-
allographic image using different loss functions. Sym-
metry, 12(4):639.
Cohn, R., Anderson, I., Prost, T., Tiarks, J., White, E., and
Holm, E. (2021). Instance segmentation for direct
measurements of satellites in metal powders and au-
tomated microstructural characterization from image
data. JOM, pages 1–14.
De Cesaro J
´
unior, T. et al. (2020). Insectcv: um sistema
para detecc¸
˜
ao de insetos em imagens digitais.
Girshick, R. (2015). Fast r-cnn. In Proceedings of the IEEE
international conference on computer vision, pages
1440–1448.
He, K., Gkioxari, G., Doll
´
ar, P., and Girshick, R. (2017).
Mask r-cnn. In Proceedings of the IEEE international
conference on computer vision, pages 2961–2969.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 770–778.
Johnson, J. W. (2018). Adapting mask-rcnn for au-
tomatic nucleus segmentation. arXiv preprint
arXiv:1805.00500.
LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learn-
ing. nature, 521(7553):436–444.
Long, J., Shelhamer, E., and Darrell, T. (2015). Fully con-
volutional networks for semantic segmentation. In
Proceedings of the IEEE conference on computer vi-
sion and pattern recognition, pages 3431–3440.
Ren, S., He, K., Girshick, R., and Sun, J. (2015). Faster
r-cnn: Towards real-time object detection with region
proposal networks. Advances in neural information
processing systems, 28:91–99.
Soloy, A., Turki, I., Fournier, M., Costa, S., Peuziat, B.,
and Lecoq, N. (2020). A deep learning-based method
for quantifying and mapping the grain size on pebble
beaches. Remote Sensing, 12(21):3659.
Tu, S., Yuan, W., Liang, Y., Wang, F., and Wan, H.
(2021). Automatic detection and segmentation for
group-housed pigs based on pigms r-cnn. Sensors,
21(9):3251.
Vuola, A. O., Akram, S. U., and Kannala, J. (2019). Mask-
rcnn and u-net ensembled for nuclei segmentation. In
2019 IEEE 16th International Symposium on Biomed-
ical Imaging (ISBI 2019), pages 208–212. IEEE.
Wu, Q., Feng, D., Cao, C., Zeng, X., Feng, Z., Wu, J.,
and Huang, Z. (2021). Improved mask r-cnn for air-
craft detection in remote sensing images. Sensors,
21(8):2618.
Xi, D., Qin, Y., and Wang, Y. (2020). Vision measurement
of gear pitting under different scenes by deep mask
r-cnn. Sensors, 20(15):4298.
Xie, S., Girshick, R., Doll
´
ar, P., Tu, Z., and He, K. (2017).
Aggregated residual transformations for deep neural
networks. In Proceedings of the IEEE conference on
computer vision and pattern recognition, pages 1492–
1500.
Xu, B., Wang, W., Falzon, G., Kwan, P., Guo, L., Sun, Z.,
and Li, C. (2020). Livestock classification and count-
ing in quadcopter aerial images using mask r-cnn. In-
ternational Journal of Remote Sensing, 41(21):8121–
8142.
Zhang, Q., Liu, Y., Gong, C., Chen, Y., and Yu, H. (2020).
Applications of deep learning for dense scenes analy-
sis in agriculture: A review. Sensors, 20(5):1520.
Zhang, R., Cheng, C., Zhao, X., and Li, X. (2019).
Multiscale mask r-cnn–based lung tumor detec-
tion using pet imaging. Molecular imaging,
18:1536012119863531.
Zhao, K., Kang, J., Jung, J., and Sohn, G. (2018). Building
extraction from satellite images using mask r-cnn with
building boundary regularization. In Proceedings of
the IEEE Conference on Computer Vision and Pattern
Recognition Workshops, pages 247–251.
Mask R-CNN Applied to Quasi-particle Segmentation from the Hybrid Pelletized Sinter (HPS) Process
469