Authors:
Mirco Gallazzi
;
Anwar Ur Rehman
;
Silvia Corchs
and
Ignazio Gallo
Affiliation:
Department of Theoretical and Applied Science, University of Insubria, 21100 Varese, Italy
Keyword(s):
YOLO, Swin Transformer, Object Detection and Segmentation, Medical Imaging, Dermatology, Skin Cancer.
Abstract:
Deep Learning plays a vital role in medical imaging, especially in classification and segmentation tasks essential for diagnosing diseases from images. However, current methods often struggle to differentiate visually similar classes and accurately delineate lesion boundaries. This study builds on prior findings of classification limitations, investigating whether segmentation can improve classification performance for skin lesion analysis with Transformer-based models. We benchmarked the segmentation capabilities of the Swin Transformer, YOLOv8, and DeepLabV3 architectures on the HAM dataset, which contains 10,015 images across seven skin lesion classes. Swin outperformed others in segmentation, achieving an intersection over union of 82.75%, while YOLOv8 achieved 77.0%. However, classification experiments using classification datasets after segmenting and cropping the lesion of interest did not produce the expected improvements, with classification accuracy showing slight drops in
the segmented data. For example, on the original HAM dataset, the model achieved a Test Accuracy (TA) of 84.64%, while Swin trained on segmented data showed a slight decline to a TA of 84.13%. These findings suggest that segmentation alone may not effectively support classification. Based on this, we propose future research into a sequential transfer learning approach, where segmentation knowledge could be progressively transferred to improve classification.
(More)