Elsevier

Pattern Recognition

Volume 120, December 2021, 108075
Pattern Recognition

Knowledge-aware deep framework for collaborative skin lesion segmentation and melanoma recognition

https://doi.org/10.1016/j.patcog.2021.108075Get rights and content

Highlights

  • We propose a novel end-to-end deep framework that is able to perform skin lesion segmentation and melanoma recognition jointly, where the clinical knowledge is exploited and transferred with the mutual guidance between these two tasks.

  • We design a lesion-based pooling and shape extraction module to transfer the lesion structure information from the skin lesion segmentation task to the melanoma recognition task, which assists the network to learn more informative feature representation for melanoma recognition.

  • We propose a diagnosis guided feature fusion scheme to pass the lesion class information from the melanoma recognition task into the skin lesion segmentation task, which generates discriminative representations for different types of skin lesions.

  • We design a recursive mutual learning method that further enhances the joint learning ability of the proposed model for both skin lesion segmentation and melanoma recognition.

Abstract

Deep learning techniques have shown their superior performance in dermatologist clinical inspection. Nevertheless, melanoma diagnosis is still a challenging task due to the difficulty of incorporating the useful dermatologist clinical knowledge into the learning process. In this paper, we propose a novel knowledge-aware deep framework that incorporates some clinical knowledge into collaborative learning of two important melanoma diagnosis tasks, i.e., skin lesion segmentation and melanoma recognition. Specifically, to exploit the knowledge of morphological expressions of the lesion region and also the periphery region for melanoma identification, a lesion-based pooling and shape extraction (LPSE) scheme is designed, which transfers the structure information obtained from skin lesion segmentation into melanoma recognition. Meanwhile, to pass the skin lesion diagnosis knowledge from melanoma recognition to skin lesion segmentation, an effective diagnosis guided feature fusion (DGFF) strategy is designed. Moreover, we propose a recursive mutual learning mechanism that further promotes the inter-task cooperation, and thus iteratively improves the joint learning capability of the model for both skin lesion segmentation and melanoma recognition. Experimental results on two publicly available skin lesion datasets show the effectiveness of the proposed method for melanoma analysis.

Introduction

Melanoma is one of the most malignant skin cancer that increases rapidly throughout the world [1], [2], [3], [4]. Timely treatment of the melanoma can efficiently improve the survive rate of the patients. Dermoscopic images captured by digital imaging devices, offer the magnified visualization of the melanoma, and thus assist the dermatologists in examining the melanoma based on a set of complex visual characteristic of the lesion. Computer-aided diagnosis (CAD) system provides an effective way that allows dermatologists’ clinical inspection of the skin lesion in dermoscopic images. A CAD system for melanoma analysis generally contains two crucial functions: lesion segmentation and melanoma recognition. Specifically, the task of segmentation aims to divide a dermoscopic image into the skin lesion parts and background parts, i.e., it is a pixel-wise classification process to generate more conceptual saliency information for melanoma analysis [5], [6], [7], [8], [9]. Meanwhile, melanoma recognition is an image-level classification task that aims to identify the skin lesion types, such as melanoma, seborrheic keratosis, and benign nevi [10], [11], [12], [13], [14].

Recent deep learning-based methods have shown their promising achievements in both lesion segmentation [15], [16], [17], [18], [19] and melanoma recognition [13], [14], [20], [21], [22], [23], [24], [25], [26], [27]. Existing approaches usually train the task-specific models to perform skin lesion segmentation and melanoma recognition separately, and do not explicitly consider the dermoscopists’ clinical criteria for melanoma inspection. For example, most of the existing methods [13], [14], [20], [21], [22], [23], [24], [25], [26], [27] heavily depend on the abstract feature representation obtained at the high layer of the network, and they often leverage the global information of the whole feature maps for melanoma identification, i.e., melanoma recognition is often conducted in a similar fashion as other general object classification tasks [28], [29]. However, in the domain of melanoma inspection, such a fashion has a high potential of losing important pathological patterns and morphology knowledge that are crucial for melanoma recognition. Generally, for dermatologists, they identify melanoma via certain prior criteria, like the statistical information of the lesion color’s variation and the border’s irregularity [25], [30], [31]. Moreover, clinical studies [11], [32], [33] have confirmed that the characteristics of the peripheral lesion are very useful morphological expressions for identifying melanoma, as shown in Fig. 1. However, these types of knowledge have not been explicitly considered yet by current deep learning-based methods [21], [22], [23], [24], [25], [26], [27].

In this paper, we propose a novel knowledge-aware framework that is able to exploit the clinical knowledge within the deep feature learning process for melanoma analysis. Specifically, in our framework, both the skin lesion segmentation task and the melanoma recognition task are learned and performed via a joint deep network, such that the clinical knowledge can be exploited and transferred with the mutual guidance and assistance of these two tasks.

Concretely, to promote the performance of the deep learning framework on melanoma recognition, we propose to incorporate the clinical knowledge by explicitly considering the morphological expression of the lesion area and also the periphery region. To achieve this, we design a novel lesion-based pooling and shape extraction (LPSE) module that transfers the lesion structure information obtained from the skin lesion segmentation task to the melanoma recognition task, and thus we are able to embed the morphological operation into our deep network. With the integration of morphological analysis of the skin lesion structure, our network is thus able to selectively learn the informative features containing useful statistical information from both the lesion center region and also the border region. Compared with the features produced by direct global average pooling in most of the existing deep learning-based methods [20], [21], [22], [23], [24], [25], [26], our network generates more discriminative lesion representation for melanoma recognition.

Melanoma and non-melanoma lesions generally have very different pathological feature representations (e.g., more border irregularity and inhomogeneous textures for melanoma lesions). With lesion class information, the segmentation network can generate more discriminative feature representation for detecting different types of skin lesions from dermoscopic images. To improve skin lesion segmentation performance for both the melanoma and non-melanoma classes, we propose a new unit called diagnosis guided feature fusion (DGFF) that incorporates the lesion diagnosis information from melanoma recognition task into skin lesion segmentation task. With the guidance of skin lesion class information learned from the melanoma recognition task, our DGFF achieves more discriminative feature representation for each lesion class and thus enhances the capability of our network in segmenting the skin lesion regions from the dermoscopic images.

Moreover, to seamlessly achieve both the aforementioned skin lesion segmentation task and the melanoma recognition task within our end-to-end network, we also propose a recursive mutual learning scheme that enables effective inter-task cooperation and recurrently improves the joint leaning capability of the model on both tasks. The proposed recursive mutual learning scheme systematically exploits the mutual guidance signals generated between skin lesion segmentation and melanoma recognition, and thus simultaneously boosts the performance of both tasks. The main contributions of this paper are summarized as follows:

We propose a novel end-to-end deep framework that is able to perform skin lesion segmentation and melanoma recognition jointly, where the clinical knowledge is exploited and transferred with the mutual guidance between these two tasks.

We design a lesion-based pooling and shape extraction module to transfer the lesion structure information from the skin lesion segmentation task to the melanoma recognition task, by embedding the morphological analysis into the feature learning process of our network, which helps generate informative clinically interested features for melanoma recognition.

We propose a diagnosis guided feature fusion scheme to pass the lesion class information from the melanoma recognition task into the skin lesion segmentation task, which generates discriminative representations for different types of skin lesions.

We design a recursive mutual learning method that further enhances the joint learning ability of the proposed model for both skin lesion segmentation and melanoma recognition.

Section snippets

Related work

Deep learning-based methods have been used for melanoma diagnosis during recent years. Skin lesion segmentation and melanoma recognition are both important research tasks for melanoma inspection. In this section, we review the deep learning-based methods that address these two tasks.

Skin lesion segmentation: The work of [3] introduced a deep fully convolutional network with Jaccard distance for skin lesion segmentation. Bi et al. [15] refined the skin lesion segmentation performance by

Proposed network

In this section, we describe the proposed knowledge-aware deep framework for joint skin lesion segmentation and melanoma recognition in detail. First, we introduce our lesion-based pooling and shape extraction (LPSE) method that transfers the skin lesion structure information from the skin lesion segmentation task to the melanoma recognition task. It is capable of selectively learning the informative clinically interested features from the lesion and its border regions. Second, we propose a

Materials

We evaluate our proposed framework on two public benchmark datasets (ISBI 2016 [48] and ISBI 2017 [49]). They are provided by the International Skin Imaging Collaboration (ISIC) for the International Symposium on Biomedical Imaging challenges named “Skin Lesion Analysis toward Melanoma Detection”. ISBI 2016 dataset includes a training set with 900 dermoscopic images (727 non-melanoma cases and 173 melanoma cases), and a testing set with 379 dermoscopic images (304 non-melanoma cases and 75

Conclusion

In this paper, we propose to integrate dermatologists clinical knowledge into the learning process of a knowledge-aware deep framework for collaborative skin lesion segmentation and melanoma recognition. In particular, we propose a lesion-based pooling and shape extraction method to transfer the lesion structure information from the skin lesion segmentation into the melanoma recognition. It embeds the morphological operation within the deep network, which assists the network to learn more

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Xiaohong Wang received the Ph.D. degree from Nanyang Technological University, Singapore, in 2020, and the M.Sc. degree from Central South University, China, in 2014. She is currently a Research Scientist in A*STAR, Singapore. Her research interests include image processing, computer vision, and pattern recognition.

References (55)

  • J. Kawahara et al.

    Seven-point checklist and skin lesion classification using multitask multimodal neural nets

    IEEE J. Biomed. Health Inf.

    (2018)
  • Z. Hu et al.

    Deep learning for image-based cancer detection and diagnosis- a survey

    Pattern Recognit.

    (2018)
  • Y. Yuan et al.

    Automatic skin lesion segmentation using deep fully convolutional networks with jaccard distance

    IEEE Trans. Med. Imaging

    (2017)
  • C. Barata et al.

    Explainable skin lesion diagnosis using taxonomies

    Pattern Recognit.

    (2020)
  • M. Zortea et al.

    A simple weighted thresholding method for the segmentation of pigmented skin lesions in macroscopic images

    Pattern Recognit.

    (2017)
  • P. Schmid

    Segmentation of digitized dermatoscopic images by two-dimensional color clustering

    IEEE Trans. Med. Imaging

    (1999)
  • H. Zhou et al.

    Anisotropic mean shift based fuzzy c-means segmentation of deroscopy images

    IEEE J. Sel. Top. Signal Process.

    (2009)
  • H. Zhou et al.

    Gradient vector flow with mean shift for skin lesion segmentation

    Comput. Med. Image Grap.

    (2011)
  • Z. Ma et al.

    A novel approach to segment skin lesions in dermoscopic images based on a deformable model

    IEEE J. Biomed. Health Inf.

    (2016)
  • L. Bi et al.

    Multi-label classification of multi-modality skin lesion via hyper-connected convolutional neural network

    Pattern Recognit.

    (2020)
  • M. Sadeghi et al.

    Detection and analysis of irregular streaks in dermoscopic images of skin lesions

    IEEE Trans. Biomed. Eng.

    (2013)
  • A. Saez et al.

    Model-based classification methods of global patterns in dermoscopic images

    IEEE Trans. Med. Imaging

    (2014)
  • Z. Yu et al.

    Melanoma recognition in dermoscopy images via aggregated deep convolutional features

    IEEE Trans. Biomed. Eng.

    (2018)
  • J. Yang et al.

    Classification for dermoscopy images using convolutional neural networks based on region average pooling

    IEEE Access

    (2018)
  • L. Bi et al.

    Dermoscopic image segmentation via multi-stage fully convolutional networks

    IEEE Trans. Biomed. Eng

    (2017)
  • Y. Yuan et al.

    Improving dermoscopic image segmentation with enhanced convolutional-deconvolutional networks

    IEEE J. Biomed. Health Inf.

    (2017)
  • X. Wang et al.

    Dermoscopic image segmentation through the enhanced high-level parsing and class weighted loss

    Proceedings of the IEEE International Conference on Image Processing

    (2019)
  • G. Zhang et al.

    DSM : a deep supervised multi-scale network learning for skin cancer segmentation

    IEEE Access

    (2019)
  • V. Singh et al.

    Fca-net: adversarial learning for skin lesion segmentation based on multi-scale features and factorized channel attention

    IEEE Access

    (2019)
  • B. Harangi

    Skin lesion classification with ensembles of deep convolutional neural networks

    J. Biomed. Inf.

    (2018)
  • A. Mahbod et al.

    Fusing fine-tuned deep features for skin lesion classification

    Comput. Med. Imaging Graph.

    (2019)
  • R. Liang et al.

    Multi-pooling attention learning for melanoma recognition

    2019 Digital Image Computing: Techniques and Applications (DICTA)

    (2019)
  • N. Gessert et al.

    Skin lesion classification using CNNS with patch-based attention and diagnosis-guided loss weighting

    IEEE Trans. Biomed. Eng.

    (2019)
  • N. Sultana et al.

    Deep residual network with regularised fisher framework for detection of melanoma

    IET Comput. Vis.

    (2018)
  • J. Hagerty et al.

    Deep learning and handcrafted method fusion: higher diagnostic accuracy for melanoma dermoscopy images

    IEEE J. Biomed. Health Inf.

    (2019)
  • J. Zhang et al.

    Attention residual learning for skin lesion classification

    IEEE Trans. Med. Imaging

    (2019)
  • L. Yu et al.

    Automated melanoma recognition in dermoscopy images via very deep residual networks

    IEEE Trans. Med. Imaging

    (2017)
  • Cited by (0)

    Xiaohong Wang received the Ph.D. degree from Nanyang Technological University, Singapore, in 2020, and the M.Sc. degree from Central South University, China, in 2014. She is currently a Research Scientist in A*STAR, Singapore. Her research interests include image processing, computer vision, and pattern recognition.

    Xudong Jiang received the B.Eng. and M.Eng. degrees from the University of Electronic Science and Technology of China and the Ph.D. degree from Helmut Schmidt University, Hamburg, Germany, all in electrical engineering. He received two Science and Technology Awards from the Ministry for Electronic Industry of China. From 1998 to 2004, he was with the Institute for Infocomm Research, A*STAR, Singapore, as a lead Scientist, and the Head of the Biometrics Laboratory, where he developed a system that achieved the highest efficiency and the second highest accuracy at the International Fingerprint Verification Competition in 2000. He joined Nanyang Technological University (NTU), Singapore, as a Faculty Member in 2004, and served as the Director of the Centre for Information Security from 2005 to 2011. He is currently a Tenured Associate Professor with the School of EEE, NTU. He holds 7 patents and has authored over 100 papers with 28 papers in the IEEE journals, including 11 TIP papers and 5 TPAMI papers. He has published 12 papers in Pattern Recognition. His research interests include image processing, pattern recognition, computer vision, machine learning, and biometrics. He is IFS TC member of the IEEE Signal Processing Society, and serves as an Associate Editor of the IEEE Transactions on Image Processing, IEEE Signal Processing Letters and IET Biometrics.

    Henghui Ding received the B.E. degree from Xi’an Jiaotong University, Xi’an, China, in 2016. He is currently pursuing the Ph.D. degree with the RapidRich Object Search (ROSE) Laboratory, School of Electrical and Electronic Engineering, Nanyang Technological University (NTU), Singapore. His research interests include computer vision and machine learning.

    Yuqian Zhao received Ph.D. degree from Central South University, Changsha, China, in 2006. He engaged in postdoctoral research at Xiangya School of Medicine, Changsha, China, from May 2007 to June 2009, and at New Jersey Institute of Technology, Newark, New Jersey, USA, from July 2009 to July 2010. He is currently a Professor in the School of Automation, Central South University. His research interests include image processing, pattern recognition, computer vision, image forensics, and computer-aided diagnosis.

    Jun Liu is an Assistant Professor with Singapore University of Technology and Design. He received the Ph.D. degree from Nanyang Technological University, Singapore, in 2019, the M.Sc. degree from Fudan University, China, in 2014, and the B.Eng. degree from Central South University, China, in 2011. His research interests include computer vision and machine learning.

    View full text