Open Access
27 November 2020 Deep learning-level melanoma detection by interpretable machine learning and imaging biomarker cues
Daniel S. Gareau, James Browning, Joel Correa Da Rosa, Mayte Suarez-Farinas, Samantha Lish , Amanda M. Zong, Benjamin Firester, Charles Vrattos, Yael Renert-Yuval, Mauricio Gamboa, María G. Vallone, Zamira F. Barragán-Estudillo, Alejandra L. Tamez-Peña, Javier Montoya, Miriam A. Jesús-Silva, Cristina Carrera, Josep Malvehy, Susana Puig, Ashfaq Marghoob, John A. Carucci, James G. Krueger
Author Affiliations +
Abstract

Significance: Melanoma is a deadly cancer that physicians struggle to diagnose early because they lack the knowledge to differentiate benign from malignant lesions. Deep machine learning approaches to image analysis offer promise but lack the transparency to be widely adopted as stand-alone diagnostics.

Aim: We aimed to create a transparent machine learning technology (i.e., not deep learning) to discriminate melanomas from nevi in dermoscopy images and an interface for sensory cue integration.

Approach: Imaging biomarker cues (IBCs) fed ensemble machine learning classifier (Eclass) training while raw images fed deep learning classifier training. We compared the areas under the diagnostic receiver operator curves.

Results: Our interpretable machine learning algorithm outperformed the leading deep-learning approach 75% of the time. The user interface displayed only the diagnostic imaging biomarkers as IBCs.

Conclusions: From a translational perspective, Eclass is better than convolutional machine learning diagnosis in that physicians can embrace it faster than black box outputs. Imaging biomarkers cues may be used during sensory cue integration in clinical screening. Our method may be applied to other image-based diagnostic analyses, including pathology and radiology.

1.

Introduction

Melanoma is the most dangerous skin cancer and the leading cause of death from skin disease. There are over 96,000 new cases in the USA annually with nearly 10,000 deaths attributed to melanoma. Worldwide, annual melanoma mortality is over 60,000 people. Gaps exist in our ability to diagnose melanomas versus nevi. Gaps also exist in providing specialty care for patients with suspicious lesions and for increasing the numbers of patients seeking potentially life-saving melanoma diagnosis from primary care providers. Despite the fact that many patients with skin lesions first present to their primary care physicians, these physicians often lack the knowledge to differentiate benign from malignant lesions. Despite evidence that early detection increases survival, and despite the need for technology to enhance screening to an expert level on a wide scale, there is uncertainty regarding the effectiveness of state-of-the-art technological methodology in clinical dermatologist screening.1 Improved screening may prevent melanoma deaths (7230 in the USA in 20192) while decreasing unnecessary invasive procedures because screening guides the binary decision for or against surgical biopsy. Technology can translate down the expertise hierarchy structure of clinical practitioners to enhance diagnosis in less-specialized medical practices. Table 1 shows that for each American to be evaluated by an expert dermoscopist, the experts would need to see millions of patients, which is not feasible, whereas if technology could enable common providers, such as general practitioners to screen with expert precision, the screening diagnostic net could be extended because providers would screen hundreds (not millions) of patients. Recent advances in machine learning have shown promise for high-performing computational diagnostic tools in multiple clinical areas,3,4 but the noninterpretability of data-intensive deep learning models remains a barrier to widespread deployment. Because the skin is accessible to relatively inexpensive and noninvasive diagnostic imaging, and because clinicians rely heavily on visual inspection, melanoma is an ideal case study.

Table 1

The Americans/Provider ratio illustrates that there are not enough Top Dermatologists with expert dermoscopy training to evaluate the entire US population. All dermatologists (20,000 in USA) automatically qualify as nonexpert dermoscopist screeners with board certification. An unmet healthcare need is to address the screening accuracy gap between top dermatologists and the broader medical network. Eclass potentially translates the top dermatologists’ pattern recognition skills (and associated diagnostic precision) to general dermatology and the broader community of nonexpert screeners.

Provider typeUS population per provider
Top dermoscopists6,480,000
Dermatologists32,400
In-pharmacy evaluation4830
General practitioner379

Melanoma growth rate is variable and multifactorial. Screening is typically done with dermoscopy using a dermatoscope, which is a low-magnification, illuminated, polarized5 contact imaging device widely used in dermatology practice.6 A recent observational study7 regarding the rate of growth in vertical and radial growth phase of superficially spreading melanomas found that melanomas in the vertical growth phase invade by 0.13±0.16  mm/month. A delay in detection of a few months can negatively impact prognosis since growth beneath the 0.3-mm-deep basal layer basement membrane constitutes invasion toward metastasis. It would be a significant medical advancement if diagnostic imaging technology could improve early detection with machine learning and associated automated image processing devices to predict underlying pathology from noninvasive dermoscopy images. Yet careful consideration of biases in training data suggests that there are ethical concerns, such as the fact that training data does not always represent all skin types. Convolutional neural networks (CNN)8 may be inappropriate for stand-alone diagnostic medical decision making because physicians cannot have confidence in a computer-derived diagnostic risk score without understanding the underlying computational diagnostic process.

There exists an unmet need to better understand diagnostic processes utilized by machine learning-derived tools. As a more interpretable alternative9 to CNN, we considered a diagnostic ensemble classifier (Eclass)10 of traditional (i.e., “nondeep”) machine learning approaches. It used a set of imaging biomarkers that quantify medically relevant features rather than brute force pixel analysis to freely choose salient features. Because it is “transparent” in that imaging biomarkers are visual features, implementation of Eclass could result in more medical accountability and confidence than CNN. Herein, our context was melanoma detection, but digital imaging biomarkers based on visual sensory cues can be applied to any image-based diagnostic analysis including pathology and radiology. Figure 1 shows a sample run for each machine learning method and Fig. 2 shows a graphic user interface application (App) capable of displaying imaging biomarker cues (IBCs) to illustrate clinical and pathological features, aiding in visual interpretation of imaging diagnoses.

Fig. 1

Receiver-operator characteristic curve for the convolutional neural network (CNN) versus the ensemble classifier (Eclass). In this comparison trial run, as in the case of 75% of our trial runs, Eclass out-performed CNN, with a greater area under the receiver-operator characteristic curve (AUROC). Although Eclass outperformed CNN in this study, both the Eclass and CNN predictive models are expected to improve with larger training data sets. A theoretical curve, with minimum AUROC diagnostic performance for translation, shows various screeners, susceptible to using the technology at different parametric ROC curve values (red dots). These range from the patient—whose high-specificity App would more accurately diagnose benign lesions that do not require escalation—to trained professional dermoscopists, who would value seeing the imaging biomarker cues in a high-sensitivity App that helps them be sure they aren’t missing rare or difficult-type lesions.

JBO_25_11_112906_f001.png

Fig. 2

Example visualization of melanoma imaging biomarkers and machine learning diagnostic App “Eclass Imaging Biomarkers” downloaded from Mac App Store. (a) Identification of abnormally long finger-like projections in the pigmented network at the peripheral border of the lesion. (b), (c) Screen captures from the Eclass Imaging Biomarkers, freely available at Mac App Store; (b) shows a radial brightness analysis as an example of a searchable imaging biomarker set (c) where the imaging biomarkers highlighted in red (Asymmetry and Colors) indicate quantitative values that fall on the statistically malignant range of the spectrum. Since no imaging biomarkers are highlighted green, users know that the algorithm found no normal features. Green circles connected to arrows indicate analytic flow with the App. Clicking on “Show Diagnostic” (b) leads to the (d) diagnostics screen, where clicking on “Show Diagnosis” reveals the standard histopathological diagnosis. This clinical diagnostic workflow from dermoscopy image (a) to imaging biomarkers (a–b) and to Eclass and CNN scores (d) could inform screening of new images acquired in the by enabling clinicians to access automated digital diagnostics.

JBO_25_11_112906_f002.png

2.

Methods

2.1.

Imaging

Dermoscopy is a mainstream clinical imaging method for melanoma screening used in dermatology practice. This report analyzes two different cohorts of demoscopy images in different countries using different dermatoscope imaging systems. The first cohort10 consisted of alcohol-coupled, nonpolarized dermoscopy images of primary melanoma skin cancers versus abnormal nevi acquired with the EpiFlash™ (Canfield Inc., New Jersey) dermatoscope attached to a Nikon D80 camera, where each image contained 1 to 5 megapixels after cropping in New York. The second cohort, presented for the first time here, consisted of digital dermoscopy images acquired with the Dermlite Foto (3Gen Inc., San Juan Capistrano, California), sized 5.9±2.7 (mean ± standard deviation) megapixels depending on lesion size in Barcelona, Spain. All lesions were pigmented lesions that did not demonstrate a benign pattern11 under dermoscopy. The current data set of 668 images was reduced to 349 images (one lesion per patient and one lesion per image) by filtering out images with hair or surgical ink markings,12 lesion borders that extend beyond the image field of view, or other features that prevented imaging biomarker computation, like cases with extreme atypia such as those that were ulcerated, nodular/palpable, or did not fit within the field of view of the dermatoscope were excluded. Both the CNN classifier3 and Eclass classifier10 were trained on the same set of 668 images (113 melanomas and 236 nevi) and the diagnostic performances of the resulting models were compared. CNN is a leading deep learning approach while Eclass is our approach that implements a “wisdom of the crowd” approach of sampling the prediction of a broad range of machine learning predictive models. Eclass has no convolutional aspect so it has the benefit of being more easily interpretable by clinicians.

2.2.

Eclass and Imaging Biomarkers Versus Deep Learning

The CNN used the raw pixels in the images as input features whereas Eclass operated on the set of 38 imaging biomarkers, which were engineered to automatically quantify visual features that dermatologists use during sensory cue integration in manual inspection of suspicious legions. Imaging biomarkers were either binary, such as the presence [0 1] of a blue or gray dermoscopic color (Fig. 3), integers such as the number of dermoscopic colors present [0–6], or continuous such as the variation coefficient of branch length in pigmented networks, but all imaging biomarkers were numbers that were high for melanoma and low for nevus. The detailed mathematical formulas for the imaging biomarkers used in Eclass can be found in the Appendix and in Sec. S5 of the Supplementary Material13 of our previous publication10 and have been adapted and reproduced in the methods section below. Both CNN and Eclass models predicted a melanoma probability (between 0 and 1) of the invasive histopathological diagnoses for each skin lesion using the noninvasive dermoscopy image (for CNN) or the imaging biomarkers derived from that image (for Eclass). But only Eclass involved dimensional reduction to intuitive, visual IBCs.

Fig. 3

Blue gray color (a simple imaging biomarker). Referred to as a “blue-white veil” in dermoscopy, this manifestation of the Tindal effect is a statistically-significant melanoma discriminant. The lesion (a) is a melanoma. The ratio of the blue pixel intensity to the red pixel intensity (b) is the first of three steps to quantitative visualization. A simple automatic (Otsu’s method) threshold (c) is a second step to annotate the dermoscopy image with an overlay of a binary presence of a blue color (d). Displaying this image to a medical professional as a visual sensory cue can augment the cognitive process of visual sensory cue integration during melanoma screening.

JBO_25_11_112906_f003.png

Fig. 4

Updated imaging biomarkers statistical significance in published data versus cross-validation data sets. In this imaging biomarker re-analysis of our published data and unpublished cross validation studies, the height of the vertical bars represents the diagnostic significance of the imaging biomarkers in discriminating melanomas from nevi. Magenta bars are spectral imaging biomarkers (ones that are derived from all color channels simultaneously) and red-green-blue bars are gray-scale imaging biomarkers evaluated in the respective color channels. The imaging biomarkers are sorted by discriminant p-value between melanoma and nevus. They are categorized as those most significant in the blue channel (left), the green channel, the red channel and full spectral (right). The total number significant (p<0.05) is printed in black and the sum is tabulated for each data set (color coded numbers). The colored imaging biomarker code (eg. B7 on top left) on top of each bar, for each imaging biomarker references the written description, the mathematical derivation. This data shows that the imaging biomarkers are diagnostic across multiple screening sites because there are a similar number of significant biomarkers in cross validation sets 1 and 2 as there are in the original published data set (35, 38 and 35, respectively).

JBO_25_11_112906_f004.png

2.3.

Statistical Methods: IBC Combination to Form the Melanoma Eclass Score

Both CNN and Eclass models were trained to predict a melanoma probability (between 0 and 1) of the invasive histopathological diagnoses for each skin lesion using the noninvasive dermoscopy image (for CNN) or the imaging biomarkers derived from that image (for Eclass). But only Eclass involved dimensional reduction to intuitive, visual IBCs described mathematically above. Eclass was trained and cross-validated within a Monte Carlo simulation as previously described in Sec. S3 in the Supplementary Materials13 of our previous publication10 and as discussed below, included a hold-out test set for each Monte Carlo iteration that was not used for training. Briefly, the 38 IBCs achieving diagnostic statistical significance (p<0.05), 4 multicolor (MC), and 34 single-color (SC) (using the most significant RGB color channel version), were input into the 12 statistical/machine learning algorithms as predictive informatics programs given in Table 2. We selected the SC IBC from the best color channel when there was any color channel for an IBC that achieved statistical significance, including that SC IBC, and also including any significant MC IBC.

Table 2

Broad universe of classification algorithms applied to melanoma discrimination.

MethodDescription
NNETFeed-forward neural networks with a single hidden layer14
SVM (linear and radial)Support vector machines15,16
GLMLogistic regression within the framework of generalized linear models17
GLMnetLogistic regression with the elastic-net penalization18
GLMboostLogistic regression with gradient boosting19
RFRandom forests20
RPClassification and regression trees (CART) algorithm for classification problems21
KNNK-nearest neighbors algorithm developed for classification22
MARSMultiple adaptive regression splines23
C50C5.0 decision tree algorithm for classification problems24
PLSPartial least squares25
LDALinear discriminant analysis26

The statistical classification methods for machine learning were chosen to represent the broad universe of base classifiers.27 The C5.0 decision tree prioritized a subset of the IBCs and conducted a series of comparisons between these IBCs and a set of thresholds ultimately leading to a classification of each lesion. Each method output the melanoma likelihood for each lesion. The likelihoods produced by all methods were combined into the overall endpoint of the analysis, the ultimate best estimate of melanoma probability between zero and one, and the melanoma Eclass score. Thus, we built a predictive model that combined the IBCs into a risk score for probability of melanoma.

Our framework was set first to identify the most discriminative IBCs upon which the predictive model will be built. To this end, we first evaluate the differences between melanoma and nevus for each one of the seven multicolor IBCs and also for the RGB channel-specific IBCs (41 IBCs for each red, green, and blue channels). For this univariate assessments, two-sided unpaired t-tests, Wilcoxon–Mann–Whitney, and chi-square tests were used for continuous (e.g., IBC B1), ordinal (e.g., IBC MC1), and categorical (e.g., IBC MC4) IBCs, respectively. Of the total 130 IBCs evaluated in our current data set (Fig. 4, Cross-Validation Set 1), 38 (4 multicolor, 9 red, 7 green, and 18 blue IBCs) were selected as the most significant discriminators (p<0.05) between melanoma and nevi to continue to the multivariate discrimination stage. When significant differences for a given IBC were found in more than one channel, the most discriminative channel regarding its p-value was selected. This set of 38 discriminative IBCs measured from all the 349 lesions were used as inputs for our predictive model.

Fig. 5

This double-blinded study retrospectively tested prediction of the histopathological biopsy diagnosis using only the dermoscopy images acquired just prior to biopsy. Co-author S. L. assembled and distributed the latter to Co-author D. G. and the former as an input to the Eclass algorithm.

JBO_25_11_112906_f005.png

Fig. 6

Diagnostic performance results vs. published techniques. The receiver-operator characteristic (ROC) curves for the individual machine learning approaches (thin colored lines) are outperformed by the compound Eclass score (thick black line) which is the median of the individual risk scores. Literature data points are marked with symbols that indicate if they are mostly human (square) or computer (circle)-derived.

JBO_25_11_112906_f006.png

Figure 4 compares our current results (validation set 1) to the imaging biomarkers computed on the data set from our previous publication. As such, Fig. 4 (Top) can be compared to Fig. 2 in the previous publication,10 to show that our algorithm has been minorly updated to achieve slightly better performance on the previously published data than previously published, with two more imaging biomarkers achieving statistical significance than before and incremental increase in area under the receiver operator characteristic curve (AUROC). Figure 4 arranges the imaging biomarkers in their order of statistical significance in discrimination between melanomas and nevi that are clinically dysplastic. The middle figure, on data from the clinic of Dr. Puig, is the data reported in this publication. Above each imaging biomarker’s significance is an alphanumeric code that correlates the imaging biomarker to its mathematical derivation below. More imaging biomarkers were statistically significant in the blue channel in cross-validation set 1, due to factors such as the differences in imaging systems and patient populations. The imaging biomarker that was the most significant blue channel imaging biomarker (B1) in our original publication had greater significance in the green channel in the current study. The meaning of imaging biomarkers that do not have a label is that these imaging biomarkers were not statistically significant in our original study (p<0.05) but are in the current study. The description of the statistical methods of Eclass has been modified from the methods in the Supplementary Material13 of our original publication10 to reflect the statistical significance breakdown in the current study. The data in Fig. 4 show that the imaging biomarkers behave similarly across data sets and that a similar number of imaging biomarkers are statistically significant is similar across study sets.

2.4.

Clinical Study -to- Predictive Model

Figure 5 shows the overall study design by which we build a model with our experimental data including paired dermoscopy and histopathological diagnoses. The double-blind study attempted prediction of the biopsy-based histological diagnosis (melanoma or atypical nevus) using only the pre-biopsy dermoscopy image and the IBCs derived from it.

Fig. 7

GUI/APP minimalist IBC visualization strategy. We designed IBC visualization methodology such that the display reproduced the dermoscopic image with a visual representation of the most important IBCs used by the predictive models in generating a melanoma risk score. When IBCs are less than a standard deviation below their mean (across a population of lesions), their names are visualized with green background as an assurance whereas if they are more than a standard deviation above the mean value, they are visualized with red background as a warning. Examples of warnings include the Asymmetry IBC and the number of Colors IBC in Fig. 2(c).

JBO_25_11_112906_f007.png

After dermoscopy imaging and surgical excision of the imaged lesion, the standard diagnostic method of histopathological evaluation was carried out as part of routine clinical care to yield a diagnosis (melanoma or nevus) for each lesion in the study cohort.

The 668 binary diagnoses along with 668 correlating dermoscopic images comprised the study data. No information about the patient’s age, sex, state of sun damage, or anatomical location of the lesion was used. Dermoscopy images were randomized and coded to remove all patient identifiers, then injected into the blind study arm that generated the image-processing algorithm targeting melanoma features by extracting diagnostic IBCs without knowledge of the histopathological diagnosis.

IBCs fed a collection of 12 classification methods that range from simple to sophisticated and altogether cover different data structures. The collection of classification algorithms is given in Table 2 and includes the K-nearest neighbors (KNN),22 a simple and efficient nonparametric distance-based method that has been successfully applied for more than 60 years. Artificial neural networks14 and support vector machines (SVM)15,16 were included to represent high-dimensional complex nonlinear functions. To accommodate complex interactions between predictors, we incorporated four methods following the decision tree/recursive partitioning paradigm: classification and regression trees (CART),21 C5.0,24 multiple adaptive regression splines (MARS),23 and random forests (RF).20 Logistic regression18 and linear discriminant analysis (LDA)26 are based on solid statistical foundations. The former permits inference upon parameters by a probabilistic approach and the latter is one of the oldest techniques for dimensionality reduction. Partial least squares regression25,28 is of more recent development and simultaneously performs regression/classification and dimensionality reduction.

2.5.

Model Estimation/Training

To estimate each of the classifiers’ parameters and to evaluate the distribution of the prediction error empirically, we created a Monte Carlo experiment. During each training iteration, the set of lesions was randomly partitioned into training (75%) and test (25%) sets. For each classifier, model parameters were estimated by maximizing a partial area under the ROC curve obtained by limiting the specificity to be within the range 0% to 40% and tuning parameters were estimated by 10-fold Cross-Validation. The best configuration for each classifier was used to predict the 25% hold-out lesions in the test set.

Ensemble of predictive algorithms likely generate more accurate predictions than single algorithms.29,30 The melanoma Eclass score is a diagnostic for melanoma discrimination obtained by evaluating the median probability across K available classifiers

Eclass Score=median{Probi(Melanoma|M)};i=1,2,,k,
where Probi{0,1} is the probability of the lesion being a melanoma, as predicted by the i’th classifier based on a set of IBCs M. Monte Carlo simulations obtain the empirical distribution of the Eclass score for each lesion. The Eclass score distribution shows that the number of false-positives (melanomas classified as nevi) is lower than the false-negatives; indicating that our classification strategy is more sensitive than specific.

2.6.

Convolutional Neural Network

The CNN was based on a widely used ResNet-50 architecture instantiated with weights pretrained on the ImageNet database for transfer learning and modified with output layers designed for binary classification. Image augmentation (flip, zoom, and rotate) and minority class (melanoma) oversampling was used during CNN training to prevent overfitting to the training data and predictive bias toward the majority class (nevus) respectively. Pixel values for training and test images were normalized to have zero mean and standard deviation of 1. During oversampling, augmented versions of minority class images were overrepresented in the training data such that the model was trained on an equivalent number of melanoma and nevi images. Test time augmentation was used during inference wherein class predictions were generated for five randomly augmented versions of each test image and the majority vote was used as the final predicted class. The CNN model trained until validation data set accuracy had not improved for ten epochs and the resulting model, with highest validation accuracy, was saved.

2.7.

Performance Analysis

The receiver operator characteristic (ROC) curve and the area under the curve (AUROC) were used to evaluate diagnostic performance. Each method (CNN and Eclass) produced a distribution of scores from melanoma images and a second distribution of scores for benign images and we swept the diagnosis criterion across those two distributions, plotting proportion of hits as a function of proportion of false alarm (i.e., the ROC) and calculated the area under the curve. Different cross-validation runs (10 for CNN and 1000 for Eclass) were used to generate a distribution of AUROCs for each method. We compared the methods by comparing a randomly drawn AUROCs from each method’s AUROC distribution and repeating that process to determine what percentage of the time Eclass outperformed CNN.

The visualization potential of Eclass is developed as an App for imaging biomarker visual sensory cue integration31 that was developed based on the results of our institutional review board (IRB)-approved human subjects research (RU DGA-0923) on clinicians using the App. Our human subjects research was used to collect data regarding specific features that the clinicians found useful and whether they were likely to implement such a technology if it was available.

3.

Results

Eclass trained several independent machine learning algorithms (see Table 2) 1000 times in 150 s compared to the CNN model, which trained 10 times in 52 h on an Nvidia Quadro M5000 GPU. The final Eclass risk score for each lesion was the median risk score produced by the eight independent machine learning algorithms. Figure 6 shows the original analysis on the published data along with the performance of the various dermoscopy algorithms used in medical practice, which are abbreviated in the legend and numbered with their respective literature references (see Table 3). CNN was computationally intensive, taking more time to learn diagnostically relevant features than the time required by Eclass. This comparison does not include the computational time required to calculate the imaging biomarkers that fed Eclass, which was about 3 h.

Fig. 8

Decision tree built with the C5.0 algorithm. The algorithm was applied to predict lesions type (melanoma vs. nevus) with the full originally published data set that included 112 lesions and 33 IBCs. The decision tree has 10 decision nodes (#1, #2, #3, #5, #9, #10, #11, #14, #15, and #16) and 11 terminal nodes (#4, #6, #7, #8, #12, #13, #17, #18, #19, #20, and #21). The algorithm selected decision nodes based on four IBCs from the blue channel (B1, B6, B7, and B15), five IBCs from the red channel (R4, R6, R8, R12, and R13) and one multicolor IBC (MC1). At the terminal nodes the proportion of melanomas (light gray) and nevi (dark gray) are shown with stacked bar plots. The final classification has yielded 7 pure terminal nodes (#4, #6, #7, #8, #13, #19, and #21) where melanoma or nevi have 100% prevalence. The nodes #4 and #20 together have 59.8% of the lesions and they perfectly discriminate nevi and melanoma, respectively.

JBO_25_11_112906_f008.png

Table 3

Diagnostic sensitivity and specificity for melanoma detection by human pattern recognition (*) and by machine-augmented pattern recognition (**). The final two listed other techniques (***) represent the current state of commercially available clinical machine-vision systems.

MethodSensitivity (%)Specificity (%)
*GP referral325171
*GP referral, dermoscopy327972
*Pattern analysis338579
*ABCD338475
*ABCD345692
*7-Point checklist357865
*7-Point checklist345892
*CASH369868
*Menzies358585
*Malignancy grading347381
*Total body photography377574
*Dermoscopy389090
*Confocal microscopy399086
**Prelim. SIAscopy405084
**Prelim. SIAscopy404495
**Solar Scan419168
**Prelim. Melafind 1429844
**Prelim. Melafind 2429138
**Image processing437787
***SIAscopy448665
***Melafind459810
Eclass109836

= referral to a dermatologist by a general practitioner nonexpert dermoscopist

= averaged over in situ and stage I melanoma.

Performance on the current data set (Validation Set 1) was characterized as the mean and standard deviation of the AUROC. For Eclass, the AUROC was 0.71±0.07 with a 95% confidence interval of [0.56 0.85] while the CNN achieved an AUROC of 0.67 with a 95% confidence interval of [0.63 0.71]. In a Monte Carlo simulation that randomly drew ROCs from the 10 CNN ROCs to compare to ROCs randomly drawn from the 1000 Eclass ROCs, the AUROC was greater for Eclass 74.88% of the time. The AUROC for both models is lower than in other studies with larger numbers of more typical nevi. This performance comparison suggests that codifying dermoscopy features into imaging biomarkers distills information from the image, enabling the Eclass model to operate more efficiently than the CNN and without access to the original image pixels.

The diagnostic score was the least favorite feature of clinicians who tested the IBC App. 10 clinicians, aged 26 to 64 years old, scored the App an average of 2.3 out of 4 for utility in their respective clinical settings. Scores ranged from 2 to 4 with the App being favored by younger clinicians (score=0.037×Age+4.17, R2=0.41). Figures 2(b)2(d) shows screen captures from the App after revisions based on the clinicians’ feedback in the study data. The major result we achieved with our human subjects’ research on dermatologists is that the dermatologists have a very small bandwidth to process IBCs compared to the analytical bandwidth of the computer. Therefore, we devised a scheme by which we showed only diagnostically significant imaging biomarkers (ones that were very low indicating nevus or very high indicating melanoma) as visual aids. Figure 7 shows this approach by which we compared each IBC value to a population of values, which we verified to be normally distributed, to display it in a clinically-appropriate way so as not to distract the App user with too much unimportant detail.

Fig. 9

Coordinate transformation and illustration of IBC derivation using angular clock sweep analysis. In images of a nevus (a) and a melanoma (e), lesion border and center (b, f). (c, g) show the blue channel data under a coordinate transformation from x–y to R-θ such that the bottom row of pixels in (g) is the same pixel in f, namely the center pixel (white circle) and the top row of pixels in (g) traces out the lesion border clockwise. (d, h) analyze the pixel brightness statistics (mean in black and standard deviation in blue) of (c, g) in the vertical direction which is along the radial in (b, f). In (d, h), IBC B12, for example, is derived from the radial variation range, which is the vertical separation of the horizontal dashed lines (d, h).

JBO_25_11_112906_f009.png

4.

Discussion

There is an ongoing need to develop digital imaging biomarker libraries in melanoma detection and other diagnostic fields. In contrast to deep learning, which has a convolutional aspect and many features that are hard for clinicians to understand, pre-existing image analytical frameworks (e.g., dermoscopy) can be codified to provide a transparent link between machine and human intelligence. The App we provide31 is an example for demonstration that visualizes a subset of the imaging biomarkers on random images drawn from a previously published10 study set, highlighting the biomarkers red if they are statistically malignant or in green if they are statistically benign, as defined as falling 1.5 standard deviations above or below the mean value for that imaging biomarker across a reference set of training images, respectively. App users can thus be directed to imaging biomarker visual sensory cues [Fig. 2(c)] of particular diagnostic importance to form a mental analysis before selecting “show diagnostic” to show [Fig. 2(d)] both the CNN and Eclass risk scores as additional inputs in the cue integration process. After making a test diagnosis, selecting “show diagnosis” reveals [Fig. 2(e)] the gold standard biopsy-based diagnosis.

Though we report the median melanoma likelihood produced by the various machine learning approaches as the Eclass score, one approach (the C5.0 decision tree approach) outperformed the Eclass score at 98% sensitivity, yielding sensitivity/specificity = 98%/44% in our originally published study. This result was produced using branching logic. Figure 8 shows an illustration of this branch choice approach, which may be the most promising approach as instructive to visual examination. This analysis may be able to “teach back” to dermatologists both new visual dermoscopic features and new ways to combine IBC evaluations sequentially. Our full diagnostic, which involves calculating a melanoma probability for each machine learning approach and taking the median of those probabilities as the Eclass, may be reduced for translation to visual screening. Thus, automated, unconstrained visual-aided screening with optimized decision trees shows feasibility for human use without computer vision. Also, it may be desirable from the point of view of decreasing computation time and/or complexity during evaluation to reduce the number of statistical classification approaches from the 13 used here to 1 or 2 (e.g., the C5.0 tree in Fig. 8) in cases where such a small subset continues to outperform the Eclass ensemble approach that uses the median result of all the classifiers.

Fig. 10

Fitting for edge demarcation. Edge demarcation was quantified as the slope of the transitioning from dark pixels inside the lesion to bright pixels outside the lesion. Increased slope of the fitting mathematical function resulted from increased lesion border demarcation. The two radial lines (Line 1, Line 2) drawn on the lesion include the lesion border from inside the lesion where the pixels are dark inside the lesion to outside the lesion where the pixels are bright in normal skin indicate illustrate two locations where the demarcation is gradual (Line 1) and sharp (Line 2). The pixel brightness extracted along these two lines (x for Line 1 and o for Line 2), pe(r) was fit to a mathematical model), se(r) to yield the fitting parameters, which were used to produce IBCs B3, B4, B9, B13, B14, R1, R5, and R10. This includes the edge demarcation slope, which is the slope of the solid line at the lesion border between normal skin and lesion and the error in the fit, which is the sum of the squared differences between the data points, pe(r) and the error function fit (solid line) se(r). Melanomas had a sharper border, a higher degree in variability of border sharpness and a greater fitting error.

JBO_25_11_112906_f010.png

4.1.

Analysis in Context

Our findings have implications for frequent machine learning scenarios where the available training set is too small to train high performing deep learning models. Although deep learning systems for breast cancer46 screening and melanoma screening3 have surpassed human experts in narrow prediction tasks, they have relied on tens or hundreds of thousands of training examples, respectively. Our EClass model, on the other hand, has a model parameter size to data size ratio in the underparameterized “classical” regime of deep learning allowing it to outperform the deep learning model.47 Our work provides a head-to-head comparison of CNN versus Eclass on a limited data set and our AUROC of 0.71 is less than the 0.91 published3 for melanoma detection based on different training data. Eclass must be evaluated head-to-head against other methods in larger studies in the future, but what makes the present study significant is that the Eclass model was trained on 10× more training images (349) than the number of imaging biomarkers17 used as free parameters (349>10×30). By comparison, the number of CNN model free parameters was 20 million. This means that the utilization of imaging biomarkers was a preprocessing step in Eclass (but not CNN) that distilled diagnostic image content. In future work, EClass should be compared to CNN for larger data sets in which the training data size results in an overparameterized EClass model. Although required computational resources were minimal for CNN training on this small dataset, increasing the size of the dataset to take advantage of the high-parameter CNN would significantly increase CNN training resources. For example, increasing the data set size from the current 349 images to the approximately 130,000 images used by Esteva et al.3 would result in a significant increase in training cost.

4.2.

Spectral Properties of IBCs

IBCs generally show that melanoma exhibits spectral and structural irregularity versus benign nevi. This can occur at increasing depth in the skin and be shown in the blue, green, and red channels. Not surprisingly, IBCs exhibit spectrally variant diagnostic statistical significance, and we expect that result to play out in our ongoing hyperspectral imaging study.48

One hypothetical mechanism for the spectral dependence of diagnostic importance as a function of wavelength is deeper penetration by longer wavelengths (e.g., red), and thus, the ability to differentially visualize differing three-dimensional tissue or chromophore characteristics of melanoma invading the dermis and superficial epidermal imaging by the shorter (e.g., blue) wavelengths of basal layer atypia or junctional nests of melanocytes associated with melanoma. While more IBCs are needed to cover more clinical presentations, IBCs also still need to be related to underlying tissue structure, including proliferative and invasion patterns of melanoma cells, and molecular pathways impacting pigment distribution. In the green channel, where saturation/desaturation of metabolically active areas of active tumor growth impacts the image, contrast captures polymorphic vasculature associated with melanoma and other skin cancers (basal cell and squamous cell carcinoma).

Within a hyperspectral image of a pigmented lesion, one can include measures of hemoglobin saturation and desaturation, which may help to identify metabolically active regions within lesions. The basis for the “steeper edge slope” (IBC R5) in melanomas does not yet have a cellular basis, but we speculate that it might represent growth of melanocytes in nests at the dermal-epidermal junction at the edge of a melanoma, whereas atypical nevi tend to have only individual junctional melanocytes (nevus cells) that are decreasing in number at the edge of this kind of lesion, whereas deeper nests of melanocytes/nevus cells are organized in the central or “body” region of an atypical nevus.49

4.3.

Strengths and Limitations

A limitation of Eclass using imaging biomarkers in this study was that there were 319 discarded images. At present, our method requires analysis of images that show the complete lesion with full borders and some adjacent normal skin, and the images cannot include hair, markings on the skin, or bubbles in the immersion media optically coupling the imaging device to the skin. In these “defective” images, one or more imaging biomarkers did not successfully compute and Eclass analysis failed but CNN analysis still worked. A number of potential solutions for this exclusion problem could enable practical clinical use, including indications to re-image and remove hair over the lesion. Although this problem must be reduced, clinicians already operate by the “when in doubt, cut it out” rule, so providing that exclusions can be reduced to <10%, we expect clinical utility. A second limitation of the current approach is that it needs retraining for each population (e.g. New York versus Barcelona in this report), so any best future case for the machine learning needs to be indicated for the population on which it was trained.

Since the number of imaging biomarkers is much less than the number of features created by CNNs, the incremental processing cost of added imaging biomarkers is small compared to the overall processing cost difference between Eclass and CNN. A strength of machine learning diagnostics using Eclass and IBCs (that may be challenging from the regulatory perspective) is that the approach can be developed by collaborating researchers combining complementary IBC sets. A repository of executable MATLAB® functions (for IBCs) is needed. Another need is for the exploration of compound models. For instance, the CNN risk score itself could be added as an additional single imaging biomarker in the Eclass approach or even downstream as an added member of the ensemble. The cost would be computational, and the benefit would be that the deep learning could produce images of similar lesions with known diagnoses and similar deep learning fingerprints to which a clinician could confirm visual similarity and thus infer probable diagnosis similarity.

4.4.

Future Directions

Needed are imaging biomarkers that recognize and correct for image defects to reduce necessary exclusions (a potential bias). This will ensure that more imaging biomarkers compute successfully, potentially leading to diagnostic confidence that the lesion contains only known features and those features have been successfully analyzed. Thus, defects in images that prevent Eclass analysis may be automatically identified and corrected, whether implementing deep learning or not. In contrast, defects in imaging biomarkers present opportunities to improve pathological interpretation of light-tissue interactions and defects in machine learning algorithms arise from biases in training data.

Although Eclass and imaging biomarkers aren’t sufficiently generalized (i.e., Fig. 4 differences across sets), which would be required to analyze lesions in real time from mobile phone-coupled dermatoscopes, Fig. 2 and the App illustrates how the data pipeline from dermatoscope image to guided biopsy decisions could work. This simulated clinical workflow highlights the translational perspective that the dermatologist can be presented a manageable amount of machine learning-augmented reality to make a decision based in part by that input but also by their medical discretion.

5.

Appendix: : Mathematical Formulas for Melanoma Imaging Biomarker Cues

5.1.

Multi-color IBCs

Multicolor IBCs were those derived from all the color channels of the red/green/blue image channels simultaneously, as opposed to single-channel IBCs [Eqs. (7)–(36)] presented later that were calculated from individual color channels. Thus, multicolor IBCs only had one version whereas each single color-IBC had three versions (one for each color channel).

Dermoscopy includes analysis of the colors present in any given lesion and there are six colors that dermatologists generally identify in lesions [light brown, dark brown, black, red, blue-gray, and white]. As direct examples, segments of dermoscopic colors were hand-segmented within the published data10 set blind to the gold standard diagnoses. At least three images containing segments of each dermoscopic color were chosen for pixel extraction and storage as exemplary for each of the six dermoscopic colors of accepted dermoscopy practice. We generated a simple color metric to assign pixels a categorical color: if the pixel ratio of red to blue was within one standard deviation of the mean for that color, and the same was true for red to green, and blue to green, then the pixel was assigned that color. For each pixel, a sequential check was made for the presence of colors in the order (light brown, dark brown, black, red, blue-gray, and white). In this manner, the two most common colors (light brown, dark brown), were first identified and labeled as the least suspicious group. Next, black and red were identified and labeled (as more suspicious) using the same color identification logic. Finally, blue-gray and white were identified as most suspicious. Thus, the algorithm checked each pixel for each color, leaving the color automatically detected as the last checked (most suspicious) color for that pixel.

A color list (CL) was produced for each lesion indicating the presence or absence of each color. For instance, CL = [1 1 1 0 0] would result from a dermoscopic image where the lesion contained light brown, dark brown, and red but no black or blue-gray/white. MC1 is then the number of dermoscopic colors identified in the lesion.

Eq. (1)

MC1=i=15CL(i).
Let L(y,x) denote an image mask of the lesion segment with value 1 inside the lesion and value 0 outside the lesion. Let Lred(y,x), Lgreen(y,x), and Lblue(y,x) be masks derived from the red, green, and blue channels of the color image, respectively. MC2 is then the normalized difference in lesion size between the red and blue color channels

Eq. (2)

MC2=x=1Nxy=1NyLred(y,x)x=1Nxy=1NyLblue(y,x)x=1Nxy=1NyLred(y,x).
Let R(θ) be the length of the radial between the geometric center of the lesion and a point on the lesion border that sweeps over the angle θ from θ=0 to θ=2π radians. Let RR(θ), RG(θ), and RB(θ) be three versions where the geometric centers and the borders are those extracted from Lred(y,x), Lgreen(y,x), and Lblue(y,x), respectively.

Eq. (3)

Rvar(θ)=σ(RR(θ),RG(θ),RB(θ))RR(θ),RG(θ),RB(θ).
MC3 is then the mean coefficient of variation of lesion radii among the color channels, where denotes the expectation value or mean operator.

Eq. (4)

MC3=Rvar(θ)|θ=0θ=2π,
where, as an illustration of the definition of the mean value, for a set x that contains n elements

Eq. (5)

x=i=1nxin.
MC4 is the binary presence of blue-gray or white in the image. Figure 3 shows MC4, which is likely the simplest IBC to visualize.

Eq. (6)

MC4=CL(5).

5.2.

IBCs with blue-channel diagnostic significance

A large set of IBCs were created based on our angular sweep analysis, shown in Fig. 9. We quantified brightness variation on an angular sweeping arm that connected the geometric center of the lesion and a point on the border tracing that border clockwise. From the center, radial arms projected to the lesion border and rotating clockwise were used as regions of interest to quantify image characteristics along the arc of rotation. The series of arcs created by radial sweep around the center covering the entire 360-degree view of the lesion, was analogous to the sweep of hands around an analog clock. The IBC-producing mathematical operations (given in Sec. S5 in the Supplementary Materials13 for Ref. 10) either produced direct transformations of the actual data (i.e., Fig. 9) or quantified differences between the data and mathematical models used to estimate the data’s deviation from smoothly transitioning functions (i.e., Fig. 10).

Let p(r1) be the pixel brightness along a radial line r1 connecting the center point of the lesion and a point on the peripheral edge of the lesion. Let Rm(θ) be the mean pixel brightness p(r1) along a set of lines that vary as specified by the angle θ. As θ varies in increments of dθ one full rotation from zero to 2π radians (360 degrees), the set of lines r1 sweep the lesion like a clock arm sweeping an analog clock.

Eq. (7)

Rm(θ)=p(r1)|θ=0θ=2π,

Eq. (8)

Rstd(θ)=σ(p(r1))|θ=0θ=2π,
where an illustration of the definition of the standard deviation, for a set x that contains n elements

Eq. (9)

σ(x)=(1n1i=1i=n(xix)2).

B1 is then the average of the absolute value of the derivative of Rm(θ) over the angular sweep is the mean instantaneous brightness shift from one angular analysis position to the next over the entire 360-degree angular range.

Eq. (10)

B1=(|Rm(θn)Rm(θn+1)|)|θ=0θ=2π.

B2 is the variance over the angular sweep of the variance in pixel brightness over the radial sampling arm. This variable is increased when there are some angles at which the lesion contains even pigmentation but others that contain variable pigmentation such as in reticular or globular patterns of bright and dark areas.

Eq. (11)

B2=σ(Rstd(θ))|θ=0θ=2π.

Let pe(r2) be the pixel brightness along a second radial line r2 of the same length as r2 and at the same angular sweep angle θ but extending from half-to-1.5 times the lesions radius R(θ) instead of 0-to-1 such as to be centered on the border between lesion and normal skin. Pe(r) has the characteristic that half of its pixels (within the lesion) are darker than the other half of its pixels (outside the lesion). Let se® be a mathematical model error function across the lesion border with three fitting parameters: Min, Max, and Slope that are iteratively adjusted to minimize the least squares difference between pe(r), the data and se(r). erf(x) is defined as twice the integral of the Gaussian distribution with 0 mean and variance of 1/2, as shown below with the dummy variable t. Considering rb as the lesion border pixel with approximately the mean pixel brightness in pe(r) and exactly the mean brightness of se(r), se(r) is defined as

Eq. (12a)

erf(x)=2π0xet2dt,

Eq. (12b)

fe(r)=erf(rrbSlope)2,

Eq. (12c)

se(r)=Min+{fe(r)min[fe(r)]}×Max.

B3 is then the mean error between the model se(r) and the data pe(r) evaluated over a range equal to the distance between the center and the lesion border but centered on the edge of the lesion. This error measurement is high if the lesion brightness does smoothly transition between dark inside the lesion and bright outside the lesion. The fitting algorithm, fminsearch() in MATLAB® (Mathworks Inc., Natick, Massachusetts), was limited to 200 fitting iterations. If convergence was reached before the 200-iteration limit, the result was flagged as one type while fits that were cut off at the 200-iteration limit were flagged as a second type. B3 included only results of the second type, that did not converge by the time the iteration limit was reached.

Eq. (13)

B3=R=0.5DR=1.5D[pe(r)se(r)]2|θ=0θ=2π.

B4 is the mode error, calculated the same as B3 but with the mode() operator instead of the mean operator, calculated for only the data that exceeded the number (200) of fitting iterations allowed.

Eq. (14)

B4=mode(R=0.5DR=1.5D(pe(r)erf(r))2)|θ=0θ=2π.

B5 is the standard deviation of the set of derivative values of the mean brightness. The variance of the derivative of brightness describes how much variability in the instantaneous change in brightness there is over the angular sweep. If some angular ranges are flat (low intra-range brightness derivative) and some ranges vary wildly, this variable will have a high value.

Eq. (15)

B5=σ(dRmdθ)=σ(|Rm(θn)Rm(θn+1)|)|θ=0θ=2π.

B6 was calculated like B3 except that it used all data and was not restricted to the data requiring more fitting iterations than MATLAB® was allowed to execute. Similarly, B7 used only the fits that did not require more iterations than (200) the maximum number of fitting iterations allowed.

A watershed analysis was developed to identify pigmented network branches. First, gray-scale images extracted from individual channels were passed through a rank filter which reset the gray-scale value of each pixel to the rank in brightness of the pixel under consideration with its group of neighboring pixels. This step was needed prior to the watershed analysis to act as a high-pass spatial filter and eliminate overall brightness variations in the lesion, leaving the local variations such as those caused by pigmented networks to be identified by the watershed analysis. Branches, which were skeletonized to a single pixel width down their spine, were characterized by three features: their length, their mean brightness, and their angle with respect to the lesion centroid. The MR clock sweep scored the mean pixel intensity of the branches Ibranch(θ), the standard deviation of intrabranch pixel intensity variation σbranch, the mean length of the branches Lbranch(θ) and the total number of branches Nbranch(θ) within a differential angle element that traced with the clock MR clock sweep. B8 is then the normalized inter-branch pixel intensity variation.

Eq. (16)

B8=σ(Ibranch(θ)|θ=0θ=2π)Ibranch(θ)|θ=0θ=2π.

B9 is the standard deviation of the error measurement like in B3, except that the standard deviation operator σ is used instead of the mean operator. B9 was evaluated only for fits requiring more fitting iterations than the 200 iterations allowed.

Eq. (17)

B9=σ{R=0.5DR=1.5D[pe(r)erf(r)]2}|θ=0θ=2π.

B10 is the normalized angular coefficient of brightness variation.

Eq. (18)

B10=σ[Rm(θ)]Rm(θ).

B11 The standardized variance of branch lengths.

Eq. (19)

B11=σ(Lbranch)|θ=0θ=2π)Lbranch|θ=0θ=2π.

B12 is the normalized range of angular brightness.

Eq. (20)

B12=max[Rm(θ)]min[Rm(θ)]Rm(θ).

B13 is calculated as is B6 except the standard deviation operator σ is used instead of the mean operator. Like B6, B13 used all the data.

Eq. (21)

B13=σ(R=0.5DR=1.5D[pe(r)erf(r)]2)|θ=0θ=2π.

B14 Is the standard deviation σ() of the error measurement as in B13 except that B14 was evaluated only for the fits that completed within the allowed number (200) of fitting iterations.

B15 Is the mean intrabranch coefficient of variation.

Eq. (22)

B15=σ(Ibranch(θ))Ibranch(θ)|θ=0θ=2π.

5.3.

IBCs with green-channel diagnostic significance

Let PerimG be the length of the perimeter of the lesion segment in the green channel Lgreen. G1 is the length of the lesion segment border normalized by the square root of the area of the lesion segment.

Eq. (23)

G1=PerimGx=1Nxy=1NyLgreen2ππ.

5.4.

IBCs with red-channel diagnostic significance

The fitting algorithm yielded a slope S for the sigmoidal edge fit. R1 was the standard deviation of the slope fit values

Eq. (24)

R1=σ(S)|θ=0θ=2π.

R2 is the fractal dimension of the lesion segment binary image as defined as50

Eq. (25)

R2=D[Lred(y,x)].

Each branch segment in terminated on two ends in either a branch point or an end point. R3 is the connectedness of the pigmented network, defined as the ratio of the number of branch points NBP to the number of endpoints NEP.

Eq. (26)

R3=NBPNEP.

R4 is the size of the lesion segment Lred, which is the sum of the binary mask valued at one inside the lesion segment and zero outside the lesion segment.

Eq. (27)

R4=x=1Nxy=1NyLred.

R5 is the mean slope (S) for the edge fit function se(r) [as used in Eq. (13)] evaluated only for the fits that did not require more iterations of the fminsearch() operator than the 200 allowed.

Eq. (28)

R5=S|θ=0θ=2π.

Let the instantaneous radius of the lesion, as in Eq. (3), be denoted by RR(θ) over the angular sweep of θ. R6 is then the coefficient of variation in the lesion radius over the angular sweep

Eq. (29)

R6=σ(Rred(θ)|θ=0θ=2π)Rred(θ)|θ=0θ=2π.

Let Nb(θ,dθ) be the number of pigmented network branches identified in a differentiual angle emelment dθ as a function of angle θ over the angular sweep. R7 is then the range in number of branches detected as a function of angle.

Eq. (30)

R7=max[Nbranch(θ,dθ)]min[Nbranch(θ,dθ)].

R8 is the range in the standard deviation of pixel brightness on the angular sweep arm over the angular sweep.

Eq. (31)

R8=max(Rstd(θ)|θ=0θ=2π)min(Rstd(θ)|θ=0θ=2π).

Pixels with the lesion segment were scored as a set Plesion. The coefficient of variation for pixels within the lesion segment was calculated by dividing the standard deviation of intralesional pixel brightness by the mean lesional pixel brightness. R9 is then the coefficient of variation in pixel brightness within the lesion.

Eq. (32)

R9=σ(Plesion)Plesion.

R10 is the mode error, calculated the same as B4, but evaluated only for the fits that did not exceed the number of fitting iterations (200) allowed.

Eq. (33)

R10=mode(R=0.5DR=1.5D[pe(r)erf(r)]2)|θ=0θ=2π.

The maximum asymmetry of the lesion was normalized by the eccentricity of the lesion E as calculated using the stats. Ecentricity function in MATLAB®. This normalization enabled de-emphasis of uniform ovals as asymmetric. R11 is then the maximum asymetry of the lesion silhouette

Eq. (34)

R11=max(AE).

R12 is the sum of the normalized derivitive in lesion radius D over the angular sweep

Eq. (35)

R12=θ=0θ=2πabs[Rred(θ,dθ)Rred(θ1,dθ)].

R13 is the asymetry of the lesion silhouette evaluated in the standard technique (Fig. S10 in the Supplementary Material for Ref. 10)13

Eq. (36)

R13=A|θsymπ2.

Disclosures

All authors have no relevant financial conflicts of interest to disclose

Acknowledgments

The authors would like to acknowledge Sarah Yagerman, who was involved on the initial study (Ref. 9) and was helped establish this clinical methodology. Funding information: Primary Funding from National Institutes of Health under Grant No. R21CA240254. Secondary funding from National Institutes of Health (Grant No. UL1TR001866), the Paul and Irma Milstein Family Foundation, The Robertson Therapeutics Development Fund, and American Skin Association.

References

1. 

M. K. Tripp et al., “State of the science on prevention and screening to reduce melanoma incidence and mortality: the time is now,” CA Cancer J. Clin., 66 460 –480 (2016). https://doi.org/10.3322/caac.21352 CAMCAM 0007-9235 Google Scholar

2. 

The American Cancer Society, “Cancer.org. Cancer Facts & Figures 2019: American Cancer Society,” (2019) https://www.cancer.org Google Scholar

3. 

A. Esteva et al., “Dermatologist-level classification of skin cancer with deep neural networks,” Nature, 542 (7639), 115 –118 (2017). https://doi.org/10.1038/nature21056 Google Scholar

4. 

E. Axpe and M. L. Oyen, “Applications of alginate-based bioinks in 3D bioprinting,” Int. J. Mol. Sci., 17 (12), 1976 (2016). https://doi.org/10.3390/ijms17121976 1422-0067 Google Scholar

5. 

Y. Pan et al., “Polarized and nonpolarized dermoscopy: the explanation for the observed differences,” Arch. Dermatol., 144 (6), 828 –829 (2008). https://doi.org/10.1001/archderm.144.6.828 Google Scholar

6. 

A. M. Forsea et al., “Eurodermoscopy Working Group, Argenziano G., Geller A. C. The impact of dermoscopy on melanoma detection in the practice of dermatologists in Europe: results of a pan-European survey,” J. Eur. Acad. Dermatol. Venereol., 31 (7), 1148 –1156 (2017). https://doi.org/10.1111/jdv.14129 JEAVEQ 0926-9959 Google Scholar

7. 

R. Betti et al., “An observational study regarding the rate of growth in vertical and radial growth phase superficial spreading melanomas,” Oncol. Lett., 12 (3), 2099 –2102 (2016). https://doi.org/10.3892/ol.2016.4813 Google Scholar

8. 

R. Yamashita et al., “Convolutional neural networks: an overview and application in radiology,” Insights Imaging, 9 (4), 611 –629 (2018). https://doi.org/10.1007/s13244-018-0639-9 Google Scholar

9. 

D. Gareau, “Digital imaging biomarkers feed machine learning for melanoma screening,” 2017). https://youtu.be/-hw_3HLcaSA Google Scholar

10. 

D. S. Gareau et al., “Digital imaging biomarkers feed machine learning for melanoma screening,” Exp. Dermatol., 26 (7), 615 –618 (2017). https://doi.org/10.1111/exd.13250 Google Scholar

11. 

A. M. Ashfaq, J. Malvehy and R. P. Braun, Atlas of Dermoscopy, Memorial Sloan-Kettering Cancer Center(2012). Google Scholar

12. 

J. K. Winkler et al., “Association between surgical skin markings in dermoscopic images and diagnostic performance of a deep learning convolutional neural network for melanoma recognition,” JAMA Dermatol., 155 1135 –1141 (2019). https://doi.org/10.1001/jamadermatol.2019.1735 Google Scholar

13. 

D. S. Gareau et al., Supplemental Materials Including Eclass Methology and Imaging Biomarkers, Wiley,2016). https://onlinelibrary.wiley.com/action/downloadSupplement?doi=10.1111%2Fexd.13250&file=exd13250-sup-0001-SupInfo.pdf Google Scholar

14. 

S. Haykin, Neural Networks and Learning Machines, 3rd ed.Prentice Hall, New York (2008). Google Scholar

15. 

C. Cortes and V. Vapnik, “Support-vector networks,” Mach. Learn., 20 (3), 273 –297 (1995). https://doi.org/10.1007/BF00994018 MALEEZ 0885-6125 Google Scholar

16. 

B. Scholkopf and A. Smola, Learning with Kernels, MIT Press(2002). Google Scholar

17. 

P. McCullagh and J. A. Nelder, Generalized Linear Models, 2nd ed.Chapman and Hall/CRC(1989). Google Scholar

18. 

J. Friedman, T. Hastie and R. Tibshirani, “Regularization paths for generalized linear models via coordinate descent,” J. Stat. Software, 33 (1), 1 –22 (2010). https://doi.org/10.18637/jss.v033.i01 Google Scholar

19. 

B. Hofner et al., “Model-based boosting in R: a hands-on tutorial using the R package mboost,” Comput. Stat., 29 (1–2), 3 –35 (2014). https://doi.org/10.1007/s00180-012-0382-5 CSTAEB 0943-4062 Google Scholar

20. 

L. Breiman, Random Forests, 5 –32 Kluwer Academic Publishers, pp. 2001). Google Scholar

21. 

L. Breiman et al., Classification and Regression Trees (Wadsworth Statistics/Probability), 1st ed.Chapman and Hall/CRC(1984). Google Scholar

22. 

E. Fix and J. L. Hodges, “Discriminatory analysis, nonparametric discrimination: consistency properties,” (1951). Google Scholar

23. 

J. H. Friedman, “Multivariate adaptive regression splines,” Ann. Stat., 19 (1), 1 –67 (1991). https://doi.org/10.1214/aos/1176347963 ASTSC7 0090-5364 Google Scholar

24. 

S. L. Salzberg, “C4.5: Programs for Machine Learning by J. Ross Quinlan. Morgan Kaufmann Publishers, Inc., 1993,” Mach. Learn., 16 235 –240 (1994). https://doi.org/10.1007/BF00993309 Google Scholar

25. 

H. Wold, Soft Modeling by Latent Variables; the Nonlinear Iterative Partial Least Squares Approach. Perspectives in Probability and Statistics Papers in Honour of M S Bartlett, (1975). Google Scholar

26. 

R. A. Fisher, “The use of multiple measurements in taxonomic problems,” Ann. Eugen., 7 179 –188 (1936). https://doi.org/10.1111/j.1469-1809.1936.tb02137.x Google Scholar

27. 

L. Kuncheva, Combining Pattern Classifiers. Methods and Algorithms, 2nd ed.Wiley(2014). Google Scholar

28. 

R. Rosipal, N. Krämer, “Overview and recent advances in partial least squares,” Subspace, Latent Structure and Feature Selection, 34 –51 Springer, Berlin Heidelberg (2006). Google Scholar

29. 

R. Polikar, “Ensemble based system in decision making,” IEEE Circuits Syst. Mag., 6 21 –45 (2006). https://doi.org/10.1109/MCAS.2006.1688199 ICSYDV 0163-6812 Google Scholar

30. 

L. Rokach, “Ensemble-based classifiers,” Artif. Intell. Rev., 33 1 –39 (2010). https://doi.org/10.1007/s10462-009-9124-7 AIREV6 Google Scholar

31. 

C. Vrattos et al., “Eclass Imaging Biomarkers,” https://apple.co/375vPQJ Google Scholar

32. 

G. Argenziano et al., “Dermoscopy improves accuracy of primary care physicians to triage lesions suggestive of skin cancer,” J. Clin. Oncol., 24 (12), 1877 –1882 (2006). https://doi.org/10.1200/JCO.2005.05.0864 JCONDN 0732-183X Google Scholar

33. 

F. Nachbar et al., “The ABCD rule of dermatoscopy. High prospective value in the diagnosis of doubtful melanocytic skin lesions,” J. Am. Acad. Dermatol., 30 (4), 551 –559 (1994). https://doi.org/10.1016/S0190-9622(94)70061-3 JAADDB 0190-9622 Google Scholar

34. 

J. Malvehy et al., “Clinical performance of the Nevisense system in cutaneous melanoma detection: an international, multicentre, prospective and blinded clinical trial on efficacy and safety,” Br. J. Dermatol., 171 (5), 1099 –1107 (2014). https://doi.org/10.1111/bjd.13121 BJDEAZ 0007-0963 Google Scholar

35. 

D. S. Rigel, J. Russak and R. Friedman, “The evolution of melanoma diagnosis: 25 years beyond the ABCDs,” CA Cancer J. Clin., 60 (5), 301 –316 (2010). https://doi.org/10.3322/caac.20074 CAMCAM 0007-9235 Google Scholar

36. 

J. S. Henning et al., “The CASH (color, architecture, symmetry, and homogeneity) algorithm for dermoscopy,” J. Am. Acad. Dermatol., 56 (1), 45 –52 (2007). https://doi.org/10.1016/j.jaad.2006.09.003 JAADDB 0190-9622 Google Scholar

37. 

R. J. Drugge et al., “Melanoma screening with serial whole body photographic change detection using Melanoscan technology,” Dermatol. Online J., 15 (6), 1 (2009). Google Scholar

38. 

M. E. Vestergaard et al., “Dermoscopy compared with naked eye examination for the diagnosis of primary melanoma: a meta-analysis of studies performed in a clinical setting,” Br. J. Dermatol., 159 (3), 669 –676 (2008). https://doi.org/10.1111/j.1365-2133.2008.08713.x BJDEAZ 0007-0963 Google Scholar

39. 

R. Hofmann-Wellenhof et al., “Reflectance confocal microscopy--state-of-art and research overview,” Semin. Cutan Med. Surg., 28 (3), 172 –179 (2009). https://doi.org/10.1016/j.sder.2009.06.004 Google Scholar

40. 

J. D. Emery et al., “Accuracy of SIAscopy for pigmented skin lesions encountered in primary care: development and validation of a new diagnostic algorithm,” BMC Dermatol., 10 9 (2010). https://doi.org/10.1186/1471-5945-10-9 Google Scholar

41. 

S. W. Menzies et al., “The performance of SolarScan: an automated dermoscopy image analysis instrument for the diagnosis of primary melanoma,” Arch. Dermatol., 141 (11), 1388 –1396 (2005). https://doi.org/10.1001/archderm.141.11.1388 Google Scholar

42. 

R. J. Friedman et al., “The diagnostic performance of expert dermoscopists vs a computer-vision system on small-diameter melanomas,” Arch. Dermatol., 144 (4), 476 –482 (2008). https://doi.org/10.1001/archderm.144.4.476 Google Scholar

43. 

M. Ramezani, A. Karimian and P. Moallem, “Automatic detection of malignant melanoma using macroscopic images,” J. Med. Signals Sens., 4 (4), 281 –290 (2014). https://doi.org/10.4103/2228-7477.144052 Google Scholar

44. 

D. Sgouros et al., “Assessment of SIAscopy in the triage of suspicious skin tumours,” Skin Res. Technol., 20 440 –444 (2014). https://doi.org/10.1111/srt.12138 Google Scholar

45. 

G. Monheit et al., “The performance of MelaFind: a prospective multicenter study,” Arch. Dermatol., 147 (2), 188 –194 https://doi.org/10.1001/archdermatol.2010.302 Google Scholar

46. 

S. M. McKinney et al., “International evaluation of an AI system for breast cancer screening,” Nature, 577 (7788), 89 –94 (2020). https://doi.org/10.1038/s41586-019-1799-6 Google Scholar

47. 

M. Belkin et al., “Reconciling modern machine-learning practice and the classical bias-variance trade-off,” Proc. Natl. Acad. Sci. U. S. A., 116 (32), 15849 –15854 (2019). https://doi.org/10.1073/pnas.1903070116 Google Scholar

48. 

“The colors of cancer: hyperspectral melanoma imaging,” 2018). https://youtu.be/olWpZrWHbvc Google Scholar

49. 

D. E. Elder, “Dysplastic naevi: an update,” Histopathology, 56 (1), 112 –120 (2010). https://doi.org/10.1111/j.1365-2559.2009.03450.x HISTDD 1365-2559 Google Scholar

50. 

A. F. Costa, Hausdorff (Box-Counting) Fractal Dimension, MathWorks Inc.(2013). Google Scholar

Biography

Daniel S. Gareau is a Clinical Instructor at the Rockefeller University, develops hyper-spectral and confocal imaging biomarkers in natural and 3D-printed biology. He develops and commercializes optical diagnostics, typically leveraging nondilutive funds. He completed his PhD in biomedical engineering at Oregon Health & Science University under Steven L. Jacques and postdoctoral studies at Memorial Sloan Kettering Cancer Center in Dr. Milind Rajadhyaksha’s laboratory. He continues fusing classic electro-optical engineering technique with modern biology, clinically translating technology.

James Browning is a research scientist at Covera Health where he develops predictive models for pathology detection. He received his PhD from the University of Colorado where he investigated 3D cardiac coherent flow structures under the direction of Jean Hertzberg, PhD. He completed a postdoc at The Rockefeller University where he investigated automated skin cancer screening and treatment assays under the direction of Daniel Gareau, PhD, and James Krueger, MD, PhD.

Amanda M. Zong is an undergrad at Columbia University pursuing a BS degree in computer science. She is interested in developing computational diagnostic tools for medical imaging and has conducted research at Rockefeller University and Columbia Medical Center.

Cristina Carrera is a consultant specialist at the Dermatology Department in Hospital Clinic of Barcelona, in charge of clinical and research tasks in Melanoma and Skin Cancer. She has a developed interest in imaging diagnostic techniques (dermoscopy, confocal microscopy, OCT) and immuno-oncology. She is author of more than 150 publications and coinvestigator in more than 20 competitive research projects, also holds the post of Associate Professor at the University of Barcelona.

Josep Malvehy is a clinician-scientist focusing on skin cancer diagnostics at The University of Barcelona and the Hospital Clinic Of Barcelona. He is the leader of the Technology Group of the International Skin Imaging Collaboration Group and a consultant of the Technical Committee of the Association Espanola contra el Cancer.

Susana Puig is a professor with the Department of Dermatology, University of Barcelona. She has authored many publications and presented works globally. As the Head of the Dermatology Service at the Hospital Clinic in Barcelona, she has worked with melanoma imaging and genetics for more than 20 years, focusing on epigenetics and therapy. She leads the research team “Melanoma: imaging, genetics and immunology” at the August Pi i Sunyer Biomedical Research Institute.

Ashfaq Marghoob is an attending physician specializing in the early detection of skin cancer by leveraging technology to help improve the morbidity and mortality associated with cutaneous malignancies.

John A. Carucci directs the Mohs and Dermatologic Surgery Unit at New York University, which interconnects head and neck oncologic surgery, surgical oncology, plastic surgery, radiation oncology, medical oncology and dermatopathology. Interests in the molecular biology and immunology of skin cancer underpin his role as physician-scientist.

James G. Krueger is the D. Martin Carter Professor in Clinical Investigation and codirector of the Center For Clinical And Translational Science at The Rockefeller University. He is interested in skin imaging and investigates the underlying molecular biology, using psoriasis as a model to study inflammatory diseases that involve Th17 cells, a set of T cells. His work has implications for cancer diagnosis and inflammatory diseases, such as rheumatoid arthritis and inflammatory bowel disease.

Biographies of the other authors are not available.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Daniel S. Gareau, James Browning, Joel Correa Da Rosa, Mayte Suarez-Farinas, Samantha Lish , Amanda M. Zong, Benjamin Firester, Charles Vrattos, Yael Renert-Yuval, Mauricio Gamboa, María G. Vallone, Zamira F. Barragán-Estudillo, Alejandra L. Tamez-Peña, Javier Montoya, Miriam A. Jesús-Silva, Cristina Carrera, Josep Malvehy, Susana Puig, Ashfaq Marghoob, John A. Carucci, and James G. Krueger "Deep learning-level melanoma detection by interpretable machine learning and imaging biomarker cues," Journal of Biomedical Optics 25(11), 112906 (27 November 2020). https://doi.org/10.1117/1.JBO.25.11.112906
Received: 25 May 2020; Accepted: 28 October 2020; Published: 27 November 2020
Lens.org Logo
CITATIONS
Cited by 11 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Melanoma

Diagnostics

Machine learning

Data modeling

Visualization

Skin

Sensors

RELATED CONTENT

Dermoscopy analysis of RGB-images based on comparative features
Proceedings of SPIE (September 22 2015)
Self-structuring data learning approach
Proceedings of SPIE (June 22 2016)
Introduction to Bayesian image analysis
Proceedings of SPIE (September 14 1993)

Back to Top