Next Article in Journal
The Relationship between Carotid Doppler Ultrasound and EEG Metrics in Healthy Preschoolers and Adults
Next Article in Special Issue
The Relationship between Oscillations in Brain Regions and Functional Connectivity: A Critical Analysis with the Aid of Neural Mass Models
Previous Article in Journal
Understanding Novel Metaphors: A Milestone in the Developmental Trajectory of Children with Agenesis of the Corpus Callosum?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Three-Stage Teacher, Student Neural Networks and Sequential Feed Forward Selection-Based Feature Selection Approach for the Classification of Autism Spectrum Disorder

by
Naseer Ahmed Khan
1,†,
Samer Abdulateef Waheeb
1,†,
Atif Riaz
2 and
Xuequn Shang
1,*,†
1
School of Computer Science and Technology, Northwestern Polytechnical University, Xi’an 710072, China
2
Department of Computer Science, University of London, London WC1E 7HU, UK
*
Author to whom correspondence should be addressed.
Current address: Northwestern Polytechnical University, Changan Campus, Changan, Xi’an 710072, China.
Brain Sci. 2020, 10(10), 754; https://doi.org/10.3390/brainsci10100754
Submission received: 5 September 2020 / Revised: 11 October 2020 / Accepted: 12 October 2020 / Published: 19 October 2020
(This article belongs to the Special Issue Neural Networks and Connectivity among Brain Regions)

Abstract

:
Autism disorder, generally known as Autism Spectrum Disorder (ASD) is a brain disorder characterized by lack of communication skills, social aloofness and repetitions in the actions in the patients, which is affecting millions of the people across the globe. Accurate identification of autistic patients is considered a challenging task in the domain of brain disorder science. To address this problem, we have proposed a three-stage feature selection approach for the classification of ASD on the preprocessed Autism Brain Imaging Data Exchange (ABIDE) rs-fMRI Dataset. In the first stage, a large neural network which we call a “Teacher ” was trained on the correlation-based connectivity matrix to learn the latent representation of the input. In the second stage an autoencoder which we call a “Student” autoencoder was given the task to learn those trained “Teacher” embeddings using the connectivity matrix input. Lastly, an SFFS-based algorithm was employed to select the subset of most discriminating features between the autistic and healthy controls. On the combined site data across 17 sites, we achieved the maximum 10-fold accuracy of 82% and for the individual site-wise data, based on 5-fold accuracy, our results outperformed other state of the art methods in 13 out of the total 17 site-wise comparisons.

1. Introduction

The human brain is considered the most complex organ in the body due to it’s structural and functional variations across the temporal and spatial domain, variety of cognitive functions based on the interaction of functional regions and the intrinsic modularity that is present in those regions [1]. Although, there are influential studies that have shown the existence of small-world-networks [2], modular networks [3] and hierarchical organization of the different modules [4] in the human brain but still much needs to be done in the field of brain sciences to grasp the complexity of the brain. New theories and understanding of the coordination and synchronization of various brain regions are being explored to solve the mysteries of the functions of the brain networks [5].
Autism Spectrum Disorder (ASD) is a brain disorder with characteristics such as social aloofness, repetitive actions and inability to communicate effectively. Although, there are many neuroscience and genetic related markers identified already but still a lot of progress needs to be done for the diagnosis of this mental disorder [6].The prevalence rate of parents reporting autism in children in the USA between the age of 3 years to 17 years, using a sample size of more than 78,000, was estimated to be 110 per 10,000 [7]. Gender is also believed to be a factor when it comes to autism disorder and studies have proved that the autism disorder is more common in men as compared to women as discussed in [8] and [9]. More recent studies have also proved that the gender factor has significance in ASD and cannot be taken as a trivial issue. In this regard meta-analysis of 54 studies with a combined size population of 13,784,284 discovered that among the 53,712 participants that had ASD, male population of 43,972 and female of 9740 were found to have ASD resulting in 3:1 prevalence in the males as compared to females, thereby, providing strong evidence that autism more likely affects men [10] and verifying an earlier work which found that autism was more common among men [11].
Many imaging modalities are used to study ASD and each of these modalities have their own peculiar characteristics. These modalities include DTI (Diffusion Tensor Imaging), EEG (Electroencephalography), fMRI (functional Magnetic Resonance Imaging), MEG (Magnetoencephalography), PET (Positron Emission Tomography) and SPECT (Single-Photon Emission Computed Tomography) are all used in practice for the study of autism [12,13,14,15,16].
Functional Magnetic Resonance Imaging (fMRI) is currently very popular as it is a non-invasive method to acquire neural activity information from the human brain. Moreover, recent advancements in the fMRI technology has made considerable improvements in the quality and reliability of the haemodynamic BOLD (Blood-Oxygen-Level-Dependent) signal at both the spatial as well as the temporal level that is obtained from the powerful magnetic scanner [17,18]. The primary task of an fMRI experiment which is based on the MRI (Magnetic Resonance Imaging) is to measure the signal using 3D scans of the brain which are stacked in the time dimension that makes them a 4D modality. After that, brain volumes are divided into small cubes called “voxels” and then the average BOLD signal or neural activity in those voxels is measured [19].
In many brain disorders, functional connectivity based analysis is getting more and more attention due to its simplicity in calculation and understanding as it is a measure that tells how close functionally related two brain regions are with each other. Studies have shown alteration patterns in the functional connectivity and existence of significant and interpretable markers from the brain region wise connectivity matrix that are discriminating between a subject of the disease and a healthy control in a variety of brain disorders such as ADHD (Attention Deficit Hyper Disorder), Alzheimer, ASD, Epilepsy and Schizophrenia [20,21,22,23,24].
In this study, we have proposed a deep learning-based Teacher-Student feature selection method to extract the most discriminating features between the autistic subjects and the healthy controls. Firstly, we built the functional connectivity matrices for each subject in the dataset, as this would be the input to our neural network-based models. Secondly, we constructed a large neural network which we call a “Teacher” neural network and trained it on the connectivity matrix so that it could learn insightful patterns on the training dataset. Thirdly, we extracted the codes from the trained Teacher neural network and normalized them for convergence, these codes can be called as a learned embedding on the trained Teacher model. In the next step, we built a one layer hidden neural network, which we call as “Student” neural network and trained it again on the connectivity matrix input and the task of this neural network was to reproduce the codes or embeddings that we extracted from the trained teacher neural network. After this, we extracted the weights from the hidden layer of the Student neural network and sorted them in descending order as higher magnitude of the weights correspond to the indices of the connectivity matrix features with greatest discrimination power. Finally, we gave those features to various classifiers and tested the sensitivity, specificity and accuracy of the classifier. Our contribution in this study has three key characteristics, (i) first we used the 3-Stage Teacher, Student and SFFS based feature selection approach, which is a novel idea for the fMRI domain, (ii) our features are of a short size and are verifiable from the supplementary material which is mentioned in our study and lastly we have outperformed several state of the art algorithms in the literature which validates the effectiveness and usefulness of our work.
The rest of the paper is as follows: in Section 2 we will discuss the related research that has been conducted on the current problem; in Section 3 will discuss dataset and the proposed methodology; in Section 4, we will discuss our experiments, results and comparison with the state of the art methods; in Section 5 we will discuss our feature selection process; in Section 6 we will discuss our selected features and their significance in the autism disorder, and lastly in Section 7 we conclude our study and propose future work for the research community on this problem.

2. Related Research Work

ASD has grabbed the attention of many researchers due to the challenging task of finding discriminating neural markers that could help differentiate an autistic person from a healthy control. In the following subsections, we will discuss various techniques that are being used for the study of classification related to fMRI imaging dataset available on ASD.

2.1. Signal Processing Based Approaches

A study explored the Generalized Autoregressive Conditional Hetroscedasiticty model (GARCH) [25] that extracted features after decomposing the subject region data into sub-bands using Double Density Dual-Tree Discrete Wavelet Transform (D3TDWT). After that, the extracted features were given to Support Vector Machine (SVM) that resulted in 71% accuracy for male and 83% accuracy for female ASD data. Fast-Entropy based algorithm [26] was used and the results were reported using the area under the receiver (AUC) metric value of 0.62 for the classification of ASD, Discriminating features on 21 autistic and 26 healthy patients were selected using two-sample t-test and were fed to SVM classifier for training. In [27] authors applied fluctuation entropy using a single functional near-infrared spectroscopy channel on 25 ASD and 21 control subjects resulted in that 97.8% accuracy. A variation of Fourier Transform (FT) called Graph Fourier Transform (GFT) was applied on 172 subjects and control data where the author first computed the statistical measures from the healthy time series of a subject then projected that to a structural graph which was computed from the healthy connectome graph that resulted in better classification. In another study [28] using ICA (Independent Component Analysis) and their associated time series, wavelet-based coherence maps were built which when given to classifier for training, resulted in 86.7% and 80% testing accuracies on two different datasets containing 12 ASD, 18 controls and 12 ASD and 12 controls respectively. In [29] authors proposed a spatial filter approach that projected the covariance matrix of the BOLD signal of the ASD and control subjects to the orthogonal directions to make the two type of subjects highly separable.

2.2. Functional Connectivity Based Approaches

Functional Connectivity (FC), a pair-wise relationship between two brain regions, is considered an important step in the search for the neuromarkers for the ASD subject. A study applied ICA to study altered FC in the brain Default Mode Network (DMN) on the 16 ASD patients with matched 16 control subjects showed promising biomarkers for the classification of ASD subjects [30], as their results showed a decreased FC among the ASD subjects as compared to control subjects. Disruption in FC was found in task unrelated neural activity between the 23 ASD and 20 control subjects, 17 Regions of Interest (ROI) were used to form a 17 × 17 correlation matrix that was used to study the altered FC by using two sample t-test between the two connectivity matrices [31]. In [32] authors showed that altered FC is important for discriminating ASD subjects from the healthy control after controlling the confounding effect of head motion, their findings concluded altered FC in 19 ASD and 20 control subjects. In another study a Probabilistic Neural Network (PNN) [33] consisting of an input layer, pattern layer, summation layer and output layer showed promising results on the large data consisting of 312 ASD and 328 control subjects. The authors used a correlation-based connectivity matrix in the input layer resulting in 90% test accuracy. In another study, FC was also found to be predicting the future traits of ASD based on three brain networks involving Default Model Network (DMN), Fronto Parietal Task Control Network (FPTN) and Salience Network (SN) but their results showed promising results only in SN with 100% sensitivity and 70% precision [34]. In [35] authors analyzed the aberrant connections in the FC correlation matrix on the ROI made from the selected ICA components. They constructed the connectivity matrix of size 54 × 54 for each type of subject and clustered those connectivity matrices into a different number of groups using the varying value of the parameter k in the k-means clustering algorithm. Finally, a two-sample t-test was performed on each of the Mean Dwell Time (MDT) for all k values to extract significant regions between ASD and control. In [36], the authors incorporated both intra-site and inter-site variability to validate their results. They used multiple brain atlases to first estimate the time-series of the ROIs and then constructed the features from the estimated connectivity matrix. Features were given to the ridge-classifies resulting in 67 % accuracy on the 1112 subjects data compiled from multiple sites. Adapted Signal Change (ASC) [37] approach was introduced to improve the specificity of results obtained using the connectivity matrix-based analysis. Authors showed that alterations in the time series signal can be decomposed into the prevalent classes of change that are more useful in the subsequent analysis done using connectivity matrices. Support Vector Machine-based Recursive Feature Elimination (SVM-RFE) was proposed by [38], where the correlation-based connectivity matrix was recursively pruned for discriminating features using SVM classifier, resulting in 90% accuracy on the dataset combined from all sites. Eigen features corresponding to 256 brain regions using the Laplacian matrix were proposed in [39] with the accuracy of 77%.

2.3. Deep Learning-Based Approaches

Deep learning [40] is based on the architecture of neural network where neurons are attached in a layer-wise fashion called feed-forward connection with a non-linear activation at the final processing unit on each layer and training of each layer based on the back-propagation algorithm. In this way, a complex relationship is modelled and a representation or latent dimension of the input model is attained that could better explain the underlying relationship. Long Short-Term Memory (LSTM)-based classification of the ASD dataset was proposed in [41]. In the first step, data augmentation was first performed based on the minimum time length of 90. Instead of calculating the connectivity matrix and then feature selection, the authors directly classified the subjects based on the individual times series resulting in 68% accuracy on the combined dataset. Deep Neural Network-based Feature Selection (DNN-FS) was proposed in [42]. First, the connectivity matrix features were fed to the sparse autoencoders to reduce the dimensionality of the dataset. After this step, data were transformed to a lower-dimensional representation. Finally, it was fine-tuned by attaching a neural network thereby extracting the features with high discriminating power resulting in 86% classification accuracy. Phenotypic information on the ASD and control data has also found to be effective in discriminating ASD subjects from the control. Phenotypic data is concatenated with the features data for the input of LSTM network resulting in 67% accuracy [43]. Layer wise pre-trained autoencoders were used in [44], two autoencoders of nodes 1000 and 600 were separately trained on the input features consisting of 19900 nodes (CC200 brain atlas) and the 1000 nodes respectively. Then they were fine-tuned by attaching a classification layer consisting of two nodes resulting in an overall accuracy of 70% after cross-validation. Deep Belief Network (DBN) was used to classify the ASD subjects from the healthy controls, both structural fMRI and resting-state fMRI were used to extract the features that were fed to stacked Restricted Boltzmann Machine (RBN)-based DBN. It resulted in 65% accuracy in the discriminating classification of ASD and healthy controls [45]. A Convolution Neural Network (CNN) consisting of two sub-networks, that is one CNN to extract the time-related components from the time series data and the other to extract the spatial features from the data using a 3D CNN was proposed in [46] resulting in F-score of 89%. A two-stage Deep Neural Network (DNN) classifier was proposed in [47]. First, a CNN-based model was built to extract important features from the data then images were corrupted using frequency normalized approach and finally, the extracted features were again given to DNN that was used to get the discriminating features from the ROIs. Auto-ASD-Network was proposed in [48] to classify the ASD subjects from the healthy control data. Firstly, the dataset was augmented using the Synthetic Minority Oversampling Technique (SMOTE) and a neural network was built for classification. Finally, the hidden layer of the neural network was connected to the ATM (Auto Tune Model)-based SVM to classify the data. ASD-Diagnet was proposed in [49], first, the data augmentation was performed using a linear interpolation method. Then a single layer perceptron and an autoencoder was used to classify the data on 1035 participants that resulted in 80% overall accuracy. A CNN based deep neural network and a Multichannel based attentional neural network for the classification of ASD were proposed in [50] and [51] respectively with promising results.

3. Materials and Methods

3.1. Dataset

We have used ASD preprocessed dataset from the Autism Brain Imaging Data Exchange (ABIDE) [52] consortium where data from 17 sites with phenotypic information related to age, sex and Autism Diagnostic Observation Schedule (ADOS) [53] scores are maintained in a pre-preprocessed state for researchers. It is pertinent to mention here that resting-state rfMRI imaging data is computationally expensive to download, store and process. The original ABIDE dataset claimed to have 1112 subjects consisting of 539 ASD and 573 Healthy controls subjects. However, after removing subjects with missing information we have a total of 1035 subjects which is consistent with the total subjects reported in earlier studies of [37,44]. A dataset with the total number of participant’s condition in each site is summarized in Table 1.

3.2. Preprocessing the Dataset

We have used the ABIDE dataset that is pre-processed using Configuration Pipeline for the Analysis of Connectome (CPAC) [54], CPAC consists of open source tools that involve both structural and functional processing of the rsfMRI-based imaging data. Where the structural preprocessing involves steps like skull stripping, segmenting brain into tissue types and normalizing them to Montreal Neurological Institute-based template. Whereas, the functional processing includes steps like slice time correction, motion correction, band-pass filtering and registering the functional images to anatomical space.

3.3. Methodology

Our feature selection approach is inspired by the deep learning-based feature selection method proposed in [55], in which they developed a knowledge distillation-based deep learning approach to extract the features from the dataset. Our proposed approach consists of 3 stages for the Feature Selection process that is shown in Figure 1.

3.3.1. Stage 1, Teacher Neural Network

In the Stage 1, a Teacher Neural Network (TNN) which is a large neural network that consists an input layer, a number of hidden layers which is much larger than the Student Neural Network (SNN) and a binary classification layer. The dimension of the input layer is the same as the dimension of the feature vector and neurons in each subsequent layers are gradually reduced to avoid over-fitting. Finally, the last layer is a classification layer as our dataset is binary class data with autistic subjects labels set to 1 and healthy controls set to 0, so the last layer contains two neurons for each of the mentioned class.

3.3.2. Stage 2, Student Neural Network

In the second stage, After Training the TNN we extracted the trained codes of the hidden layer of the TNN and gave Student Neural Network (SNN) a task of reproducing these codes as in the case of simple autoencoder-based neural network model. We used a lower dimension for the embedding layer as the higher the size of the embedding layer, the more parameters it will contain in the SNN resulting in poor training and higher loss. It is pertinent to mention here that the SNN just contains one hidden layer with customized regularity conditions on the weights of the hidden layer that is explained in [55]. After training the SNN we extracted the weight matrix corresponding to the input feature layer and hidden layer as our objective is to select features based on this weight matrix.

3.3.3. Stage 3, Feature Extraction Module

In the last stage, Feature extraction is on the hidden layer’s weights that are extracted from the trained SNN. Selection of discriminating and meaningful features is done, by first, sorting the trained weights in order of increasing magnitude. Secondly, the feature indices from the input feature vector are extracted using the corresponding ranked weight vector. Lastly, a Sequential Forward Feature Selection (SFFS) approach is used to select a set of the most discriminating features in a linear stepwise way, that is we started from one feature and stopped the feature selection when the maximum accuracy got attained on the trained data using a selected classifier.

3.3.4. Algorithm

Consider that S i , i = 1 , 2 , N is a subject matrix denoting ith autistic or healthy subject’s information where N denotes the number of subjects. Moreover, each S i = X j k where X j k denotes the time series data for each subject such that j = 1 , 2 , R and k = 1 , 2 , T , where j corresponds to the brain region out of R total regions and k denotes the time point out of the T total time points for each subject. Our algorithm’s input is the subject connectivity matrix which is obtained by calculating a pairwise correlation between two brain regions. Therefore, first, we need to convert the brain region time series data from each subject to the region-wise correlation matrix that is, C i = C o r r e l a t i o n ( S i ) where C i is a R × R matrix, computed by taking Pearson correlation of each brain region time series with the rest of the regions resulting in a symmetric square connectivity matrix for each subject. F i = U p p e r T r i a n g u l e r ( C i ) is a feature vector that contains the upper-triangular elements of the connectivity matrix as only the unique elements of this matrix will be useful for further analysis. The number of elements in the feature vector are of size R ( R 1 ) 2 where R depends upon the selected brain parcellation template like AAL, CC200 and dosenbach160. Details of the working of our 3-Stage Feature selection method that converts the connectivity features F i to the indices vector and the SFFS method that on getting F i , index vector, a specified threshold accuracy level and a classifier, sequentially selects the features are described in Algorithm 1 and Algorithm 2, respectively.
Algorithm 1 Ranking Discriminating Features Algorithm
Input: C O N N m × n     Output: I n d i c e s S × 1
1:
C O D E S m × d T e a c h e r N e u r a l N e t w o r k ( CONN )
2:
W m × h S t u d e n t N e u r a l N e t w o r k ( CONN , CODES )
3:
I n d i c e s S × 1 D i a g ( W . W T )
4:
I n d i c e s S × 1 A r g S o r t ( I n d i c e s )
Algorithm 2 Sequential Forward Feature Selection Algorithm
Input: C O N N m × n , I n d i c e s L × 1 , T H R , CLF
Output: F S × 1
1:
f e a t u r e B o x e m p t y
2:
A C C L × 1 0
3:
for i n d e x in Indicesdo
4:
f e a t u r e s CONN[ 0 to  i n d e x ]
5:
A C C A c c u r a c y S c o r e ( CLF[ f e a t u r e s ] )
6:
if A C C is = T H R then
7:
   f e a t u r e B o x . a p p e n d ( features)
8:
end if
9:
end for
10:
m a x I n d e x ArgMax(ACC)
11:
F (featureBox[maxIndex])

4. Experimentation and Results

We have done our experiments on both the combined dataset and on the site-wise data to prove the effectiveness of the selected features based on our 3-Stage-based feature selection approach.

Experimental Settings

We have used the brain Automated anatomical labelling (AAL) template which divides the brain into 90 regions. Specific details about the AAL regions see Table S5 in Supplementary Materials, the reason for choosing the AAL template is the fewer number of correlation features, which are 4005 unique paired regions, thereby, avoiding the curse of dimensionality as explained in another work on the ADHD brain disorder classification [56]. For each autistic subject and healthy control, we first calculated the functional connectivity between the brain regions. The FC is of size 90 × 90 correlation matrix where the correlation is computed using Equation (1), where x i and y i correspond to the time series region values corresponding to region x and y, x i ¯ and y i ¯ correspond to the average values against those regions and lastly, n is the total number of time points.
c o r r ( x i , y i ) = i = 1 n ( x i x i ¯ ) ( y i y i ¯ ) i = 1 n ( x i x i ¯ ) 2 i = 1 n ( y i y i ¯ ) 2
For TNN we have made a relatively large neural network as compared to the SNN, Our TNN is a 17 layer neural network including the input and output layer, neurons in each layer are gradually reduced from the original 4005 input dimension to the final binary layer corresponding to autistic and healthy controls. TNN is trained with “Adadelta” optimizer with the learning rate set to “0.001” and the loss is set to ”binary-cross-entropy”. After training the TNN with the defined parameters, we extracted the trained TNN until the layer of size “5” and the complete dataset was transformed to the latent dimension of size 5 instead of the original 4005 input dimension, for specific details about the architecture and parameters of the TNN please see the Tables S1 and S2 in the Supplementary Materials. The justification for choosing the embedding layer of size 5 is that we did experiments with embedding layer of size greater than 5 and size lesser than 5. For the larger embedding dimension, the overall accuracy was not good which led to poor results in site-wise accuracy as well. Similarly, a lower dimension embedding size gave lesser accurate results in site-wise accuracy and could not learn the complexity of the dataset. All those experiments proved that embedding size of 5 was optimum for our study. SNN is a shorter neural network with only one hidden layer of size 100, where the size of 100 in the hidden is chosen after doing experiments with other sizes, the purpose of the SNN is to learn the embeddings of size “5”. Therefore, the SNN input is the 4005 input dimension, 100 in the hidden layer and the final layer is of size “5”. Training of the TNN is done by setting the optimizer to “Adadelta”, learning rate of “0.01” and loss function set to “mse”. We now give the justification for choosing the specific hyper-parameters values that we used after experiments. The number of hidden layers of the neural network is still a challenging problem that needs much attention. We used a 17 layer neural network for the TNN after experiments with 10-layer, 15-layer and 20-layer neural networks as possibilities for choosing a specific number of neural network’ layers are just endless. The gradual decrease in the number of neurons for the TNN and in the dropout is to avoid over-fitting, as the large dropout value ignored many useful features. The lower learning rate of 0.001 in the TNN was used as the higher learning rate resulted in much larger variations in the loss values during training. The number of epochs was chosen as 500 with a batch size of 16 as lower learning rate converges at a slower speed and a smaller batch will not ignore too many subjects or examples. The binary-cross entropy loss function in the TNN and mean squared error loss function in the SNN is used as the former is the classification task and the latter is the reproduction task. The optimizers “Adadelta” and “Adam” did not have any difference in the results, so to make our experiments consistent, we used the same optimizer in TNN and SNN. A comparatively higher learning rate of 0.01 was used in the SNN as it is one layer network with the task of reproducing the embedding values and with thousands of parameters in the first two layers so it’s loss values were converging very slow which prompted us to increase the learning rate. The activation function of “tanh” was used in the TNN as the dataset contains negative values so using the “relu” activation resulted in the loss of information and resulted in poor convergence during training. On the other hand activation function of “relu” was used in the SNN as the embedding layer was first standardized for convergence during the training of SNN. Lastly, for the training of SNN whole dataset was used because just using a feature set of size 4005 will lead to over-fitting and we wanted the latent lower dimension representation of the whole dataset. Specific details about the architecture and parameters of the SNN, please see Tables S3 and S4 in the Supplementary Materials. Both the TNN and SNN get the input in the form of correlation matrix-based connectivity features, where the number of layers is chosen after doing experiments.

5. Feature Selection

We have employed the sequential forward feature selection (SFFS) approach to get the most discriminant features that differentiate an autistic subject from the healthy control. In forward feature selection, features are appended to the training vector in a forward direction and are fed to the classifier to measure the accuracy. We have tested our feature selection approach using five different classifiers, which are Support Vector Machine (SVM), Random Forests, Decision Trees, Logistic Regression-based classifier and Linear Discriminant classifier. After obtaining the indices of the most discriminant features using the trained SNN, we selected a sequential group of these indices to form a set of features that were fed to the classifier. We have evaluated the method using 10-fold Accuracy using various classifiers with the cumulative feature step of 50 features at each forward step, by cumulative feature step of 50 we mean that, each step of 1 means a set of 50 features, for example, step of 1 means features in range 1 to 50, step of 2 means features in the range 1 to 100. Here it is pertinent to mention that step larger than 1 is used to speed the classifier training process. 10-Fold accuracy of various classifiers with the cumulative step as defined above is shown in Figure 2. It can be seen from the figure that accuracy of the classifier is highest in cumulative step 1 to 5 and after that, it starts to decrease, which is a typical case of over-fitting as the selection of large features means “curse of dimensionality”. Selected cumulative set with the cumulative step size that corresponds to the highest.

5.1. Justification of Selected Features

In the following three sections we will explain why we have selected the features indices corresponding to the highest accuracy cumulative feature index.

5.1.1. UnderFitting

With reference to Figure 2, we have labelled area corresponding to the low accuracy region as “biased”. The reason we have called it a biased region is that as in the initial cumulative steps, we have only a few features that cannot explain the complex relationship that exists between the features and the condition. A pattern can be observed in all the Table 2, Table 3, Table 4 and Table 5, whereas the number of selected features are added in the classifier, the 10-fold accuracy of the classifiers starts to increase.

5.1.2. Over Fitting

With reference to Figure 2, we have labelled another area corresponding to the low accuracy region as “variance”. The reason we have called it a variance region is that after a large number of features are added for the classifier, the relationship between the features and the condition becomes too much complex and it could not well generalize on the new test data. A pattern can be observed in all the Table 6, Table 7, Table 8 and Table 9, whereas the number of selected features are increased in number, the 10-fold accuracy of the classifiers starts to decrease.

5.1.3. Selected Features

We have selected the features based on the highest accuracy value as presented in the Figure 2 “Selected Features” marked area. Details of the cumulative feature step and the number of features corresponding to each of the classifiers are described in Table 10. For specific details of the selected indices on the three highest performing classifiers, Support Vector Machine, Logistic Classifier and Linear Discriminant Classifier, please see Tables S6–S8 in the Supplementary Materials section.

5.2. Combined Dataset Accuracy Using 10-Fold Cross Validation

A comparison of our Stage feature selection-based classifiers with the state of the art methods is presented in Table 11. Where first the correlation-based connectivity matrix is arranged on all the sites, ignoring the sites intrinsic variability, then a 10-fold cross-validation-based accuracy, sensitivity and specificity is given in the mentioned table. Our 3-Stage-based feature selection results outperformed state of the art results. Specifically, Logistic classifier accuracy and sensitivity is the highest when compared to the rest of classifiers, similarly, the highest specificity is observed in the SVM classifier. Hence, based on all the presented results, Improved performance based on all the three accuracy metrics is observed using our proposed feature selection approach which proves the effectiveness and robustness of our approach.

5.3. Site Wise Accuracy Using 5-Fold Cross Validation

In this experiment, we have demonstrated the discriminating power of selected features on the individual sites using the 5-fold accuracy metric. We compared our site-wise accuracy results with the state of the art works in the literature. Our experiment validates the robustness of the selected features that we selected using the 3-Stage baed approach on the combined dataset experiment as mentioned in the previous section. We have presented the 5-fold accuracy, sensitivity and specificity of various classifiers that are shown in Table 12. A reason for choosing the 5-fold accuracy metric is consistent with the results shown by previous works on the same dataset. Our 3-Stage Features-based classifiers have shown improved accuracy in 13 of the 17 sites, specifically as with the combined sites data results, here again, logistic classifier has outperformed in the 5-fold accuracy comparison with the state of the art methods in 13 out of the 17 sites.

6. Discussion

In this section, we will explain the usefulness and importance of the learned discriminant features that we obtained using the Teacher-Student neural network-based feature selection approach. It is pertinent to mention here that our features are not specific to autistic or healthy subjects. In fact, these features are related to both the autistic and healthy control subjects, hence they are referred to as altered or discriminating features related to both the condition. The selected discriminating features are the subset of the total 4005 pair-wise connected regions, hence any feature is the pair in the lookup of 4005 which underscores that for each feature we get the connected region from the 90 × 90 connectivity matrix.

6.1. Connectogram for the Brain Region Network

We have chosen a set of 154 features as these features are common among the three highest performing classifiers that are in, Support Vector Machine, Logistic Classifier and the Linear Discriminant classifier. Each of the 4005 sets of pair-wise connected regions correspond to two regions in the original 90 × 90 connectivity matrix, therefore, for each feature we have set a value of 1 where the regions are present and 0 elsewhere to display these discriminating features in the two connectograms corresponding to interlobe and intralobe connections in Figure 3.

6.1.1. Connectivity in the Intralobe Network

Underconnectivity in the intralobe connections is visible as shown in the Figure 3 lobe features connectogram with most affected areas correspond to Temporal, Medial-Temporal and Subcorital lobes. Medial Temporal lobe has many key functionalities in the human life cycle, most importantly this is linked to the functionality of a person cognitive skills and attention that is required to perform a task [57]. Altered connectivity pattern in this part of the brain lobe and other lobes as mentioned above have also been found to be affecting the person who has autism as shown in the works of [58,59,60,61].

6.1.2. Connectivity in the Interlobe Network

Altered connectivity patterns are visible in the brain Frontal and Parietal lobes in the interlobe connection network as shown in Figure 3 intralobe section on the selected features. The frontal lobe is the largest part of the brain that consists of two-third of the brain region has a variety of functionalities related to person’s mood, personality, cognition and social behaviour [62], whereas, the parietal lobe of the brain processes sensory information from the body and processing of arithmetic information, which are the key functions that are performed in this part of the brain [63].
Due to denseness of the brain interlobe connectogram, we have shown the brain lobe wise count statistics in Figure 4 where altered connectivity counts are displayed lobe wise in the interconnection network data of the lobes. These altered connectivity pattern in the mentioned lobes are also consistent with the previous works [64,65,66,67].

6.2. Alterations in Brain’s Hemisphere Connectivity Patterns

The brain is perceived as a pair of two hemispheres consisting of a left hemisphere and a right hemisphere that form two interconnected units of the brain which are thought to have a key role in the functioning of the normal brain [68]. Set of node connections based on brain parcellation template with the inter and intra connections between the two hemispheres reveal patterns that are of key importance in many brain disorders such as ADHD [69], Alzheimer [70], Parkinson [71] and Schizophrenia [72] disorders. ASD has also proved to be exhibiting the altered connectivity patter in the brain two hemispheres as demonstrated in the works of [73,74]. To prove the altered connectivity patterns in the autistic and healthy control subjects we have used the famous tool developed by [75]. For this purpose we have selected high performing classification features that showed better results in the three mentioned classifiers as described int the feature selection section. BrainNet-viewer tool requires the set of nodes and the connected edges between those nodes, where as in our case we have set of discrimination features. For this purpose first we converted those features to the node and edge set using the 4005 lookup vector, where the node pair corresponds to feature index and for edge we set the value to 1.We have shown brain areas connectivity regions from three sides, that is left, middle and right are shown in Figure 5 using the BrainNet-viewer tool. It can be observed that the inter connectivity is higher in the brain hemisphere as opposed to the number of connection in each of the hemispheres. For better visualization using the Brain net viewer, we have changed the four critical parameters of the tool, that is the colour of each node was made thicker, node size was set to the fixed unit, weight corresponds to each of the edges was set to a fixed value and lastly, the opacity value of the surface was made lighter so that nodes and edges got more visible. Although all the nodes correspond to the selected features can be used to display on the BrainNet-viewer, but this will lead to clutter and a lack of clarity. To avoid this problem, only the top 50 features are used to select the nodes and connectivity between those nodes so that a better visualization could be achieved.
For better visualization, we have presented the connection counts between each hemisphere and in a hemisphere in Figure 6. An altered connectivity count is evident between the brain hemisphere, which is consistent with the work found in [76].

7. Conclusions

In this study, we have proposed a Teacher Student neural network-based feature selection approach to get the most discriminant features on the ABIDE preprocessed dataset. We have demonstrated the usefulness and importance of our features using various classifiers and compared our results with the state of the art methods at overall and site-wise level. We also have discussed the significance of our features using the brain anatomical analysis at the interlobe, intralobe and hemispherical level. We believe that based on our highly accurate results and detailed discussion, our methodology can be deployed in the clinical domain after discussion with the relevant experts and after checking the robustness of our methodology on more datasets on the autism disorder. In future studies, we will experiment to make this feature selection a part of an End-To-End deep learning model so that selection of features and classifiers are both packaged in one unified framework.

Supplementary Materials

The following are available at https://www.mdpi.com/2076-3425/10/10/754/s1. Table S1: Teacher Student Neural Network, Table S2: Teacher Neural Network Parameter Settings, Table S3: Student Neural Network, Table S4: Student Neural Network Parameter Settings, Table S5: Automated Anatomical Label (AAL) regions, Table S6: Support Vector Machine Features List indices in the 4005 lookup, Table S7: Logistic Classifier Features List indices in the 4005 lookup, Table S8: Linear Discriminant Features List indices in the 4005 lookup.

Author Contributions

N.A.K. conceived the idea of ASD classification problem, proposed the Teacher-Student neural network approach for classification and started experiments. S.A.W. helped in experiments and writing of the manuscript. A.R. streamlined the thought process and suggested valuable edits in the manuscript and experiments. X.S. checked the overall progress of the manuscript writing and experimentation and helped in organizing various sections of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 61772426).

Acknowledgments

We are extremely thankful to China Scholarship Council (CSC) and National Natural Science Foundation of China for giving us both the administrative and financial support to complete this study.

Conflicts of Interest

There are no conflict of interests.

Abbreviations

The following abbreviations are used in this manuscript:
ABIDEAutism Brain Imaging Data Exchange
ASDAutism Spectrum Disorder
DTDecision Trees
LDLinear Discriminant
rs-fMRIResting State Functional Magnetic Resonance Imaging
RFRandom Forests
SVMSupport Vector Machine

References

  1. Bassett, D.S.; Gazzaniga, M.S. Understanding complexity in the human brain. Trends Cogn. Sci. 2011, 15, 200–209. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Liao, X.; Vasilakos, A.V.; He, Y. Small-world human brain networks: Perspectives and challenges. Neurosci. Biobehav. Rev. 2017, 77, 286–300. [Google Scholar] [CrossRef] [PubMed]
  3. Van den Heuvel, M.P.; Sporns, O. Network hubs in the human brain. Trends Cogn. Sci. 2013, 17, 683–696. [Google Scholar] [CrossRef] [PubMed]
  4. Meunier, D.; Lambiotte, R.; Fornito, A.; Ersche, K.; Bullmore, E.T. Hierarchical modularity in human brain functional networks. Front. Neuroinform. 2009, 3, 37. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Tognoli, E.; Kelso, J. Enlarging the scope: Grasping brain complexity. Front. Syst. Neurosci. 2014, 8, 122. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Lord, C.; Elsabbagh, M.; Baird, G.; Veenstra-Vanderweele, J. Autism spectrum disorder. Lancet 2018, 392, 508–520. [Google Scholar] [CrossRef]
  7. Kogan, M.D.; Blumberg, S.J.; Schieve, L.A.; Boyle, C.A.; Perrin, J.M.; Ghandour, R.M.; Singh, G.K.; Strickland, B.B.; Trevathan, E.; van Dyck, P.C. Prevalence of parent-reported diagnosis of autism spectrum disorder among children in the US, 2007. Pediatrics 2009, 124, 1395–1403. [Google Scholar] [CrossRef] [Green Version]
  8. Werling, D.M.; Geschwind, D.H. Sex differences in autism spectrum disorders. Curr. Opin. Neurol. 2013, 26, 146. [Google Scholar] [CrossRef] [Green Version]
  9. Beggiato, A.; Peyre, H.; Maruani, A.; Scheid, I.; Rastam, M.; Amsellem, F.; Gillberg, C.I.; Leboyer, M.; Bourgeron, T.; Gillberg, C.; et al. Gender differences in autism spectrum disorders: Divergence among specific core symptoms. Autism Res. 2017, 10, 680–689. [Google Scholar] [CrossRef]
  10. Loomes, R.; Hull, L.; Mandy, W.P.L. What is the male-to-female ratio in autism spectrum disorder? A systematic review and meta-analysis. J. Am. Acad. Child Adolesc. Psychiatry 2017, 56, 466–474. [Google Scholar] [CrossRef]
  11. Brugha, T.S.; McManus, S.; Bankart, J.; Scott, F.; Purdon, S.; Smith, J.; Bebbington, P.; Jenkins, R.; Meltzer, H. Epidemiology of autism spectrum disorders in adults in the community in England. Arch. Gen. Psychiatry 2011, 68, 459–465. [Google Scholar] [CrossRef] [Green Version]
  12. Ingalhalikar, M.; Kanterakis, S.; Gur, R.; Roberts, T.P.; Verma, R. DTI based diagnostic prediction of a disease via pattern classification. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin, Germany, 2010; pp. 558–565. [Google Scholar]
  13. Yasuhara, A. Correlation between EEG abnormalities and symptoms of autism spectrum disorder (ASD). Brain Dev. 2010, 32, 791–798. [Google Scholar] [CrossRef]
  14. Kleinhans, N.M.; Richards, T.; Johnson, L.C.; Weaver, K.E.; Greenson, J.; Dawson, G.; Aylward, E. fMRI evidence of neural abnormalities in the subcortical face processing system in ASD. Neuroimage 2011, 54, 697–704. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Tsiaras, V.; Simos, P.G.; Rezaie, R.; Sheth, B.R.; Garyfallidis, E.; Castillo, E.M.; Papanicolaou, A.C. Extracting biomarkers of autism from MEG resting-state functional connectivity networks. Comput. Biol. Med. 2011, 41, 1166–1177. [Google Scholar] [CrossRef] [PubMed]
  16. Zürcher, N.R.; Bhanot, A.; McDougle, C.J.; Hooker, J.M. A systematic review of molecular imaging (PET and SPECT) in autism spectrum disorder: Current state and future research opportunities. Neurosci. Biobehav. Rev. 2015, 52, 56–73. [Google Scholar] [CrossRef] [PubMed]
  17. Glover, G.H. Overview of functional magnetic resonance imaging. Neurosurg. Clin. 2011, 22, 133–139. [Google Scholar] [CrossRef] [Green Version]
  18. Logothetis, N.K. What we can do and what we cannot do with fMRI. Nature 2008, 453, 869–878. [Google Scholar] [CrossRef]
  19. Heeger, D.J.; Ress, D. What does fMRI tell us about neuronal activity? Nat. Rev. Neurosci. 2002, 3, 142–151. [Google Scholar] [CrossRef]
  20. Tomasi, D.; Volkow, N.D. Abnormal functional connectivity in children with attention-deficit/hyperactivity disorder. Biol. Psychiatry 2012, 71, 443–450. [Google Scholar] [CrossRef] [Green Version]
  21. Sheline, Y.I.; Raichle, M.E. Resting state functional connectivity in preclinical Alzheimer’s disease. Biol. Psychiatry 2013, 74, 340–347. [Google Scholar] [CrossRef] [Green Version]
  22. Monk, C.S.; Peltier, S.J.; Wiggins, J.L.; Weng, S.J.; Carrasco, M.; Risi, S.; Lord, C. Abnormalities of intrinsic functional connectivity in autism spectrum disorders. Neuroimage 2009, 47, 764–772. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Waites, A.B.; Briellmann, R.S.; Saling, M.M.; Abbott, D.F.; Jackson, G.D. Functional connectivity networks are disrupted in left temporal lobe epilepsy. Ann. Neurol. 2006, 59, 335–343. [Google Scholar] [CrossRef] [PubMed]
  24. Lynall, M.E.; Bassett, D.S.; Kerwin, R.; McKenna, P.J.; Kitzbichler, M.; Muller, U.; Bullmore, E. Functional connectivity and brain networks in schizophrenia. J. Neurosci. 2010, 30, 9477–9487. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Sartipi, S.; Shayesteh, M.G.; Kalbkhani, H. Diagnosing of autism spectrum disorder based on GARCH variance series for rs-fMRI data. In Proceedings of the 2018 9th International Symposium on Telecommunications (IST), Geneva, Switzerland, 10–12 December 2018; pp. 86–90. [Google Scholar]
  26. Zhang, L.; Wang, X.H.; Li, L. Diagnosing autism spectrum disorder using brain entropy: A fast entropy method. Comput. Methods Programs Biomed. 2020, 190, 105240. [Google Scholar] [CrossRef]
  27. Xu, L.; Guo, Y.; Li, J.; Yu, J.; Xu, H. Classification of autism spectrum disorder based on fluctuation entropy of spontaneous hemodynamic fluctuations. Biomed. Signal Process. Control. 2020, 60, 101958. [Google Scholar] [CrossRef]
  28. Bernas, A.; Aldenkamp, A.P.; Zinger, S. Wavelet coherence-based classifier: A resting-state functional MRI study on neurodynamics in adolescents with high-functioning autism. Comput. Methods Programs Biomed. 2018, 154, 143–151. [Google Scholar] [CrossRef]
  29. Subbaraju, V.; Suresh, M.B.; Sundaram, S.; Narasimhan, S. Identifying differences in brain activities and an accurate detection of autism spectrum disorder using resting state functional-magnetic resonance imaging: A spatial filtering approach. Med. Image Anal. 2017, 35, 375–389. [Google Scholar] [CrossRef] [PubMed]
  30. Assaf, M.; Jagannathan, K.; Calhoun, V.D.; Miller, L.; Stevens, M.C.; Sahl, R.; O’Boyle, J.G.; Schultz, R.T.; Pearlson, G.D. Abnormal functional connectivity of default mode sub-networks in autism spectrum disorder patients. Neuroimage 2010, 53, 247–256. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Jones, T.B.; Bandettini, P.A.; Kenworthy, L.; Case, L.K.; Milleville, S.C.; Martin, A.; Birn, R.M. Sources of group differences in functional connectivity: An investigation applied to autism spectrum disorder. Neuroimage 2010, 49, 401–414. [Google Scholar] [CrossRef] [Green Version]
  32. Tyszka, J.M.; Kennedy, D.P.; Paul, L.K.; Adolphs, R. Largely typical patterns of resting-state functional connectivity in high-functioning adults with autism. Cereb. Cortex 2014, 24, 1894–1905. [Google Scholar] [CrossRef] [Green Version]
  33. Iidaka, T. Resting state functional magnetic resonance imaging and neural network classified autism and control. Cortex 2015, 63, 55–67. [Google Scholar] [CrossRef] [PubMed]
  34. Plitt, M.; Barnes, K.A.; Wallace, G.L.; Kenworthy, L.; Martin, A. Resting-state functional connectivity predicts longitudinal change in autistic traits and adaptive functioning in autism. Proc. Natl. Acad. Sci. USA 2015, 112, E6699–E6706. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Yao, Z.; Hu, B.; Xie, Y.; Zheng, F.; Liu, G.; Chen, X.; Zheng, W. Resting-state time-varying analysis reveals aberrant variations of functional connectivity in autism. Front. Hum. Neurosci. 2016, 10, 463. [Google Scholar] [CrossRef]
  36. Abraham, A.; Milham, M.P.; Di Martino, A.; Craddock, R.C.; Samaras, D.; Thirion, B.; Varoquaux, G. Deriving reproducible biomarkers from multi-site resting-state data: An Autism-based example. NeuroImage 2017, 147, 736–745. [Google Scholar] [CrossRef] [Green Version]
  37. Duff, E.P.; Makin, T.; Cottaar, M.; Smith, S.M.; Woolrich, M.W. Disambiguating brain functional connectivity. Neuroimage 2018, 173, 540–550. [Google Scholar] [CrossRef] [PubMed]
  38. Wang, C.; Xiao, Z.; Wu, J. Functional connectivity-based classification of autism and control using SVM-RFECV on rs-fMRI data. Phys. Medica 2019, 65, 99–105. [Google Scholar] [CrossRef] [PubMed]
  39. Mostafa, S.; Tang, L.; Wu, F.X. Diagnosis of autism spectrum disorder based on eigenvalues of brain networks. IEEE Access 2019, 7, 128474–128486. [Google Scholar] [CrossRef]
  40. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  41. Dvornek, N.C.; Ventola, P.; Pelphrey, K.A.; Duncan, J.S. Identifying autism from resting-state fMRI using long short-term memory networks. In International Workshop on Machine Learning in Medical Imaging; Springer: Berlin, Germany, 2017; pp. 362–370. [Google Scholar]
  42. Guo, X.; Dominick, K.C.; Minai, A.A.; Li, H.; Erickson, C.A.; Lu, L.J. Diagnosing autism spectrum disorder from brain resting-state functional connectivity patterns using a deep neural network with a novel feature selection method. Front. Neurosci. 2017, 11, 460. [Google Scholar] [CrossRef]
  43. Dvornek, N.C.; Ventola, P.; Duncan, J.S. Combining phenotypic and resting-state fMRI data for autism classification with recurrent neural networks. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 725–728. [Google Scholar]
  44. Heinsfeld, A.S.; Franco, A.R.; Craddock, R.C.; Buchweitz, A.; Meneguzzi, F. Identification of autism spectrum disorder using deep learning and the ABIDE dataset. Neuroimage Clin. 2018, 17, 16–23. [Google Scholar] [CrossRef]
  45. Aghdam, M.A.; Sharifi, A.; Pedram, M.M. Combination of rs-fMRI and sMRI data to discriminate autism spectrum disorders in young children using deep belief network. J. Digit. Imaging 2018, 31, 895–903. [Google Scholar] [CrossRef]
  46. Li, X.; Dvornek, N.C.; Papademetris, X.; Zhuang, J.; Staib, L.H.; Ventola, P.; Duncan, J.S. 2-channel convolutional 3D deep neural network (2CC3D) for fMRI analysis: ASD classification and feature learning. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 1252–1255. [Google Scholar]
  47. Li, X.; Dvornek, N.C.; Zhuang, J.; Ventola, P.; Duncan, J.S. Brain biomarker interpretation in asd using deep learning and fmri. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin, Germany, 2018; pp. 206–214. [Google Scholar]
  48. Eslami, T.; Saeed, F. Auto-ASD-network: A technique based on deep learning and support vector machines for diagnosing autism spectrum disorder using fMRI data. In Proceedings of the 10th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics, Niagara Falls, NY, USA, 7–10 September 2019; pp. 646–651. [Google Scholar]
  49. Saeed, F.; Eslami, T.; Mirjalili, V.; Fong, A.; Laird, A. ASD-DiagNet: A hybrid learning approach for detection of Autism Spectrum Disorder using fMRI data. Front. Neuroinformatics 2019, 13, 70. [Google Scholar]
  50. Niu, K.; Guo, J.; Pan, Y.; Gao, X.; Peng, X.; Li, N.; Li, H. Multichannel deep attention neural networks for the classification of autism spectrum disorder using neuroimaging and personal characteristic data. Complexity 2020, 2020. [Google Scholar] [CrossRef]
  51. Sherkatghanad, Z.; Akhondzadeh, M.; Salari, S.; Zomorodi-Moghadam, M.; Abdar, M.; Acharya, U.R.; Khosrowabadi, R.; Salari, V. Automated detection of autism spectrum disorder using a convolutional neural network. Front. Neurosci. 2019, 13. [Google Scholar] [CrossRef] [Green Version]
  52. Craddock, C.; Benhajali, Y.; Chu, C.; Chouinard, F.; Evans, A.; Jakab, A.; Khundrakpam, B.S.; Lewis, J.D.; Li, Q.; Milham, M.; et al. The neuro bureau preprocessing initiative: Open sharing of preprocessed neuroimaging data and derivatives. Neuroinform. 2013, 4. [Google Scholar]
  53. Lord, C.; Risi, S.; Lambrecht, L.; Cook, E.H.; Leventhal, B.L.; DiLavore, P.C.; Pickles, A.; Rutter, M. The Autism Diagnostic Observation Schedule—Generic: A standard measure of social and communication deficits associated with the spectrum of autism. J. Autism Dev. Disord. 2000, 30, 205–223. [Google Scholar] [CrossRef] [PubMed]
  54. Craddock, C.; Sikka, S.; Cheung, B.; Khanuja, R.; Ghosh, S.S.; Yan, C.; Li, Q.; Lurie, D.; Vogelstein, J.; Burns, R.; et al. Towards automated analysis of connectomes: The configurable pipeline for the analysis of connectomes (c-pac). Front. Neuroinform. 2013, 42. [Google Scholar]
  55. Mirzaei, A.; Pourahmadi, V.; Soltani, M.; Sheikhzadeh, H. Deep feature selection using a teacher-student network. Neurocomputing 2020, 383, 396–408. [Google Scholar] [CrossRef] [Green Version]
  56. Riaz, A.; Asad, M.; Alonso, E.; Slabaugh, G. DeepFMRI: End-to-end deep learning for functional connectivity and classification of ADHD using fMRI. J. Neurosci. Methods 2020, 335, 108506. [Google Scholar] [CrossRef] [PubMed]
  57. Fogassi, L.; Luppino, G. Motor functions of the parietal lobe. Curr. Opin. Neurobiol. 2005, 15, 626–631. [Google Scholar] [CrossRef] [PubMed]
  58. Welchew, D.E.; Ashwin, C.; Berkouk, K.; Salvador, R.; Suckling, J.; Baron-Cohen, S.; Bullmore, E. Functional disconnectivity of the medial temporal lobe in Asperger’s syndrome. Biol. Psychiatry 2005, 57, 991–998. [Google Scholar] [CrossRef]
  59. Nair, A.; Treiber, J.M.; Shukla, D.K.; Shih, P.; Müller, R.A. Impaired thalamocortical connectivity in autism spectrum disorder: A study of functional and anatomical connectivity. Brain 2013, 136, 1942–1955. [Google Scholar] [CrossRef]
  60. Ye, A.X.; Leung, R.C.; Schäfer, C.B.; Taylor, M.J.; Doesburg, S.M. Atypical resting synchrony in autism spectrum disorder. Hum. Brain Mapp. 2014, 35, 6049–6066. [Google Scholar] [CrossRef]
  61. Ha, S.; Sohn, I.J.; Kim, N.; Sim, H.J.; Cheon, K.A. Characteristics of brains in autism spectrum disorder: Structure, function and connectivity across the lifespan. Exp. Neurobiol. 2015, 24, 273–284. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  62. Chayer, C.; Freedman, M. Frontal lobe functions. Curr. Neurol. Neurosci. Rep. 2001, 1, 547–552. [Google Scholar] [CrossRef] [PubMed]
  63. Squire, L.R.; Stark, C.E.; Clark, R.E. The medial temporal lobe. Annu. Rev. Neurosci. 2004, 27, 279–306. [Google Scholar] [CrossRef] [Green Version]
  64. Turner, K.C.; Frost, L.; Linsenbardt, D.; McIlroy, J.R.; Müller, R.A. Atypically diffuse functional connectivity between caudate nuclei and cerebral cortex in autism. Behav. Brain Funct. 2006, 2, 34. [Google Scholar] [CrossRef] [Green Version]
  65. Kleinhans, N.M.; Richards, T.; Sterling, L.; Stegbauer, K.C.; Mahurin, R.; Johnson, L.C.; Greenson, J.; Dawson, G.; Aylward, E. Abnormal functional connectivity in autism spectrum disorders during face processing. Brain 2008, 131, 1000–1012. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Wass, S. Distortions and disconnections: Disrupted brain connectivity in autism. Brain Cogn. 2011, 75, 18–28. [Google Scholar] [CrossRef]
  67. Khan, A.J.; Nair, A.; Keown, C.L.; Datko, M.C.; Lincoln, A.J.; Müller, R.A. Cerebro-cerebellar resting-state functional connectivity in children and adolescents with autism spectrum disorder. Biol. Psychiatry 2015, 78, 625–634. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  68. Martínez, J.H.; Buldú, J.M.; Papo, D.; Fallani, F.D.V.; Chavez, M. Role of inter-hemispheric connections in functional brain networks. Sci. Rep. 2018, 8, 1–10. [Google Scholar] [CrossRef] [Green Version]
  69. Hale, T.S.; Loo, S.K.; Zaidel, E.; Hanada, G.; Macion, J.; Smalley, S.L. Rethinking a right hemisphere deficit in ADHD. J. Atten. Disord. 2009, 13, 3–17. [Google Scholar] [CrossRef] [Green Version]
  70. Yang, C.; Zhong, S.; Zhou, X.; Wei, L.; Wang, L.; Nie, S. The abnormality of topological asymmetry between hemispheric brain white matter networks in Alzheimer’s disease and mild cognitive impairment. Front. Aging Neurosci. 2017, 9, 261. [Google Scholar] [CrossRef] [Green Version]
  71. Li, P.; Ensink, E.; Lang, S.; Marshall, L.; Schilthuis, M.; Lamp, J.; Vega, I.; Labrie, V. Hemispheric asymmetry in the human brain and in Parkinson’s disease is linked to divergent epigenetic patterns in neurons. Genome Biol. 2020, 21, 1–23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  72. Angrilli, A.; Spironelli, C.; Elbert, T.; Crow, T.J.; Marano, G.; Stegagno, L. Schizophrenia as failure of left hemispheric dominance for the phonological component of language. PLoS ONE 2009, 4, e4507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Schaer, M.; Ottet, M.C.; Scariati, E.; Dukes, D.; Franchini, M.; Eliez, S.; Glaser, B. Decreased frontal gyrification correlates with altered connectivity in children with autism. Front. Hum. Neurosci. 2013, 7, 750. [Google Scholar] [CrossRef] [Green Version]
  74. Rausch, A.; Zhang, W.; Haak, K.V.; Mennes, M.; Hermans, E.J.; van Oort, E.; van Wingen, G.; Beckmann, C.F.; Buitelaar, J.K.; Groen, W.B. Altered functional connectivity of the amygdaloid input nuclei in adolescents and young adults with autism spectrum disorder: A resting state fMRI study. Mol. Autism 2016, 7, 13. [Google Scholar] [CrossRef] [Green Version]
  75. Xia, M.; Wang, J.; He, Y. BrainNet Viewer: A network visualization tool for human brain connectomics. PLOS ONE 2013, 8, e68910. [Google Scholar] [CrossRef] [Green Version]
  76. Hahamy, A.; Behrmann, M.; Malach, R. The idiosyncratic brain: Distortion of spontaneous connectivity patterns in autism spectrum disorder. Nat. Neurosci. 2015, 18, 302. [Google Scholar] [CrossRef] [PubMed]
Figure 1. 3 Stage Teacher-Student-based and SFFS-based Feature Selection Architecture.
Figure 1. 3 Stage Teacher-Student-based and SFFS-based Feature Selection Architecture.
Brainsci 10 00754 g001
Figure 2. Comparison of Cumulative Feature Selection on the 10-Fold Accuracy.
Figure 2. Comparison of Cumulative Feature Selection on the 10-Fold Accuracy.
Brainsci 10 00754 g002
Figure 3. Brain’s lobes connectogram at the interlobe and intralobe level.
Figure 3. Brain’s lobes connectogram at the interlobe and intralobe level.
Brainsci 10 00754 g003
Figure 4. Brain’s Lobe wise connection statistics comparison level.
Figure 4. Brain’s Lobe wise connection statistics comparison level.
Brainsci 10 00754 g004
Figure 5. Brain node Connectivity network in Left, Middle and Right areas Using BrainNet-viewer tool.
Figure 5. Brain node Connectivity network in Left, Middle and Right areas Using BrainNet-viewer tool.
Brainsci 10 00754 g005
Figure 6. Brain’s nodes connection network counts between inter hemisphere and Intra hemisphere.
Figure 6. Brain’s nodes connection network counts between inter hemisphere and Intra hemisphere.
Brainsci 10 00754 g006
Table 1. Autism Brain Imaging Data Exchange (ABIDE) Preprocessed Data from all sites.
Table 1. Autism Brain Imaging Data Exchange (ABIDE) Preprocessed Data from all sites.
Participants
Sr No.Site NameAutistic SubjectsHealthy Controls
1Caltech1918
2CMU1413
3KKI2028
4Leuven2934
5MaxMun2428
6NYU75100
7OHSU1214
8OLIN1915
9PITT2927
10SBL1515
11SDSU1422
12Stanford1920
13Trinity2225
14UCLA5444
15UM6674
16USM4625
17Yale2828
Total 505530
1035
Table 2. 10-Fold Results based on the Top 10 Selected Features based on Proposed 3-Stage Approach.
Table 2. 10-Fold Results based on the Top 10 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.510.560.6
2SVM0.450.40.36
3LD0.490.510.54
4RF0.540.480.51
5DT0.310.340.29
Table 3. 10-Fold Results based on the Top 30 Selected Features based on Proposed 3-Stage Approach.
Table 3. 10-Fold Results based on the Top 30 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.620.590.63
2SVM0.530.590.62
3LD0.510.580.6
4RF0.560.490.48
5DT0.50.490.5
Table 4. 10-Fold Results based on the Top 60 Selected Features based on Proposed 3-Stage Approach.
Table 4. 10-Fold Results based on the Top 60 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.680.630.58
2SVM0.650.680.7
3LD0.70.610.62
4RF0.610.630.58
5DT0.590.510.6
Table 5. 10-Fold Results based on the Top 90 Selected Features based on Proposed 3-Stage Approach.
Table 5. 10-Fold Results based on the Top 90 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.730.580.65
2SVM0.750.650.68
3LD0.730.710.69
4RF0.690.530.46
5DT0.630.510.8
Table 6. 10-Fold Results based on the Top 500 Selected Features based on Proposed 3-Stage Approach.
Table 6. 10-Fold Results based on the Top 500 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.780.690.7
2SVM0.790.750.74
3LD0.740.70.67
4RF0.650.510.59
5DT0.620.580.61
Table 7. 10-Fold Results based on the Top 1000 Selected Features based on Proposed 3-Stage Approach.
Table 7. 10-Fold Results based on the Top 1000 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.730.650.74
2SVM0.740.710.72
3LD0.70.690.68
4RF0.610.60.63
5DT0.60.610.5
Table 8. 10-Fold Results based on the Top 1500 Selected Features based on Proposed 3-Stage Approach.
Table 8. 10-Fold Results based on the Top 1500 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.60.610.65
2SVM0.70.690.73
3LD0.620.640.61
4RF0.50.560.59
5DT0.360.390.4
Table 9. 10-Fold Results based on the Top 2000 Selected Features based on Proposed 3-Stage Approach.
Table 9. 10-Fold Results based on the Top 2000 Selected Features based on Proposed 3-Stage Approach.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR0.510.550.52
2SVM0.510.550.52
3LD0.50.480.51
4RF0.310.290.33
5DT0.330.320.3
Table 10. Classifiers’ cumulative feature selection step and length.
Table 10. Classifiers’ cumulative feature selection step and length.
Sr No.Classifier NameMax Cumulative Feature StepFeatures Count
1LR5256
2RF2103
3DT152
4SVM3154
5LD3154
Table 11. 10-Fold Results Comparison of our 3-Stage approach-based selected Features on an all-Site combined dataset.
Table 11. 10-Fold Results Comparison of our 3-Stage approach-based selected Features on an all-Site combined dataset.
Sr No.Classifier NameAccuracySensitivitySpecificity
1LR (Ours)0.820.830.84
2LD (Ours)0.820.830.88
3SVM (Ours)0.810.800.92
4RF (Ours)0.700.670.80
5DT (Ours)0.570.600.64
6Heinsfeld et al., 2018 [44]0.630.580.67
7Taban Eslami et al., 2019 [48]0.670.630.71
8Ke Niu et al., 2020 [50]0.730.740.71
9Zeinab et al., 2020 [51]0.700.770.61
Table 12. Proposed 3-Stage selected features’ 5-Fold Accuracy Comparison on each individual Site dataset.
Table 12. Proposed 3-Stage selected features’ 5-Fold Accuracy Comparison on each individual Site dataset.
SrSiteSVM (Ours)RF (Ours)DT (Ours)LR (Ours)LD (Ours)Heinsfeld et al., 2018 [44]Taban Eslami et al., 2019 [48]Ke Niu et al., 2020 [50]Zeinab et al., 2020 (5 fold) [51]
1Caltech0.830.650.530.780.670.520.520.660.54
2CMU0.840.70.60.710.60.450.680.630.7
3KKI0.620.60.580.660.60.580.69 0.72
4Leuven0.660.550.630.630.760.510.610.620.65
5MaxMun0.590.490.480.610.470.540.48 0.46
6NYU0.780.690.60.780.640.640.680.70.65
7OHSU0.50.50.60.740.690.740.82 0.57
8Olin0.520.590.70.70.70.440.65 0.58
9Pitt0.750.690.510.780.720.590.670.690.69
10SBL0.660.630.560.660.590.460.51 0.56
11SDSU0.610.640.610.690.690.630.630.690.75
12Stanford0.690.690.660.710.580.480.640.610.48
13Trinity0.460.660.50.520.630.610.540.690.61
14UCLA0.690.690.460.770.660.570.730.750.69
15UM0.70.710.520.710.630.620.680.680.66
16USM0.710.760.760.80.740.570.630.80.77
17Yale0.750.60.490.80.610.530.630.690.69
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Khan, N.A.; Waheeb, S.A.; Riaz, A.; Shang, X. A Three-Stage Teacher, Student Neural Networks and Sequential Feed Forward Selection-Based Feature Selection Approach for the Classification of Autism Spectrum Disorder. Brain Sci. 2020, 10, 754. https://doi.org/10.3390/brainsci10100754

AMA Style

Khan NA, Waheeb SA, Riaz A, Shang X. A Three-Stage Teacher, Student Neural Networks and Sequential Feed Forward Selection-Based Feature Selection Approach for the Classification of Autism Spectrum Disorder. Brain Sciences. 2020; 10(10):754. https://doi.org/10.3390/brainsci10100754

Chicago/Turabian Style

Khan, Naseer Ahmed, Samer Abdulateef Waheeb, Atif Riaz, and Xuequn Shang. 2020. "A Three-Stage Teacher, Student Neural Networks and Sequential Feed Forward Selection-Based Feature Selection Approach for the Classification of Autism Spectrum Disorder" Brain Sciences 10, no. 10: 754. https://doi.org/10.3390/brainsci10100754

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop