Elsevier

Neural Networks

Volume 129, September 2020, Pages 7-18
Neural Networks

DLPNet: A deep manifold network for feature extraction of hyperspectral imagery

https://doi.org/10.1016/j.neunet.2020.05.022Get rights and content

Abstract

Deep learning has received increasing attention in recent years and it has been successfully applied for feature extraction (FE) of hyperspectral images. However, most deep learning methods fail to explore the manifold structure in hyperspectral image (HSI). To tackle this issue, a novel graph-based deep learning model, termed deep locality preserving neural network (DLPNet), was proposed in this paper. Traditional deep learning methods use random initialization to initialize network parameters. Different from that, DLPNet initializes each layer of the network by exploring the manifold structure in hyperspectral data. In the stage of network optimization, it designed a deep-manifold learning joint loss function to exploit graph embedding process while measuring the difference between the predictive value and the actual value, then the proposed model can take into account the extraction of deep features and explore the manifold structure of data simultaneously. Experimental results on real-world HSI datasets indicate that the proposed DLPNet performs significantly better than some state-of-the-art methods.

Introduction

As an effective imaging and spectroscopy technology, hyperspectral imaging technology has attracted much attention in the past decades (Du et al., 2018, Huang et al., 2019, Liang et al., 2018, Song et al., 2018, Xie et al., 2019, Zhang, Zhang, et al., 2019). Pixel-level hyperspectral image (HSI) classification utilizes rich spectral information contained in HSI, which is widely used in urban planning, precision agriculture, mineralogy, and geological research (Hong, Yokoya, Chanussot, and Zhu, 2019, Maxwell et al., 2018, Peng et al., 2019, Su et al., 2019, Xie et al., 2018). Hundreds of continuous spectral bands provide useful information for the fine classification of land covers. However, high-dimensional characteristics of spectral bands often lead to the curse-of-dimensionality, and the traditional classification methods will suffer from the Hughes phenomena due to the limited training samples (Ghamisi et al., 2017, Huang et al., 2019, Jiao et al., 2018, Kang et al., 2018, Luo et al., 2018, Qian et al., 2017).

A straightforward strategy to tackle the above problem is to reduce the dimensionality of hyperspectral data and map high-dimensional bands to a low-dimensional embedding space through feature extraction (FE), which can obtain the intrinsic features of high-dimensional data (Chen et al., 2019, Hong, Yokoya, Chanussot, Xu, and Zhu, 2019, Li, Li, and Zhang, 2019, Luo et al., 2020, Zhang, Gong, et al., 2019). FE methods can be classified into supervised, unsupervised and semi-supervised methods (Hong, Yokoya, Ge, Chanussot, & Zhu, 2019). Unsupervised FE methods obtain low-dimensional features of HSI without priori information, such methods include principal component analysis (PCA) (Tyo, Konsolakis, Diersen, & Olsen, 2003), independent component analysis (ICA) (Bayliss, Gualtieri, & Cromp, 1998), neighborhood preserving embedding (NPE) (Lu, Jin, & Zou, 2012), and locality preserving projections (LPP) (Li, Pan, He, & Liu, 2015). However, unsupervised methods fail to exploit label information to get discriminant features for classification. Supervised FE methods attempt to explore priori knowledge of training samples to compact the intraclass samples and separate the interclass samples, which brings benefits for classification. Linear discriminant analysis (LDA) (Du & Chang, 2001) and maximum margin criterion (MMC) (Datta, Ghosh, & Ghosh, 2014) are two classical supervised learning methods. To discover the intrinsic manifold structure in HSI, many supervised manifold learning algorithms are designed to explore the discriminative properties of data by using labeled samples. The marginal Fisher analysis (MFA) (Yan et al., 2007) method constructs an intrinsic graph and a penalty graph to reveal the discriminative manifold structure in hyperspectral data. The local geometric structure Fisher analysis (LGSFA) (Luo, Huang, Duan, Liu, & Liao, 2017) method was designed to explore the geometric properties in HSI, which compacts the intraclass neighbors with its reconstruction points while separates the interclass samples from its reconstruction points. Zhang, He, and Gao (2019) designed a manifold learning-based framework to utilize an explicit polynomial mapping to learn a low-dimensional embedding space. However, the above methods depend on shallow-based feature descriptors, which are sensitive to noises and difficult to extract discriminant features under a small size of training samples.

Deep learning (DL) has been regarded as one of the state-of-the-art machine learning tools, and it has shown the advantages in many fields, such as change detection, scene understanding, and image classification (Ghamisi et al., 2018, Li, Song, et al., 2019). In recent years, DL has been applied to hyperspectral imagery to extract deep abstract features through hierarchical networks, which can deal with the complex nonlinear relationships contained in HSI (Li et al., 2020, Liu et al., 2019, Mostafa, 2017, Zhou et al., 2019). Ratle, Camps-Valls, and Weston (2010) utilized artificial neural networks (ANN) to learn the deep features in hyperspectral data, and a regularizer was designed to improve the ability of feature extraction. Chen, Zhao, and Jia (2015) investigated a deep learning model based on deep belief network (DBN) to explore the intrinsic information in HSI data, and it achieved good performance through combining deep features with spectral-spatial features. Chen, Lin, Zhao, Wang, and Gu (2014) introduced the stacked autoencoder (SAE) to classify HSI data by utilizing the rich spectral bands and obtained high classification accuracy. Although deep learning methods can extract deep features of samples to deal with the nonlinear relationship in complex scenes, they ignore to reveal manifold structure in HSI.

Based on the above discussion, a novel graph-based deep learning model termed deep locality preserving neural network (DLPNet) was proposed in this paper. The main purpose of this method is to explore the manifold structure in HSI data in the process of extracting deep features, and then deep manifold features can be extracted for classification. DLPNet introduces graph embedding into neural network and it constructs the deep-manifold joint loss function to optimize the parameters of neural network. Based on the proposed joint loss function, the difference between the predictive value and the actual value of data is calculated, and an intrinsic graph and a penalty graph are constructed to explore the manifold structure in HSI data. DLPNet designs an iteration strategy to minimize the loss value, which characterizes the intraclass compactness by calculating the sum of the distance between each sample and its neighbors from the same class, and measures the interclass separability through the sum of the distance between each sample and its neighbors belonging to different classes. In the iteration process, it enhances the separability of different classes of samples and then extracts the discriminant features.

The main characteristics of the proposed method can be concluded as the following: (1) DLPNet combines graph embedding with deep learning to explore the manifold structure of HSI data; (2) An iteration strategy is designed to optimize the network adaptively by using manifold structure embedded in hyperspectral data; (3) By exploring the structure information of HSI, the dependence of deep learning on the labeled samples is reduced, and high classification accuracy can be achieved under a small number of labeled training samples.

The remainder of this paper is organized as follows. Section 2 briefly describes the related work. The details of the proposed method are introduced in Section 3. Section 4 presents experimental results and detailed analysis, and these demonstrate the effectiveness of the proposed DLPNet. Finally, Section 5 presents our concluding remarks and gives recommendations for future works.

Section snippets

Related works

Assuming that a HSI dataset X contains N samples with D bands, ith (i=1,2,,N) pixel can be defined as xiD. l(xi)={1,2,,c} denotes the class label of xi and total c classes are considered in data. The purpose of FE methods is to extract low-dimensional embedding features yid from xi, where d(dD) is the dimension of embedding features. For convenience, the symbols used in the paper are summarized in Table 1.

Deep locality preserving neural network

Traditional deep learning models mainly focus on exploring the nonlinear characteristics in data, while they ignore the intrinsic manifold structure in HSI. To address this issue, a deep locality preserving neural network (DLPNet) was proposed to extract the deep manifold features of hyperspectral data. DLPNet not only optimizes network parameters by measuring the differences between predictive value and actual value but also introduces the graph embedding framework to measure the differences

Data description

Two popular and widely used HSI datasets, PaviaU and Salinas, are used to investigate the performance of the DLPNet.

Discussion

In this section, we discuss the influence of training set size on the classification accuracy and analyze the parameter sensitivity.

Conclusion

In view of the fact that traditional deep learning methods fail to discover the manifold structure in hyperspectral imagery, this paper proposes a deep locality preserving neural network (DLPNet) for deep manifold features extraction of HSI data. In the stage of network initialization, DLPNet designed an initialization method based on manifold learning to initialize the network hierarchically. After that, a deep-manifold learning based joint loss function is constructed to combine the

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

The authors would like to thank the anonymous reviewers for their comments on this paper. This work was supported in part by the Basic and Frontier Research Programmes of Chongqing under Grant cstc2018jcyjAX0093, the Innovation Program for Chongqing Overseas Returnees under Grant cx2019144, the Scientific and Technological Research Project of Chongqing Education Commission under Grant KJZD-K201902501, and the Chongqing University Postgraduates Innovation Project under Grants CYS18035 and

References (44)

  • ZhangP. et al.

    A nonlinear and explicit framework of supervised manifold-feature extraction for hyperspectral image classification

    Neurocomputing

    (2019)
  • ZhangL.F. et al.

    Hyperspectral image unsupervised classification by robust manifold matrix factorization

    Information Sciences

    (2019)
  • Bayliss, J. D., Gualtieri, J. A., & Cromp, R. F. (1998). Analyzing hyperspectral data with independent component...
  • ChenZ.K. et al.

    Trilateral smooth filtering for hyperspectral image feature extraction

    IEEE Geoscience and Remote Sensing Letters

    (2019)
  • ChenY.S. et al.

    Deep learning-based classification of hyperspectral data

    IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing

    (2014)
  • ChenY.S. et al.

    Spectral–spatial classification of hyperspectral data based on deep belief network

    IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing

    (2015)
  • Datta, A., Ghosh, S., & Ghosh, A. (2014). Maximum margin criterion based band extraction of hyperspectral imagery. In...
  • DianR.W. et al.

    Deep hyperspectral image sharpening

    IEEE Transactions on Neural Networks and Learning Systems

    (2018)
  • DuP.J. et al.

    Multikernel adaptive collaborative representation for hyperspectral image classification

    IEEE Transactions on Geoscience and Remote Sensing

    (2018)
  • FoodyG.M.

    Thematic map comparison: Evaluating the statistical significance of differences in classification accuracy

    Photogrammetric Engineering and Remote Sensing

    (2004)
  • GaoX.H. et al.

    Hyperspectral imagery classification with deep metric learning

    Neurocomputing

    (2019)
  • GhamisiP. et al.

    New frontiers in spectral-spatial hyperspectral image classification: The latest advances based on mathematical morphology Markov random fields segmentation sparse representation and deep learning

    IEEE Geoscience and Remote Sensing Magazine

    (2018)
  • Cited by (18)

    • Rapid detection of the reducing sugar and amino acid nitrogen contents of Daqu based on hyperspectral imaging

      2021, Journal of Food Composition and Analysis
      Citation Excerpt :

      The main research objective of this paper is to realize the rapid and accurate detection of the reducing sugar and amino acid nitrogen contents of Daqu during the fermentation process using the feature learning ability of the cascade forest (CF) model to mine the inner connection between spectral information and substance content (Guo et al., 2018; Zhu et al., 2019). Compared with deep neural networks (Vidyarthi et al., 2020), CF models are simpler to create and more controllable, and their generalization performance and accuracy are also quite competitive (Al-Sarayreh et al., 2020; Guo et al., 2020; Li et al., 2020a,b). To highlight the excellent performance of the CF model, a backpropagation neural network (BPNN) model, a partial least squares regression (PLSR) model, and a support vector regression (SVR) model were established for comparative analysis.

    View all citing articles on Scopus
    View full text