Elsevier

Signal Processing

Volume 182, May 2021, 107931
Signal Processing

Joint estimation of low-rank components and connectivity graph in high-dimensional graph signals: Application to brain imaging

https://doi.org/10.1016/j.sigpro.2020.107931Get rights and content

Highlights

  • We formulate the problem of joint estimation of low-rank components and connectivity graph for distorted high-dimensional graph signal.

  • We propose two approaches to this joint estimation problem. One approach applies alternating optimization to solve the related subproblems. Another approach solves the joint estimation directly.

  • We provide analysis to understand the impact of inexact graph on the low- rank components estimation accuracy, and we apply the analysis result to inform the choice of regularization parameter when the graph is inexact.

  • We perform extensive experiments on synthetic and real data, especially brain imaging (MEG) data, to illustrate and validate our proposed algo- rithms.

Abstract

Given high-dimensional, graph-smooth and grossly-corrupted signals, this paper presents the algorithm to simultaneously estimate the intrinsic low-rank components and the underlying graph from high-dimensional noisy graph signal. We assume that the perturbation on low-rank components is sparse and signals are smooth on unknown underlying graph. The proposed algorithm learns the low-rank components by exploiting estimate of the graph, and refines the graph estimation with learned low-rank components. We propose two solutions to this problem: One applies alternating optimization to solve the subproblems. The other solves the problem directly. Furthermore, we analyze the impact of inexact graph on low-rank components estimation to justify our approach. We conduct extensive experiments on the proposed algorithm with synthetic data and real brain imaging data, Magnetoencephalography (MEG) and compare it with state-of-the-art methods. In particular, our algorithm is applied to estimate the low-rank components for classifying MEG signals evoked when a subject views a face or non-face image. We observe that our proposed algorithm is competitive in estimating low-rank components, adequately capturing intrinsic task-related information in lower-dimensional representation. This leads to improved classification performance. Furthermore, we notice that our estimated graph indicates brain active regions for the visual activity that are consistent with neuroscientific findings.

Introduction

We consider the problem of uncovering intrinsic low-rank components of high-dimensional data. This data has a large number of features per sample. We assume the intrinsic representations of the data lie on some low-dimensional subspace, referred to as low-rank components. Furthermore, we assume the data is defined on the vertices of some underlying graph and changes smoothly between the connected vertices, referred to as graph-smoothness [1], [2]. However, in many problems, the underlying graph is unknown or inexact [3], [4], [5] and needs to be estimated from the data. Therefore, in this work, we propose to apply graph-smoothness property as a bridge to combine graph learning and low-rank components estimation together. As will be discussed, our proposed algorithm estimates low-rank components and learns the underlying graph simultaneously.

High-dimensional data is common in many engineering areas such as image and video processing, biomedical imaging, computer networks, and transportation networks. As a specific application of our proposed algorithm, we are interested in automated analysis of brain imaging data. In many cases, the goal is to find the spatio-temporal neural signature of a task, by performing classification on cortical activation evoked by different stimuli [6], [7], [8], [9], [10], [11]. Common brain imaging techniques are Electroencephalography (EEG) and Magnetoencephalography (MEG). These measurements are high-dimensional spatio-temporal data. Furthermore, the measurements are degraded by various types of noise (e.g., sensor noise, ambient magnetic field noise, etc.) and some noise models are complicated (non-Gaussian). The high-dimensionality and noise limit both speed and accuracy of signal analysis, which may result in unreliable signature modeling for classification. The high-dimensionality of these signals also increases complexity of classifier. Combination of a complex classifier and availability of few data samples (due to time, cost, or study limitations) can easily lead to model overfitting. Thus, for a reliable study of brain imaging data, there is a need for a dimensionality reduction method that ensures inclusion of task-related information.

Note that it has been recognized that there are patterns of anatomical links, statistical dependencies or causal interactions between distinct units within a nervous system [12]. Some techniques have also been developed to estimate this brain connectivity graph [13], [14]. However, this task is complicated. In many cases, the estimated graph may not be accurate. Thus, it is necessary to have a method to refine the graph.

Our work addresses the low-rank components estimation and graph learning simultaneously. Previous works have studied these two problems independently [15], [16]. For low-rank components estimation, several linear or nonlinear methods have been proposed making use of the graph Laplacian of the sample affinity graphs. Specifically, given the high-dimensional datasets that consist of n p-dimensional data points, these methods make use of the sample affinity graphs of n vertices to capture the geometry of the data points and aim to preserve the structure. In Laplacian Eigenmap (LE) [17], dimensionality is reduced by making the nodes that are close in the high-dimensional space to be close in the low-dimensional space. Locality Preserving Projections (LPP) [18] is based on LE, but assumes that there is a vector of transformation. The vector is calculated as the results of LE. Local Linear Embedding (LLE) [19] aims to recover the low-dimensional structure using neighboring points relationships, which are calculated from the high-dimensional data. Multidimensional Scaling (MDS) [20] aims to recover the low-dimensional structure by minimizing the difference between the distance of two nodes in high-dimensional space and the distance of these two nodes in low-dimensional space. Isomap [21] is a nonlinear generalization of MDS. These methods can be viewed as kernel PCA on specially constructed graph [22]. In addition, various approaches have been proposed to incorporate spectral graph regularization [23], [24], [25] to uncover the low-rank components. In these works, graphs are fixed in their algorithms, precomputed from the noisy input data. These works do not consider inaccuracy in the graphs. In contrast, our proposed algorithms refine the graph to improve low-rank components estimation at the same time. Some low-rank components estimation algorithms have been applied for brain imaging data. Principal Components Analysis (PCA) and Independent Components Analysis (ICA) [26] are commonly-used algorithms to estimate the latent representations from noisy brain imaging data [27]. Maximum-likelihood estimation based algorithm [28] has been applied to estimate low-rank signals for MEG/EEG. Robust Principal Components Analysis (RPCA) has been applied in fMRI data to improve the classification accuracy [29].

Graph estimation or graph learning has been an active research topic recently [2], [30]. Different algorithms have been proposed based on various assumptions on the graph. The work of Spyrou and Escudero [31] assumes that original data is a four-mode tensor and it applies graph regularized tensor factorization of a four-mode tensor to denoise data. The work of Maretic et al. [5] assumes that graph signals are sparse and signals are a sparse linear combination of some atoms in a structured graph dictionary. We refer their method as Sparse Graph Dictionary method (SGDict) and we compare it with our proposed method in graph estimation accuracy. The work of Chepuri et al. [32] assumes that topology is sparse and is represented with a sparse edge selection vector. In particular, [32] assumes the number of edges is known and learns the connectivities from the data. It is related to [33], which will be introduced in the next paragraph. Note that [32] only estimates edge existence. However, [33] estimates both edge existence and edge weights. The work of Berger et al. [34] assumes that signals are smooth on the graph, which is also similar to [33]. But it controls the node degree in the graph. Compared with our proposed algorithms, this work only considers data with Gaussian noise. The methods proposed by Pavez and Ortega [4], Rabbat [35] and Hu et al. [36] also assume that signal is smooth on the graph. In these works, data, considered as random vectors, is modeled by a Gaussian Markov Random Field (GMRF) distribution, and its corresponding precision matrix is estimated as graph matrix. The works of Kao et al. [37] and Kalofolias [3] combine sparsity and smoothness assumption together and assume that signals are smooth on the sparse graph. The work of Bars et al. [38] assumes signals are smooth on the graph and bandlimited, which imply sparsity of signal representation in the spectral domain. Study on dynamic graph (topology changes with time) estimation is another active research topic. The work of Villafañe-Delgado et al. [39] proposes the dynamic Graph Fourier Transform (DGFT) to deal with dynamic graph signals with different Laplacian matrices. The work of Kalofolias et al. [40] proposes an algorithm to learn the dynamic graphs based on the assumption that changes in graph are smooth along the time. The work of Yamada et al. [41] infers the time-varying graphs based on the assumption that changes in graph are sparse along the time. The work of Shen et al. [42] proposes a data model based on the autoregressive model to model the time-varying graph signals and learns the graph from the data. Another data model is used by Shafipour et al. [43] and Wai et al. [44]. In this model, graph signals are generated by local diffusion dynamics on the graph. Shafipour et al. [43] infer the graph from non-stationary graph signals. Wai et al. [44] apply low-pass graph filter to detect communities from data with Gaussian noise.

The work of Dong et al. [33] is most related to our work. While its focus is to learn the connectivity graph topology, it also estimates some noise-free data of the input data as output. Gaussian noise model and Frobenius norm optimization are employed in [33] to make it suitable for the problem when noise is Gaussian with small amplitudes. As will be discussed, our work assumes another signal model with a low-rank component plus a sparse perturbation. Experiment results suggest that our method outperforms other methods for high-dimensional graph data corrupted by complicated noise, such as brain imaging signals. Furthermore, our work analyzes the impact of inexact graph on low-rank components estimation.

Graph signal processing tools have been applied to some brain imaging tasks [45]. In [46], graph Fourier transform is applied to decompose brain signals into low, medium, and high-frequency components for analysis of functional brain networks properties. The work of Liu et al. [47] uses eigenvectors of graph Laplacian matrix to approximate intrinsic subspace of high-dimensional brain imaging data. They evaluate different brain connectivity estimations to compute the graph. Estimation of the brain connectivity graph using a Gaussian noise model has been proposed in [36]. The work of Hu et al. [48] introduces a graph regression model for learning structural brain connectivity of Alzheimer’s disease from PiB-PET imaging data. The work of Griffa et al. [49] proposes a method to capture the network of spatio-temporal connectivity integrating neural oscillations and anatomical connectivity from diffusion fMRI. The work of Medaglia et al. [50] applies the graph Fourier transform to distill functional brain signals into aligned and liberal components. The work of Guo et al. [51] applies deep convolutional networks on graph signals for brain signal analysis. Besides, some works [52], [53], [54], [55] make use of graph smoothness regularization to learn the latent manifold in magnetic resonance imaging (dMRI) reconstruction application. The goal of reconstruction is to recover high-dimensional or high resolution data from undersampled inputs. Graph signal processing has also been shown to be useful in image compression [56], temperature data [57], wireless sensor data [58] and Internet traffic analysis [59]. A few signal features motivated by graph signal processing have also been proposed [60], [61].

In this paper, we propose algorithms to estimate low-rank components and learn the underlying graph simultaneously. Based on a signal model with sparse perturbation, the proposed algorithm estimates low-rank components of data using graph smoothness assumption and refines the graph with estimated low-rank components. Our specific contributions are:

  • We formulate the problem of joint estimation of low-rank components and connectivity graph for distorted high-dimensional graph signal.

  • We propose two approaches to this joint estimation problem. One approach applies alternating optimization to solve the related subproblems. Another approach solves the joint estimation directly.

  • We provide analysis to understand the impact of inexact graph on the low-rank components estimation accuracy, and we apply the analysis result to inform the choice of regularization parameter when the graph is inexact.

  • We perform extensive experiments on synthetic and real data, especially brain imaging (MEG) data, to illustrate and validate our proposed algorithms.

The joint estimation was first formulated and studied in our previous conference paper [62]. This work is an extension of [62]. Compared with [62], we have included substantial new contributions: a new method to address joint estimation, a study of the impact of graph distortion on low-rank component estimation, extensive and new experiments to understand the subtleties of joint estimation, and MEG experiments with consideration of field spreading and relevant frequency bands.

The rest of the paper is organized as follows. The proposed algorithm and its two solutions are presented in Section 2. Then, a detailed analysis on the impact of graph distortion is described in Section 3. Detailed experiment results on synthetic data are provided in Section 4 and results on real brain imaging data as well as MNIST image data are presented in Section 5. Section 6 concludes the work.

Section snippets

Simultaneous low-rank components and graph estimation

Table 1 lists the important notations used in this work. We consider X=(x1,,xn)Rp×n, the high-dimensional data matrix that consists of n p-dimensional data points. Based on the assumption that data points have low intrinsic dimensionality and lie near some low-dimensional subspaces, the following mathematical model for the data is proposed:X=L0+ML0Rp×n represents the low-rank components of the data matrix which is of primary interest in this paper, and MRp×n is a perturbation matrix that

Impact of inexact graphs

In many graph signal processing problems, the underlying graph needs to be estimated from data and it is inexact. In this section, we investigate the impact of inexact graph. Specifically, we focus on graph-assisted low-rank components estimation, i.e., (5), and analyze the performance under different graph distortion. We compare the low-rank components estimation performance of RPCAG and LGE2 on a 12-node Erdos-Renyi graph dataset, referring to Section 4.1 for synthetic data generation

Synthetic data generation

For synthetic data, we generate low-rank, graph-smooth and grossly-corrupted datasets with the following procedure. Our signal model is:X=L0+Mwhere L0Rp×n is a ground-truth low-rank components matrix with rank r and M is sparse matrix. The low-rank components matrix is a product L0=PYT where PRp×r contains r basis vectors and YRn×r is coefficient matrix. L0 is smooth on the graph built with following steps.

We generate two types of graphs as ground-truth: Erdos-Renyi model (ER model) and

Experiments with real data

To further illustrate our proposed algorithms, in this section, we apply our proposed algorithms on two datasets: 1) MNIST image dataset and 2) a high-dimensional brain imaging dataset, compare them with other methods. MNIST image dataset is a benchmark image dataset which RPCAG and RPCA are tested on. So it is interesting to see the low-rank components estimation performance of the proposed algorithm on it. Furthermore, we examine performances of the proposed algorithms in estimating low-rank

Conclusion

We have presented new algorithms to learn low-rank components and graph simultaneously for high-dimensional, graph-smooth and noisy datasets. We have analyzed the impacts of inexact graphs on the estimation of low-rank components. We have evaluated the proposed algorithms using both synthetic and real brain imaging data. Comparisons with other state-of-the-art methods suggest that the proposed algorithms are competitive in both low-rank components estimation and graph learning.

CRediT authorship contribution statement

Rui Liu: Conceptualization, Methodology, Software, Data curation, Writing - original draft, Writing - review & editing, Visualization, Investigation, Validation. Ngai-Man Cheung: Conceptualization, Supervision, Writing - review & editing.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

The authors would like to thank Dr. Hossein Nejati of KroniKare Pte Ltd, Dr. Pavitra Krishnaswamy of Institute for Infocomm Research, A*STAR and Dr. Sheraz Khan of Massachusetts General Hopital, Harvard Medical School and Massachusetts Institute of Technology for helpful discussion.

References (71)

  • S. Bentin et al.

    Electrophysiological studies of face perception in humans

    J. Cognit. Neurosci.

    (1996)
  • Y. Kamitani et al.

    Decoding the visual and subjective contents of the human brain

    Nat. Neurosci.

    (2005)
  • H. Nejati et al.

    Towards perception awareness: perceptual event detection for brain computer interfaces

    2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)

    (2015)
  • H. Wen et al.

    Neural encoding and decoding with deep learning for dynamic natural vision

    Cereb. Cortex

    (2017)
  • E. Bullmore et al.

    Complex brain networks: graph theoretical analysis of structural and functional systems

    Nat. Rev. Neurosci.

    (2009)
  • A. Brovelli et al.

    Beta oscillations in a large-scale sensorimotor cortical network: directional influences revealed by Granger causality

    Proc. Natl. Acad. Sci. U. S. A.

    (2004)
  • W. Chen et al.

    Nonconvex robust low-rank tensor reconstruction via an empirical Bayes method

    IEEE Trans. Signal Process.

    (2019)
  • W. Chen

    Simultaneously sparse and low-rank matrix reconstruction via nonconvex and nonseparable regularization

    IEEE Trans. Signal Process.

    (2018)
  • M. Belkin et al.

    Laplacian eigenmaps for dimensionality reduction and data representation

    Neural Comput.

    (2003)
  • X. He et al.

    Locality preserving projections

    Advances in Neural Information Processing Systems

    (2004)
  • S.T. Roweis et al.

    Nonlinear dimensionality reduction by locally linear embedding

    Science

    (2000)
  • T.F. Cox et al.

    Multidimensional Scaling

    (2000)
  • J.B. Tenenbaum et al.

    A global geometric framework for nonlinear dimensionality reduction

    Science

    (2000)
  • J. Ham et al.

    A kernel view of the dimensionality reduction of manifolds

    Proceedings of the Twenty-First International Conference on Machine Learning

    (2004)
  • B. Jiang et al.

    Graph-Laplacian PCA: Closed-form solution and robustness

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    (2013)
  • N. Shahid et al.

    Robust principal component analysis on graphs

    Proceedings of International Conference on Computer Vision

    (2015)
  • N. Shahid et al.

    Fast robust PCA on graphs

    IEEE J. Sel. Top. Signal Process.

    (2016)
  • A. Hyvrinen et al.

    Independent component analysis: algorithms and applications

    Neural Netw.

    (2000)
  • J. Gross et al.

    Good practice for conducting and reporting meg research

    NeuroImage

    (2013)
  • P. Georgieva et al.

    Robust principal component analysis for brain imaging

    International Conference on Artificial Neural Networks

    (2013)
  • X. Dong et al.

    Learning graphs from data: a signal representation perspective

    IEEE Signal Process. Mag.

    (2019)
  • L. Spyrou et al.

    Graph regularised tensor factorisation of eeg signals based on network connectivity measures

    Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on

    (2017)
  • S.P. Chepuri et al.

    Learning sparse graphs under smoothness prior

    Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on

    (2017)
  • X. Dong et al.

    Learning Laplacian matrix in smooth graph signal representations

    IEEE Trans. Signal Process.

    (2016)
  • P. Berger et al.

    Graph learning based on total variation minimization

    2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)

    (2018)
  • View full text