This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: BASIRA Lab, Faculty of Computer and Informatics, Istanbul Technical University, Istanbul, Turkey

Deep Graph Normalizer: A Geometric Deep Learning Approach for Estimating Connectional Brain Templates

Mustafa Burak Gurbuz    Islem Rekik ID corresponding author: [email protected], http://basira-lab.com.
Abstract

A connectional brain template (CBT) is a normalized graph-based representation of a population of brain networks —also regarded as an ‘average’ connectome. CBTs are powerful tools for creating representative maps of brain connectivity in typical and atypical populations. Particularly, estimating a well-centered and representative CBT for populations of multi-view brain networks (MVBN) is more challenging since these networks sit on complex manifolds and there is no easy way to fuse different heterogeneous network views. This problem remains unexplored with the exception of a few recent works rooted in the assumption that the relationship between connectomes are mostly linear. However, such an assumption fails to capture complex patterns and non-linear variation across individuals. Besides, existing methods are simply composed of sequential MVBN processing blocks without any feedback mechanism, leading to error accumulation. To address these issues, we propose Deep Graph Normalizer (DGN), the first geometric deep learning (GDL) architecture for normalizing a population of MVBNs by integrating them into a single connectional brain template. Our end-to-end DGN learns how to fuse multi-view brain networks while capturing non-linear patterns across subjects and preserving brain graph topological properties by capitalizing on graph convolutional neural networks. We also introduce a randomized weighted loss function which also acts as a regularizer to minimize the distance between the population of MVBNs and the estimated CBT, thereby enforcing its centeredness. We demonstrate that DGN significantly outperforms existing state-of-the-art methods on estimating CBTs on both small-scale and large-scale connectomic datasets in terms of both representativeness and discriminability (i.e., identifying distinctive connectivities fingerprinting each brain network population). Our DGN code is available at https://github.com/basiralab/DGN111Paper YouTube video: https://www.youtube.com/watch?v=Q_WLY2ZNxRk&t=3s222Code demo video: https://www.youtube.com/watch?v=8tDNYV0NYpY&t=1s.

Keywords:
Connectional brain templates \cdot Deep graph normalizer \cdot Population multiview brain network integration

1 Introduction

The field of network neuroscience has made substantial advances in characterizing the human brain network (or connectome modeling the pairwise relationship between brain regions of interest (ROIs)) by means of large-scale connectomic datasets collected through various projects such as Human Connectome Project (HCP) [1] and Connectome Related to Human Disease (CRHD) [2]. These rich and multimodal brain datasets allow us to map brain connectivity and efficiently detect atypical deviations from the healthy brain connectome. Particularly, learning how to normalize a population of brain networks by estimating a well-centered and representative connectional brain template (CBT) is an essential step for group comparison studies as well as discovering the integral signature of neurological disorders [3]. Intuitively, a CBT can be defined as a ‘normalized connectome’ of a population of brain networks, which can be simply produced by linear averaging. However, such a normalization technique is very sensitive to outliers and overlooks non-linear relationships between subjects. This normalization process can be regarded as an ‘integration’ or ‘standardization’ of brain networks.

More broadly, estimating a CBT of a population of heterogeneous multi-view brain networks, where each view captures particular traits of the brain construct (e.g., cortical thickness, function, cognition), is even a more challenging task since such connectomic data might lie on complex high-dimensional manifolds. To address this challenge, [3] proposed a clustering-based approach based on similarity network fusion (SNF) [4] to fuse multi-view brain networks in each cluster. Fused networks are then linearly averaged to produce a CBT for a population of multi-view brain networks (MVBN). Despite its promising results, [3] heavily depends on the selection of the number of clusters. In order to overcome this problem, [5] introduced the netNorm framework, which instead of clustering, constructs a high-order graph using cross-view connectional features as nodes and their Euclidean distance as dissimilarity measure to select the most centered brain connections in a population of MVBN. Next, the selected connections are integrated into a single network using SNF [4]. Currently, netNorm presents the state-of-the-art method by outperforming SCA [3] and other baseline methods in the CBT estimation task. However, netNorm [5] has several limitations. First, it uses Euclidean distance as a pre-defined metric for selecting the most representative brain connection which might fail to capture complex non-linear patterns in the brain connectome across subjects. Second, netNorm also uses SNF for fusing different views. Even though SNF is a powerful tool since it is a generic unsupervised technique, it comes with assumptions such as emphasizing top kk local connections for each node (i.e., brain region) and equally averaging the global topology of complementary networks for each iterative update to ultimately merge them. Instead of relying on such general assumptions, this MVBN normalization process can instead be learned to decide which information provided by the networks is important for the target CBT estimation. Third, netNorm consists of independent feature extraction, feature selection, and fusion steps. These fully independent steps cannot provide feedback to each other in order to globally optimize the CBT estimation process. Therefore errors might accumulate throughout the estimation pipeline.

To address all these limitations, we propose Deep Graph Normalizer (DGN): an unprecedented approach capitalizing on geometric deep learning (GDL) to learn how to normalize a population of heterogeneous MVBNs and generate a well-representative and centered CBT in an end-to-end fashion. Although GDL achieved remarkable results in several recent biomedical data analysis works such as disease classification [6] and protein interface prediction [7], to the best of our knowledge, no previous works used GDL to address the problem of integrating a population of multi-view networks [8, 9]. To fill this gap, we present several major contributions to the state-of-the-art as follows. First, we design a GDL architecture that takes multi-view brain networks and maps them into a normalized CBT. Specifically, we propose a GDL architecture that maps MVBN of a training subject to a population-representative CBT. Brain networks of each training subject passes through several layers of graph convolutional neural networks that are consecutively applied to learn hidden embeddings for each node (i.e., brain ROI) by locally integrating connectivities offered by different heterogeneous views and blending the previous layer’s embeddings using integrated connectivities. Next, we compute the pairwise absolute difference of the final layer’s node embeddings to derive connectivity weights of the generated CBT. Second, we introduce the Subject Normalization Loss (SNL) which is a randomized weighted loss function that evaluates the representativeness of a generated subject-biased CBT (i.e. obtained by feeding a particular subject to the model) against a random subset of brain networks in the training set to achieve subject-to-population mapping. Besides forcing the model to learn how to generate population-based representative CBTs by fusing complementary information supplied by MVBNs, SNL also acts as a regularization due to randomization and different weights assigned to each view according to their connectivity weight distributions. Third, finalized CBT can be obtained by feeding an arbitrary subject of the training population to the trained model since the model learns how to map any subject to a population-representative CBT thanks to SNL optimization. However, the choice of the subject biases the output CBT and leads to non-optimal performance. We introduce a post-training step to overcome this bias and further refine the finalized CBT.

Refer to caption
Figure 1: Proposed Deep Graph Normalizer (DGN) architecture for estimating connectional brain templates for a given population of multi-view brain networks. (A) Tensor representation of multi-view brain networks. Each subject ss is represented by 𝒯snr×nr×nv\mathcal{T}_{s}\in\mathbb{R}^{n_{r}\times n_{r}\times n_{v}}, composed of a set of undirected, fully connected graphs, each capturing single connectional feature. (B) Geometric deep learning layers. Our model includes a sequence of edge conditioned [10] graph convolutional neural network layers which are separated by ReLU non-linearity. Each layer learns deeper embeddings for ROIs by utilizing activation of the previous layer and topological structure of the brain network. (C) CBT generation and loss function. ROI embeddings that are output by the final layer are passed through a series of tensor operations to calculate the pairwise absolute difference of each pair of nodes for CBT construction. Next, the representativeness of the estimated CBT is evaluated against a random subset of training views for loss calculation. (D) CBT refinement after training. To select the most centered connections for final CBT generation, we first pass each training subject through the trained model to generate its corresponding CBT. Finally, we produce the final CBT by selecting the element-wise median of all training CBTs.

2 Proposed Method

In this section, we detail the components of our DGN architecture for estimating CBTs (Fig. 1). First, each subject in a population of multi-view brain networks is represented by an undirected fully-connected graph where each node (i.e., brain ROI) is initialized with identity features (trivially set to 1) and each edge has nvn_{v} attributes that correspond to connectivity weights in different network views. We generate CBTs using a training set of MVBNs T={𝐓11,𝐓21,,𝐓iv,,𝐓Nnv}T=\{\mathbf{T}_{1}^{1},\mathbf{T}_{2}^{1},\dots,\mathbf{T}_{i}^{v},\dots,\mathbf{T}_{N}^{n_{v}}\} , where 𝐓iv\mathbf{T}_{i}^{v} denotes the vthv^{th} brain view of subject ii, and evaluate the representativeness on a testing set using 5-fold cross-validation. These training subjects pass through 3 layers of GDL layers that includes edge conditioned filter learner (a shallow neural network). From layer to layer, deeper embeddings are learned for each ROI using edge-conditioned graph convolution [10] which aggregates the information passed by its neighbours while taking into consideration the multi-view attributes of its neighboring edges. We then use ROI embeddings output by the final layer to produce the connectivity matrix of the generated CBT by calculating the between the final embeddings of each ROI pair. This generated CBT is evaluated against a random subset of the training MVBNs for loss optimization. Once the DGN architecture is fully trained using this randomized training sample selection strategy, each training subject is then fed to the trained model to produce a population representative CBTs that are biased by the given input subject. Finally, we eliminate outlier connectivities due to subject bias by taking the element-wise median of all possible CBTs. We detail all these steps in what follows.

A- Tensor representation of multi-view brain networks. Given a population of subjects, we represent each subject ss by a tensor 𝒯snr×nr×nv\mathcal{T}_{s}\in\mathbb{R}^{n_{r}\times n_{r}\times n_{v}} (Fig. 1–A) that is composed by stacking connectivity matrices of nvn_{v} brain networks where each network has nrn_{r} nodes (i.e., ROI). Note that using this subject-specific brain tensor representation, an 𝐞ijnv×1\mathbf{e}_{ij}\in\mathbb{R}^{n_{v}\times 1} connecting ROIs ii and jj encapsulates nvn_{v} attributes. We set diagonal entries in the tensor to zero to eliminate self-loops (i.e., ROI self-connectivity). In addition to inputting edge multi-view attributes, our DGN architecture also takes a node attributes matrix 𝐕0nr×d0\mathbf{V}^{0}\in\mathbb{R}^{n_{r}\times d_{0}} as an input, where d0d_{0} denotes the number of initial attributes for each ROI. As for each brain ROI, we do not have predefined attributes, we set each entry of the 𝐕0\mathbf{V}^{0} matrix to ‘11’ (i.e., identity), and through our deep model training for optimizing the SNL function, we learn these node-specific implicit attribute representations by simply using the input edge attributes. In each graph convolution layer and for each multi-view brain connection 𝐞ij\mathbf{e}_{ij} linking nodes ii and jj, we generate an edge-specific weight matrix utilizing an edge-conditioned filter learner [10]. Then these weight matrices multiplied by ROIs attributes to compute the next layer’s ROI attributes therefore after the first convolution each ROI will have a different set of attributes even though they were identical in the beginning.

B- Geometric deep learning layers. In this section, we detail the graph convolutional layers of our architecture that maps ROIs with identical attributes to a high dimensional distinctive representations by utilizing edge features 𝐞ij\mathbf{e}_{ij} between ROIs. This is achieved through 3 graph convolutional network layers (Fig. 1–B) with edge-conditioned convolution operation [10] that are separated by ReLU non-linearity. Each layer l{1,2,,L}l\in\{1,2,...,L\} includes edge-conditioned filter learner neural network Fl:nvdl×dl1F^{l}:\mathbb{R}^{n_{v}}\mapsto\mathbb{R}^{d_{l}\times d_{l-1}} that dynamically generates edge specific weights for filtering message passing between ROIs ii and jj given the features of 𝐞ij\mathbf{e}_{ij}. This operation is defined as follows:

𝐯=il𝚯l.𝐯+il11|N(i)|(jϵN(i)Fl(𝐞ij;𝐖l)𝐯jl1+𝐛l);Fl(𝐞ij;𝐖l)=𝚯ij\displaystyle\mathbf{v}{{}_{i}^{l}}=\mathbf{\Theta}^{l}.\mathbf{v}{{}_{i}^{l-1}}+\frac{1}{\left|N(i)\right|}\left(\sum_{j\epsilon N(i)}\ F^{l}(\mathbf{e}_{ij};\mathbf{W}^{l})\mathbf{v}^{l-1}_{j}+\mathbf{b}^{l}\right);F^{l}(\mathbf{e}_{ij};\mathbf{W}^{l})=\mathbf{\Theta}_{ij}

where 𝐯il\mathbf{v}{{}_{i}^{l}} is the embedding of ROI ii at layer ll, 𝚯l\mathbf{\Theta}^{l} is a learnable parameter, and N(i)N(i) denotes the neighbours of ROI ii. 𝐛ldl\mathbf{b}^{l}\in\mathbb{R}^{d_{l}} denotes a network bias and FlF^{l} is a neural network that maps nv\mathbb{R}^{n_{v}} to dl×dl1\mathbb{R}^{d_{l}\times d_{l-1}} with weights 𝐖l\mathbf{W}^{l}. 𝚯ij\mathbf{\Theta}_{ij} represents the dynamically generated edge specific weights by FlF^{l}. Note that FlF^{l} can be any type of neural network and vary in each layer depending on the characteristics and complexity of edge weights.

C- CBT construction layer and subject normalization loss function. After obtaining the output 𝐕L=[𝐯1L,𝐯2L,,𝐯nr1L,𝐯nrL]T\mathbf{V}^{L}=\left[\mathbf{v}_{1}^{L},\mathbf{v}_{2}^{L},...,\mathbf{v}_{n_{r}-1}^{L},\mathbf{v}_{n_{r}}^{L}\right]^{T} of the final DGN layer, which consists of embeddings for each ROI, we compose the output CBT by computing the pair-wise absolute difference of the learned embeddings. We formulate this process using several tensor operations (Fig. 1–C) for easy and efficient backpropagation. First, 𝐕L\mathbf{V}^{L} is replicated horizontally nrn_{r} times to obtain nr×nr×dL\mathcal{R}\in\mathbb{R}^{n_{r}\times n_{r}\times d_{L}}. Next, \mathcal{R} is transposed (replacing all elements xyz\mathcal{R}_{xyz} with yxz\mathcal{R}_{yxz}) to get T\mathcal{R}^{T}. Last, we compute the element-wise absolute difference of \mathcal{R} and T\mathcal{R}^{T}. The resulting tensor is summed along zz-axis (i.e. size of node embeddings) to estimate the final CBT 𝐂nr×nr\mathbf{C}\in\mathbb{R}^{n_{r}\times n_{r}}.

Subject Normalization Loss (SNL) optimization. We propose to evaluate the representativeness of the generated CBT using a random subset of the training subject views. This random selection procedure has two main advantages compared to evaluating against all training subjects. First, randomization has a regularization effect since it is much easier for the model to overfit if the loss is calculated against the same set of subjects in each iteration. Secondary, the sample size can be fixed to a constant number so that the magnitude of the loss and the computation time will be independent of the size of the training set. Note that since SNL compares generated CBT with a subset of training subject views, model weights updated in a way so that the generated CBT represents a population of MVBNs even though it is rooted in a single subject input. Given the generated CBT 𝐂s\mathbf{C}_{s} for subject ss and a random subset SS of training subject indices, we define our SNL for training subject ss and the optimization loss as follows:

SNLs=v=1nviS𝐂s𝐓ivF×λv;min𝐖1,𝐛1𝐖L,𝐛L1|T|s=1|T|SNLsSNL_{s}=\sum_{v=1}^{n_{v}}\sum_{i\in S}\left\|\mathbf{C}_{s}-\mathbf{T}_{i}^{v}\right\|_{F}\times\lambda_{v}{\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1};}\ \min\limits_{\mathbf{W}_{1},\mathbf{b}_{1}\dots\mathbf{W}_{L},\mathbf{b}_{L}}\frac{1}{\left|T\right|}\sum_{s=1}^{\left|T\right|}\\ SNL_{s}

The λv\lambda_{v} is a view specific normalization term that is defined as:

λv=1μvmax{1μj}j=1nv\lambda_{v}=\frac{\frac{1}{\mu_{v}}}{\max\left\{\frac{1}{\mu_{j}}\right\}_{j=1}^{n_{v}}}

where μv\mu_{v} is the mean of brain graph connectivity weights of view vv and max{1μj}j=1nv\max\left\{\frac{1}{\mu_{j}}\right\}_{j=1}^{n_{v}} is the maximum of mean reciprocals 1μ1\frac{1}{\mu_{1}} to 1μnv\frac{1}{\mu_{n_{v}}}. We use this view-specific normalization weight since brain network connectivity distribution and value range might largely vary across views. This will help avoid view-biased CBT estimation where the trained model might overfit some views and overlook others. Related problems in the literature are addressed by normalizing the connectivity matrix. For example, SNF [4] divides connectivities in each row by the sum of the entries in that row to normalize measurements; however, this breaks the symmetry in the views, therefore, it is not applicable in our case. Another simple normalization approach such as using min-max scaling can saturate our inputs at 0 and 1 while standard z-score scaling generates negative connectivities in the graph that is not suitable for representing a fully positive brain connectomes (such as structural or morphological). Therefore, we introduce λv\lambda_{v} to ensure that the model gives equal attention to each brain view regardless of their value range.

D- CBT refinement after the training. Our model learns to map multi-view brain networks of a particular subject ss to a population-based representative CBT 𝐂s\mathbf{C}_{s}. Although all of the CBTs generated by the model are representative of the randomly sampled training population, they are biased towards the given input subject ss. To eliminate this bias, we propose an additional step (Fig. 1–D) to obtain a more refined and representative CBT for the whole training set. First, each subject is fed through the trained model to obtain corresponding CBTs. Next, most centered connections are selected from these CBTs by calculating the element-wise median. The median operation produces a valid view with a non-negative symmetrical adjacency matrix. This operation could also be replaced with other measures of central tendency; however, we used the representativeness score to verify that the median is the most suitable for our case.

3 Results and Discussion

Connectomic datasets and model hyperparameter setting. We benchmarked our DGN against state-of-the-art method netNorm for CBT estimation [5] on two small-scale and large-scale connectomic datasets using 5-fold cross-validation. The first dataset (AD/LMCI dataset) consists of 77 subjects (41 subjects diagnosed with Alzheimer’s diseases (AD) and 36 with Late Mild Cognitive Impairment (LMCI)) from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database GO public dataset [11]. Each subject is represented by 4 cortical morphological brain networks derived from maximum principal curvature, the mean cortical thickness, the mean sulcal depth, and the average curvature as in [12, 13, 14]. The second dataset (NC/ASD dataset) is collected from the Autism Brain Imaging Data Exchange ABIDE I public dataset [15] and includes 310 subjects (155 normal control (NC) and 155 subjects with autism spectral disorder (ASD)) with 6 cortical morphological brain networks extracted from the 4 aforementioned cortical measures in addition to cortical surface area and minimum principle area. For each hemisphere, the cortical surface is reconstructed from T1-weighted MRI using FreeSurfer pipeline [16] and parcellated into 35 ROIs using Desikan-Killiany atlas [17] and its corresponding brain network is derived by computing the pairwise absolute difference in cortical measurements between pairs of ROIs.

We trained 8 different models to generate CBTs for both hemispheres of 4 populations namely; AD, LMCI, NC, and ASD. We empirically set all hyperparameters for the DGN models using grid search. Each model includes 3 edge-conditioned convolutional neural network layers with an edge-conditioned filter learner neural network that maps 4 (for AD/LMCI dataset) or 6 (for NC/ASD dataset) attributes obtained from heterogeneous views to dl×dl1\mathbb{R}^{d_{l}\times d_{l-1}}. These layers are separated by ReLU activation function and output embeddings with 36, 24 and 5 (for AD/LMCI dataset) or 8 (for NC/ASD dataset) dimensions for each ROI in the MVBN, respectively. We trained all models using gradient descent with Adam optimizer with a learning rate of 0.00050.0005. We fixed the number of random samples in our SNL function to 1010.

Refer to caption
Figure 2: Representativeness comparison between CBTs generated by the proposed model and netNorm [5]. Charts illustrate the average Frobenius distance between the CBTs generated using the training set and the network views in the testing set. Also, p-values obtained by two-tailed t-test are reported for each population. LH: left hemisphere. RH: right hemisphere.

CBT representativeness test. To evaluate the representativeness of generated CBT, we computed mean Frobenius distance which is calculated as dF(A,B)=ij|AijBij|2d_{F}(A,B)=\sqrt{\sum_{i}\sum_{j}\left|A_{ij}-B_{ij}\right|^{2}} between the estimated CBT and the different views in the testing set. We split both datasets into training and testing sets using 5-fold cross-validation for reproducibility and generalizability. Fig. 2 depicts the average Frobenius distance between CBTs generated by DGN and netNorm [5] using the training set and the views in the left out test population. We note that our proposed model significantly (p<0.001p<0.001) outperforms netNorm in terms of representativeness across all left-out folds and evaluation datasets.

CBT discriminativeness reproducibility test. We hypothesize that a well-representative CBT can capture the most discriminative traits of a population of MVBNs, acting as a connectional brain fingerprint. To test this hypothesis, we first spot the top kk most discriminative brain connectivities where a class-A CBT largely differs from a class-B CBT. To do so, we compute the absolute difference between both estimated CBTs using respectively A and B populations. Next, we sum the columns of this difference matrix to obtain discriminability score for each brain ROI. We then pick the top kk most discriminative ROIs with the highest score. To evaluate the reproducibility of CBT-based discriminative ROIs, for each brain view vv, we independently train a support vector machine (SVM) and with a supervised feature selection method. Specifically, we extract connectional features from each brain network view by vectorizing the upper triangle entries. Next, for each network view, we use 5-fold randomized partitioning to divide each population pAp^{A} and pBp^{B} into 5 subpopulations. For each brain view vv and a combination of piAp^{A}_{i} and pjBp^{B}_{j} a SVM is trained and a weight vector 𝐰ijv\mathbf{w}_{ij}^{v} that scores the discriminativeness of each feature (i.e., ROI) is learned using Multiple Kernel Learning (MKL), which is a wrapper method assigning weights to features according to their distinctiveness for the given classification task. The final feature weight vector is computed by summing up the weight vectors for all views and all possible A-B combinations of their 5 subpopulations as follows: ω=𝐰=1nvi,j=15𝐰i,jv\mathbf{\omega}=\sum_{\mathbf{w}=1}^{n_{v}}\sum_{i,j=1}^{5}\mathbf{w}_{i,j}^{v}. Next, we anti-vectorize ω\mathbf{\omega} vector and obtain matrix 𝐌nr×nr\mathbf{M}\in\mathbb{R}^{n_{r}\times n_{r}}. By summing the columns of the resulting matrix we get ROIs discriminability scores. Finally, we picked the top kk ROIs with the highest score. Table 1 reports the overlap between the top k=15k=15 ROIs identified using CBT-based method (netNorm and DGN) and MKL-based SVM method. Remarkably, our proposed model not only generates more representative and centered CBTs but is significantly more reproducible in discriminability than netNorm [5].

Discovery of most discriminative brain ROIs for each disordered population. DGN also revealed left insula cortex, left superior temporal sulcus (STS) and right frontal pole as most discriminative regions of ASD population, which resonates with existing findings on autism. [18] reports that the alteration of the left insula in the ASD population might be the cause of abnormalities in emotional and affective functions. Next, by comparing activation of STS in different social scenarios, [19] shows that the dysfunction of STS is the essential factor of social perception impairment in autism. For instance, in contrast to NC subjects, individuals with autism show hypoactivation in the STS when exposed to matched visual and auditory information. Furthermore, [19] demonstrates that healthy children had greater response in STS triggered by biological (e.g. human movement) than by non-biological motions (e.g. clock). However, STS activation’s of children with autism do not differ significantly depending on the nature of the motion. Lastly, [20] shows that the faces of boys with ASD have atypical right dominant asymmetry and suggests that the asymmetric growth of the right frontal pole can explain this facial anomaly. As for the AD-LMCI dataset, DGN picked the left temporal pole (TP) and right entorhinal cortex (EC) as the most discriminative regions of the brain. [21] highlights that the pathological changes in TP are a common trait among all AD patients. Moreover, [22] confirms that the alteration of the EC is a good biomarker of AD and LMCI and indicates that the AD patients show greater atrophy in the right EC which supports DGN’s choice.

Overlap Rate netNorm [5] DGN
AD-LMCI Left Hem. 0.60 0.73
AD-LMCI Right Hem. 0.33 0.40
Overlap Rate netNorm [5] DGN
NC-ASD Left Hem. 0.53 0.53
NC-ASD Right Hem. 0.33 0.40
Table 1: Overlap rate between ROIs selected by MKL and CBT-based methods.

4 Conclusion

In this paper, we introduced Deep Graph Normalizer for estimating connectional brain templates for a given population of multi-view brain networks. Beside capturing non-linear patterns across subjects, the proposed method also learns how to fuse complementary information offered by MVBNs in an end-to-end fashion. We showed that the proposed DGN outperformed state-of-the-art method for estimating CBTs in terms of both representativeness and discriminability. In our future work, we will evaluate our architecture on multi-modal brain networks such as functional and structural brain networks while capitalizing on geometric deep learning for estimating holistic CBTs. Also, we will introduce topological loss constraints such as Kullback-Leibler divergence of node degree distributions of generated CBTs and population brain networks to further ensure that the generated CBTs are topologically sound.

5 Acknowledgments

I. Rekik is supported by the European Union’s Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie Individual Fellowship grant agreement No 101003403 (http://basira-lab.com/normnets/).

References

  • [1] Essen, D., Ugurbil, K., Auerbach, E., Barch, D., Behrens, T., Bucholz, R., Chang, A., Chen, L., Corbetta, M., Curtiss, S., Della Penna, S., Feinberg, D., Glasser, M., Harel, N., Heath, A., Larson-Prior, L., Marcus, D., Michalareas, G., Moeller, S., Yacoub, E.: The human connectome project: A data acquisition perspective. NeuroImage 62 (2012) 2222–31
  • [2] Van Essen, D., Glasser, M.: The human connectome project: Progress and prospects. Cerebrum: the Dana Forum on Brain Science 2016 (2016)
  • [3] Dhifallah, S., Rekik, I., Initiative, A.D.N., et al.: Clustering-based multi-view network fusion for estimating brain network atlases of healthy and disordered populations. Journal of neuroscience methods 311 (2019) 426–435
  • [4] Wang, B., Mezlini, A., Demir, F., Fiume, M., Tu, Z., Brudno, M., Haibe-Kains, B., Goldenberg, A.: Similarity network fusion for aggregating data types on a genomic scale. Nature methods 11 (2014)
  • [5] Dhifallah, S., Rekik, I.: Estimation of connectional brain templates using selective multi-view network normalization. Medical Image Analysis 59 (2019) 101567
  • [6] Rhee, S., Seo, S., Kim, S.: Hybrid approach of relation network and localized graph convolutional filtering for breast cancer subtype classification. (2018) 3527–3534
  • [7] Fout, A., Byrd, J., Shariat, B., Ben-Hur, A.: Protein interface prediction using graph convolutional networks. In Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., eds.: Advances in Neural Information Processing Systems 30. Curran Associates, Inc. (2017) 6530–6539
  • [8] Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., Yu, P.: A comprehensive survey on graph neural networks (2019)
  • [9] Zhou, J., Cui, G., Zhang, Z., Liu, Z., Sun, M.: Graph neural networks: A review of methods and applications (2018)
  • [10] Simonovsky, M., Komodakis, N.: Dynamic edge-conditioned filters in convolutional neural networks on graphs. (2017) 29–38
  • [11] Mueller, S., Weiner, M., Thal, L., Petersen, R., Jack, C., Jagust, W., Trojanowski, J., Toga, A., Beckett, L.: The alzheimer’s disease neuroimaging initiative. Neuroimaging clinics of North America 15 (2005) 869–77, xi
  • [12] Raeper, R., Lisowska, A., Rekik, I.: Cooperative correlational and discriminative ensemble classifier learning for early dementia diagnosis using morphological brain multiplexes. IEEE Access 6 (2018) 43830–43839
  • [13] Lisowska, A., Rekik, I., for the ADNI: Joint pairing and structured mapping of convolutional brain morphological multiplexes for early dementia diagnosis. Brain connectivity 9 (2019) 22–36
  • [14] Nebli, A., Rekik, I.: Gender differences in cortical morphological networks. Brain Imaging and Behavior (2019) 1–9
  • [15] Di Martino, A., Yan, C., Li, Q., Denio, E., Castellanos, F., Alaerts, K., Anderson, J., Assaf, M., Bookheimer, S., Dapretto, M., Deen, B., Delmonte, S., Dinstein, I., Ertl-Wagner, B., Fair, D., Gallagher, L., Kennedy, D., Keown, C., Keysers, C., Lainhart, J., Lord, C., Luna, B., Menon, V., Minshew, N., Monk, C., Mueller, S., Müller, R., Nebel, M., Nigg, J., O’Hearn, K., Pelphrey, K., Peltier, S., Rudie, J., Sunaert, S., Thioux, M., Tyszka, J., Uddin, L., Verhoeven, J., Wenderoth, N., Wiggins, J., Mostofsky, S., Milham, M.: The autism brain imaging data exchange: Towards a large-scale evaluation of the intrinsic brain architecture in autism. Molecular Psychiatry 19 (2014) 659–667
  • [16] Fischl, B.: Freesurfer. NeuroImage 62 (2012) 774–81
  • [17] Fischl, B., Kouwe, A., Destrieux, C., Halgren, E., Segonne, F., Salat, D., Busa, E., Seidman, L., Goldstein, J., Kennedy, D., Caviness, V., Makris, N., Rosen, B., Dale, A.: Automatically parcellating the human cerebral cortex. Cerebral cortex (New York, N.Y. : 1991) 14 (2004) 11–22
  • [18] Yamada, T., Itahashi, T., Nakamura, M., Watanabe, H., Kuroda, M., Ohta, H., Kanai, C., Kato, N., Hashimoto, R.i.: Altered functional organization within the insular cortex in adult males with high- functioning autism spectrum disorder: evidence from connectivity-based parcellation. Molecular Autism 7 (2016)
  • [19] Pelphrey, K., Carter, E.: Brain mechanisms for social perception lessons from autism and typical development. Annals of the New York Academy of Sciences 1145 (2009) 283–99
  • [20] Hammond, P., Forster-Gibson, C., Chudley, A., Allanson, J., Hutton, T., Farrell, S., McKenzie, J., Holden, J., Lewis, M.E.S.: Face-brain asymmetry in autism spectrum disorders. Molecular psychiatry 13 (2008) 614–23
  • [21] Arnold, S.E., Hyman, B.T., Van Hoesen, G.W.: Neuropathologic Changes of the Temporal Pole in Alzheimer’s Disease and Pick’s Disease. Archives of Neurology 51 (1994) 145–150
  • [22] Zhou, M., Zhang, F., Zhao, L., Qian, J., Dong, C.: Entorhinal cortex: A good biomarker of mild cognitive impairment and mild alzheimer’s disease. Reviews in the neurosciences 27 (2015)