This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

scBiGNN: Bilevel Graph Representation Learning for Cell Type Classification from Single-cell RNA Sequencing Data

Rui Yang1,2, Wenrui Dai1, Chenglin Li1, Junni Zou1, Dapeng Wu2, Hongkai Xiong1
1Shanghai Jiao Tong University,   2City University of Hong Kong
{rui_yang, daiwenrui, xionghongkai}@sjtu.edu.cn,   [email protected]
Abstract

Single-cell RNA sequencing (scRNA-seq) technology provides high-throughput gene expression data to study the cellular heterogeneity and dynamics of complex organisms. Graph neural networks (GNNs) have been widely used for automatic cell type classification, which is a fundamental problem to solve in scRNA-seq analysis. However, existing methods do not sufficiently exploit both gene-gene and cell-cell relationships, and thus the true potential of GNNs is not realized. In this work, we propose a bilevel graph representation learning method, named scBiGNN, to simultaneously mine the relationships at both gene and cell levels for more accurate single-cell classification. Specifically, scBiGNN comprises two GNN modules to identify cell types. A gene-level GNN is established to adaptively learn gene-gene interactions and cell representations via the self-attention mechanism, and a cell-level GNN builds on the cell-cell graph that is constructed from the cell representations generated by the gene-level GNN. To tackle the scalability issue for processing a large number of cells, scBiGNN adopts an Expectation Maximization (EM) framework in which the two modules are alternately trained via the E-step and M-step to learn from each other. Through this interaction, the gene- and cell-level structural information is integrated to gradually enhance the classification performance of both GNN modules. Experiments on benchmark datasets demonstrate that our scBiGNN outperforms a variety of existing methods for cell type classification from scRNA-seq data.

1 Introduction

Single-cell RNA sequencing (scRNA-seq), which allows the measurement of gene expression at the resolution of individual cells, has revolutionized transcriptomic analysis and greatly improved the understanding of biomedical science over the past decade. Accurate cell type annotation is essential for elucidating cellular states and dynamics in scRNA-seq data mining comparison and contributes significantly to a broad range of downstream analyses, such as cancer biology cancer and drug development drug . In early studies, the cluster-then-annotate paradigm is commonly used to categorize cell types clustering1 ; clustering2 . These methods rely on manually selected marker genes and require prior knowledge of cell types, which is highly subjective and prone to errors due to the unknown quality of clustering results. As copious annotated scRNA-seq data become publicly available, many classification methods have been developed for automatic cell type labeling.

Existing classification methods can be grouped into three categories: 1) traditional machine learning algorithms, 2) similarity-based measurement and 3) deep learning models. The first category of methods applies classic machine learning approaches to scRNA-seq data analysis, such as random forest, linear discriminant analysis and support vector machine. Typical works include CasTLe castle , scPred scpred and scID scid . These methods commonly have limited model capacity and do not scale well to large datasets. The second category is based on some similarity criteria to measure the correlation between the unlabeled cells and the reference dataset. For example, scmap scmap calculates the similarity between the query cell and the median gene expression values for each cell type in the reference dataset, while SingleR singler computes Spearman correlation between cells using variable genes. Similarity-based methods are heavily affected by the batch effect caused by variant experimental conditions batch , and the adopted similarity measurement may not be suitable for the high-dimensional and sparse scRNA-seq data scbert . To enable scalable and robust cell type identification, deep learning models have been increasingly explored and achieved superior performance. For instance, ACTINN actinn first uses fully connected neural networks for cell type identification. scCapsNet sccapsnet employs capsule networks capsule for interpretable gene feature selection in cell representation learning, which improves the reliability of deep learning models. Furthermore, graph neural networks (GNNs) have emerged as widely used tools for cell type annotation siggcn ; scgraph ; improving ; scbert ; adversarial , as they can leverage the intrinsic biological networks (e.g., gene-gene interaction network) within gene expression profiles for more expressive representation learning on scRNA-seq data.

However, several problems still impede GNN-based classification methods from reaching their full potential. Firstly, although it is generally acknowledged that gene-gene and cell-cell relationships are both beneficial to scRNA-seq analyses improving ; scbert , none of the existing models consider these two kinds of structural information simultaneously. To be concrete, gene-level GNNs classify cells from the local biological perspective of each cell’s own gene expression profile, which leverages informative gene interactions to improve cell representations but relationships between different cells are ignored. Cell-level GNNs consider the connectivity between cells from the global view of the whole database but cannot well capture the fine-grained gene-level biological contexts. Secondly, most existing works rely on predetermined biological networks for graph representation learning. For example, sigGCN siggcn utilizes the STRING database string to construct gene-gene interaction network, while Li et al. improving consider to precompute the cell-cell graph by kk-nearest neighbors analysis based on the gene expression data. Nevertheless, it is hard to evaluate whether these graphs constructed by task-irrelevant knowledge are optimal for cell type classification. Learning task-driven adaptive graphs may better serve GNN-based models for more accurate supervised single-cell classification.

To address these issues, we propose scBiGNN, a novel bilevel graph representation learning method that takes advantage of the above two kinds of biological networks in scRNA-seq classification. Specifically, scBiGNN consists of two GNN modules for cell type classification. A gene-level GNN based on the gene-gene interaction graph is designed to produce a representation for each individual cell, in which the interactions are adaptively learned via self-attention transformer ; gat . Then the cell-cell graph is constructed by linking cells that have similar representations generated by the gene-level GNN (i.e., neighboring cells are more likely from the same class), upon which a cell-level GNN is built to improve cell representations by aggregating features from the neighborhood. Such a bilevel framework can integrate both gene- and cell-level structural information for more comprehensive scRNA-seq graph representation learning.

In real-world scRNA-seq datasets, the dimension of the gene expression data is commonly very high. Therefore, it is challenging to optimize our scBiGNN in an end-to-end manner when both the gene-gene and cell-cell graphs are large. Inspired by previous studies that optimize two neural networks in a whole framework gmnn ; text ; emgcn , we employ the Expectation Maximization (EM) method to alternately train the two GNN modules via the E-step and M-step, which theoretically maximizes the evidence lower bound of the log-likelihood of the observed cell labels. In each step, one GNN is fixed to generate pseudo-labels for optimizing the other. In this alternating training fashion, scBiGNN can scale to large datasets, and gradually enhance the classification performance of both GNN modules by reinforcing each other. To validate the effectiveness of our proposed method, we conduct cell type classification on several popular benchmark datasets. Experimental results demonstrate that our scBiGNN achieves superior performance over a number of baselines.

2 Background

Data Description and Pre-processing. The scRNA-seq data can be represented as a gene expression matrix 𝐗N×M\mathbf{X}\in\mathbb{R}^{N\times M}, where NN is the number of cells, MM is the number of genes, and each element XijX_{ij} indicates the expression counts of the jj-th gene in the ii-the cell. First, genes that have zero expression values across all the cells are removed. Then, the gene expression data of each cell is normalized by X~ij=log(1+sXijmXim)\tilde{X}_{ij}=\log\left(1+s\cdot\frac{X_{ij}}{\sum_{m}X_{im}}\right) in which ss is a scaling value and is set as 10610^{6} following existing work scgraph . After normalization, the variance value of each gene across all the cells is calculated, and the top-TT variable genes are selected (e.g., T=1000T=1000 siggcn ) while others are filtered out, based on which we can obtain the pre-processed gene expression data 𝐗~N×T\tilde{\mathbf{X}}\in\mathbb{R}^{N\times T}.

Graph Neural Networks (GNNs). The goal of GNNs is to learn effective node and graph representations by iteratively aggregating information from the topological neighborhoods. Formally, let 𝒢=(𝒱,𝐀)\mathcal{G}=(\mathcal{V},\mathbf{A}) denote a graph where 𝒱={v1,,vn}\mathcal{V}=\{v_{1},\dots,v_{n}\} is the node set and 𝐀n×n\mathbf{A}\in\mathbb{R}^{n\times n} is the adjacency matrix with each non-zero value AijA_{ij} representing the edge between nodes viv_{i} and vjv_{j}. We also have a feature matrix 𝐅=[𝐟1,,𝐟n]n×f\mathbf{F}=[\mathbf{f}_{1},\dots,\mathbf{f}_{n}]^{\top}\in\mathbb{R}^{n\times f} with 𝐟if\mathbf{f}_{i}\in\mathbb{R}^{f} being the feature vector of node viv_{i}. The ll-th feature aggregration layer of GNNs can take the form of 𝐅(l)=MLP(l)(R(l)(𝐀)𝐅(l1)𝐖(l))\mathbf{F}^{(l)}=\operatorname{MLP}^{(l)}\left(\operatorname{R}^{(l)}\left(\mathbf{A}\right)\mathbf{F}^{(l-1)}\mathbf{W}^{(l)}\right), in which 𝐅(l)=[𝐟1(l),,𝐟n(l)]n×fl\mathbf{F}^{(l)}=[\mathbf{f}^{(l)}_{1},\dots,\mathbf{f}^{(l)}_{n}]^{\top}\in\mathbb{R}^{n\times f_{l}} is the matrix of node representations after ll aggregation layers, 𝐅(0)=𝐅\mathbf{F}^{(0)}=\mathbf{F}, 𝐖(l)fl1×fl\mathbf{W}^{(l)}\in\mathbb{R}^{f_{l-1}\times f_{l}} is a learnable weight matrix, R(l)(𝐀)\operatorname{R}^{(l)}\left(\mathbf{A}\right) is some operator on the adjacency matrix such as normalization gcn and self-attention mechansim gat , and MLP(l)()\operatorname{MLP}^{(l)}\left(\cdot\right) is the multi-layer perception (MLP) for nonlinear feature transformation. After LL aggregation layers, we can obtain the node representations 𝐅(l)\mathbf{F}^{(l)} (l=0,,Ll=0,\dots,L). To produce a graph-level representation, we can employ a read-out function that pools all the node representations into a single vector 𝐟𝒢=ReadOut({𝐅l}l=0L)\mathbf{f}_{\mathcal{G}}=\operatorname{Read-Out}\left(\{\mathbf{F}_{l}\}_{l=0}^{L}\right), which is permutation invariant to the order of graph nodes, e.g., sum-pooling and max-pooling.

GNN-based Single-cell Classification. Existing GNN-based methods for cell type identification can be categorized into two groups. The first group leverages gene-gene relationships. Specifically, Yin et al. scgraph collect several well-known gene interaction networks such as the STRING database string , on which GNNs are applied to aggregate information from interacting genes for each expressed gene to improve cell representations. Wang et al. siggcn also utilize the STRING database and propose sigGCN, in which a GNN-based autoencoder is used to reconstruct gene expression data and a fully connected neural network extracts features by taking the scRNA-seq data as input. The features generated by the encoder part of the autoencoder and the fully connected neural network are concatenated for cell type classification. HNNVAT adversarial further introduces virtual adversarial training that adds noise to input data to make the model robust against noise perturbations. In addition, Yang et al. scbert propose scBERT, which follows BERT’s pretraining and fine-tuning paradigm bert with vast amounts of unlabelled scRNA-seq data and learns gene-gene interactions via Performers performer . The second group of methods focuses on cell-level relationships. As more and more research demonstrates that cell-cell graph provides valuable structural information to learn effective cell representations for scRNA-seq analyses imputing ; scgnn ; cellvgae ; zinb ; scgac , Li et al. improving benchmarks several GNNs for cell type classification based on the cell-cell graph and outperforms traditional machine learning methods. These results motivate us to combine the merits of these two types of methods to integrate gene- and cell-level structural information for more accurate cell type classification.

3 The Proposed Method: scBiGNN

In this section, we develop the scBiGNN framework for bilevel graph representation learning on scRNA-seq data, in which both gene- and cell-level relationships are exploited for more accurate cell type classification. We first introduce some preliminaries about the problem statement and the workflow of gene- and cell-level GNNs. Then, we overview our EM-based learning framework and elaborate the optimization procedures of the E-step and M-step respectively.

3.1 Preliminaries

Problem Statement. In our work, two biological networks are constructed based on the pre-processed gene expression data 𝐗~N×T\tilde{\mathbf{X}}\in\mathbb{R}^{N\times T}. One is the gene-gene interaction graph 𝒢g=(𝒱g,𝐀g)\mathcal{G}^{\text{g}}=(\mathcal{V}^{\text{g}},\mathbf{A}^{\text{g}}) where 𝒱g\mathcal{V}^{\text{g}} is the set of TT genes {v1g,,vTg}\{v_{1}^{\text{g}},\ldots,v_{T}^{\text{g}}\} and 𝐀gT×T\mathbf{A}^{\text{g}}\in\mathbb{R}^{T\times T} is the associated adjacency matrix representing the interactions between genes, the other is the cell-cell graph 𝒢c=(𝒱c,𝐀c)\mathcal{G}^{\text{c}}=(\mathcal{V}^{\text{c}},\mathbf{A}^{\text{c}}) where 𝒱c={v1c,,vNc}\mathcal{V}^{\text{c}}=\{v_{1}^{\text{c}},\ldots,v_{N}^{\text{c}}\} indicates the set of NN cells in the scRNA-seq dataset and 𝐀cN×N\mathbf{A}^{\text{c}}\in\mathbb{R}^{N\times N} is its adjacency matrix encoding the relationships between cells. Given 𝐗~\tilde{\mathbf{X}} and the labels 𝐘L|𝒱L|×C\mathbf{Y}^{L}\in\mathbb{R}^{|\mathcal{V}^{L}|\times C} for a subset of annotated cells 𝒱L𝒱c\mathcal{V}^{L}\subset\mathcal{V}^{\text{c}}, the goal is to predict the labels 𝐘U|𝒱U|×C\mathbf{Y}^{U}\in\mathbb{R}^{|\mathcal{V}^{U}|\times C} for the the unlabeled cells 𝒱U=𝒱c\𝒱L\mathcal{V}^{U}=\mathcal{V}^{\text{c}}\backslash\mathcal{V}^{L}, where CC is the number of cell types.

Gene-level GNNs. The gene-level GNNs aim to predict the label of each cell vicv_{i}^{\text{c}} using its own gene expression data 𝐱~i\tilde{\mathbf{x}}_{i} (the ii-th row of 𝐗~\tilde{\mathbf{X}}) and the gene-gene interaction network. Formally, they model the following label distribution qϕ(𝐲i|𝐱~i)q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i}) for each vic𝒱cv_{i}^{\text{c}}\in\mathcal{V}^{\text{c}} with parameters ϕ\phi:

{𝐟i,t}t=1Ti\displaystyle\left\{\mathbf{f}_{i,t}\right\}_{t=1}^{T_{i}} =GNNg(𝐱~i,𝐀g),\displaystyle=\operatorname{GNN}^{\text{g}}\left(\tilde{\mathbf{x}}_{i},\mathbf{A}^{\text{g}}\right),
𝐡i\displaystyle\mathbf{h}_{i} =ReadOut({𝐟i,t}t=1Ti),\displaystyle=\operatorname{Read-Out}\left(\left\{\mathbf{f}_{i,t}\right\}_{t=1}^{T_{i}}\right),
qϕ(𝐲i|𝐱~i)\displaystyle q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i}) =Cat(𝐲i|Clfg(𝐡i)),\displaystyle=\operatorname{Cat}\left(\mathbf{y}_{i}\Big{|}\operatorname{Clf}^{\text{g}}\left(\mathbf{h}_{i}\right)\right), (1)

where GNNg()\operatorname{GNN}^{\text{g}}(\cdot) denotes the node representation learning function of the gene-level GNN, {𝐟i,t}t=1Ti\left\{\mathbf{f}_{i,t}\right\}_{t=1}^{T_{i}} is the set of gene representations learned by GNNg()\operatorname{GNN}^{\text{g}}(\cdot), TiT_{i} is the number of genes that have non-zero expression values in 𝐱~i\tilde{\mathbf{x}}_{i}, 𝐡i\mathbf{h}_{i} is the representation of cell vicv_{i}^{\text{c}}, Clfg()\operatorname{Clf}^{\text{g}}(\cdot) denotes a classifier, and Cat()\operatorname{Cat}(\cdot) stands for the categorical distribution. In this way, gene-gene structural information can be exploited from the local biological view of each individual cell to enhance cell representation learning.

Cell-level GNNs. The cell-level GNNs classify cell types by using the relationships among all the cells in the dataset. Specifically, the workflow can be characterized as below:

{𝐡i}i=1N\displaystyle\left\{\mathbf{h}^{\prime}_{i}\right\}_{i=1}^{N} =GNNc({𝐱~i,𝐡i}i=1N,𝐀c),\displaystyle=\operatorname{GNN}^{\text{c}}\left(\left\{\tilde{\mathbf{x}}_{i},\mathbf{h}_{i}\right\}_{i=1}^{N},\mathbf{A}^{\text{c}}\right),
pθ(𝐲i|𝐗~)\displaystyle p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}}) =Cat(𝐲i|Clfc(𝐡i)),\displaystyle=\operatorname{Cat}\left(\mathbf{y}_{i}\Big{|}\operatorname{Clf}^{\text{c}}\left(\mathbf{h}^{\prime}_{i}\right)\right), (2)

where GNNc()\operatorname{GNN}^{\text{c}}(\cdot) is the node representation learning function, {𝐡i}i=1N\left\{\mathbf{h}^{\prime}_{i}\right\}_{i=1}^{N} is the set of cell representations learned by GNNc()\operatorname{GNN}^{\text{c}}(\cdot), Clfc()\operatorname{Clf}^{\text{c}}(\cdot) is a classifier, and θ\theta is the model parameters. By aggregating features from neighboring cells with similar characteristics, cell-level GNNs extract the cell-cell structural information from a global view of the scRNA-seq dataset to improve single-cell classification.

Both gene- and cell-level GNNs have proved effective in scRNA-seq analyses and achieved superior performance for automatic cell type identification. Therefore, we propose scBiGNN, a bilevel graph representation learning method that aims to integrate these two levels of structural information to have a more comprehensive view of scRNA-seq data.

3.2 EM Framework

Directly cascading the above gene- and cell-level GNNs and training them end-to-end would face the scalability issue when the size of the scRNA-seq dataset is large. Moreover, as these two types of GNNs are complementary, end-to-end learning does not allow them to interact and enhance each other. Therefore, inspired by several previous studies gmnn ; text ; emgcn , we instead employ the EM algorithm to train these two GNN modules alternately and make them gradually reinforce each other for more accurate cell type classification.

To be concrete, our proposed scBiGNN maximizes the evidence lower bound (ELBO) of the log-likelihood of the observed cell labels:

logpθ(𝐘L|𝐗~)𝔼qϕ(𝐘U|𝐗~)[logpθ(𝐘L,𝐘U|𝐗~)logqϕ(𝐘U|𝐗~)]ELBO(θ,ϕ;𝐘L,𝐗~),\displaystyle\log p_{\theta}(\mathbf{Y}^{L}|\tilde{\mathbf{X}})\geq\mathbb{E}_{q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})}\left[\log p_{\theta}(\mathbf{Y}^{L},\mathbf{Y}^{U}|\tilde{\mathbf{X}})-\log q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})\right]\triangleq\mathcal{L}_{\text{ELBO}}(\theta,\phi;\mathbf{Y}^{L},\tilde{\mathbf{X}}), (3)

where qϕ(𝐘U|𝐗~)q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}}) can be arbitrary distribution over 𝐘U\mathbf{Y}^{U} (s.t. qϕ(𝐘U|𝐗~)>0q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})>0 if pθ(𝐘L,𝐘U|𝐗~)>0p_{\theta}(\mathbf{Y}^{L},\mathbf{Y}^{U}|\tilde{\mathbf{X}})>0), and the equality holds when qϕ(𝐘U|𝐗~)q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}}) equals to the true posterior distribution pθ(𝐘U|𝐘L,𝐗~)p_{\theta}(\mathbf{Y}^{U}|\mathbf{Y}^{L},\tilde{\mathbf{X}}). As can be seen, Eq. (3) formalizes the objective function with two distributions qϕq_{\phi} and pθp_{\theta}, which can be modeled by the abovementioned gene- and cell-level GNNs respectively. According to the EM framework, we alternately train qϕq_{\phi} in the E-step (with pθp_{\theta} fixed) and pθp_{\theta} in the M-step (with qϕq_{\phi} fixed) to maximize ELBO(θ,ϕ;𝐘L,𝐗~)\mathcal{L}_{\text{ELBO}}(\theta,\phi;\mathbf{Y}^{L},\tilde{\mathbf{X}}). Thus, we can separately optimize these two GNN modules through several iterations, leading to better scalability and making them interact and enhance each other. In what follows, we introduce the structures of these two GNN modules and how they reinforce each other in the training procedure.

3.3 E-step

In the E-step, the cell-level GNN (i.e., pθp_{\theta}) is fixed and the gene-level GNN (i.e., qϕq_{\phi}) is optimized to maximize ELBO(θ,ϕ;𝐘L,𝐗~)\mathcal{L}_{\text{ELBO}}(\theta,\phi;\mathbf{Y}^{L},\tilde{\mathbf{X}}). In this section, we first present the detailed structure of our employed gene-level GNN, and then we introduce the optimization of qϕq_{\phi} in the E-step.

Structure of Gene-level GNN. The input of the gene-level GNN module includes two embeddings. The first is gene embedding, which assigns a unique and learnable vector 𝐞jgenedg\mathbf{e}^{\text{gene}}_{j}\in\mathbb{R}^{d_{g}} to each gene vjgv_{j}^{\text{g}}. The second is count embedding, which transforms every non-zero gene expression value X~ij\tilde{X}_{ij} of cell vicv_{i}^{\text{c}} to a dgd_{g}-dimensional vector through an MLP, i.e., 𝐞ijcount=MLP(X~ij)dg\mathbf{e}^{\text{count}}_{ij}=\operatorname{MLP}(\tilde{X}_{ij})\in\mathbb{R}^{d_{g}}. The input embedding of gene vjgv_{j}^{\text{g}} in cell vicv_{i}^{\text{c}} is then defined as 𝐞ij=𝐞jgene+𝐞ijcount\mathbf{e}_{ij}=\mathbf{e}^{\text{gene}}_{j}+\mathbf{e}^{\text{count}}_{ij}. By gathering and stacking all the input embeddings of genes that have non-zero expression values in cell vicv_{i}^{\text{c}}, we have the input embedding matrix 𝐄iTi×dg\mathbf{E}_{i}\in\mathbb{R}^{T_{i}\times d_{g}} for cell vicv_{i}^{\text{c}}.

Inspired by attention-based models transformer ; gat , the gene-level GNN of scBiGNN adaptively learns gene-gene interactions via the self-attention mechanism. Specifically, the feature aggregation layer of GNNg()\operatorname{GNN}^{\text{g}}(\cdot) is defined as

R(l)(𝐀g)\displaystyle\operatorname{R}^{(l)}\left(\mathbf{A}^{\text{g}}\right) =softmax((𝐅i(l1)𝐖g(l))(𝐅i(l1)𝐖g(l))),\displaystyle=\operatorname{softmax}\left((\mathbf{F}_{i}^{(l-1)}\mathbf{W}_{\text{g}}^{(l)})(\mathbf{F}_{i}^{(l-1)}\mathbf{W}_{\text{g}}^{(l)})^{\top}\right),
𝐅i(l)\displaystyle\mathbf{F}_{i}^{(l)} =MLPg(l)(R(l)(𝐀g)𝐅i(l1)𝐖g(l)),\displaystyle=\operatorname{MLP}_{\text{g}}^{(l)}\left(\operatorname{R}^{(l)}\left(\mathbf{A}^{\text{g}}\right)\mathbf{F}_{i}^{(l-1)}\mathbf{W}_{\text{g}}^{(l)}\right), (4)

where softmax()\operatorname{softmax}\left(\cdot\right) is the row-wise softmax operation, 𝐖g(l)\mathbf{W}_{\text{g}}^{(l)} is a learnable weight matrix, MLPg(l)()\operatorname{MLP}_{\text{g}}^{(l)}(\cdot) is an MLP, 𝐅i(l)\mathbf{F}_{i}^{(l)} is the output feature matrix of the ll-th layer, and 𝐅i(0)=𝐄i\mathbf{F}_{i}^{(0)}=\mathbf{E}_{i} for each cell vicv_{i}^{\text{c}}. Moreover, in each layer of GNNg()\operatorname{GNN}^{\text{g}}(\cdot), we can also employ the multi-head attention mechanism transformer , which performs several different attention functions of Eq. (3.3) and concatenates their outputs as the feature matrix. After LL layers, we can obtain the set of gene representations {𝐟i,tj}t=1Ti\{\mathbf{f}_{i,t}^{j}\}_{t=1}^{T_{i}} in which 𝐟i,tj\mathbf{f}_{i,t}^{j} is the tt-th row of 𝐅i(L)\mathbf{F}_{i}^{(L)} and the superscript jj indicates that it corresponds to the gene vjgv_{j}^{\text{g}}. The read-out function in Eq. (3.1) can take the form of the weighted sum of all the gene representations to produce the cell representation for vicv_{i}^{\text{c}}:

𝐡i=ReadOut({𝐟i,tj}t=1Ti)=jαj𝐟i,tj,\displaystyle\mathbf{h}_{i}=\operatorname{Read-Out}\left(\{\mathbf{f}_{i,t}^{j}\}_{t=1}^{T_{i}}\right)=\sum\nolimits_{j}\alpha_{j}\mathbf{f}_{i,t}^{j}, (5)

where αj\alpha_{j} can be a learnable weight for gene vjgv_{j}^{\text{g}} (j=1,,Tj=1,\dots,T) or set as 1/Ti1/T_{i} which performs average pooling. Then, qϕ(𝐲i|𝐱~i)q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i}) can be obtained by feeding 𝐡i\mathbf{h}_{i} to an MLP classifier, as shown in Eq. (3.1).

Optimization of Gene-level GNN. Maximizing ELBO(θ,ϕ;𝐘L,𝐗~)\mathcal{L}_{\text{ELBO}}(\theta,\phi;\mathbf{Y}^{L},\tilde{\mathbf{X}}) in the E-step is equivalent to making qϕ(𝐘U|𝐗~)q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}}) approximate the true posterior distribution pθ(𝐘U|𝐘L,𝐗~)p_{\theta}(\mathbf{Y}^{U}|\mathbf{Y}^{L},\tilde{\mathbf{X}}). According to the EM framework, the standard operation is to minimize the Kullback-Leibler (KL) divergence KL(qϕ(𝐘U|𝐗~)pθ(𝐘U|𝐘L,𝐗~))\operatorname{KL}(q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})\|p_{\theta}(\mathbf{Y}^{U}|\mathbf{Y}^{L},\tilde{\mathbf{X}})). However, directly optimizing the KL divergence is nontrivial, as it depends on the entropy of qϕ(𝐘U|𝐗~)q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}}), which is hard to deal with text . Moreover, for vic𝒱Uv_{i}^{\text{c}}\in\mathcal{V}^{U}, when pθ(𝐲i|𝐘L,𝐗~)p_{\theta}(\mathbf{y}_{i}|\mathbf{Y}^{L},\tilde{\mathbf{X}}) is close to the one-hot categorical distribution, it would face the instability issue caused by logy\log y with yy approaching zero. Therefore, following several EM-based optimization methods gmnn ; text ; emgcn , we instead treat the fixed pθ(𝐘U|𝐘L,𝐗~)p_{\theta}(\mathbf{Y}^{U}|\mathbf{Y}^{L},\tilde{\mathbf{X}}) as the target and minimize the reverse KL divergence KL(pθ(𝐘U|𝐘L,𝐗~)qϕ(𝐘U|𝐗~))\operatorname{KL}(p_{\theta}(\mathbf{Y}^{U}|\mathbf{Y}^{L},\tilde{\mathbf{X}})\|q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})) to make qϕ(𝐘U|𝐗~)q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}}) approximate the target. Assuming that the labels of different cells are independent, which is reasonable as the organization and function of each cell are not determined by other cells, we can derive the following loss function for unlabeled cells:

EU=𝔼pθ(𝐘U|𝐘L,𝐗~)[logqϕ(𝐘U|𝐗~)]=vic𝒱U𝔼pθ(𝐲i|𝐘L,𝐗~)[logqϕ(𝐲i|𝐱~i)],\displaystyle\mathcal{L}_{E}^{U}=-\mathbb{E}_{p_{\theta}(\mathbf{Y}^{U}|\mathbf{Y}^{L},\tilde{\mathbf{X}})}\left[\log q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})\right]=-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{V}^{U}}\mathbb{E}_{p_{\theta}(\mathbf{y}_{i}|\mathbf{Y}^{L},\tilde{\mathbf{X}})}\left[\log q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i})\right], (6)

where pθ(𝐲i|𝐘L,𝐗~)=pθ(𝐲i|𝐗~)p_{\theta}(\mathbf{y}_{i}|\mathbf{Y}^{L},\tilde{\mathbf{X}})=p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}}) is the categorical distribution predicted by the cell-level GNN for vic𝒱Uv_{i}^{\text{c}}\in\mathcal{V}^{U} in the previous M-step. That is, the pseudo-labels predicted by the cell-level GNN are used for training the gene-level GNN. Additionally, the gene-level GNN can also be trained by the ground-truth labels of cells in 𝒱L\mathcal{V}^{L}. Therefore, we have the loss function for labeled cells:

EL=vic𝒱Llogqϕ(𝐲i|𝐱~i),\displaystyle\mathcal{L}_{E}^{L}=-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{V}^{L}}\log q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i}), (7)

where 𝐲i\mathbf{y}_{i} is the ground-truth label for cell vic𝒱Lv_{i}^{\text{c}}\in\mathcal{V}^{L}. Combining Eqs. (6) and (7), the overall objective in the E-step for optimizing ϕ\phi is to minimize E=EU+βEL\mathcal{L}_{E}=\mathcal{L}_{E}^{U}+\beta\mathcal{L}_{E}^{L} with β\beta being a balancing hyper-parameter (we set β=1\beta=1 in this work), which can be solved via stochastic gradient descent (SGD). To be more specific, in each update step of SGD, we sample a batch of labeled cells L\mathcal{B}^{L} and a batch of unlabeled cells U\mathcal{B}^{U} to perform gradient descent ϕϕγgϕ~E\phi\leftarrow\phi-\gamma_{\text{g}}\nabla_{\phi}\tilde{\mathcal{L}}_{E}, where ~E=vicUlogqϕ(𝐲^i|𝐱~i)βvicLlogqϕ(𝐲i|𝐱~i)\tilde{\mathcal{L}}_{E}=-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{B}^{U}}\log q_{\phi}(\hat{\mathbf{y}}_{i}|\tilde{\mathbf{x}}_{i})-\beta\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{B}^{L}}\log q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i}) with 𝐲^i\hat{\mathbf{y}}_{i} sampled by 𝐲^ipθ(𝐲i|𝐗~)\hat{\mathbf{y}}_{i}\sim p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}}) for unlabeled cells and γg\gamma_{\text{g}} is the learning rate.

3.4 M-step

In the M-step, the gene-level GNN (i.e., qϕq_{\phi}) is fixed and the cell-level GNN (i.e., pθp_{\theta}) is updated to maximize ELBO(θ,ϕ;𝐘L,𝐗~)\mathcal{L}_{\text{ELBO}}(\theta,\phi;\mathbf{Y}^{L},\tilde{\mathbf{X}}). Also, before we introduce the optimization of pθp_{\theta}, we present the detailed workflow of our cell-level GNN.

Structure of Cell-level GNN. Existing cell-level GNNs predetermine the cell-cell graph based on 𝐗~\tilde{\mathbf{X}} improving . By contrast, we construct the cell-cell graph by performing kk-nearest neighbors analysis on the cell representations {𝐡i}i=1N\{\mathbf{h}_{i}\}_{i=1}^{N} generated by the gene-level GNN in each EM iteration, i.e., 𝐀c=kNNgraph({𝐡i}i=1N)\mathbf{A}^{\text{c}}=\operatorname{kNN-graph}(\{\mathbf{h}_{i}\}_{i=1}^{N}), since {𝐡i}i=1N\{\mathbf{h}_{i}\}_{i=1}^{N} should have better clustering characteristics than 𝐗~\tilde{\mathbf{X}} for identifying cell types and its quality is enhanced during the optimization of qϕq_{\phi}.

Given the gene expression data 𝐱~i\tilde{\mathbf{x}}_{i} and the cell representation 𝐡i\mathbf{h}_{i}, the input feature for each cell vicv_{i}^{\text{c}} is defined as 𝐳i=Concat(𝐡i,MLP(𝐱~i))dc\mathbf{z}_{i}=\operatorname{Concat}(\mathbf{h}_{i},\operatorname{MLP}(\tilde{\mathbf{x}}_{i}))\in\mathbb{R}^{d_{c}}, where Concat()\operatorname{Concat}(\cdot) denotes the concatenation operation for the input vectors. By stacking the input features of all the cells, we have the input feature matrix 𝐙N×dc\mathbf{Z}\in\mathbb{R}^{N\times d_{c}}. And the structure of GNNc()\operatorname{GNN}^{\text{c}}(\cdot) can be modeled as:

R(l)(𝐀c)=𝐃1𝐀c,𝐃=diag(𝐀c𝟏N),\displaystyle\operatorname{R}^{(l)}\left(\mathbf{A}^{\text{c}}\right)=\mathbf{D}^{-1}\mathbf{A}^{\text{c}},\quad\mathbf{D}=\operatorname{diag}\left(\mathbf{A}^{\text{c}}\mathbf{1}_{N}\right),
𝐙(l)=MLPc(l)(R(l)(𝐀c)𝐙(l1)𝐖c(l)),\displaystyle\mathbf{Z}^{(l)}=\operatorname{MLP}_{\text{c}}^{(l)}\left(\operatorname{R}^{(l)}\left(\mathbf{A}^{\text{c}}\right)\mathbf{Z}^{(l-1)}\mathbf{W}_{\text{c}}^{(l)}\right), (8)

where 𝟏N\mathbf{1}_{N} is the all-ones vector of length NN, diag()\operatorname{diag}(\cdot) returns a diagonal matrix with the input vector as the diagonal, 𝐖c(l)\mathbf{W}_{\text{c}}^{(l)} is a learnable weight matrix, MLPc(l)()\operatorname{MLP}_{\text{c}}^{(l)}(\cdot) is an MLP, 𝐙(l)\mathbf{Z}^{(l)} is the output feature matrix of the ll-th layer, and 𝐙(0)=𝐙\mathbf{Z}^{(0)}=\mathbf{Z}. As for 𝐡i\mathbf{h}^{\prime}_{i} that is fed to an MLP classifier in Eq. (3.1), we find that setting 𝐡i=Concat(𝐳i(0),,𝐳i(L))\mathbf{h}^{\prime}_{i}=\operatorname{Concat}(\mathbf{z}_{i}^{(0)},\dots,\mathbf{z}_{i}^{(L)}) with 𝐳i(l)\mathbf{z}_{i}^{(l)} being the ii-th row of 𝐙(l)\mathbf{Z}^{(l)} practically achieves better classification performance for pθ(𝐲i|𝐗~)p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}}), which is similar to the jumping knowledge networks jknet .

Optimization of Cell-level GNN. Maximizing ELBO(θ,ϕ;𝐘L,𝐗~)\mathcal{L}_{\text{ELBO}}(\theta,\phi;\mathbf{Y}^{L},\tilde{\mathbf{X}}) in the M-step is equivalent to minimizing the following loss function:

M\displaystyle\mathcal{L}_{M} =𝔼qϕ(𝐘U|𝐗~)[logpθ(𝐘L,𝐘U|𝐗~)]\displaystyle=-\mathbb{E}_{q_{\phi}(\mathbf{Y}^{U}|\tilde{\mathbf{X}})}\left[\log p_{\theta}(\mathbf{Y}^{L},\mathbf{Y}^{U}|\tilde{\mathbf{X}})\right]
=vic𝒱U𝔼qϕ(𝐲i|𝐱~i)[logpθ(𝐲i|𝐗~)]vic𝒱Llogpθ(𝐲i|𝐗~).\displaystyle=-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{V}^{U}}\mathbb{E}_{q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i})}\left[\log p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}})\right]-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{V}^{L}}\log p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}}). (9)

Here, the first term is the loss for unlabeled cells, which trains pθp_{\theta} using the pseudo-labels generated by the gene-level GNN in the previous E-step. The second term is the loss for labeled cells, which is the standard supervised learning with ground-truth labels. Minimizing Eq. (3.4) w.r.t. pθp_{\theta} can also be solved via SGD. Concretely, in each update step of SGD, we sample 𝐲^iqϕ(𝐲i|𝐱~i)\hat{\mathbf{y}}_{i}\sim q_{\phi}(\mathbf{y}_{i}|\tilde{\mathbf{x}}_{i}) for unlabeled cells and perform gradient descent θθγcθ~M\theta\leftarrow\theta-\gamma_{\text{c}}\nabla_{\theta}\tilde{\mathcal{L}}_{M}, in which ~M=vic𝒱Ulogpθ(𝐲^i|𝐗~)vic𝒱Llogpθ(𝐲i|𝐗~)\tilde{\mathcal{L}}_{M}=-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{V}^{U}}\log p_{\theta}(\hat{\mathbf{y}}_{i}|\tilde{\mathbf{X}})-\sum\nolimits_{v_{i}^{\text{c}}\in\mathcal{V}^{L}}\log p_{\theta}(\mathbf{y}_{i}|\tilde{\mathbf{X}}) and γc\gamma_{\text{c}} is the learning rate.

3.5 Overall Optimization

To optimize the overall scBiGNN framework, we first pre-train the gene-level GNN with labeled cells to obtain the initial qϕq_{\phi}. Then, the EM algorithm alternately optimizes pθp_{\theta} (i.e., the cell-level GNN) and qϕq_{\phi} (i.e., the gene-level GNN) to reinforce each other. In the M-step, the cell representations generated by the gene-level GNN are used to construct the cell-cell graph, and the pseudo-labels predicted by qϕq_{\phi} together with the given cell labels are utilized to train the cell-level GNN. In the E-step, the pseudo-labels produced by pθp_{\theta} and the ground-truth cell labels are used to train the gene-level GNN, in which the gene-gene interactions are adaptively learned. After several EM iterations, the performance of both cell- and gene-level GNNs is enhanced in cell type classification.

4 Experiments

Table 1: Statistics of the five scRNA-seq benchmark datasets.
Dataset Description Protocol Species #Genes #Cells #Cell types
Zheng68K PBMC 10X Chromium Homo sapiens 20387 65943 11
Zhengsorted FACS-sorted PBMC 10X Chromium Homo sapiens 21952 20000 10
BaronHuman Human pancreas inDrop Homo sapiens 17499 8569 14
BaronMouse Mouse pancreas inDrop Mus musculus 14861 1886 13
AMB Primary mouse visual cortex SMART-Seq v4 Mus musculus 42625 12832 22
Table 2: Classification accuracy of all the methods on the five datasets.
Method Zheng68K Zhengsorted BaronHuman BaronMouse AMB
SVM 0.701 0.829 0.978 0.973 0.992
LDA 0.662 0.692 0.978 0.952 0.901
NMC 0.597 0.724 0.912 0.919 0.976
RF 0.674 0.796 0.962 0.968 0.985
scID 0.525 0.743 0.535 0.338 0.906
CHETAH 0.298 0.645 0.925 0.895 0.939
SingleR 0.673 0.718 0.968 0.908 0.962
ACTINN 0.724 0.825 0.977 0.978 0.992
GCN-C 0.710 0.831 0.978 0.977 0.992
scGraph 0.729 0.835 0.983 0.976 0.991
sigGCN 0.733 0.844 0.979 0.978 0.992
HNNVAT 0.734 0.846 0.982 0.979 0.992
scBiGNN (qϕq_{\phi}) 0.757 0.860 0.981 0.979 0.993
scBiGNN (pθp_{\theta}) 0.760 0.867 0.983 0.983 0.994

4.1 Benchmark Datasets

In this work, we conduct cell type classification experiments on five benchmark datasets to evaluate the performance of our scBiGNN, i.e., Zheng68K, Zhengsorted, BaronHuman, BaronMouse and AMB, which are widely used in many published papers for single-cell classification. These datasets can be directly downloaded from Zenodo (https://doi.org/10.5281/zenodo.3357167), and the detailed information is depicted in Table 1.

4.2 Implementations

Refer to caption
Figure 1: The convergence curves of qϕq_{\phi} and pθp_{\theta} on the Zhengsorted dataset.

As for the gene-level GNN module, we use a one-layer model with four attention heads for the BaronMouse dataset, while a two-layer model with a single head in each layer is employed for the other datasets. Additionally, for Zheng68K and Zhengsorted datasets, the read-out function is set as the simple average pooling operation, while for the others we find that using learnable αj\alpha_{j} in Eq. (5) performs better. As for the cell-level GNN module, a three-layer model is employed for cell graph representation learning. The feature dimension of each layer’s output is set as 32 for both GNN modules. All the MLPs used in scBiGNN have one hidden layer with 32 neurons and ReLU activation. The maximum number of EM iterations is set as 3, which is commonly enough for scBiGNN to converge, as shown in Figure 1.

4.3 Comparison Methods

We compare with a variety of baseline models to demonstrate the superior performance of scBiGNN. Specifically, following the previous study scgraph , eight methods are considered: support vector machine (SVM), linear discriminant analysis (LDA), nearest mean classifier (NMC), random forest (RF), scID scid , CHETAH chetah , SingleR singler and ACTINN actinn . In addition, three recent works that use gene-level GNNs are included: scGraph scgraph , sigGCN siggcn and HNNVAT adversarial . We also compare with the cell-level GNN, termed GCN-C, in which the cell-cell graph is constructed by principal component analysis and kk-nearest neighbors analysis on the raw gene expression data improving .

4.4 Classification Results

We benchmark scBiGNN against all the baselines by performing five-fold cross validation on each dataset. The results of cell type classification accuracy are summarized in Table 2, in which the best performance in each column is highlighted in bold. As can be seen, among all the baselines, graph-based deep learning methods are the most accurate classifiers, verifying that the biological structural information is beneficial to identifying cell types from scRNA-seq data. Out of all the graph-based models, our scBiGNN consistently achieves the highest accuracy on all the benchmark datasets. Noticeably, it outperforms all the baselines with more than 2.5% accuracy improvement on the largest scRNA-seq dataset (i.e., Zheng68K). Moreover, from Table 2 and Figure 1 we can observe that pθp_{\theta} always performs better than qϕq_{\phi}. The reason is that pθp_{\theta} explicitly integrates the bilevel graph representations of 𝐡i\mathbf{h}_{i} that summarizes gene-level structural information and 𝐀c\mathbf{A}^{\text{c}} that contains cell-level structural information. Overall, these results validate that the gene- and cell-level relationships are effectively mined by pθp_{\theta} of scBiGNN to facilitate scRNA-seq classification.

4.5 Network Analysis

Refer to caption
Figure 2: Network analysis of gene-gene and cell-cell graphs. (A) Venn plot of five important gene lists extracted from gene-gene interactions with five-fold cross validation on the BaronMouse dataset. (B) Ablation study on kk of cell-cell kkNN graphs. (C) Homophily analysis on the cell-cell graphs.

Gene-gene Graph. Our scBiGNN adaptively learns gene-gene interactions by the self-attention function R(l)(𝐀g)\operatorname{R}^{(l)}\left(\mathbf{A}^{\text{g}}\right) in Eq. (3.3). We integrate all the attention matrices into a gene-gene interaction matrix 𝐀~\tilde{\mathbf{A}} by taking the average across all the corresponding elements in all attention matrices for every gene-gene pair, where each element A~ij\tilde{A}_{ij} measures the attention of gene vigv_{i}^{\text{g}} paid to gene vjgv_{j}^{\text{g}}. Then, we can obtain the importance score sjs_{j} for each gene vjgv_{j}^{\text{g}}, which is defined as sj=iA~ijs_{j}=\sum_{i}\tilde{A}_{ij}. We sort the importance scores and get five lists of top 50 important genes from five-fold cross validation. Figure 2(A) shows that most genes appear more than once in the five lists, indicating that scBiGNN is trustworthy and effective in learning consistent important genes across different training procedures.

Cell-cell Graph. The cell-cell graph in scBiGNN is constructed as the kk-nearest neighbor (kkNN) graph based on the cell representations learned from qϕq_{\phi}. In Figure 2(B), we can observe that our model is insensitive to kk. A small value of kk (e.g., k=5)k=5) is enough for pθp_{\theta} to achieve good performance, verifying the high quality of cell representations generated by qϕq_{\phi} in finding close neighborhoods. In Figure 2(C), we measure the homophily ratio (i.e., the proportion of intra-class edges in all the edges) of cell-cell graphs constructed by scBiGNN and the raw data 𝐗~\tilde{\mathbf{X}} improving respectively. As can be seen, cell-cell graphs built by scBiGNN link more cells of the same type, which is critical to learning separable cell representations across different classes to ease GNN-based node classification emgcn .

5 Conclusion

In this paper, we propose a novel bilevel graph representation learning framework termed scBiGNN to improve GNN-based cell type classification in scRNA-seq analysis. scBiGNN consists of two GNN modules, namely gene- and cell-level GNNs, to adaptively extract structural information at both biological levels. The EM algorithm is employed to optimize the two GNN modules by alternating between the E-step and M-step. In each step, the pseudo-labels predicted by one module are used to train the other, making them gradually reinforce each other. Experiments on five scRNA-seq datasets verify the effectiveness of scBiGNN compared to competing cell type classification methods.

Acknowledgments and Disclosure of Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62250055, Grant 62371288, Grant 62320106003, Grant 61932022, Grant 61971285, Grant 62120106007, and in part by the Program of Shanghai Science and Technology Innovation Project under Grant 20511100100.

References

  • [1] Tamim Abdelaal, Lieke Michielsen, Davy Cats, Dylan Hoogduin, Hailiang Mei, Marcel JT Reinders, and Ahmed Mahfouz. A comparison of automatic cell identification methods for single-cell rna sequencing data. Genome biology, 20:1–19, 2019.
  • [2] Jose Alquicira-Hernandez, Anuja Sathe, Hanlee P Ji, Quan Nguyen, and Joseph E Powell. scpred: accurate supervised method for cell-type classification from single-cell rna-seq data. Genome biology, 20(1):1–17, 2019.
  • [3] Dvir Aran, Agnieszka P Looney, Leqian Liu, Esther Wu, Valerie Fong, Austin Hsu, Suzanna Chak, Ram P Naikawadi, Paul J Wolters, Adam R Abate, et al. Reference-based analysis of lung single-cell sequencing reveals a transitional profibrotic macrophage. Nature immunology, 20(2):163–172, 2019.
  • [4] Katerina Boufea, Sohan Seth, and Nizar N Batada. scid uses discriminant analysis to identify transcriptionally equivalent cell types across single-cell rna-seq data with batch effect. iScience, 23(3), 2020.
  • [5] David Buterez, Ioana Bica, Ifrah Tariq, Helena Andrés-Terré, and Pietro Liò. Cellvgae: an unsupervised scrna-seq analysis workflow with graph attention networks. Bioinformatics, 38(5):1277–1286, 2022.
  • [6] Krzysztof Marcin Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tamas Sarlos, Peter Hawkins, Jared Quincy Davis, Afroz Mohiuddin, Lukasz Kaiser, David Benjamin Belanger, Lucy J Colwell, and Adrian Weller. Rethinking attention with performers. In International Conference on Learning Representations, 2021.
  • [7] Jurrian K De Kanter, Philip Lijnzaad, Tito Candelli, Thanasis Margaritis, and Frank CP Holstege. Chetah: a selective, hierarchical cell type identification method for single-cell rna sequencing. Nucleic acids research, 47(16):e95–e95, 2019.
  • [8] Christopher T Fincher, Omri Wurtzel, Thom de Hoog, Kellie M Kravarik, and Peter W Reddien. Cell type transcriptome atlas for the planarian schmidtea mediterranea. Science, 360(6391):eaaq1736, 2018.
  • [9] Laleh Haghverdi, Aaron TL Lun, Michael D Morgan, and John C Marioni. Batch effects in single-cell rna-sequencing data are corrected by matching mutual nearest neighbors. Nature biotechnology, 36(5):421–427, 2018.
  • [10] Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186, 2019.
  • [11] Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations, 2017.
  • [12] Vladimir Yu Kiselev, Andrew Yiu, and Martin Hemberg. scmap: projection of single-cell rna-seq data across data sets. Nature methods, 15(5):359–362, 2018.
  • [13] Han-Bo Li, Ramon Viñas Torné, and Pietro Lio. Improving classification and data imputation for single-cell transcriptomics with graph neural networks. In NeurIPS 2022 AI for Science: Progress and Promises, 2022.
  • [14] Yuval Lieberman, Lior Rokach, and Tal Shay. Castle–classification of single cells by transfer learning: harnessing the power of publicly available single cell rna sequencing experiments to annotate new experiments. PloS one, 13(10):e0205499, 2018.
  • [15] Feiyang Ma and Matteo Pellegrini. Actinn: automated identification of cell types in single cell rna sequencing. Bioinformatics, 36(2):533–538, 2020.
  • [16] Anghui Peng, Xiying Mao, Jiawei Zhong, Shuxin Fan, and Youjin Hu. Single-cell multi-omics and its prospective application in cancer biology. Proteomics, 20(13):1900271, 2020.
  • [17] Mireya Plass, Jordi Solana, F Alexander Wolf, Salah Ayoub, Aristotelis Misios, Petar Glažar, Benedikt Obermayer, Fabian J Theis, Christine Kocks, and Nikolaus Rajewsky. Cell type atlas and lineage tree of a whole complex animal by single-cell transcriptomics. Science, 360(6391):eaaq1723, 2018.
  • [18] Meng Qu, Yoshua Bengio, and Jian Tang. Gmnn: Graph markov neural networks. In International Conference on Machine Learning, pages 5241–5250. PMLR, 2019.
  • [19] Jiahua Rao, Xiang Zhou, Yutong Lu, Huiying Zhao, and Yuedong Yang. Imputing single-cell rna-seq data by combining graph convolution and autoencoder neural networks. iScience, 24(5), 2021.
  • [20] Sara Sabour, Nicholas Frosst, and Geoffrey E Hinton. Dynamic routing between capsules. In Advances in Neural Information Processing Systems, volume 30, 2017.
  • [21] Damian Szklarczyk, Annika L Gable, David Lyon, Alexander Junge, Stefan Wyder, Jaime Huerta-Cepas, Milan Simonovic, Nadezhda T Doncheva, John H Morris, Peer Bork, et al. String v11: protein–protein association networks with increased coverage, supporting functional discovery in genome-wide experimental datasets. Nucleic acids research, 47(D1):D607–D613, 2019.
  • [22] Bram Van de Sande, Joon Sang Lee, Euphemia Mutasa-Gottgens, Bart Naughton, Wendi Bacon, Jonathan Manning, Yong Wang, Jack Pollard, Melissa Mendez, Jon Hill, et al. Applications of single-cell rna sequencing in drug discovery and development. Nature Reviews Drug Discovery, pages 1–25, 2023.
  • [23] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in Neural Information Processing Systems, volume 30, 2017.
  • [24] Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph attention networks. In International Conference on Learning Representations, 2018.
  • [25] Juexin Wang, Anjun Ma, Yuzhou Chang, Jianting Gong, Yuexu Jiang, Ren Qi, Cankun Wang, Hongjun Fu, Qin Ma, and Dong Xu. scgnn is a novel graph neural network framework for single-cell rna-seq analyses. Nature communications, 12(1):1882, 2021.
  • [26] Kangwei Wang, Zhengwei Li, Zhu-Hong You, Pengyong Han, and Ru Nie. Adversarial dense graph convolutional networks for single-cell classification. Bioinformatics, 39(2):btad043, 2023.
  • [27] Lifei Wang, Rui Nie, Zeyang Yu, Ruyue Xin, Caihong Zheng, Zhang Zhang, Jiang Zhang, and Jun Cai. An interpretable deep-learning architecture of capsule networks for identifying cell-type gene expression programs from single-cell rna-sequencing data. Nature Machine Intelligence, 2(11):693–703, 2020.
  • [28] Tianyu Wang, Jun Bai, and Sheida Nabavi. Single-cell classification using graph convolutional networks. BMC bioinformatics, 22(1):1–23, 2021.
  • [29] Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken-ichi Kawarabayashi, and Stefanie Jegelka. Representation learning on graphs with jumping knowledge networks. In International Conference on Machine Learning, pages 5453–5462. PMLR, 2018.
  • [30] Fan Yang, Wenchuan Wang, Fang Wang, Yuan Fang, Duyu Tang, Junzhou Huang, Hui Lu, and Jianhua Yao. scbert as a large-scale pretrained deep language model for cell type annotation of single-cell rna-seq data. Nature Machine Intelligence, 4(10):852–866, 2022.
  • [31] Rui Yang, Wenrui Dai, Chenglin Li, Junni Zou, and Hongkai Xiong. Tackling over-smoothing in graph convolutional networks with em-based joint topology optimization and node classification. IEEE Transactions on Signal and Information Processing over Networks, 9:123–139, 2023.
  • [32] Qijin Yin, Qiao Liu, Zhuoran Fu, Wanwen Zeng, Boheng Zhang, Xuegong Zhang, Rui Jiang, and Hairong Lv. scgraph: a graph neural network-based approach to automatically identify cell types. Bioinformatics, 38(11):2996–3003, 2022.
  • [33] Zhuohan Yu, Yifu Lu, Yunhe Wang, Fan Tang, Ka-Chun Wong, and Xiangtao Li. Zinb-based graph embedding autoencoder for single-cell rna-seq interpretations. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pages 4671–4679, 2022.
  • [34] Dai-Jun Zhang, Ying-Lian Gao, Jing-Xiu Zhao, Chun-Hou Zheng, and Jin-Xing Liu. A new graph autoencoder-based consensus-guided model for scrna-seq cell type detection. IEEE Transactions on Neural Networks and Learning Systems, 2022.
  • [35] Jianan Zhao, Meng Qu, Chaozhuo Li, Hao Yan, Qian Liu, Rui Li, Xing Xie, and Jian Tang. Learning on large-scale text-attributed graphs via variational inference. In International Conference on Learning Representations, 2022.