This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

ProteinRPN: Towards Accurate Protein Function Prediction with Graph-Based Region Proposals

Shania Mitra\equalcontrib1, Lei Huang\equalcontrib1, 2, Manolis Kellis1
Abstract

Protein function prediction is a crucial task in bioinformatics, with significant implications for understanding biological processes and disease mechanisms. While the relationship between sequence and function has been extensively explored, translating protein structure to function continues to present substantial challenges. Various models, particularly, CNN and graph-based deep learning approaches that integrate structural and functional data, have been proposed to address these challenges. However, these methods often fall short in elucidating the functional significance of key residues essential for protein functionality, as they predominantly adopt a retrospective perspective, leading to suboptimal performance.

Inspired by region proposal networks in computer vision, we introduce the Protein Region Proposal Network (ProteinRPN) for accurate protein function prediction. Specifically, the region proposal module component of ProteinRPN identifies potential functional regions (anchors) which are refined through the hierarchy-aware node drop pooling layer favoring nodes with defined secondary structures and spatial proximity. The representations of the predicted functional nodes are enriched using attention mechanisms and subsequently fed into a Graph Multiset Transformer, which is trained with supervised contrastive (SupCon) and InfoNCE losses on perturbed protein structures. Our model demonstrates significant improvements in predicting Gene Ontology (GO) terms, effectively localizing functional residues within protein structures. The proposed framework provides a robust, scalable solution for protein function annotation, advancing the understanding of protein structure-function relationships in computational biology.

Introduction

Advancements in genomics technology have illuminated the study of protein functions, enabling researchers to uncover the roles and interactions of proteins within living systems, making this a pivotal task in modern biology. Despite the vast number of proteins available, only a few of them have been reviewed by human curators. Among these reviewed proteins, less than 19.4% are substantiated by wet-lab experimental evidence (uni 2023). Precise functional annotations of proteins are crucial for tasks such as pinpointing drug targets, unraveling disease mechanisms, and enhancing biotechnological applications across industries (Kulmanov et al. 2024).

Currently, Gene Ontology (GO) (Aleksander et al. 2023; gen 2021) stands out as the most comprehensive resource, embodying all the essential attributes of an ideal functional classification system. The GO consortium delineates the functional attributes of genomic products, including genes, proteins, and RNA. Specifically, GO utilizes three subontologies to organize function terms according to each product: Biological Process (BP), Molecular Function (MF), and Cellular Component (CC). Although the UniProtKB/Swiss-Prot database records manually curated GO annotations that are verified by wet-lab experiments, there are still a significant number of protein sequences lacking functional annotations due to the high costs and limited throughput of experimental studies.

Fortunately, machine learning methods have emerged as a promising tool to address this challenge. Recently developed machine learning methods leverage different protein information for function prediction, including protein sequential information, protein tertiary structure, protein-protein interaction (PPI) networks, phylogenetic analysis, and literature information (You et al. 2018, 2021, 2019; Gligorijević et al. 2021; Lai and Xu 2022; Kulmanov and Hoehndorf 2022; Kulmanov, Khan, and Hoehndorf 2018; Pan et al. 2023; Kulmanov et al. 2024; Gu et al. 2023). Specifically, early studies focused on learning the similarities of homologous proteins by utilizing sequence alignment tools (Gong, Ning, and Tian 2016). This idea was then extended to harness additional protein information, such as PPI networks and biophysical properties, to predict protein function (Cho, Berger, and Peng 2016; You et al. 2021, 2019; Pan et al. 2023; Cho, Berger, and Peng 2016). However, the sequential similarity of proteins alone cannot fully determine protein function. Furthermore, these knowledge-based models heavily rely on selected features and cannot be generalized to new proteins due to the absence of prior knowledge.

Subsequent studies leverage primary sequence as the main feature for function prediction (Kulmanov, Khan, and Hoehndorf 2018; Kulmanov and Hoehndorf 2022). While the relationship between sequence and function has been extensively investigated, translating protein structure into function remains a significant challenge. Various models, notably CNNs and graph-based deep learning approaches that incorporate both structural and functional information, have been proposed to tackle these hurdles (Gligorijević et al. 2021; Lai and Xu 2022; Gu et al. 2023). However, these methods often fall short in elucidating the functional significance of key residues essential for protein functionality. Most of these approaches employ post-hoc techniques, such as Gradient-based Class Activation Maps (Gu et al. 2023; Gligorijević et al. 2021), to provide visual explanations of which residues contribute most to the predicted function. Yet, this retrospective analysis lacks biological insight, as it relies solely on what the model has learned during training without accounting for prior knowledge about functional residues. Moreover, these methods often result in a selection of numerous scattered residues with low specificity, diluting the focus on the truly important regions and leading to suboptimal performance.

To address these limitations, we introduce ProteinRPN, a novel model for accurate protein function prediction. ProteinRPN intentionally incorporates functional residue detection, enabling it to prioritize critical regions where groups of residues work together to perform specific functions. Inspired by region proposal networks in computer vision (Ren et al. 2015; Tang et al. 2018), ProteinRPN incorporates a graph-based Region Proposal submodule to identify potential functional regions within proteins. The model starts by detecting regions which contain functional residues, focusing on kk-hop subgraphs (anchors) surrounding each node. These identified functional regions are then refined harnessing a node drop pooling layer, which prioritizes nodes with defined secondary structures and spatial proximity, employing hierarchy-aware attention to assess functionality. The representations of these functional nodes are further enriched through a functional attention layer. Finally, the Graph Multiset Transformer (GMT) converts node-level representations into comprehensive graph-level embeddings, integrating locally emphasized interactions while preserving the global graph structure. Additionally, we utilize contrastive learning to generate similar representations for functionally related proteins while ensuring that distinct proteins have distinct representations.

The region proposal module is initially pretrained on the PDBSite dataset (Ivanisenko, Grigorovich, and Kolchanov 2000), containing functional residue annotations sourced from the Protein Data Bank (PDB) (Berman et al. 2000), which is a popular database known for its experimentally derived structural data on proteins. Then, we conduct experiments on the same dataset as baseline models (Gligorijević et al. 2021; You et al. 2021; Gu et al. 2023) for a fair comparison with baselines. The experimental results indicate significant improvements in predicting protein functions compared to state-of-the-art (SOTA) models. Remarkably, the proposed model achieves a \sim7% improvement in protein-centric Fmax on BP and MF ontologies compared to SOTA models. We also visualize the predicted functional residues, demonstrating that our model can identify essential functional structures and regions, which are meaningful for biological analysis.

Related Work

Computational methods have been proposed for protein function prediction, offering a more efficient and less resource-intensive alternative to wet-lab experimental assays. The task is framed as a multiclass multilabel classification problem, where each protein can be associated with multiple GO terms. Due to the hierarchical structure of GO terms within an ontology, predicting a given term also implies predicting all its ancestor terms, adding an additional layer of complexity. Early studies (Tian, Arakaki, and Skolnick 2004; Gong, Ning, and Tian 2016) leveraged query sequence-based Multiple Sequence Alignments (MSA) to predict protein GO terms. Based on the Position-Specific Scoring Matrix (PSSM), these models could identify query sequences that are more similar to sequences in the homo-functional MSA. Consequently, the protein sequence is more likely to be annotated with the target GO term.

Machine learning models have since emerged for more accurate protein function prediction by utilizing a broader range of biological features. Some methods (You et al. 2018, 2019) rely on external knowledge or even the hierarchical structure of GO terms, including GO term frequency, sequence alignment, amino acid trigram, domains and motifs, biophysical properties, and PPI networks. These approaches often employ a learning to rank (LTR) (Li 2011) framework for automatic function prediction. Sequence-based methods (Fa et al. 2018; Kulmanov, Khan, and Hoehndorf 2018; Wang et al. 2023) utilize sequential models like 1D CNNs and Transformers to derive protein sequence representations. Given that Graph Neural Networks (GNNs) are well-suited for learning the topology of PPI networks, subsequent studies (Zhao, Liu, and Wang 2022) have combined hybrid features from protein sequences and PPI networks, embedded using GNN modules, for function prediction.

Since protein structures determine essential biological and chemical properties (Jeffery 2023), relying exclusively on sequence-based methodologies may present a significant limitation. Therefore, several studies have incorporated protein structures for more accurate predictions (Gligorijević et al. 2021; Lai and Xu 2022; Gu et al. 2023). Specifically, these models derive contact maps from protein structures to construct residue graphs. Additionally, as protein amino acid sequences are similar to natural language sentences, recent studies (Gu et al. 2023) utilize advanced protein language models like ESM-1b (Rives et al. 2021) to obtain richer sequence representations. However, there remains a gap in models that accurately detect and predict constellations of amino acids in protein active sites and leverage these for structural and functional insights (Jeffery 2023).

Methodology

Refer to caption
Figure 1: The ProteinRPN model predicts protein function by converting protein sequences into residue graphs, processing them through a k-layer GCN to identify functional subgraphs (anchors), refining these subgraphs via domain knowledge and hierarchy-aware attention mechanisms, and categorizing them into GO terms using a GMT layer

In this section, we introduce ProteinRPN, a novel model for protein function prediction. As illustrated in Figure 1, ProteinRPN operates on protein graphs where nodes represent individual residues and edges are defined by the contact map which reflects residue proximity within the three-dimensional structure. The architecture is composed of three primary components. The first component, Region Proposal Network, is responsible for processing the protein graphs and proposing subgraphs which contain functionally relevant regions. These subgraphs are then fed through Functional Attention Layer which enhances the region proposals and selectively amplifies the representations of functional regions through a learned attention mechanism. The refined representations are subsequently passed to the Function Prediction block, consisting of a Graph Multiset Transformer (GMT) pooling layer and an MLP readout layer which generates predictions for GO terms. The entire framework is optimized through a combination of Supervised Contrastive (SupCon) loss and a self-supervised Information Noise-Contrastive Estimation (InfoNCE) loss, ensuring robust and effective protein representation learning.

Motivation

Our architecture is motivated by an analysis of 603 protein structures from PDBSite (Ivanisenko, Grigorovich, and Kolchanov 2000), which reveals that functional residues tend to cluster in three-dimensional space, even when they are not sequentially adjacent. Furthermore, in the studied sequences, each with hundreds of residues, the number of functional nodes ranged between 1 and 30. These observations, firstly, highlight the need to consider subgraphs, rather than individual nodes, in protein graphs, as protein function is influenced by the local environment and is usually carried by a cluster of residues, rather than isolated ones. It also suggests that aggressive pruning is necessary to accurately identify these few functional residues within graphs containing hundreds of nodes, necessitating a multi-stage pruning and refinement process. Finally, it is crucial that the pruning process preserves the subgraph structure, ensuring that the selected nodes form coherent clusters rather than being randomly scattered.

Preliminaries

Protein sequences are represented as graphs G(V,E)G(V,E), where the vertices VV correspond to the protein residues, and the edges EE represent the proximity of residues in three-dimensional space. The adjacency matrix AN×NA\in\mathbb{R}^{N\times N} for an NN-residue protein graph is defined by calculating the contact map, where an edge is added between two nodes if the distance between their CαC_{\alpha} atoms is less than 10 Å. In this work, we use G(V,E)G(V,E) and G(Z,A)G(Z,A) interchangeably, where VV and EE denote the set of vertices and edge list, while Z|V|×DZ\in\mathbb{R}^{|V|\times D} and A|V|×|V|A\in\mathbb{R}^{|V|\times|V|} represent the node features and adjacency matrices, respectively, and DD is the chosen dimension for residue features. The goal of ProteinRPN is to predict a probability vector 𝐲^i(j)lj\hat{\mathbf{y}}_{i}^{(j)}\in\mathbb{R}^{l_{j}}, where ljl_{j} denotes the number of GO terms associated with subontology j{BP,CC,MF}j\in\{\text{BP},\text{CC},\text{MF}\}. The vector 𝐲^i(j)\hat{\mathbf{y}}_{i}^{(j)} represents the predicted probabilities for the ljl_{j} GO terms, reflecting the likelihood of each protein being associated with multiple GO terms across all subontologies.

Residue Features

Residue features for NN nodes in any protein residue graph are derived through a two-step process. First, each node is assigned ESM-1b (Rives et al. 2021) embeddings ZEN×DEZ_{E}\in\mathbb{R}^{N\times D_{E}} to capture the intrinsic sequence-based information of the residues. In parallel, the residues are also label encoded according to their amino acid identities and transformed into embeddings ZRN×DRZ_{R}\in\mathbb{R}^{N\times D_{R}}. These two feature sets are subsequently projected onto a common DD-dimensional space and combined to form the final node embeddings Z=ZE+ZRN×DZ=Z_{E}+Z_{R}\in\mathbb{R}^{N\times D}, effectively integrating both deep sequence information and basic residue identity.

Enhanced Domain Knowledge

To enhance the graph representation with domain-specific knowledge, we further extract the secondary structure of each residue for each protein using DSSP (Dictionary of Secondary Structure in Proteins (Touw et al. 2015; Kabsch and Sander 1983), which is a database of secondary structure assignments for all protein entries in PDB (Berman et al. 2000). Experimental evidence suggests that functional residues are more likely to be found in regions with defined secondary structures, such as alpha helices and beta sheets (Bartlett et al. 2002). To align the residue coordinate information with the secondary structure data, we perform sequence alignments between DSSP-processed variants and residues with available PDB coordinates, addressing any discrepancies that arise between these data sources.

Functional Region Proposal Network

Inspired by object detection models in computer vision, we propose a strategy analogous to region proposal networks in Faster R-CNN (Ren et al. 2015), adapted for protein function prediction in graphs. By targeting regions containing functional residues, which are often a small subset of the protein, this approach improves functional understanding. To the best of our knowledge, this is the first work to introduce graph region proposals, applied specifically to protein function prediction.

The proposed Region Proposal Network employs kk layers of Graph Convolutional Networks (GCNs) (Kipf and Welling 2017) to process protein graphs G(Z,A)G(Z,A). In particular, let H(0)=ZH^{(0)}=Z represent the initial hidden node embedding matrix. The hidden embeddings HH are updated iteratively as:

H(i+1)=ReLU(D~0.5A~D~0.5H(i)W(i))H^{(i+1)}=\text{ReLU}\left(\tilde{D}^{-0.5}\tilde{A}\tilde{D}^{-0.5}H^{(i)}W^{(i)}\right)

where A~=A+I\tilde{A}=A+I is the adjacency matrix with self-loops included, and D~\tilde{D} is the diagonal degree matrix used for normalization. After kk message-passing layers, the final node embedding matrix Z1=H(k)N×D1Z_{1}=H^{(k)}\in\mathbb{R}^{N\times D_{1}} encapsulate information from their respective kk-hop neighborhoods, effectively extending each node’s receptive field to encompass its kk-hop subgraph. Each node can, now, be designated as the representative of its corresponding kk-hop subgraph, termed as an anchor. Consequently, this procedure transforms the original graph GG into a new graph G(Z1,A)G^{\prime}(Z_{1},A), where each node in GG^{\prime} corresponds to a subgraph in GG. Empirical results indicate that setting k=2k=2 is sufficient to capture functional residues within proteins.

The second step in the region proposal module involves localizing regions that are likely to contain functional residues. This is formulated as a node classification task, where the goal is to predict whether the anchor centred around each node contains a functionally relevant region. More precisely, given the node embeddings Z1Z_{1} after kk GCN layers, the classification of each node viv_{i} in the transformed graph G(Z1,A)G^{\prime}(Z_{1},A) is performed using a Graph Attention Network (GAT) convolution (Veličković et al. 2018). The output for each node viv_{i} can be formulated as:

y^i=σ(j𝒩(i)αijWHj(k))\hat{y}_{i}=\sigma\left(\sum_{j\in\mathcal{N}(i)}\alpha_{ij}{W}H_{j}^{(k)}\right)

where, y^i\hat{y}_{i} is the predicted probability that the node viv_{i} in GG^{\prime}, which represents the kk-hop subgraph SiS_{i} in GG, contains functional residues, α\alpha and WW are the attention scores and weight matrix, respectively, learnt by the GAT layer, 𝒩(i)\mathcal{N}(i), represents the neighbors of node ii in the graph G(Z1,A)G^{\prime}(Z_{1},A)

Nodes predicted as functional are selected, and kk-hop subgraphs (anchors) centered around these nodes are extracted. This results in a collection of anchors enriched for functional regions, ensuring high recall but with room for precision improvement. To address this, we introduce a pruning step that selectively retains the most functionally relevant subgraphs within the larger anchors. This pruning leverages a novel node-drop pooling layer that incorporates domain knowledge alongside a hierarchy-aware attention mechanism. Rather than relying on conventional dot product attention, we utilize penumbral cone attention (Tseng et al. 2023) for modeling the inherent hierarchical relationships in proteins. These hierarchies span multiple levels, from the arrangement of secondary and tertiary structures to the organization of functional domains and motifs, all the way up to the interactions of subunits within protein complexes. By capturing these complex dependencies, penumbral cone attention enables a more precise focus on critical functional regions within the protein graph, refining our predictions and improving precision without compromising recall.

Node drop pooling layers are commonly used to reduce graph size by selectively removing lower-scoring nodes while retaining higher-scoring ones based on their importance or features. Instead of removing low-scoring nodes, we evaluate the functionality of each node in the Node Drop Pooling layer. Subsequently, we leverage the Functional Attention Layer to enrich representations of high-scoring nodes, ensuring that lower-scoring nodes contributing valuable contextual information are retained to preserve the overall graph structure.

Node Drop Pooling/Node Scoring

In order to obtain scores in our case, the feature embeddings extracted from these subgraphs are passed through GCN layers to obtain query and key representations.

q=LeakyReLU(GCN1(G(Z1,A))q=\text{LeakyReLU}(GCN_{1}(G(Z_{1},A))
k=LeakyReLU(GCN2(G(Z1,A))k=\text{LeakyReLU}(GCN_{2}(G(Z_{1},A))

These representations are then fed into a hierarchy-aware attention layer to decide which nodes to prune.

Penumbral Cone Attention

We employ cone attention (Gulcehre et al. 2018; Tseng et al. 2023) which serves as a seamless alternative to dot product attention, relying on hyperbolic entailment cones to model the hierarchies between the residue nodes. Specifically, we utilize hyperbolic distance attention, which defines the similarity as S(qi,ki)=exp(βd(qi,ki)c)S\left(q_{i},k_{i}\right)=\exp\left(-\beta d_{\mathbb{H}}\left(q_{i},k_{i}\right)-c\right) where dd_{\mathbb{H}} is the hyperbolic distance. Following the previous work (Tseng et al. 2023), we use the Poincaré half-space model to calculate the hyperbolic distance.

Proximity Scores

To compute proximity scores, we first measure the pairwise distances between each residue and all other residues, akin to constructing a contact map. Rather than applying a threshold to these distances, the proximity score PiP_{i} for residue ii is computed by summing the inverse distances between residue ii and all other residues jj, i.e, Pi=αpsji1dijP_{i}=\alpha_{ps}\sum_{j\neq i}\frac{1}{d_{ij}}, where dijd_{ij} represents the distance between residues ii and jj, and αps\alpha_{ps} is a scaling factor that determines the influence of proximity on the final node score. This method prioritizes residues that are closely clustered with a few others, resulting in higher scores compared to residues that are moderately close to many others, aligning with our insights from PDBSite (Ivanisenko, Grigorovich, and Kolchanov 2000).

Secondary Structure Scores

Certain functional residues have been observed to preferentially reside in regions of defined secondary structure. For instance, Bartlett et al. (Bartlett et al. 2002) reports that catalytic residues are frequently located in alpha helices (39%) and beta sheets (28%), with a lower prevalence in loops and unstructured regions. To reflect this, we assign higher predicted scores to residues within alpha helices and beta sheets.

The final node scores, derived from the combination of the three components, are converted into probabilities using a sigmoid function. Residues with the highest probabilities are identified as functional for subsequent processing.

Functional Attention Layer

Once candidate functional residues are identified, their representations are refined through a functional attention layer. This layer assigns weights to edges based on their connectivity to predicted functional nodes, allowing the model to emphasize relationships critical to protein function. By incorporating multistage refinement, we iteratively enhance the accuracy of functional node identification. The edge-centric approach helps preserve the structural integrity of selected subgraphs, avoiding the fragmentation that can occur when individual nodes are selected in isolation, in line with insights from PDBSite.

We feed the original residue features ZN×DZ\in\mathbb{R}^{N\times D} as the node feature matrix for enrichment. For each edge (i,j)(i,j) in the graph, the model computes an attention score eij{e}_{ij} using the concatenation of the feature vectors Zi{Z}_{i} and ZjZ_{j}, followed by a learnable weight vector a2D1×1{a}\in\mathbb{R}^{2D_{1}\times 1} and a ReLU activation function, that would reduce all negative scores to zero, i.e., eij=ReLU(a[ZiZj]){e}_{ij}=\text{ReLU}\left({a}^{\top}[{Z}_{i}\,\|\,{Z}_{j}]\right). This attention score is then adjusted based on the node type zj{0,1}{z}_{j}\in\{0,1\} of the target node jj, modifying the score as follows:

eij=αFAeijzj+βFAeij(1zj){e}_{ij}=\alpha_{FA}\cdot{e}_{ij}\cdot{z}_{j}+\beta_{FA}\cdot{e}_{ij}\cdot(1-{z}_{j})\cdot

where αFA1\alpha_{FA}\geq 1 and βFA<1\beta_{FA}<1. This adjustment increases the attention for functional nodes while reducing it for contextual ones. For the purpose if this study, we use αFA=1,βFA=0.5\alpha_{FA}=1,\beta_{FA}=0.5 in order to explicitly ensure focus on functional nodes. The attention coefficients αij\alpha_{ij} are obtained by normalizing eije_{ij} across all neighbors. They determine how much influence a neighboring node ii has on the target node jj.

Finally, the updated feature vector for node jj, Z2j{Z}_{2j}, is computed by aggregating the messages from its neighbors applying, weighted by the corresponding attention coefficients, i.e., Z2j=i𝒩(j)αijWZi{Z_{2j}}=\sum_{i\in\mathcal{N}(j)}\alpha_{ij}\cdot W\cdot{Z}_{i}, where the transformation matrix WD2×DW\in\mathbb{R}^{D_{2}\times D} is a learnable parameter as in the GAT layer and helps transform the initial extracted features of the nodes (residues) into a new space where relationships between residues can be more effectively captured. As a result of this operation, subgraphs surrounding functional residues—those likely to be critical for protein function—get more attention and influence the final node representations more significantly. This approach enhances the model’s ability to capture the rich, context-aware interactions between residues, leading to a more comprehensive understanding of the protein’s functional regions.

Graph Multiset Transformer

In the final step, the enriched representations Z2Z_{2} are fed into a Graph Multiset Transformer (GMT) layer, which transforms node-level embeddings into a comprehensive graph-level representation by capturing both local interactions and global structure. The GMT layer introduces learnable super-nodes to capture long-distance structural information and aggregates this information into a unified graph representation.

Optimization Framework

Our model is optimized using a multi-component loss function that integrates cross-entropy loss CE\mathcal{L}_{\text{CE}} for multilabel classification, contrastive loss con\mathcal{L}_{\text{con}}, and a penalty term penalty\mathcal{L}_{\text{penalty}} to minimize the number of disconnected components in the functional attention layer.

The contrastive loss, con\mathcal{L}_{\text{con}}, is a combination of supervised contrastive (SupCon) loss (Khosla et al. 2021) and self-supervised noise contrastive estimation (InfoNCE) loss (van den Oord, Li, and Vinyals 2019). SupCon encourages the model to cluster representations of proteins with similar GO terms, while InfoNCE ensures that the representations are robust to noise by maximizing the similarity between original and perturbed embeddings. The combined contrastive loss for a batch of BB proteins is defined as:

con=\displaystyle\mathcal{L}_{\text{con}}=\ (1Bi=1Bji𝟏{yiyj}logexp(sim(zi,zj)/τ)kiexp(sim(zi,zk)/τ))\displaystyle\left(-\frac{1}{B}\sum_{i=1}^{B}\sum_{j\neq i}\mathbf{1}\{y_{i}\cap y_{j}\neq\emptyset\}\cdot\log\frac{\exp(\text{sim}(z_{i},z_{j})/\tau)}{\sum_{k\neq i}\exp(\text{sim}(z_{i},z_{k})/\tau)}\right)
αSupCon+(logexp(sim(zi,zi)/τ)zjexp(sim(zi,zj)/τ))αNCE\displaystyle\cdot\alpha_{\text{SupCon}}+\left(-\log\frac{\exp(\text{sim}(z_{i},z_{i}^{\prime})/\tau)}{\sum_{z_{j}}\exp(\text{sim}(z_{i},z_{j})/\tau)}\right)\alpha_{\text{NCE}}

where ziz_{i} and zjz_{j} are the embeddings of proteins ii and jj, sim(zi,zj)\text{sim}(z_{i},z_{j}) represents their cosine similarity, and τ\tau is a temperature parameter. The indicator function 𝟏{yiyj}\mathbf{1}\{y_{i}\cap y_{j}\neq\emptyset\} ensures that only pairs with shared GO terms contribute to the SupCon loss, im order to adapt it to the multilabel case. The InfoNCE loss optimizes the similarity between the original and perturbed embeddings ziz_{i} and ziz_{i}^{\prime}. The hyperparameters αSupCon\alpha_{\text{SupCon}} and αNCE\alpha_{\text{NCE}} control the contributions of the SupCon and InfoNCE losses.

To ensure that identified functional regions are structurally cohesive, reflecting the biological reality of interconnected functional residues, we introduce a connected components penalty. This discourages the formation of disconnected components in the functional attention layer and is defined as:

penalty=αccκ(Gi′′)j=1Myij\mathcal{L}_{\text{penalty}}=\alpha_{cc}\cdot\frac{\kappa(G^{\prime\prime}_{i})}{\sum_{j=1}^{M}y_{ij}}

where Gi′′=(Vf,Ef)G^{\prime\prime}_{i}=(V_{f},E_{f}) denotes the subgraph of the ii-th protein graph GiG_{i} induced by functional nodes, and κ(Gi′′)\kappa(G^{\prime\prime}_{i}) represents the number of connected components in the subgraph. Finally, αcc\alpha_{cc} is a hyperparameter that controls the strength of the penalty.

The overall loss function is formulated as:

=CE+con+penalty\mathcal{L}=\mathcal{L}_{\text{CE}}+\mathcal{L}_{\text{con}}+\mathcal{L}_{\text{penalty}}

This comprehensive loss function guides the model toward producing biologically meaningful predictions by leveraging robust, context-aware graph representations while maintaining the structural coherence of the predicted functional subgraphs.

Experiments

In this section, we elabroate on the datasets, training setup, and evaluation criteria used for training the model.

Datasets

PDBSite (Ivanisenko, Grigorovich, and Kolchanov 2000) is a comprehensive dataset comprising biologically active sites derived from the Protein Data Bank (PDB) (Berman et al. 2000). The dataset encompasses 4,723 active sites belonging to 197 different functions located within 603 proteins. PDBSite stands out among annotation databases due to its diverse representation of functional categories, enabling broad analysis across various protein functions. We leverage PDBSite to guide our model architecture and pretrain the model on predicting functional sites.

For protein function prediction, we utilize a dataset curated by (Gu et al. 2023), originally developed to train their model, HEAL, which serves as our baseline. This dataset is an adapted version of the DeepFRI dataset (Gligorijević et al. 2021), comprising 36,629 sequences sourced from the PDB database (Berman et al. 2000) and 42,994 from the SWISS-MODEL repository (Bienert et al. 2016). Further details can be found in the Appendix.

Experimental Setup

We begin by training ProteinRPN on the PDBSite which is split into training and validation sets with an 80:20 ratio, with the goal of predicting all functional sites within a protein. The details of the pretraining can be found in the Appendix. Then we train the entire framework on the comprehensive protein function prediction task using the HEAL dataset.

We have conducted comprehensive experimennts to comapre ProteinRPN’s [erformance to SOTA models. Those methods encompass sequence-based models such as BLAST (Altschul et al. 1990) and FunFams (Das et al. 2015), sequence and PPI-based models like DeepGO (Kulmanov, Khan, and Hoehndorf 2017), and sequence and structure-based models such as DeepFRI (Gligorijević et al. 2021) and HEAL (Gu et al. 2023).

We conduct ablation studies to assess the significance of each model component, including the impact of secondary structure, coordinate information, and contrastive learning losses. Additional studies to test the efficacy of other modules can be found in the Appendix.

Model predictions are evaluated using the standard Critical Assessment of Functional Annotation (CAFA) evaluator (Jiang et al. 2016). Protein-centric Fmax, the maximum F1 score over all prediction thresholds ranging from 0 to 1 with a step size of 0.1, is utilized. Smin, representing the semantic distance between predicted and actual annotations, considers the information content of each function. The function-centric AUPR is employed as a robust measure for situations with high class imbalance. Further details on the formulas and implementation are available in (Jiang et al. 2016), and comprehensive information on model training and hyperparameters can be found in the Appendix.

Results and Analysis

Method Fmax (\uparrow) AUPR (\uparrow) Smin (\downarrow)
BP CC MF BP CC MF BP CC MF
Blast 0.336 0.448 0.328 0.067 0.097 0.136 0.651 0.628 0.632
FunFams 0.500 0.627 0.572 0.260 0.288 0.367 0.579 0.503 0.531
DeepGO 0.493 0.594 0.577 0.182 0.263 0.391 0.577 0.550 0.472
DeepFRI 0.540 0.613 0.625 0.261 0.274 0.495 0.543 0.527 0.437
HEAL 0.581 0.673 0.708 0.298 0.415 0.630 0.504 0.462 0.369
ProteinRPN 0.618 0.691 0.754 0.344 0.459 0.683 0.495 0.458 0.335
Table 1: Baseline Comparison: Fmax, AUPR, and Smin of different methods on the designated test set; best performances are highlighted in bold, i.e., for Fmax and AUPR, we consider the highest, while for Smin we consider the lowest value

GO term Prediction

Table 1 presents the performance metrics of ProteinRPN in comparison to all baseline models on the HEAL dataset. ProteinRPN consistently outperforms the baselines across all metrics, showing notable improvements over the HEAL model. Specifically, ProteinRPN achieves higher Fmax scores, with gains of 6.4% in Biological Process (BP), 2.7% in Cellular Component (CC), and 7.1% in Molecular Function (MF) ontologies. Beyond Fmax, ProteinRPN also demonstrates superior performance in Smin and Area Under the Precision-Recall Curve (AUPR), highlighting its effectiveness in predicting protein function GO terms.

Moreover, as shown in Table 2, the ablation study reveals that both contrastive learning and the incorporation of domain knowledge positively contribute to the model’s overall performance.

During pretraining, the region proposal module exhibits strong performance, achieving an ROC of 0.95 in the anchor functionality prediction task and 0.85 in the pruning task. Although direct comparison is limited due to the absence of established baselines, the module’s effectiveness is evident in downstream functional prediction tasks.

Overall, the results demonstrate that enabling the model to detect and focus on residues critical for function significantly enhances its performance. This improvement is primarily driven by the multistage refinement approach, which efficiently localizes functional residues within protein structures.

Model Fmax (\uparrow) AUPR (\uparrow) Smin (\downarrow)
BP CC MF BP CC MF BP CC MF
ProteinRPN CL 0.6175 0.6906 0.7542 0.3438 0.4527 0.6833 0.4948 0.4576 0.3350
ProteinRPN w/o CL 0.6009 0.6878 0.7408 0.3223 0.4166 0.6479 0.5062 0.4587 0.3557
ProteinRPN w/o SS w CL 0.6114 0.6894 0.7498 0.3426 0.4591 0.6778 0.4984 0.4576 0.3421
ProteinRPN w/o SS w/o CL 0.5975 0.6801 0.7364 0.3161 0.4242 0.6446 0.5088 0.4674 0.3547
Table 2: Ablation Studies: Fmax, AUPR, and Smin of different variants of ProteinRPN, where CL: Contrastive Learning, SS: secondary structure and proximity scoring; best performances are highlighted in bold, i.e., for Fmax and AUPR, we consider the highest, while for Smin we consider the lowest value. On removing the Contrastive Learning module, there is a moderate decrease in performance across all three GO domains; further, removing Domain Knowledge from the node pooling layer is also seen to impact performance negatively.
Refer to caption
Refer to caption
Figure 2: Visual Demonstration of Region Proposal Network detected residues in proteins (a) 2BCC-B and (b) 2CHG-A

Functional Residue Visualization

We evaluate the functional residue predictor in ProteinRPN by analyzing specific proteins. For example, on protein 2BCC (B chain, 422 residues, 10 functional), ProteinRPN accurately identifies 8 functional residues, with region proposals covering subgraphs of 28 residues, as shown in Fig. 2(a). Functional residues predicted correctly are highlighted in green, while missed ones are marked in red.

Similarly, for protein 2CHG (A chain, 226 residues, 11 functional), the model successfully identifies 9 functional residues, with region proposals covering 43 residues. As shown in Fig. 2(b), the correctly identified residues are closely clustered within the structure, while the missed residues are located farther from the cluster. These results demonstrate its ability to accurately identify and localize constellations of functional residues.

Conclusion

In this work, we introduced ProteinRPN, a novel graph-based model equipped with graph region proposal networks which is designed to identify and refine functional regions within protein residue graphs. By leveraging hierarchical attention mechanisms, domain-specific knowledge, and multistage refinement, through a combination of supervised contrastive learning and self-supervised InfoNCE loss, ProteinRPN significantly improves the accuracy of protein function prediction across GO terms. Our results demonstrate substantial gains over SOTA methods, with enhanced precision in identifying functional residues and preserving structural integrity in predicted subgraphs.

While our model provides generalized insights across a range of protein functions, the current analysis is based on a limited set of protein structures. Future work will focus on extending the model’s capabilities by incorporating diverse knowledge sources and exploring additional mechanisms to further enhance the accuracy and scalability of protein function prediction.

References

  • gen (2021) 2021. The Gene Ontology resource: enriching a GOld mine. Nucleic acids research, 49(D1): D325–D334.
  • uni (2023) 2023. UniProt: the universal protein knowledgebase in 2023. Nucleic acids research, 51(D1): D523–D531.
  • Aleksander et al. (2023) Aleksander, S. A.; Balhoff, J.; Carbon, S.; Cherry, J. M.; Drabkin, H. J.; Ebert, D.; Feuermann, M.; Gaudet, P.; Harris, N. L.; et al. 2023. The gene ontology knowledgebase in 2023. Genetics, 224(1): iyad031.
  • Altschul et al. (1990) Altschul, S. F.; Gish, W.; Miller, W.; Myers, E. W.; and Lipman, D. J. 1990. Basic local alignment search tool. Journal of Molecular Biology, 215(3): 403–410.
  • Bartlett et al. (2002) Bartlett, G. J.; Porter, C. T.; Borkakoti, N.; and Thornton, J. M. 2002. Analysis of Catalytic Residues in Enzyme Active Sites. Journal of Molecular Biology, 324(1): 105–121.
  • Berman et al. (2000) Berman, H. M.; Westbrook, J.; Feng, Z.; Gilliland, G.; Bhat, T. N.; Weissig, H.; Shindyalov, I. N.; and Bourne, P. E. 2000. The Protein Data Bank. Nucleic Acids Research, 28(1): 235–242.
  • Bienert et al. (2016) Bienert, S.; Waterhouse, A.; deBeer, T. A.; Tauriello, G.; Studer, G.; Bordoli, L.; and Schwede, T. 2016. The SWISS-MODEL Repository—new features and functionality. Nucleic Acids Research, 45(D1): D313–D319.
  • Cho, Berger, and Peng (2016) Cho, H.; Berger, B.; and Peng, J. 2016. Compact integration of multi-network topology for functional analysis of genes. Cell systems, 3(6): 540–548.
  • Das et al. (2015) Das, S.; Lee, D.; Sillitoe, I.; Dawson, N. L.; Lees, J. G.; and Orengo, C. A. 2015. Functional classification of CATH superfamilies: a domain-based approach for protein function annotation. Bioinformatics, 31(21): 3460–3467.
  • Fa et al. (2018) Fa, R.; Cozzetto, D.; Wan, C.; and Jones, D. T. 2018. Predicting human protein function with multi-task deep neural networks. PloS one, 13(6): e0198216.
  • Gligorijević et al. (2021) Gligorijević, V.; Renfrew, P. D.; Kosciolek, T.; Leman, J. K.; Berenberg, D.; Vatanen, T.; Chandler, C.; Taylor, B. C.; Fisk, I. M.; Vlamakis, H.; et al. 2021. Structure-based protein function prediction using graph convolutional networks. Nature communications, 12(1): 3168.
  • Gong, Ning, and Tian (2016) Gong, Q.; Ning, W.; and Tian, W. 2016. GoFDR: a sequence alignment based method for predicting protein functions. Methods, 93: 3–14.
  • Gu et al. (2023) Gu, Z.; Luo, X.; Chen, J.; Deng, M.; and Lai, L. 2023. Hierarchical graph transformer with contrastive learning for protein function prediction. Bioinformatics, 39(7): btad410.
  • Gulcehre et al. (2018) Gulcehre, C.; Denil, M.; Malinowski, M.; Razavi, A.; Pascanu, R.; Hermann, K. M.; Battaglia, P.; Bapst, V.; Raposo, D.; Santoro, A.; et al. 2018. Hyperbolic attention networks. arXiv preprint arXiv:1805.09786.
  • Ivanisenko, Grigorovich, and Kolchanov (2000) Ivanisenko, V.; Grigorovich, D.; and Kolchanov, N. 2000. PDBSite: a database on biologically active sites and their spatial surroundings in proteins with known tertiary structure. In The Second International Conference on Bioinformatics of Genome Regulation and Structure (BGRS’2000), volume 2, 171–174. Novosibirsk, Russia.
  • Jeffery (2023) Jeffery, C. J. 2023. Current successes and remaining challenges in protein function prediction. Frontiers in Bioinformatics, 3.
  • Jiang et al. (2016) Jiang, Y.; Oron, T. R.; Clark, W. T.; et al. 2016. An expanded evaluation of protein function prediction methods shows an improvement in accuracy. Genome Biology, 17(1): 184.
  • Kabsch and Sander (1983) Kabsch, W.; and Sander, C. 1983. Dictionary of protein secondary structure: pattern recognition of hydrogen-bonded and geometrical features. Biopolymers, 22: 2577–2637.
  • Khosla et al. (2021) Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; and Krishnan, D. 2021. Supervised Contrastive Learning. arXiv:2004.11362.
  • Kipf and Welling (2017) Kipf, T. N.; and Welling, M. 2017. Semi-Supervised Classification with Graph Convolutional Networks. arXiv:1609.02907.
  • Kulmanov et al. (2024) Kulmanov, M.; Guzmán-Vega, F. J.; Duek Roggli, P.; Lane, L.; Arold, S. T.; and Hoehndorf, R. 2024. Protein function prediction as approximate semantic entailment. Nature Machine Intelligence, 6(2): 220–228.
  • Kulmanov and Hoehndorf (2022) Kulmanov, M.; and Hoehndorf, R. 2022. DeepGOZero: improving protein function prediction from sequence and zero-shot learning based on ontology axioms. Bioinformatics, 38(Supplement_1): i238–i245.
  • Kulmanov, Khan, and Hoehndorf (2017) Kulmanov, M.; Khan, M. A.; and Hoehndorf, R. 2017. DeepGO: predicting protein functions from sequence and interactions using a deep ontology-aware classifier. Bioinformatics, 34(4): 660–668.
  • Kulmanov, Khan, and Hoehndorf (2018) Kulmanov, M.; Khan, M. A.; and Hoehndorf, R. 2018. DeepGO: predicting protein functions from sequence and interactions using a deep ontology-aware classifier. Bioinformatics, 34(4): 660–668.
  • Lai and Xu (2022) Lai, B.; and Xu, J. 2022. Accurate protein function prediction via graph attention networks with predicted structure information. Briefings in Bioinformatics, 23(1): bbab502.
  • Li (2011) Li, H. 2011. A short introduction to learning to rank. IEICE TRANSACTIONS on Information and Systems, 94(10): 1854–1862.
  • Pan et al. (2023) Pan, T.; Li, C.; Bi, Y.; Wang, Z.; Gasser, R. B.; Purcell, A. W.; Akutsu, T.; Webb, G. I.; Imoto, S.; and Song, J. 2023. PFresGO: an attention mechanism-based deep-learning approach for protein annotation by integrating gene ontology inter-relationships. Bioinformatics, 39(3): btad094.
  • Ren et al. (2015) Ren, S.; He, K.; Girshick, R.; and Sun, J. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28.
  • Rives et al. (2021) Rives, A.; Meier, J.; Sercu, T.; Goyal, S.; Lin, Z.; Liu, J.; Guo, D.; Ott, M.; Zitnick, C. L.; Ma, J.; et al. 2021. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proceedings of the National Academy of Sciences, 118(15): e2016239118.
  • Tang et al. (2018) Tang, P.; Wang, X.; Wang, A.; Yan, Y.; Liu, W.; Huang, J.; and Yuille, A. 2018. Weakly supervised region proposal network and object detection. In Proceedings of the European conference on computer vision (ECCV), 352–368.
  • Tian, Arakaki, and Skolnick (2004) Tian, W.; Arakaki, A. K.; and Skolnick, J. 2004. EFICAz: a comprehensive approach for accurate genome-scale enzyme function inference. Nucleic acids research, 32(21): 6226–6239.
  • Touw et al. (2015) Touw, W. G.; Baakman, C.; Black, J.; te Beek, T. A. H.; Krieger, E.; Joosten, R. P.; and Vriend, G. 2015. A series of PDB related databases for everyday needs. Nucleic Acids Research, 43(Database issue): D364–D368.
  • Tseng et al. (2023) Tseng, A.; Yu, T.; Liu, T.; and De Sa, C. M. 2023. Coneheads: Hierarchy Aware Attention. In Oh, A.; Naumann, T.; Globerson, A.; Saenko, K.; Hardt, M.; and Levine, S., eds., Advances in Neural Information Processing Systems, volume 36, 51421–51433. Curran Associates, Inc.
  • van den Oord, Li, and Vinyals (2019) van den Oord, A.; Li, Y.; and Vinyals, O. 2019. Representation Learning with Contrastive Predictive Coding. arXiv:1807.03748.
  • Veličković et al. (2018) Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; and Bengio, Y. 2018. Graph Attention Networks. arXiv:1710.10903.
  • Wang et al. (2023) Wang, S.; You, R.; Liu, Y.; Xiong, Y.; and Zhu, S. 2023. NetGO 3.0: protein language model improves large-scale functional annotations. Genomics, Proteomics & Bioinformatics, 21(2): 349–358.
  • You et al. (2021) You, R.; Yao, S.; Mamitsuka, H.; and Zhu, S. 2021. DeepGraphGO: graph neural network for large-scale, multispecies protein function prediction. Bioinformatics, 37(Supplement_1): i262–i271.
  • You et al. (2019) You, R.; Yao, S.; Xiong, Y.; Huang, X.; Sun, F.; Mamitsuka, H.; and Zhu, S. 2019. NetGO: improving large-scale protein function prediction with massive network information. Nucleic acids research, 47(W1): W379–W387.
  • You et al. (2018) You, R.; Zhang, Z.; Xiong, Y.; Sun, F.; Mamitsuka, H.; and Zhu, S. 2018. GOLabeler: improving sequence-based large-scale protein function prediction by learning to rank. Bioinformatics, 34(14): 2465–2473.
  • Zhao, Liu, and Wang (2022) Zhao, C.; Liu, T.; and Wang, Z. 2022. PANDA2: protein function prediction using graph neural networks. NAR genomics and bioinformatics, 4(1): lqac004.