This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Polyp-PVT: Polyp Segmentation with Pyramid Vision Transformers

Bo Dong, Wenhai Wang, Deng-Ping Fan, Jinpeng Li, Huazhu Fu, and Ling Shao B. Dong and D.-P. Fan are with the Nankai University, Tianjin, 300350, China. * Corresponding author ([email protected]). W. Wang is with Shanghai Artificial Intelligence Laboratory, Shanghai, 200232, ChinaJ. Li is with Computer Vision Lab, Inception Institute of Artificial Intelligence, Abu Dhabi 00000, UAEH. Fu is with Institute of High Performance Computing, Agency for Science, Technology and Research, Singapore 138632, SingaporeL. Shao is with UCAS-Terminus AI Lab, Terminus Group, Chongqing 400042, China
Abstract

Most polyp segmentation methods use CNNs as their backbone, leading to two key issues when exchanging information between the encoder and decoder: 1) taking into account the differences in contribution between different-level features and 2) designing an effective mechanism for fusing these features. Unlike existing CNN-based methods, we adopt a transformer encoder, which learns more powerful and robust representations. In addition, considering the image acquisition influence and elusive properties of polyps, we introduce three standard modules, including a cascaded fusion module (CFM), a camouflage identification module (CIM), and a similarity aggregation module (SAM). Among these, the CFM is used to collect the semantic and location information of polyps from high-level features; the CIM is applied to capture polyp information disguised in low-level features, and the SAM extends the pixel features of the polyp area with high-level semantic position information to the entire polyp area, thereby effectively fusing cross-level features. The proposed model, named Polyp-PVT, effectively suppresses noises in the features and significantly improves their expressive capabilities. Extensive experiments on five widely adopted datasets show that the proposed model is more robust to various challenging situations (e.g., appearance changes, small objects, rotation) than existing representative methods. The proposed model is available at https://github.com/DengPingFan/Polyp-PVT.

Index Terms:
Polyp segmentation, pyramid vision transformer, colonoscopy, computer vision

I Introduction

Colonoscopy is the gold standard for detecting colorectal lesions since it enables colorectal polyps to be identified and removed in time, thereby preventing further spread. As a fundamental task in medical image analysis, polyp segmentation (PS) aims to locate polyps accurately in the early stage, which is of great significance in the clinical prevention of rectal cancer. Traditional PS models mainly rely on low-level features, e.g., texture [1], geometric features [2], simple linear iterative clustering superpixels [3]. However, these methods yield low-quality results and suffer from poor generalization ability. With the development of deep learning, PS has achieved promising progress. In particular, the U-shaped [4] has attracted significant attention due to its ability to adopt multi-level features for reconstructing high-resolution results. PraNet [5] employs a two-stage segmentation approach, adopting a parallel decoder to predict rough regions and an attention mechanism to restore a polyp’s edges and internal structure for fine-grained segmentation. ThresholdNet [6] is a confidence-guided data enhancement method based on a hybrid manifold for solving the problems caused by limited annotated data and imbalanced data distributions.

\begin{overpic}[width=403.26341pt]{Imgs/firstMap-min.pdf} \put(7.0,-4.5){Image} \put(35.0,-4.5){GT} \put(58.5,-4.5){\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}Ours} \put(81.5,-4.5){SANet} \end{overpic}
Figure 1: The segmentation examples of our model and SANet [7] with different challenge cases, e.g., camouflage (1st1^{st} and 2nd2^{nd} rows) and image acquisition influence (3rd3^{rd} row). The images from top to bottom are from ClinicDB [8], ETIS [9], and ColonDB [10], which show that our model has better generalization ability.

Although these methods have greatly improved accuracy and generalization ability compared to traditional methods, it is still challenging for them to locate the boundaries of polyps, as shown in Fig. 1, for several reasons: (1) Image noise. During the data collection process, the lens rotates in the intestine to obtain polyp images from different angles, which also causes motion blur and reflector problems. As a result, this greatly increases the difficulty of polyp detection; (2) Camouflage. The color and texture of polyps are very similar to surrounding tissues, with low contrast, providing them with powerful camouflage properties [11, 12], and making them difficult to identify; (3) Polycentric data. Current models struggle to generalize to multicenter (or unseen) data with different domains/distributions. To address the above issues, our contributions in this paper are as follows:

  • We present a novel polyp segmentation framework, termed Polyp-PVT. Unlike existing CNN-based methods, we adopt the pyramid vision transformer as an encoder to extract more robust features.

  • To support our framework, we introduce three simple modules. Specifically, the cascaded fusion module (CFM) collects polyps’ semantic and location information from the high-level features through progressive integration. Meanwhile, the camouflage identification module (CIM) is applied to capture polyp cues disguised in low-level features, using an attention mechanism to pay more attention to potential polyps, reducing incorrect information in the lower features. We further introduce the similarity aggregation module (SAM) equipped with a non-local and convolutional graph layer to mine local pixels and global semantic cues from the polyp area.

  • Finally, we conduct extensive experiments on five challenging benchmark datasets, including Kvasir-SEG [13], ClinicDB [8], ColonDB [10], Endoscene [14], and ETIS [9], to evaluate the performance of the proposed Polyp-PVT. On ColonDB, our method achieves a mean Dice (mDic) of 0.808, which is 5.5% higher than the existing cutting-edge method SANet [7]. On the ETIS dataset, our model achieves a mean Dice (mDic) of 0.787, which is 3.7% higher than SANet [7].

II Related Works

II-A Polyp Segmentation

Traditional Methods. Computer-aided detection is an effective alternative to manual detection, and a detailed survey has been conducted on detecting ulcers, polyps, and tumors in wireless capsule endoscopy imaging [15]. Early solutions for polyp segmentation were mainly based on low-level features, such as texture [2], geometric features [2], or simple linear iterative clustering superpixels [3]. However, these methods have a high risk of missed or false detection due to the high similarity between polyps and surrounding tissues.

Deep Learning-Based Methods. Deep learning techniques [16, 17, 18, 19, 20, 21, 22, 23, 24, 25] have greatly promoted the development of polyp segmentation tasks. Akbari et al. [26] proposed a polyp segmentation model using a fully convolutional neural network, whose segmentation results are significantly better than traditional solutions. Brandao et al. [27] used the shape from the shading strategy to restore depth, merging the result into an RGB model to provide richer feature representations. More recently, encoder-decoder-based models, such as U-Net [4], UNet++ [28], and ResUNet++ [29], have gradually come to dominate the field with excellent performance. Sun et al. [30] introduced a dilated convolution to extract and aggregate high-level semantic features with resolution retention to improve the encoder network. Psi-Net [31] introduced a multi-task segmentation model that combines contour and distance map estimation to assist segmentation mask prediction. Hemin et al. [32] first attempted to use a deeper feature extractor to perform polyp segmentation based on Mask R-CNN [33].

Different from the methods based on U-Net [4, 28, 34], PraNet [5] uses reverse attention modules to mine boundary information with a global feature map, which is generated by a parallel partial decoder from high-level features. Polyp-Net [35] proposed a dual-tree wavelet pooling CNN with a local gradient-weighted embedding level set, effectively avoiding erroneous information in high signal areas, thereby significantly reducing the false positive rate. Rahim et al. [36] proposed to use different convolution kernels for the same hidden layer for deeper feature extraction with MISH and rectified linear unit activation functions for deep feature propagation and smooth non-monotonicity. In addition, they adopted joint generalized intersections, which overcome scale invariance, rotation, and shape differences. Jha et al. [37] designed a real-time polyp segmentation method called ColonSNet. For the first time, Ahmed et al. [38] applied the generative adversarial network to the field of polyp segmentation. Another interesting idea proposed by Thambawita et al. [39] is introducing pyramid-based augmentation into the polyp segmentation task. Further, Tomar et al. [40] designed a dual decoder attention network based on ResUNet++ for polyp segmentation. More recently, MSEG [41] improved the PraNet and proposed a simple encoder-decoder structure. Specifically, they used Hardnet [42] to replace the original backbone network Res2Net50 backbone network and removed the attention mechanism to achieve faster and more accurate polyp segmentation. As an early attempt, Transfuse [43] was the first to employ a two-branch architecture combining CNNs and transformers in a parallel style. DCRNet [44] uses external and internal context relations modules to separately estimate the similarity between each location and all other locations in the same and different images. MSNet [45] introduced a multi-scale subtraction network to eliminate redundancy and complementary information between the multi-scale features. Providing a comprehensive review of polyp segmentation is beyond the scope of this paper. In Tab. I, however, we briefly survey representative works related to ours.

TABLE I: A survey on polyp segmentation. CL = CVC-CLINIC, EL = ETIS-Larib, C6 = CVC-612, AM = ASU-Mayo [46, 47], ES = EndoScene, DB = ColonDB, CV = CVC-VideoClinicDB, C = Colon, ED = Endotect 2020, KS = Kvasir-SEG, KCS = Kvasir Capsule-SEG, PraNet = same to datasets used in PraNet [5], IS = image segmentation, VS = video segmentation, CF = classfication, OD = object detection, Own = private data. CSCPD [1], APD [2], SBCP [3], FCN [26], D-FCN [27], UNet++ [28], Psi-Net [31], Mask R-CNN [32], UDC [30], ThresholdNet [6], MI2GAN [48], ACSNet [49], PraNet [5], GAN [38], APS [50], PFA [39], MMT [51], U-Net-ResNet50 [34], Survey [15], Polyp-Net [35], Deep CNN [36], EU-Net [52], DSAS [53], U-Net-MobileNetV2 [54], DCRNet [44], MSEG [41], FSSNet [55], AG-CUResNeSt [56], MPAPS [57], ResUNet++ [58], NanoNet [59], ColonSegNet [37], Segtran [60], DDANet [40], UACANet [61], DivergentNet [62], DWHieraSeg [63], Transfuse [43], SANet [7], PNS-Net [64].
No. Model Publication Code Type Dataset Core Components
1 CSCPD IJPRAI N/A IS Own Adaptive-scale candidate
2 APD TMI N/A IS Own Geometrical analysis, binary classifier
3 SBCP SPMB N/A IS Own Superpixel
4 FCN EMBC N/A IS DB FCN and patch selection
5 D-FCN JMRR N/A IS CL, EL, AM, and DB FCN and Shape-from-Shading (SfS)
6 UNet++ DLMIA PyTorch IS AM Skip pathways and deep supervision
7 Psi-Net EMBC PyTorch IS Endovis Shape and boundary aware
8 Mask R-CNN ISMICT N/A IS C6, EL, and DB Deep feature extractors
9 UDC ICMLA N/A IS C6 and EL Dilation convolution
10 ThresholdNet TMI PyTorch IS ES and WCE Learn to threshold Confidence-guided manifold mixup
11 MI2GAN MICCAI N/A IS C6 and EL GAN based model
12 ACSNet MICCAI PyTorch IS ES and KS Adaptive context selection
13 PraNet MICCAI PyTorch IS PraNet Parallel partial decoder attention
14 GAN MediaEval N/A IS KS Image-to-image translation
15 APS MediaEval N/A IS KS Variants of U-shaped structure
16 PFA MediaEval PyTorch IS KS Pyramid focus augmentation
17 MMT MediaEval N/A IS KS Competition introduction
18 U-Net-ResNet50 MediaEval N/A IS KS Variants of U-shaped structure
19 Survey CMIG N/A CF Own Classification
20 Polyp-Net TIM N/A IS DB and CV Multimodel fusion network
21 Deep CNN BSPC N/A OD EL Convolutional neural network
22 EU-Net CRV PyTorch IS PraNet Semantic information enhancement
23 DSAS MIDL Matlab IS KS Stochastic activation selection
24 U-Net-MobileNetV2 arXiv N/A IS KS Variants of U-shaped structure
25 DCRNet ISBI PyTorch IS ES, KS, and PICCOLO Within-image and cross-image contextual relations
26 MSEG arXiv PyTorch IS PraNet Hardnet and partial decoder
27 FSSNet arXiv N/A IS C6 and KS Meta-learning
28 AG-CUResNeSt RIVF N/A IS PraNet ResNeSt, attention gates
29 MPAPS JBHI PyTorch IS DB, KS, and EL Mutual-prototype adaptation network
30 ResUNet++ JBHI PyTorch IS, VS PraNet and AM ResUNet++, CRF and TTA
31 NanoNet CBMS PyTorch IS, VS ED, KS, and KCS Real-Time polyp segmentation
32 ColonSegNet Access PyTorch IS KS Residual block and SENet
33 Segtran IJCAI PyTorch IS C6 and KS Transformer
34 DDANet ICPR PyTorch IS KS Dual decoder attention network
35 UACANet ACM MM PyTorch IS PraNet Uncertainty augmented Context attention network
36 DivergentNet ISBI PyTorch IS EndoCV 2021 Combine multiple models
37 DWHieraSeg MIA PyTorch IS ES Dynamic-weighting
38 Transfuse MICCAI PyTorch IS PraNet Transformer and CNN
39 SANet MICCAI PyTorch IS PraNet Shallow attention network
40 PNS-Net MICCAI PyTorch VS C6, KS, ES, and AM Progressively normalized self-attention network
\begin{overpic}[width=411.93767pt]{Imgs/Polyp-PVT-min.pdf} \end{overpic}
Figure 2: Framework of our Polyp-PVT, which consists of a pyramid vision transformer (PVT) (a) as the encoder network, (b) cascaded fusion module (CFM) for fusing the high-level feature, (c) camouflage identification module (CIM) to filter out the low-level information, and (d) similarity aggregation module (SAM) for integrating the high- and low-level features for the final output.

II-B Vision Transformer

Transformers use multi-head self-attention (MHSA) layers to model long-term dependencies. Unlike the convolutional layer, the MHSA layer has dynamic weights and a global receptive field, making it more flexible and effective. The transformer [65] was first proposed by Vaswani et al. for the machine translation task and has since extensively influenced the natural language processing field. To apply transformers to computer vision tasks, Dosovitskiy et al. [66] proposed a vision transformer (ViT), which was the first pure transformer for image classification. ViT divides an image into multiple patches, which are sequentially sent to a transformer encoder after being encoded, and then an MLP is used to perform image classification. HVT [67] is based on a hierarchical progressive pooling method to compress the sequence length of a token and reduce the redundancy and number of calculations in ViT. The pooling-based vision transformer [68] draws on the principle of CNNs whereby, as the depth increases, the number of feature map channels increases, and the spatial dimension decreases. Yuan et al. [69] pointed out that the simple token structure in ViT cannot capture important local features, such as edges and lines, which reduces the training efficiency and leads to redundant attention mechanisms. T2T ViT was thus proposed to use layer-by-layer tokens-to-token transformation to gradually merge neighboring tokens and model local features while reducing the token’s length. TNT [70] employs a transformer suitable for fine-grained image tasks, which divides the original image patch and conducts self-attention mechanism calculations in smaller units. Meanwhile, external and internal transformers are used to extract global and local features.

To adapt to dense prediction tasks such as semantic segmentation, several methods [71, 72, 73, 74, 75, 76, 77] have also introduced the pyramid structure of CNNs to the design of transformer backbones. For instance, PVT-based models [71, 72] use a hierarchical transformer with four stages, showing that a pure transformer backbone can be as versatile as its CNN counterparts, and performs better in detection and segmentation tasks. In this work, we design a new transformer-based polyp segmentation framework, which can accurately locate the boundaries of polyps even in extreme scenarios.

III Proposed Polyp-PVT

III-A Overall Architecture

As shown in Fig. 2, our Polyp-PVT consists of 4 key modules: namely, a pyramid vision transformer (PVT) encoder, cascaded fusion module (CFM), camouflage identification module (CIM), and similarity aggregation module (SAM). Specifically, the PVT extracts multi-scale long-range dependencies features from the input image. The CFM is employed to collect semantic cues and locate polyps by aggregating high-level features progressively. The CIM is designed to remove noise and enhance low-level representation information of polyps, including texture, color, and edges. The SAM is adopted to fuse the low- and high-level features provided by the CIM and CFM, effectively transmitting the information from the pixel-level polyp to the entire polyp.

Given an input image IH×W×3I\in\mathbb{R}^{H\times W\times 3}, we use the transformer-based backbone [71] to extract four pyramid features XiH2i+1×W2i+1×CiX_{i}\in\mathbb{R}^{\frac{H}{2^{i+1}}\times\frac{W}{2^{i+1}}\times C_{i}}, where Ci{64,128,320,512}C_{i}\in\{64,128,320,512\} and i{1,2,3,4}i\in\{1,2,3,4\}. Then, we adjust the channel of three high-level features X2X_{2}, X3X_{3} and X4X_{4} to 32 through three convolutional units and feed them (i.e., X2X_{2}^{{}^{\prime}}, X3X_{3}^{{}^{\prime}}, and X4X_{4}^{{}^{\prime}}) to CFM to fuse, leading a feature map T1H8×W8×32T_{1}\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 32}. Meanwhile, low-level features X1X_{1} are converted to T2H4×W4×64T_{2}\in\mathbb{R}^{\frac{H}{4}\times\frac{W}{4}\times 64} by the CIM. After that, the T1T_{1} and T2T_{2} are aligned and fused by SAM, yielding the final feature map FH8×W8×32F\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 32}. Finally, FF is fed into a 1×11\times 1 convolutional layer to predict the polyp segmentation result P2P_{2}. We use the sum of P1P_{1} and P2P_{2} as the final prediction. During training, we optimize the model with a main loss main\mathcal{L}_{\rm main} and an auxiliary loss aux\mathcal{L}_{\rm aux}. The main loss is calculated between the final segmentation result P2P_{2} and the ground truth (GT), which is used to optimize the final polyp segmentation result. Similarly, the auxiliary loss is used to supervise the intermediate result P1P_{1} generated by the CFM.

III-B Transformer Encoder

Due to uncontrolled factors in their acquisition, polyp images tend to contain significant noise, such as motion blur, rotation, and reflection. Some recent works [78, 79] have found that the vision transformer [66, 71, 72] demonstrates stronger performance and better robustness to input disturbances than CNNs [17, 16]. Inspired by this, we use a vision transformer as our backbone network to extract more robust and powerful features for polyp segmentation. Different from [73, 66] that uses a fixed “columnar” structure or shifted windowing manner, the PVT [71] is a pyramid architecture whose representation is calculated with spatial-reduction attention operations; thus it enables to reduce the resource consumption. Note that the proposed model is backbone-independent; other famous transformer backbones are feasible in our framework. Specifically, we adopt the PVTv2 [72], which is the improved version of PVT with a more powerful feature extraction ability. To adapt PVTv2 to the polyp segmentation task, we remove the last classification layer and design a polyp segmentation head on top of four multi-scale feature maps (i.e., X1X_{1}, X2X_{2}, X3X_{3}, and X4X_{4}) generated by different stages. Among these feature maps, X1X_{1} gives detailed appearance information of polyps, and X2X_{2}, X3X_{3}, and X4X_{4} provide high-level features.

III-C Cascaded Fusion Module

To balance the accuracy and computational resources, we follow recent popular practices [80, 5] to implement the cascaded fuse module (CFM). Specifically, we define ()\mathcal{F}(\cdot) as a convolutional unit composed of a 3×33\times 3 convolutional layer with padding set to 1, batch normalization [81] and ReLU [82]. As shown in Fig. 2 (b), the CFM mainly consists of two cascaded parts, as follows:

(1) In part one, we up-sample the highest-level feature map X4X_{4}^{{}^{\prime}} to the same size as X3X_{3}^{{}^{\prime}} and then pass the result through two convolutional units 1()\mathcal{F}_{1}(\cdot) and 2()\mathcal{F}_{2}(\cdot), yieldings: X41X_{4}^{1} and X42X_{4}^{2}. Then, we multiply X41X_{4}^{1} and X3X_{3}^{{}^{\prime}} and concatenate the result with X42X_{4}^{2}. Finally, we use a convolution unit 3()\mathcal{F}_{3}(\cdot) to smooth the concatenated feature, yielding fused feature map X34H16×W16×32X_{34}\in\mathbb{R}^{\frac{H}{16}\times\frac{W}{16}\times 32}. The process can be summarized as Eqn. 1.

X34=3(Concat(1(X4)X3,2(X4))),X_{34}=\mathcal{F}_{3}(\text{Concat}(\mathcal{F}_{1}({{{X}}_{4}^{{}^{\prime}}})\odot{X_{3}^{{}^{\prime}}},\mathcal{F}_{2}({{{X}}_{4}^{{}^{\prime}}}))), (1)

where “\odot” denotes the Hadamard product, and Concat()\text{Concat}(\cdot) is the concatenation operation along the channel dimension.

(2) As shown Eqn. 2, the second part follows a similar process to part one. Firstly, we up-sample X4X_{4}^{{}^{\prime}}, X3X_{3}^{{}^{\prime}}, X34X_{34} to the same size as X2X_{2}^{{}^{\prime}}, and smooth them using convolutional units 4()\mathcal{F}_{4}(\cdot), 5()\mathcal{F}_{5}(\cdot), and 6()\mathcal{F}_{6}(\cdot), respectively. Then, we multiply the smoothed X4X_{4}^{{}^{\prime}} and X3X_{3}^{{}^{\prime}} with X2X_{2}^{{}^{\prime}}, and concatenate the resulting map with up-sampled and smoothed X34X_{34}. Finally, we feed the concatenated feature map into two convolutional units (i.e., 7()\mathcal{F}_{7}(\cdot) and 8()\mathcal{F}_{8}(\cdot)) to reduce the dimension, and obtain T1H8×W8×32T_{1}\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 32}, which is also the output of the CFM.

T1=8(7(Concat(4(X4)5(X3)X2,6(X34)))),T_{1}=\mathcal{F}_{8}(\mathcal{F}_{7}(\text{Concat}(\mathcal{F}_{4}(X_{4}^{{}^{\prime}})\odot\mathcal{F}_{5}(X_{3}^{{}^{\prime}})\odot{X_{2}^{{}^{\prime}}},\mathcal{F}_{6}(X_{34})))), (2)

III-D Camouflage Identification Module

Low-level features often contain rich detail information, such as texture, color, and edges. However, polyps tend to be very similar in appearance to the background. Therefore, we need a powerful extractor to identify the polyp details.

As shown in Fig. 2 (c), we introduce a camouflage identification module (CIM) to capture the details of polyps from different dimensions of the low-level feature map X1X_{1}. Specifically, the CIM consists of a channel attention operation [83] Attc()\text{Att}_{c}(\cdot) and a spatial attention operation [84] Atts()\text{Att}_{s}(\cdot), which can be formulated as:

T2=Atts(Attc(X1)),T_{2}=\text{Att}_{s}\left(\text{Att}_{c}\left(X_{1}\right)\right), (3)

The channel attention operation Attc()\text{Att}_{c}(\cdot) can be written as follow:

Attc(x)=σ(1(Pmax(x))+2(Pavg(x)))x,\text{Att}_{c}(x)=\sigma\left(\mathcal{H}_{1}\left({P_{\rm max}}\left(x\right)\right)+\mathcal{H}_{2}\left({P_{\rm avg}}\left(x\right)\right)\right)\odot x, (4)

where xx is the input tensor and σ()\sigma(\cdot) is the Softmax function. Pmax()P_{\rm max}(\cdot) and Pavg()P_{\rm avg}(\cdot) denote adaptive maximum pooling and adaptive average pooling functions, respectively. i(),i{1,2}{\mathcal{H}_{i}(\cdot)},i\in\{1,2\} shares parameters and consists of a convolutional layer with 1×11\times 1 kernel size to reduce the channel dimension 16 times, followed by a ReLU layer and another 1×11\times 1 convolutional layer to recover the original channel dimension. The spatial attention operation Atts()\text{Att}_{s}(\cdot) can be formulated as:

Atts(x)=σ(𝒢(Concat(Rmax(x),Ravg(x))))x,\text{Att}_{s}(x)=\sigma(\mathcal{G}(\text{Concat}({R_{\rm max}}(x),{R_{\rm avg}}(x))))\odot x, (5)

where Rmax()R_{\rm max}(\cdot) and Ravg()R_{\rm avg}(\cdot) represent the maximum and average values obtained along the channel dimension, respectively. 𝒢()\mathcal{G}(\cdot) is a 7×77\times 7 convolutional layer with padding set to 3.

\begin{overpic}[width=424.94574pt]{Imgs/SAM.pdf} \end{overpic}
Figure 3: Details of the introduced SAM. It is composed of GCN and non-local, which extend the pixel features of polyp regions with high-level semantic location cues to the entire region.

III-E Similarity Aggregation Module

To explore high-order relations between the lower-level local features from CIM and higher-level cues from CFM. We introduce the non-local [85, 86] operation under graph convolution domain [87] to implement our similarity aggregation module (SAM). As a result, SAM can inject detailed appearance features into high-level semantic features using global attention.

Given the feature map T1{T_{1}}, which contains high-level semantic information, and T2{T_{2}} with rich appearance details, we fuse them through self-attention. First, two linear mapping functions Wθ()W_{\theta}(\cdot) and Wϕ()W_{\phi}(\cdot) are applied on T1{T_{1}} to reduce the dimension and obtain feature maps QH8×W8×16Q\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 16} and KH8×W8×16K\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 16}. Here, we take a convolution operation with a kernel size of 1×11\times 1 as the linear mapping process. This process can be expressed as follows:

Q=Wθ(T1),K=Wϕ(T1).Q=W_{\theta}(T_{1}),K=W_{\phi}(T_{1}). (6)

For T2{T_{2}}, we use a convolutional unit Wg()W_{g}(\cdot) to reduce the channel dimension to 32 and interpolate it to the same size as T1{T_{1}}. Then, we apply a Softmax function on the channel dimension and choose the second channel as the attention map, leading to T2H8×W8×1{T^{{}^{\prime}}_{2}}\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 1}. These operations are represented as F()\text{F}(\cdot) in Fig. 3. Next, we calculate the Hadamard product between KK and T2{T^{{}^{\prime}}_{2}}. This operation assigns different weights to different pixels, increasing the weight of edge pixels. After that, we use an adaptive pooling operation to reduce the displacement of features and apply a center crop on it to obtain the feature map V4×4×16V\in\mathbb{R}^{4\times 4\times 16}. In summary, the process can be formulated as follows:

V=AP(KF(Wg(T2))),V=\text{AP}(K\odot\text{F}(W_{g}({T_{2}}))), (7)

where AP()\text{AP}(\cdot) denotes the pooling and crop operations.

Then, we establish the correlation between each pixel in VV and KK through an inner product, which is written as follows:

f=σ(V𝖳K),f=\sigma(V^{\mathsf{T}}\otimes K), (8)

where ”\otimes” denotes the inner product operation. V𝖳V^{\mathsf{T}} is the transpose of VV and ff is the correlation attention map.

After obtaining the correlation attention map ff, we multiply it with the feature map QQ, and the result features are fed to the graph convolutional layer [86] GCN()\text{GCN}(\cdot), leading to G4×4×16G\in\mathbb{R}^{4\times 4\times 16}. Same to [86], we calculate the inner product between ff and GG as Eqn. 9, reconstructing the graph domain features into the original structural features:

Y=f𝖳GCN(f𝖳Q).Y^{\prime}=f^{\mathsf{T}}\otimes\text{GCN}(f^{\mathsf{T}}\otimes Q). (9)

The reconstructed feature map YY^{\prime} is adjusted to the same channel sizes with YY by a convolutional layer Wz()W_{z}(\cdot) with 1×11\times 1 kernel size, and then combined with the feature T1{T_{1}} to obtain the final output ZH8×W8×32Z\in\mathbb{R}^{\frac{H}{8}\times\frac{W}{8}\times 32} of the SAM. Eqn. 10 summarizes the details of this process:

Z=T1+Wz(Y).Z=T_{1}+W_{z}(Y^{\prime}). (10)
\begin{overpic}[width=433.62pt]{./Imgs/loss.pdf} \end{overpic}
Figure 4: Loss curves under different training parameter settings.

III-F Loss Function

Our loss function can be formulated as Eqn. 11:

=main+aux,\mathcal{L}=\mathcal{L}_{\rm main}+\mathcal{L}_{\rm aux}, (11)

where main\mathcal{L}_{\rm main} and aux\mathcal{L}_{\rm aux} are the main loss and auxiliary loss, respectively. The main loss main\mathcal{L}_{\rm main} is calculated between the final segmentation result P2P_{2} and ground truth GG, which can be written as:

main=IoUw(P2,G)+BCEw(P2,G).\mathcal{L}_{\rm main}=\mathcal{L}^{w}_{\text{IoU}}(P_{2},G)+\mathcal{L}^{w}_{\text{BCE}}(P_{2},G). (12)

The auxiliary loss aux\mathcal{L}_{\rm aux} is calculated between the intermediate result P1P_{1} from the CFM and ground truth GG, which can be formulated as:

aux=IoUw(P1,G)+BCEw(P1,G).\mathcal{L}_{\rm aux}=\mathcal{L}^{w}_{\text{IoU}}(P_{1},G)+\mathcal{L}^{w}_{\text{BCE}}(P_{1},G). (13)

IoUw()\mathcal{L}^{w}_{\text{IoU}}(\cdot) and BCEw()\mathcal{L}^{w}_{\text{BCE}}(\cdot) are the weighted intersection over union (IoU) loss [88] and weighted binary cross entropy (BCE) loss [88], which restrict the prediction map in terms of the global structure (object-level) and local details (pixel-level) perspectives. Unlike the standard BCE loss function, which treats all pixels equally, BCEw()\mathcal{L}^{w}_{\text{BCE}}(\cdot) considers the importance of each pixel and assigns higher weights to hard pixels. Furthermore, compared to the standard IoU loss, IoUw()\mathcal{L}^{w}_{\text{IoU}}(\cdot) pays more attention to the hard pixels.

III-G Implementation Details

We implement our Polyp-PVT with the PyTorch framework and use a Tesla P100 to accelerate the calculations. Considering the differences in the sizes of each polyp image, we adopt a multi-scale strategy [5, 41] in the training stage. The hyperparameter details are as follows. To update the network parameters, we use the AdamW [89] optimizer, which is widely used in transformer networks [71, 73, 72]. The learning rate is set to 1e-4 and the weight decay is adjusted to 1e-4 too. Further, we resize the input images to 352×352352\times 352 with a mini-batch size of 16 for 100 epochs. More details about the training loss cures, parameter setting, and network parameters are shown in Fig. 4, Tab. II, and Tab. III, respectively. The total training time is nearly 3 hours to achieve the best (e.g., 30 epochs) performance. For testing, we only resize the images to 352×352352\times 352 without any post-processing optimization strategies.

TABLE II: Parameter setting during the training stage.
Optimizer Learning Rate (lr) Multi-scale Clip
AdamW 1e-4 [0.75,1,1.25] 0.5
Decay rate Weight decay Epochs Input Size
0.1 1e-4 100 352×352352\times 352

IV Experiments

IV-A Evaluation Metrics

We employ six widely-used evaluation metrics, including Dice [90], IoU, mean absolute error (MAE), weighted F-measure (Fβw{F}_{\beta}^{w}[91], S-measure (Sα{S}_{\alpha}[92], and E-measure (Eξ{E}_{\xi}[93, 94] to evaluate the model performances. Among these metrics, Dice and IoU are similarity measures at the regional level, which mainly focus on the internal consistency of segmented objects. Here, we report the mean value of Dice and IoU, denoted as mDic and mIoU, respectively. MAE is a pixel-by-pixel comparison indicator that represents the average value of the absolute error between the predicted value and the true value. Weighted F-measure (Fβw{F}_{\beta}^{w}) comprehensively considers the recall and precision and eliminates the effect of considering each pixel equally in conventional indicators. S-measure (Sα{S}_{\alpha}) focuses on the structural similarity of target prospects at the region and object level. E-measure (Eξ{E}_{\xi}) is used to evaluate the segmentation results at the pixel and image level. We report the mean and max value of E-measure, denoted as mEξm{E}_{\xi} and maxEξmax{E}_{\xi}, respectively. The evaluation toolbox is derived from https://github.com/DengPingFan/PraNet.

TABLE III: Network parameters of each module. Note that the encoder parameters are the same as PVT without any changes. BasicConv2d and Conv2d with the parameters [in_channel, out_channel, kernel_size, padding] and GCN [num_state, num_node].
Encoder SAM
patch_size [4] AvgPool2d [6]
embed_dims [64, 128, 320, 512] Conv2d [32,16,1,1]
num_heads [1, 2, 5, 8] Conv2d [32,16,1,1]
mlp_ratios [8, 8, 4, 4] Conv2d [16,32,1,1]
depths [3, 4, 18, 3] GCN [16,16]
sr_ratios [8, 4, 2, 1] BasicConv2d [64,32,1,0]
drop_rate [0]
drop_path_rate [0.1]
CFM CIM
BasicConv2d [32,32,3,1] AvgPool2d [1]
BasicConv2d [32,32,3,1] AvgPool2d [1]
BasicConv2d [32,32,3,1] Conv2d [64,4,1,0]
BasicConv2d [32,32,3,1] ReLU
BasicConv2d [64,64,3,1] Conv2d [4,64,1,0]
BasicConv2d [64,64,3,1] Sigmoid
BasicConv2d [96,96,3,1] Conv2d [2,1,7,3]
BasicConv2d [96,32,3,1] Sigmoid

IV-B Datasets and Compared Models

Datasets. Following the experimental setups in PraNet [5], we adopt five challenging public datasets, including Kvasir-SEG [13], ClinicDB [8], ColonDB [10], Endoscene [14] and ETIS [9] to verify the effectiveness of our framework.

Models. We collect several open source models from the field of polyp segmentation, for a total of nine comparative models, including U-Net [4], UNet++ [28], PraNet [5], SFA [95], MSEG [41], ACSNet [49], DCRNet [44], EU-Net [52] and SANet [7]. For a fair comparison, we use their open-source codes to evaluate the same training and testing sets. Note that the SFA results are generated using the released test model.

IV-C Analysis of Learning Ability

Settings. We use the ClinicDB and Kvasir-SEG datasets to evaluate the learning ability of the proposed model. ClinicDB contains 612 images, which are extracted from 31 colonoscopy videos. Kvasir-SEG is collected from the polyp class in the Kvasir dataset and includes 1,000 polyp images. Following PraNet, we adopt the same 900 and 548 images from ClinicDB and Kvasir-SEG datasets as the training set, and the remaining 64 and 100 images are employed as the respective test sets.

TABLE IV: Quantitative results of the test datasets, i.e., Kvasir-SEG and ClinicDB.
Kvasir-SEG [13] ClinicDB [8]
Model mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE
MICCAI’15 U-Net 0.818 0.746 0.794 0.858 0.881 0.893 0.055 0.823 0.755 0.811 0.889 0.913 0.954 0.019
DLMIA’18 UNet++ 0.821 0.743 0.808 0.862 0.886 0.909 0.048 0.794 0.729 0.785 0.873 0.891 0.931 0.022
MICCAI’19 SFA 0.723 0.611 0.670 0.782 0.834 0.849 0.075 0.700 0.607 0.647 0.793 0.840 0.885 0.042
arXiv’21 MSEG 0.897 0.839 0.885 0.912 0.942 0.948 0.028 0.909 0.864 0.907 0.938 0.961 0.969 0.007
arXiv’21 DCRNet 0.886 0.825 0.868 0.911 0.933 0.941 0.035 0.896 0.844 0.890 0.933 0.964 0.978 0.010
MICCAI’20 ACSNet 0.898 0.838 0.882 0.920 0.941 0.952 0.032 0.882 0.826 0.873 0.927 0.947 0.959 0.011
MICCAI’20 PraNet 0.898 0.840 0.885 0.915 0.944 0.948 0.030 0.899 0.849 0.896 0.936 0.963 0.979 0.009
CRV’21 EU-Net 0.908 0.854 0.893 0.917 0.951 0.954 0.028 0.902 0.846 0.891 0.936 0.959 0.965 0.011
MICCAI’21 SANet 0.904 0.847 0.892 0.915 0.949 0.953 0.028 0.916 0.859 0.909 0.939 0.971 0.976 0.012
Polyp-PVT (Ours) 0.917 0.864 0.911 0.925 0.956 0.962 0.023 0.937 0.889 0.936 0.949 0.985 0.989 0.006
TABLE V: Quantitative results of the test datasets ColonDB and ETIS. The SFA result is generated using the published code.
ColonDB [10] ETIS [9]
Model mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE
MICCAI’15 U-Net 0.512 0.444 0.498 0.712 0.696 0.776 0.061 0.398 0.335 0.366 0.684 0.643 0.740 0.036
DLMIA’18 UNet++ 0.483 0.410 0.467 0.691 0.680 0.760 0.064 0.401 0.344 0.390 0.683 0.629 0.776 0.035
MICCAI’19 SFA 0.469 0.347 0.379 0.634 0.675 0.764 0.094 0.297 0.217 0.231 0.557 0.531 0.632 0.109
MICCAI’20 ACSNet 0.716 0.649 0.697 0.829 0.839 0.851 0.039 0.578 0.509 0.530 0.754 0.737 0.764 0.059
arXiv’21 MSEG 0.735 0.666 0.724 0.834 0.859 0.875 0.038 0.700 0.630 0.671 0.828 0.854 0.890 0.015
arXiv’21 DCRNet 0.704 0.631 0.684 0.821 0.840 0.848 0.052 0.556 0.496 0.506 0.736 0.742 0.773 0.096
MICCAI’20 PraNet 0.712 0.640 0.699 0.820 0.847 0.872 0.043 0.628 0.567 0.600 0.794 0.808 0.841 0.031
CRV’21 EU-Net 0.756 0.681 0.730 0.831 0.863 0.872 0.045 0.687 0.609 0.636 0.793 0.807 0.841 0.067
MICCAI’21 SANet 0.753 0.670 0.726 0.837 0.869 0.878 0.043 0.750 0.654 0.685 0.849 0.881 0.897 0.015
Polyp-PVT (Ours) 0.808 0.727 0.795 0.865 0.913 0.919 0.031 0.787 0.706 0.750 0.871 0.906 0.910 0.013
\begin{overpic}[width=346.89731pt]{Imgs/excelmap-min.pdf} \end{overpic}
Figure 5: Evaluation of model generalization ability. We provide the max Dice results on ColonDB and ETIS.

Results. As can be seen in Tab. IV, our model is superior to the current methods, demonstrating that it has a better learning ability. On the Kvasir-SEG dataset, the mDic score of our model is 1.3% higher than that of the second-best model, SANet, and 1.9% higher than that of PraNet. On the ClinicDB dataset, the mDic score of our model is 2.1% higher than that of SANet and 3.8% higher than that of PraNet.

TABLE VI: Quantitative results of the test dataset Endoscene. The SFA result is generated using the published code.
Endoscene [14]
Model mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE
U-Net 0.710 0.627 0.684 0.843 0.847 0.875 0.022
UNet++ 0.707 0.624 0.687 0.839 0.834 0.898 0.018
SFA 0.467 0.329 0.341 0.640 0.644 0.817 0.065
MSEG 0.874 0.804 0.852 0.924 0.948 0.957 0.009
ACSNet 0.863 0.787 0.825 0.923 0.939 0.968 0.013
DCRNet 0.856 0.788 0.830 0.921 0.943 0.960 0.010
PraNet 0.871 0.797 0.843 0.925 0.950 0.972 0.010
EU-Net 0.837 0.765 0.805 0.904 0.919 0.933 0.015
SANet 0.888 0.815 0.859 0.928 0.962 0.972 0.008
Polyp-PVT 0.900 0.833 0.884 0.935 0.973 0.981 0.007
TABLE VII: The standard deviation (SD) of the mean dice (mDic) of our model and the comparison models.
     Datasets      Kvasir-SEG      ClinicDB      ColonDB      ETIS      Endoscene
     Metrics      mDic ±\pm SD      mDic ±\pm SD      mDic ±\pm SD      mDic ±\pm SD      mDic ±\pm SD
     MICCAI’15 U-Net      .818 ±\pm .039      .823 ±\pm .047      .483 ±\pm .034      .398 ±\pm .033      .710 ±\pm .049
     DLMIA’18 UNet++      .821 ±\pm .040      .794 ±\pm .044      .456 ±\pm .037      .401 ±\pm .057      .707 ±\pm .053
     MICCAI’19 SFA      .723 ±\pm .052      .701 ±\pm .054      .444 ±\pm .037      .297 ±\pm .025      .468 ±\pm .050
     arXiv’21 MSEG      .897 ±\pm .041      .910 ±\pm .048      .735 ±\pm .039      .700 ±\pm .039      .874 ±\pm .051
     MICCAI’20 ACSNet      .898 ±\pm .045      .882 ±\pm .048      .716 ±\pm .040      .578 ±\pm .035      .863 ±\pm .055
     arXiv’21 DCRNet      .886 ±\pm .043      .896 ±\pm .049      .704 ±\pm .039      .556 ±\pm .039      .857 ±\pm .052
     MICCAI’20 PraNet      .898 ±\pm .041      .899 ±\pm .048      .712 ±\pm .038      .628 ±\pm .036      .871 ±\pm .051
     CRV’21 EU-Net      .908 ±\pm .042      .902 ±\pm .048      .756 ±\pm .040      .687 ±\pm .039      .837 ±\pm .049
     MICCAI’21 SANet      .904 ±\pm .042      .916 ±\pm .049      .752 ±\pm .040      .750 ±\pm .047      .888 ±\pm .054
     Polyp-PVT (Ours)      .917 ±\pm .042      .937 ±\pm .050      .808 ±\pm .043      .787 ±\pm .044      .900 ±\pm .052

IV-D Analysis of Generalization Ability

Settings. To verify the generalization performance of the model, we test it on three unseen (i.e., Polycentric) datasets, namely ETIS, ColonDB, and EndoScene. There are 196 images in ETIS, 380 images in ColonDB, and 60 images in EndoScene. It is worth noting that the images in these datasets belong to different medical centers. In other words, the model has not seen their training data, which is different from the verification methods of ClinicDB and Kvasir-SEG.

\begin{overpic}[width=416.27809pt]{Imgs/visualization1-min.pdf} \small\put(5.0,-1.5){Image} \put(20.0,-1.5){GT} \put(33.0,-1.5){\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}Ours} \put(48.0,-1.5){SANet} \put(62.0,-1.5){PraNet} \put(75.0,-1.5){ACSNet} \put(90.0,-1.5){DCRNet} \end{overpic}
Figure 6: Visualization results with the current models. Green indicates a correct polyp. Yellow is the missed polyp. Red is the wrong prediction. As we can see, the proposed model can accurately locate and segment polyps, regardless of size.
\begin{overpic}[width=416.27809pt]{Imgs/visualization2-min.pdf} \small\put(4.0,-1.5){Image} \put(17.0,-1.5){GT} \put(30.0,-1.5){\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}Ours} \put(41.0,-1.5){EU-Net} \put(53.0,-1.5){HarDNet} \put(67.0,-1.5){SFA} \put(79.0,-1.5){U-Net} \put(91.0,-1.5){UNet++} \end{overpic}
Figure 7: Visualization results with the current models.

Results. The results are shown in Tab. VI and Tab. V. As can be seen, our Polyp-PVT achieves a good generalization performance compared with the existing models. And our model generalizes easily to multicentric (or unseen) data with different domains/distributions. On ColonDB, it is ahead of the second-best SANet and classical PraNet by 5.5% and 9.6%, respectively. On ETIS, we exceed the SANet and PraNet by 3.7% and 15.9%, respectively. In addition, on EndoScene, our model is better than SANet and PraNet by 1.2% and 2.9%, respectively. Moreover, to prove the generalization ability of Polyp-PVT, we present the max Dice results in Fig. 5, where our model shows a steady improvement on both ColonDB and ETIS. In addition, we show the standard deviation (SD) of the mean dice (mDic) between our model and others in Tab. VII. As seen, there is not much difference in SD between our model and the comparison model, and they are both stable and balanced.

IV-E Qualitative Analysis

\begin{overpic}[width=429.28616pt]{./Imgs/ColonDB.pdf} \end{overpic}
Figure 8: FROC curves of different methods on ColonDB.

Fig. 6 and Fig. 7 show the visualization results of our model and the compared models. We find that our results have two advantages.

  • Our model is able to adapt to data under different conditions. That is, it maintains a stable recognition and segmentation ability under different acquisition environments (different lighting, contrast, reflection, motion blur, small objects, and rotation).

  • The model segmentation results have internal consistency and predicted edges are closer to the ground-truth labels. We also provide FROC curves on ColonDB in Fig. 8, and our result is at the top, indicating that our effect achieves the best.

IV-F Ablation Study

We describe in detail the effectiveness of each component on the overall model. The training, testing, and hyperparameter settings are the same as mentioned in Sec. III-G. The results are shown in Tab. VIII.

\begin{overpic}[width=433.62pt]{./Imgs/polyp-cim.pdf} \end{overpic}
Figure 9: Visualization of the feature map in the CIM module.
TABLE VIII: Quantitative results for ablation studies.
Dataset Metric Bas. w/o CFM w/o CIM w/o SAM Final
Endoscene mDic 0.869 0.892 0.882 0.874 0.900
mIoU 0.792 0.826 0.808 0.801 0.833
ClinicDB mDic 0.903 0.915 0.930 0.930 0.937
mIoU 0.847 0.865 0.881 0.877 0.889
ColonDB mDic 0.796 0.802 0.805 0.779 0.808
mIoU 0.707 0.721 0.724 0.696 0.727
ETIS mDic 0.759 0.771 0.785 0.778 0.787
mIoU 0.668 0.690 0.711 0.693 0.706
Kvasir-SEG mDic 0.910 0.922 0.910 0.910 0.917
mIoU 0.856 0.872 0.858 0.853 0.864

Components. We use PVTv2 [72] as our baseline (Bas.) and evaluate module effectiveness by removing or replacing components from the complete Polyp-PVT and comparing the variants with the standard version. The standard version is denoted as “Polyp-PVT (PVT+CFM+CIM+SAM)”, where “CFM”, “CIM” and “SAM” indicate the usage of the CFM, CIM, and SAM, respectively.

Effectiveness of CFM. To analyze the effectiveness of the CFM, a version of “Polyp-PVT (w/o CFM)” is trained. Tab. VIII shows that the model without the CFM drops sharply on all five datasets compared to the standard Polyp-PVT. In particular, the mDic is reduced from 0.937 to 0.915 on ClinicDB.

Effectiveness of CIM. To demonstrate the ability of the CIM, we also remove it from Polyp-PVT, denoting this as “Polyp-PVT (w/o CIM)”. As shown in Tab. VIII, this variant performs worse than the overall Polyp-PVT. Specifically, removing the CIM causes the mDic to decrease by 1.8% on Endoscene. Meanwhile, it is obvious that the lack of the CIM introduces significant noise (please refer to Fig. 10). In order to further explore the internal of CIM, the feature visualizations of the two main configurations inside the CIM are shown in Fig 9. It can be seen that the low-level features have a large amount of detailed information. Still, the differences between polyps and other normal tissues cannot be mined directly from this information. Thanks to the channel attention and spatial attention mechanism, information such as details and edges of polyps can be discerned from a large amount of redundant information.

\begin{overpic}[width=433.62pt]{Imgs/ablation.pdf} \small\put(6.0,-2.5){Image} \put(21.0,-2.5){GT} \put(35.0,-2.5){Bas.} \put(46.0,-2.5){w/o CFM} \put(60.0,-2.5){w/o CIM} \put(75.0,-2.5){w/o SAM} \put(91.0,-2.5){\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}Ours} \end{overpic}
Figure 10: Visualization of the ablation study results, which are converted from the output into heat maps. As can be seen, removing any module leads to missed or incorrectly detected results.
TABLE IX: Ablation study of GCN in the SAM module.
Setting Endoscene ClinicDB ColonDB ETIS Kvasir-SEG
w/o GCN 0.876 0.928 0.784 0.725 0.894
w/ Conv 0.894 0.919 0.787 0.742 0.909
w/ GCN 0.900 0.937 0.808 0.787 0.917
TABLE X: About the ablation experiments of the powerful rotation adaptability. All experiments are under the condition of large rotation (15 degrees).
Setting Endoscene ClinicDB ColonDB ETIS Kvasir-SEG
w/o GCN 0.857 0.909 0.756 0.667 0.894
w/ Conv 0.865 0.898 0.789 0.719 0.893
w/ GCN 0.874 0.929 0.806 0.744 0.915
\begin{overpic}[width=433.62pt]{./Imgs/polyp-P1P2.pdf} \end{overpic}
Figure 11: visualization of the P1 and P2 predictions.
TABLE XI: The result of video polyp segmentation on the i.e., CVC-612-T and CVC-612-V.
CVC-612-T [8] CVC-612-V [8]
Model mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE
MICCAI’15 U-Net 0.711 0.618 0.694 0.810 0.836 0.853 0.058 0.709 0.597 0.680 0.826 0.855 0.872 0.023
TMI’19 UNet++ 0.697 0.603 0.688 0.800 0.817 0.865 0.059 0.668 0.557 0.642 0.805 0.830 0.846 0.025
ISM’19 ResUNet++ 0.616 0.512 0.604 0.727 0.758 0.760 0.084 0.750 0.646 0.717 0.829 0.877 0.879 0.023
MICCAI’20 ACSNet 0.780 0.697 0.772 0.838 0.864 0.866 0.053 0.801 0.710 0.765 0.847 0.887 0.890 0.054
MICCAI’20 PraNet 0.833 0.767 0.834 0.886 0.904 0.926 0.038 0.857 0.793 0.855 0.915 0.936 0.965 0.013
MICCAI’21 PNS-Net 0.837 0.765 0.838 0.903 0.903 0.923 0.038 0.851 0.769 0.836 0.923 0.944 0.962 0.012
Polyp-PVT (Ours) 0.846 0.776 0.850 0.895 0.908 0.926 0.037 0.882 0.810 0.874 0.924 0.963 0.967 0.012
\begin{overpic}[width=433.62pt]{Imgs/limitation.pdf} \small\put(6.0,-2.5){Image} \put(28.0,-2.5){GT} \put(48.0,-2.5){\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}Ours} \put(67.0,-2.5){SANet} \put(87.0,-2.5){PraNet} \end{overpic}
Figure 12: Visualization of some failure cases. Green indicates a correct polyp. Yellow is the missed polyp. Red is the wrong prediction.

Effectiveness of SAM. Similarly, we test the effectiveness of the SAM module by removing it from the overall Polyp-PVT and replacing it with an element-wise addition operation, which is denoted as “Polyp-PVT (w/o SAM)”. The performance of the complete Polyp-PVT shows an improvement of 2.9% and 3.1% in terms of mDic and mIoU, respectively, on ColonDB. Fig. 10 shows the benefits of SAM more intuitively. It is found that the lack of the SAM leads to more detailed errors or even missed inspections. As reported in Tab IX, we add more results on the GCN in the SAM module. The experimental results further illustrate that GCN plays a key role. The effect of the lack of GCN is significantly reduced, and the effect is improved after replacing it with convolution. Still, GCN can significantly exceed the capabilities of the convolution module. The experimental results also verified the importance of GCN’s large receptive field and rotation insensitivity to polyp segmentation. The rotational robustness of GCN is stronger than convolutions. As shown in Tab X, under the condition of large rotation (15 degrees), GCN has better adaptability to image rotation than convolutions. To further explore the role of SAM, we visualized P1 and P2, and the results of P1 and P2 are shown in Fig 11. Compared with P1, P2 has higher reliability in error recognition and identification of uncertain regions. This is mainly due to the large number of low-level details collected by CIM and mining local pixels and global semantic cues from the polyp area of SAM.

TABLE XII: Video polyp segmentation results on the CVC-300-TV.
CVC-300-TV [96]
Model mDic mIoU Fβw{F}_{\beta}^{w} Sα{S}_{\alpha} mEξm{E}_{\xi} maxEξmax{E}_{\xi} MAE
U-Net 0.631 0.516 0.567 0.793 0.826 0.849 0.027
UNet++ 0.638 0.527 0.581 0.796 0.831 0.847 0.024
ResUNet++ 0.533 0.410 0.469 0.703 0.718 0.720 0.052
ACSNet 0.732 0.627 0.703 0.837 0.871 0.875 0.016
PraNet 0.716 0.624 0.700 0.833 0.852 0.904 0.016
PNS-Net 0.813 0.710 0.778 0.909 0.921 0.942 0.013
Ours 0.880 0.802 0.869 0.915 0.961 0.965 0.011

IV-G Video Polyp Segmentation

To validate the superiority of the proposed model, we conduct experiments on the video polyp segmentation datasets. For a fair comparison, we re-train our model with the same training datasets and use the same testing set as PNS-Net [64, 97]. We compare our model on three standard benchmarks (i.e., CVC-300-TV [96], CVC-612-T [8], and CVC-612-V [8]) against six cutting-edge approaches, including U-Net [4], UNet++ [28], ResUNet++ [29], ACSNet [49], PraNet [5], PNS-Net [64], in Tab. XI and Tab. XII. Note that PNS-Net provides all the prediction maps of the compared methods. As seen, our method is very competitive and far ahead of the best existing model, PNS-Net, by 3.1% and 6.7% on CVC-612-V and CVC-300-TV, respectively, in terms of mDice.

IV-H Limitations

Although the proposed Polyp-PVT model surpasses existing algorithms, it still performs poorly in certain cases. We present some failure cases in Fig. 12. As can be seen, one major limitation is the inability to detect accurate polyp boundaries with overlapping light and shadow (1st1^{st} row). Our model can identify the location information of polyps (green mask in 1st1^{st} row), but it regards the light and shadow part of the edge as the polyp (red mask in 1st1^{st} row). More deadly, our model incorrectly predicts the reflective point as a polyp (red mask in 2nd2^{nd} and 3rd3^{rd} rows). We notice that the reflective points are very salient in the image. Therefore, we speculate that the prediction may be based on only these points. More importantly, we believe that a simple way is to convert the input image into a gray image, which can eliminate the reflection and overlap of light and shadow to assist the model in judgment.

V Conclusion

In this paper, we propose a new image polyp segmentation framework, named Polyp-PVT, which utilizes a pyramid vision transformer backbone as the encoder to explicitly extract more powerful and robust features. Extensive experiments show that Polyp-PVT consistently outperforms all current cutting-edge models on five challenging datasets without any pre-/post-processing. In particular, for the unseen ColonDB dataset, the proposed model reaches a mean Dice score of above 0.8 for the first time. Interestingly, we also surpass the current cutting-edge PNS-Net in terms of the video polyp segmentation task, demonstrating excellent learning ability. Specifically, we obtain the above-mention achievements by introducing three simple components, i.e., a cascaded fusion module (CFM), a camouflage identification module (CIM), and a similarity aggregation module (SAM), which effectively extract high and low-level cues separately, and effectively fuse them for the final output. We hope this research will stimulate more novel ideas for solving the polyp segmentation task.

References

  • [1] M. Fiori, P. Musé, and G. Sapiro, “A complete system for candidate polyps detection in virtual colonoscopy,” IJPRAI, vol. 28, no. 07, p. 1460014, 2014.
  • [2] A. V. Mamonov, I. N. Figueiredo, P. N. Figueiredo, and Y.-H. R. Tsai, “Automated polyp detection in colon capsule endoscopy,” IEEE TMI, vol. 33, no. 7, pp. 1488–1502, 2014.
  • [3] O. H. Maghsoudi, “Superpixel based segmentation and classification of polyps in wireless capsule endoscopy,” in IEEE SPMB, 2017.
  • [4] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in MICCAI, 2015.
  • [5] D.-P. Fan, G.-P. Ji, T. Zhou, G. Chen, H. Fu, J. Shen, and L. Shao, “Pranet: Parallel reverse attention network for polyp segmentation,” in MICCAI, 2020.
  • [6] X. Guo, C. Yang, Y. Liu, and Y. Yuan, “Learn to threshold: Thresholdnet with confidence-guided manifold mixup for polyp segmentation,” IEEE TMI, vol. 40, no. 4, pp. 1134–1146, 2020.
  • [7] J. Wei, Y. Hu, R. Zhang, Z. Li, S. K. Zhou, and S. Cui, “Shallow attention network for polyp segmentation,” in MICCAI, 2021.
  • [8] J. Bernal, F. J. Sánchez, G. Fernández-Esparrach, D. Gil, C. Rodríguez, and F. Vilariño, “Wm-dova maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians,” CMIG, vol. 43, pp. 99–111, 2015.
  • [9] J. Silva, A. Histace, O. Romain, X. Dray, and B. Granado, “Toward embedded detection of polyps in wce images for early diagnosis of colorectal cancer,” IJCARS, vol. 9, no. 2, pp. 283–293, 2014.
  • [10] N. Tajbakhsh, S. R. Gurudu, and J. Liang, “Automated polyp detection in colonoscopy videos using shape and context information,” IEEE TMI, vol. 35, no. 2, pp. 630–644, 2015.
  • [11] D.-P. Fan, G.-P. Ji, M.-M. Cheng, and L. Shao, “Concealed object detection,” IEEE TPAMI, 2021.
  • [12] D.-P. Fan, G.-P. Ji, G. Sun, M.-M. Cheng, J. Shen, and L. Shao, “Camouflaged object detection,” in CVPR, 2020.
  • [13] D. Jha, P. H. Smedsrud, M. A. Riegler, P. Halvorsen, T. de Lange, D. Johansen, and H. D. Johansen, “Kvasir-seg: A segmented polyp dataset,” in MMM, 2020.
  • [14] D. Vázquez, J. Bernal, F. J. Sánchez, G. Fernández-Esparrach, A. M. López, A. Romero, M. Drozdzal, and A. Courville, “A benchmark for endoluminal scene segmentation of colonoscopy images,” JHE, vol. 2017, 2017.
  • [15] T. Rahim, M. A. Usman, and S. Y. Shin, “A survey on contemporary computer-aided tumor, polyp, and ulcer detection methods in wireless capsule endoscopy imaging,” CMIG, p. 101767, 2020.
  • [16] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in CVPR, 2016.
  • [17] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in ICLR, 2015.
  • [18] X. Li, W. Wang, X. Hu, and J. Yang, “Selective kernel networks,” in CVPR, 2019.
  • [19] W. Wang, X. Li, T. Lu, and J. Yang, “Mixed link networks,” in IJCAI, 2018.
  • [20] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in CVPR, 2015.
  • [21] L. Cai, M. Wu, L. Chen, W. Bai, M. Yang, S. Lyu, and Q. Zhao, “Using guided self-attention with local information for polyp segmentation,” in MICCAI.   Springer, 2022.
  • [22] N. K. Tomar, D. Jha, U. Bagci, and S. Ali, “Tganet: Text-guided attention for improved polyp segmentation,” in MICCAI.   Springer, 2022.
  • [23] R. Zhang, P. Lai, X. Wan, D.-J. Fan, F. Gao, X.-J. Wu, and G. Li, “Lesion-aware dynamic kernel for polyp segmentation,” in MICCAI.   Springer, 2022.
  • [24] J.-H. Shi, Q. Zhang, Y.-H. Tang, and Z.-Q. Zhang, “Polyp-mixer: An efficient context-aware mlp-based paradigm for polyp segmentation,” IEEE TCSVT, 2022.
  • [25] X. Zhao, Z. Wu, S. Tan, D.-J. Fan, Z. Li, X. Wan, and G. Li, “Semi-supervised spatial temporal attention network for video polyp segmentation,” in MICCAI.   Springer, 2022.
  • [26] M. Akbari, M. Mohrekesh, E. Nasr-Esfahani, S. R. Soroushmehr, N. Karimi, S. Samavi, and K. Najarian, “Polyp segmentation in colonoscopy images using fully convolutional network,” in IEEE EMBC, 2018.
  • [27] P. Brandao, O. Zisimopoulos, E. Mazomenos, G. Ciuti, J. Bernal, M. Visentini-Scarzanella, A. Menciassi, P. Dario, A. Koulaouzidis, A. Arezzo et al., “Towards a computed-aided diagnosis system in colonoscopy: automatic polyp segmentation using convolution neural networks,” JMRR, vol. 3, no. 02, p. 1840002, 2018.
  • [28] Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: A nested u-net architecture for medical image segmentation,” in DLMIA, 2018.
  • [29] D. Jha, P. H. Smedsrud, M. A. Riegler, D. Johansen, T. de Lange, P. Halvorsen, and H. D. Johansen, “Resunet++: An advanced architecture for medical image segmentation,” in IEEE ISM, 2019.
  • [30] X. Sun, P. Zhang, D. Wang, Y. Cao, and B. Liu, “Colorectal polyp segmentation by u-net with dilation convolution,” in IEEE ICMLA, 2019.
  • [31] B. Murugesan, K. Sarveswaran, S. M. Shankaranarayana, K. Ram, J. Joseph, and M. Sivaprakasam, “Psi-net: Shape and boundary aware joint multi-task deep network for medical image segmentation,” in IEEE EMBC, 2019.
  • [32] H. A. Qadir, Y. Shin, J. Solhusvik, J. Bergsland, L. Aabakken, and I. Balasingham, “Polyp detection and segmentation using mask r-cnn: Does a deeper feature extractor cnn always perform better?” in ISMICT, 2019.
  • [33] K. He, G. Gkioxari, P. Dollár, and R. Girshick, “Mask r-cnn,” in ICCV, 2017.
  • [34] S. Alam, N. K. Tomar, A. Thakur, D. Jha, and A. Rauniyar, “Automatic polyp segmentation using u-net-resnet50,” in MediaEvalW, 2020.
  • [35] D. Banik, K. Roy, D. Bhattacharjee, M. Nasipuri, and O. Krejcar, “Polyp-net: A multimodel fusion network for polyp segmentation,” IEEE TIM, vol. 70, pp. 1–12, 2020.
  • [36] T. Rahim, S. A. Hassan, and S. Y. Shin, “A deep convolutional neural network for the detection of polyps in colonoscopy images,” BSPC, vol. 68, p. 102654, 2021.
  • [37] D. Jha, S. Ali, N. K. Tomar, H. D. Johansen, D. Johansen, J. Rittscher, M. A. Riegler, and P. Halvorsen, “Real-time polyp detection, localization and segmentation in colonoscopy using deep learning,” IEEE Access, vol. 9, pp. 40 496–40 510, 2021.
  • [38] A. M. A. Ahmed, “Generative adversarial networks for automatic polyp segmentation,” in MediaEvalW, 2020.
  • [39] V. Thambawita, S. Hicks, P. Halvorsen, and M. A. Riegler, “Pyramid-focus-augmentation: Medical image segmentation with step-wise focus,” in MediaEvalW, 2020.
  • [40] N. K. Tomar, D. Jha, S. Ali, H. D. Johansen, D. Johansen, M. A. Riegler, and P. Halvorsen, “Ddanet: Dual decoder attention network for automatic polyp segmentation,” in ICPRW, 2021.
  • [41] C.-H. Huang, H.-Y. Wu, and Y.-L. Lin, “Hardnet-mseg: A simple encoder-decoder polyp segmentation neural network that achieves over 0.9 mean dice and 86 fps,” arXiv preprint arXiv:2101.07172, 2021.
  • [42] P. Chao, C.-Y. Kao, Y.-S. Ruan, C.-H. Huang, and Y.-L. Lin, “Hardnet: A low memory traffic network,” in CVPR, 2019.
  • [43] Y. Zhang, H. Liu, and Q. Hu, “Transfuse: Fusing transformers and cnns for medical image segmentation,” in MICCAI, 2021.
  • [44] Z. Yin, K. Liang, Z. Ma, and J. Guo, “Duplex contextual relation network for polyp segmentation,” in IEEE ISBI, 2022.
  • [45] Z. Xiaoqi, Z. Lihe, and L. Huchuan, “Automatic polyp segmentation via multi-scale subtraction network,” in MICCAI, 2021.
  • [46] Z. Zhou, J. Shin, L. Zhang, S. Gurudu, M. Gotway, and J. Liang, “Fine-tuning convolutional neural networks for biomedical image analysis: actively and incrementally,” in CVPR, 2017.
  • [47] N. Tajbakhsh, J. Y. Shin, S. R. Gurudu, R. T. Hurst, C. B. Kendall, M. B. Gotway, and J. Liang, “Convolutional neural networks for medical image analysis: Full training or fine tuning?” IEEE TMI, vol. 35, no. 5, pp. 1299–1312, 2016.
  • [48] X. Xie, J. Chen, Y. Li, L. Shen, K. Ma, and Y. Zheng, “Mi2{}^{\mbox{2}}gan: Generative adversarial network for medical image domain adaptation using mutual information constraint,” in MICCAI, 2020.
  • [49] R. Zhang, G. Li, Z. Li, S. Cui, D. Qian, and Y. Yu, “Adaptive context selection for polyp segmentation,” in MICCAI, 2020.
  • [50] N. K. Tomar, “Automatic polyp segmentation using fully convolutional neural network,” in MediaEvalW, 2020.
  • [51] D. Jha, S. Hicks, K. Emanuelsen, H. D. Johansen, D. Johansen, T. de Lange, M. A. Riegler, and P. Halvorsen, “Medico multimedia task at mediaeval 2020: Automatic polyp segmentation,” in MediaEvalW, 2020.
  • [52] K. Patel, A. M. Bur, and G. Wang, “Enhanced u-net: A feature enhancement network for polyp segmentation,” in CRV, 2021.
  • [53] A. Lumini, L. Nanni, and G. Maguolo, “Deep ensembles based on stochastic activation selection for polyp segmentation,” in MIDL, 2021.
  • [54] M. V. Branch and A. S. Carvalho, “Polyp segmentation in colonoscopy images using u-net-mobilenetv2,” arXiv preprint arXiv:2103.15715, 2021.
  • [55] R. Khadga, D. Jha, S. Ali, S. Hicks, V. Thambawita, M. A. Riegler, and P. Halvorsen, “Few-shot segmentation of medical images based on meta-learning with implicit gradients,” arXiv preprint arXiv:2106.03223, 2021.
  • [56] D. V. Sang, T. Q. Chung, P. N. Lan, D. V. Hang, D. Van Long, and N. T. Thuy, “Ag-curesnest: A novel method for colon polyp segmentation,” in IEEE RIVF, 2021.
  • [57] C. Yang, X. Guo, M. Zhu, B. Ibragimov, and Y. Yuan, “Mutual-prototype adaptation for cross-domain polyp segmentation,” IEEE JBHI, 2021.
  • [58] D. Jha, P. H. Smedsrud, D. Johansen, T. de Lange, H. D. Johansen, P. Halvorsen, and M. A. Riegler, “A comprehensive study on colorectal polyp segmentation with resunet++, conditional random field and test-time augmentation,” IEEE JBHI, vol. 25, no. 6, pp. 2029–2040, 2021.
  • [59] D. Jha, N. K. Tomar, S. Ali, M. A. Riegler, H. D. Johansen, D. Johansen, T. de Lange, and P. Halvorsen, “Nanonet: Real-time polyp segmentation in video capsule endoscopy and colonoscopy,” in IEEE CBMS, 2021.
  • [60] S. Li, X. Sui, X. Luo, X. Xu, L. Yong, and R. S. M. Goh, “Medical image segmentation using squeeze-and-expansion transformers,” in IJCAI, 2021.
  • [61] T. Kim, H. Lee, and D. Kim, “Uacanet: Uncertainty augmented context attention for polyp semgnetaion,” in ACM MM, 2021.
  • [62] V. Thambawita, S. A. Hicks, P. Halvorsen, and M. A. Riegler, “Divergentnets: Medical image segmentation by network ensemble,” in ISBI & EndoCV, 2021.
  • [63] G. Xiaoqing, Y. Chen, and Y. Yixuan, “Dynamic-weighting hierarchical segmentation network for medical images,” MIA, p. 102196, 2021.
  • [64] G.-P. Ji, Y.-C. Chou, D.-P. Fan, G. Chen, D. Jha, H. Fu, and L. Shao, “Pns-net: Progressively normalized self-attention network for video polyp segmentation,” in MICCAI, 2021.
  • [65] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in NeurIPS, 2017.
  • [66] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” in ICLR, 2021.
  • [67] Z. Pan, B. Zhuang, J. Liu, H. He, and J. Cai, “Scalable visual transformers with hierarchical pooling,” in ICCV, 2021.
  • [68] B. Heo, S. Yun, D. Han, S. Chun, J. Choe, and S. J. Oh, “Rethinking spatial dimensions of vision transformers,” in ICCV, 2021.
  • [69] L. Yuan, Y. Chen, T. Wang, W. Yu, Y. Shi, Z. Jiang, F. E. Tay, J. Feng, and S. Yan, “Tokens-to-token vit: Training vision transformers from scratch on imagenet,” in ICCV, 2021.
  • [70] K. Han, A. Xiao, E. Wu, J. Guo, C. Xu, and Y. Wang, “Transformer in transformer,” Advances in Neural Information Processing Systems, vol. 34, pp. 15 908–15 919, 2021.
  • [71] W. Wang, E. Xie, X. Li, D.-P. Fan, K. Song, D. Liang, T. Lu, P. Luo, and L. Shao, “Pyramid vision transformer: A versatile backbone for dense prediction without convolutions,” in ICCV, 2021.
  • [72] W. Wang, E. Xie, X. Li, D.-P. Fan, K. Song, D. Liang, T. Lu, P. Luo, and L. Shao, “Pvt v2: Improved baselines with pyramid vision transformer,” CVMJ, vol. 8, no. 3, pp. 415–424, 2022.
  • [73] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in ICCV, 2021.
  • [74] H. Wu, B. Xiao, N. Codella, M. Liu, X. Dai, L. Yuan, and L. Zhang, “Cvt: Introducing convolutions to vision transformers,” in ICCV, 2021.
  • [75] W. Xu, Y. Xu, T. Chang, and Z. Tu, “Co-scale conv-attentional image transformers,” in ICCV, 2021.
  • [76] X. Chu, Z. Tian, Y. Wang, B. Zhang, H. Ren, X. Wei, H. Xia, and C. Shen, “Twins: Revisiting the design of spatial attention in vision transformers,” Advances in Neural Information Processing Systems, vol. 34, pp. 9355–9366, 2021.
  • [77] B. Graham, A. El-Nouby, H. Touvron, P. Stock, A. Joulin, H. Jégou, and M. Douze, “Levit: a vision transformer in convnet’s clothing for faster inference,” in ICCV, 2021.
  • [78] S. Bhojanapalli, A. Chakrabarti, D. Glasner, D. Li, T. Unterthiner, and A. Veit, “Understanding robustness of transformers for image classification,” in ICCV, 2021.
  • [79] E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo, “Segformer: Simple and efficient design for semantic segmentation with transformers,” Advances in Neural Information Processing Systems, vol. 34, pp. 12 077–12 090, 2021.
  • [80] Z. Wu, L. Su, and Q. Huang, “Cascaded partial decoder for fast and accurate salient object detection,” in CVPR, 2019.
  • [81] S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in ICML, 2015.
  • [82] X. Glorot, A. Bordes, and Y. Bengio, “Deep sparse rectifier neural networks,” in AISTATS, 2011.
  • [83] S. Woo, J. Park, J.-Y. Lee, and I. So Kweon, “Cbam: Convolutional block attention module,” in ECCV, 2018.
  • [84] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in CVPR, 2018.
  • [85] X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural networks,” in CVPR, 2018.
  • [86] G. Te, Y. Liu, W. Hu, H. Shi, and T. Mei, “Edge-aware graph representation learning and reasoning for face parsing,” in ECCV, 2020.
  • [87] Y. Lu, Y. Chen, D. Zhao, and J. Chen, “Graph-fcn for image semantic segmentation,” in ISNN, 2019.
  • [88] J. Wei, S. Wang, and Q. Huang, “F3net: Fusion, feedback and focus for salient object detection,” in AAAI, 2020.
  • [89] I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” in ICLR, 2019.
  • [90] F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 3DV, 2016.
  • [91] R. Margolin, L. Zelnik-Manor, and A. Tal, “How to evaluate foreground maps?” in CVPR, 2014.
  • [92] M.-M. Chen and D.-P. Fan, “Structure-measure: A new way to evaluate foreground maps,” IJCV, vol. 129, pp. 2622–2638, 2021.
  • [93] D.-P. Fan, G.-P. Ji, X. Qin, and M.-M. Cheng, “Cognitive vision inspired object segmentation metric and loss function,” SSI, 2021.
  • [94] D.-P. Fan, C. Gong, Y. Cao, B. Ren, M.-M. Cheng, and A. Borji, “Enhanced-alignment measure for binary foreground map evaluation,” in IJCAI, 2018.
  • [95] Y. Fang, C. Chen, Y. Yuan, and K.-y. Tong, “Selective feature aggregation network with area-boundary constraints for polyp segmentation,” in MICCAI, 2019.
  • [96] J. Bernal, J. Sánchez, and F. Vilarino, “Towards automatic polyp detection with a polyp appearance model,” PR, vol. 45, no. 9, pp. 3166–3182, 2012.
  • [97] G.-P. Ji, G. Xiao, Y.-C. Chou, D.-P. Fan, K. Zhao, G. Chen, and L. Van Gool, “Video polyp segmentation: A deep learning perspective,” MIR, vol. 19, no. 06, pp. 531–549, 2022.