TSI-Net: A Timing Sequence Image Segmentation Network for Intracranial Artery Segmentation in Digital Subtraction Angiography
Abstract
Cerebrovascular disease is one of the major diseases facing the world today. Automatic segmentation of intracranial artery (IA) in digital subtraction angiography (DSA) sequences is an important step in the diagnosis of vascular related diseases and in guiding neurointerventional procedures. While, a single image can only show part of the IA within the contrast medium according to the imaging principle of DSA technology. Therefore, 2D DSA segmentation methods are unable to capture the complete IA information and treatment of cerebrovascular diseases. We propose A timing sequence image segmentation network with U-shape, called TSI-Net, which incorporates a bi-directional ConvGRU module (BCM) in the encoder. The network incorporates a bi-directional ConvGRU module (BCM) in the encoder, which can input variable-length DSA sequences, retain past and future information, segment them into 2D images. In addition, we introduce a sensitive detail branch (SDB) at the end for supervising fine vessels. Experimented on the DSA sequence dataset DIAS, the method performs significantly better than state-of-the-art networks in recent years. In particular, it achieves a Sen evaluation metric of 0.797, which is a 3% improvement compared to other methods.
Index Terms:
Digital subtraction angiography sequence, intracranial artery segmentation, UNet, ConvGRUI Introduction
Cerebrovascular disease, with its high incidence, has always been a major health issue worldwide[1]. The IA is complex and fragile, and a sudden break or blockage of blood flow somewhere can lead to insufficient blood supply to the entire brain, causing damage to brain tissue. This condition is known as a stroke and has an extremely high rate of disability and death[2].
The IA comprise one of the cerebrovascular systems within the human body. Alongside the main blood vessels, the intracranial region houses densely branching thin vessels that play a crucial role in transporting blood to different brain tissues. IA are much smaller in volume than the blood vessels in the rest of the body, and when combined with the pressure put on them by the brain tissue, they significantly slow down blood flow[3]. Cerebrovascular disease usually shows up as abnormalities like vessel narrowing and occlusion, including intracranial arterial stenosis and middle cerebral artery occlusion. It is possible to extract interest regions and ignore interference from other areas by segmenting the structure of the IA, which enables a clear observation of structural abnormalities in the IA. In order to make more accurate diagnoses and prevent them, doctors can use this information to assess the extent of cerebrovascular lesions as accurately as possible[4].
Early stage of cerebrovascular disease can effectively reduce IA accidents if IA imaging abnormalities are detected in time. Currently, digital subtraction angiography (DSA) remains the gold standard for visualization and diagnosis of angiographic features in most cerebrovascular diseases[5]. DSA is a vascular imaging technique. During diagnostic and neurovascular interventions, dynamic imaging of the cerebrovascular by DSA enables observation of blood flow dynamics as well as changes in vessel appearance over time[6]. The recording usually lasts from 3 to 15 seconds, with a sampling rate of 3 to 5 frames per second[7]. To date, DSA images have mostly been visually examined by neuroradiologists, which can be cumbersome, subjective, and prone to error. Recent studies on segmentation based on DSA images largely adhere to the traditional 2D image segmentation methods, that is, processing a single image extracted from the DSA sequence[8, 9]. Intraoperative DSA examinations for IA-related diseases typically involve sequential images over time, with the position of the contrast agent in the vessel changing at different times, resulting in temporal differences in image visibility. As shown in Fig. 1, only a portion of the IA that filled with contrast agents can be displayed in a single image. In fact, successful detection of IA abnormalities depends largely on clear and accurate vascular imaging information in the DSA image sequence. 2D DSA segmentation methods alone are inadequate for the diagnosis and treatment of cerebrovascular diseases due to their failure to capture the complete information related to the IA. The segmentation of DSA sequences exhibiting 2D+time properties should be pursued. Therefore, we consider the segmentation of DSA sequences in IA as a task of dimensionality reduction, where the source consists of 2D+ time images and the target is a 2D mask.
To solve the issue, we propose a Timing Sequence Image segmentation network (TSI-Net) that takes variable frames DSA sequences as input. In TSI-Net, the bi-directional ConvGRU[10] module (BCM) are used to model temporal blood flow features, which learn the spatio-temporal correlation of the target from the features of the current image and its neighboring frames. According to the special output mode of BCM, the final output of the network is a 2D image, which realizes the dimensionality reduction segmentation. To improve the sensitivity of the segmentation, a sensitive detail branch (SDB) is employed for precision segmentation of fine vessels. In summary, the contributions in the paper can be summarized as follows:
-
•
We designed a timing sequence image segmentation network (TSI-Net) by a U-shape. There is a BCM for processing continuous time image information. In BCM, we concatenate multiple Gate Recurrent Unit cells in both forward and reverse order directions, and finally connect them head to tail to form a GRU layer. The update and reset gates of each ConvGRU cell can learn and forget the contents of the current image memory in the previous or next frame, respectively.
-
•
At the output side we employ an SDB to perform auxiliary supervision of the network, which is used to improve the sensitivity and connectivity of the segmentation.
-
•
The experimental results demonstrate the necessity of BCM and the effectiveness of SDB. Comparative results on DIAS with other state-of-the-art methods validate the promising segmentation performance of the proposed approach.
II Related Work
Thanks to the development of biomedical technology, we can easily access a large amount of biological and physiological data through modern equipment. At the same time, the rapid development of deep learning has brought new ideas to solve the problem of IA segmentation, assisting doctors to realize the extraction and quantification of vascular indicators for clinical diagnosis.
The groundbreaking U-Net developed by Ronneberger et al.[11] is considered a pioneering method in the field of segmentation and has become the focal point of medical image segmentation due to its outstanding performance. As a convolutional neural network, derivatives of U-Net have been widely applied to various vascular segmentation tasks. For example, 2D[12, 13, 14] or 3D[15, 16] methods with modifications made mainly to the U-Net backbone network to achieve better segmentation results. They use the structure of an encoder-decoder and introduce popular structural components such as attention mechanism[15, 12, 17], atrous convolution[13, 18], gate structure[17, 16]. Liu et al.[19] combined the original 3D U-Net with Maximum Intensity Projection and used a spatial attention-guided 3D Inception U-Net segmentation flow and a 2D composite multi-directional MIPs U-Net segmentation flow to segment 3D MRA IA images. Zhang et al.[13] designed a cascade of modules by combining the spatial attention module with the dense atrous convolution module and multi-kernel pooling module to obtain a larger receptive field for retinal vessel segmentation. Liu et al.[17] used convolutional long short-term memory network and deep dilated convolution in the decoder for retinal vessel segmentation. Lee et al.[16] considered the connectivity of vessels between axial slices to achieve MR modality IA segmentation and inserted long short-term memory (LSTM) into the model to capture the context of continuous data.
However, there is limited research on DSA vascular segmentation. Meng et al.[14] proposed a CNN-based segmentation framework that includes a multi-scale atlas convolution module, improved dense blocks, and redesigned skip connections for the segmentation of IA in DSA images. Zhang et al.[20] proposed a boundary enhancement and feature denoising module to improve the network’s ability to extract boundary information in coronary artery DSA segmentation. These two methods were developed on an internal DSA dataset and are only applicable to 2D DSA vascular segmentation. However, DSA is a modality with powerful spatio-temporal visualization of blood flow dynamics in continuous frames, and these methods fail to consider the temporal characteristics of DSA imaging patterns and do not fully utilize the temporal and spatial correlation in IA DSA images.
III Method
The automatic IA segmentation of DSA sequences is an unprecedented and challenging task. We propose a timing sequence image segmentation network called TSI-Net, and its structure is shown in Fig. 2
III-A TSI-Net
The TSI-Net architecture consists of an encoder-decoder network based on the U-shape design. The encoder of TSI-Net receives preprocessed IA DSA sequences as input. In contrast to the original U-Net architecture, the DSA sequences are fed to the network frame by frame. The bi-directional ConvGRU module (BCM) is utilized to incorporate temporal information features, where each group of Down-Conv shares one weights. The encoder consists of four IA down-sampling layers, and each frame of one sequence undergoes two convolution in the same layer and then connect into BCM block frame by frame. Following the bi-directional ConvGRU module (BCM), the output sequence undergoes IA down-sampling through maximum pooling with a step size of 2, and the output of the final ConvGRU cell is connected to the decoder using a skip connection.
Each IA down-sampling doubles the number of channels in the feature map, which increases the receptive field of the deep features and ensures their global nature to some extent. The decoder is symmetric to the encoder and consists of 4 up-sampling layers. Connected the up-sampling feature maps with the same resolution feature maps from encoder can reduce the loss of edge features because of up-sampling. The features of the encoder have higher resolution, while they extracted by down-sampling, this information is not retrievable at the time of up-sampling. Finally, a 1x1 fully connected layer is used to classify the background and blood vessels in the image to output a segmented image of IA. The number of channels start from 32. An important improvement over U-Net is the use of 2D+T DSA sequences that contain temporal information instead of a single 2D image as an input.
III-B Bidirectional ConvGRU Module
It is essential to depict the sequential evolution of blood flow contrast information in the temporal domain, Since a DSA sequence presents a series of IA information in chronological order. Our proposal involves utilizing the BCM to simulate the temporal progression of blood flow appearance in the sequence. ConvGRU[10] is an extension of the traditional fully connected GRU, with convolutional structures in both the input-to-state and state-to-state connections. Define the given sequence with t-frames as S = (,,…,), and represents the hidden state at time t. A ConvGRU cell consists of a reset gate and a update gate are expressed as and . Through these gates, ConvGRU cells can not only achieve selective memory but also forgetting. Based on the above definition, the overall update process at the ConvGRU cell for different time steps in a DSA sequence is as follows:
(1) |
where ‘*’ denotes the convolutional operation, ‘’ represents the hadamard product, denotes sigmoid function, , means weights that can be learned. is a candidate activation which is computed in a manner similart to that of a traditional recurrent unit in an Recurrent Neural Network. The formula simplifies the expression and ignores the bias term.
We input the DSA sequences frame-by-frame into encoder for feature extraction. Inspired by[21], we stack two sets of ConvGRU to compose BCM in forward and backward propagation, respectively, to enhance the exchange of spatio-temporal information between them. In this way, the BGM not only remembers past sequences but also predicts future sequences. This can be expressed as:
(2) |
where and represent the hidden state from forward and backward ConvGRU cells, respectively. represents the final output of BCM.
It should be noted that all blocks in the BCM share the same weights. The last layer of each BCM aggregates the features of all frames in the DSA sequence. This operation is designed to allow the network to learn the relationships between frames rather than treating them as individual frames for processing.
III-C Sensitive Detail Branch
Data imbalance is a common problem in medical image segmentation. The region of inerest in a medical image is usually much less than entire image. Moreover the proportion of voxels in the regions to be segmented is smaller than the rest of the image. If the data is unbalanced, the learning process may converge to a suboptimal local minimum, and the final prediction results could be biased towards non-relevant features. This data imbalance issue is particularly severe in vessel segmentation.
Consequently, we used sensitive detail branch (SDB) for adjunctive monitoring of the thin vessels. In this branch, we use the skeletonization operation to perform centerline extraction of the original ground truth (GT) as a new label. In this module for assisted supervision, we calculate the mean square error values between the actual and predicted values one by one. The aim is to keep reducing the gap between the two, making the network more sensitive to fine vessels and some details, and maintaining a higher connectivity between the vessels.It can be defined as:
(3) |
where represent predicted probability of network output, denotes GT, is the centerline lifting method using the skeletonization operation.
On the one hand, the most often used loss function is the cross-entropy loss, which examines each pixel individually and compares the class prediction vectors to the GT. It can be defined as:
(4) |
where N is the total pixel number.
On the other hand, the dice coefficient is a measure of the overlap area of the picture difference region and is used to compare the predicted map to the GT. It is a superior metric for minor goals. The dice loss () can be defined as:
(5) |
The final loss function is defined as:
(6) |
where and denotes the weight of and respectively.
Auxiliary supervised mean square error loss improves the sensitivity of the model. Sensitivity measures how many true positives the model correctly detects among all actual positives. As a result, it allows for better segmentation of the end vessels and fine vessels of the vascular tree, improving segmentation accuracy and vascular connectivity.
IV Experiment
IV-A Data and Evaluation
To validate the performance of our proposed TSI-Net on the IA segmentation task, we conducted experiments on the DIAS dataset[7].
The DIAS dataset contains 60 one-on-one pixel-level labeled sequences. In the same way as the data are divided in [7], 30 as training samples, 10 as validation samples, and 20 as test samples. The DSA sequences have dimensions of . While is the resolution of a single frame of the image in the sequence, and 8 is the number of frames in a single sequence.
We compare the predicted segmentation results with the corresponding ground truth to calculate the dice similarity coefficient (Dice) and evaluate the accuracy (Acc), sensitivity (Sen), specificity (Spe), and intersection over union (IOU) of the binary segmentation maps obtained using thresholding. Their definitions are as follows:
(7) |
(8) |
(9) |
(10) |
(11) |
where true positives (TP) and true negatives (TN) denote the numbers of correctly segmented vascular pixels and nonvascular pixels, respectively; false positives (FP) and false negatives (FN) denote the numbers of incorrectly segmented vascular pixels and nonvascular pixels, respectively.
Additionally, we introduced the vascular connectivity (VC) metric[7]. This metric calculates the ratio of connected components in the segmentation image to those in the ground truth, providing a measure of vessel connectivity.
IV-B Implementation Details
Firstly, all DSA sequences were z-score normalized. We randomly crop the input sequences into patches of size . We randomly crop the input image into patches of size . Randomized horizontal, vertical and 90-degree flips were used for data enhancement. The preprocessed sequences are used for training with an epoch size of 64. We optimize the model parameters using AdamW, with an initial learning rate of 5e-4. The learning rate is gradually reduced using the cosine annealing algorithm for more than 100 iterations. All experiments are conducted using PyTorch with constant hyperparameters and on a single GeForce RTX 3090 GPU.
V Result
V-A Comparisons With the State-of-The-Art Methods
We conducted IA segmentation experiments on the most popular networks, including U-Net[11], Res-UNet[22], UNet++[23] and CS2Net[24]. To address the limitation of network processing only individual 2D images, a dimension reduction module (DRM)[7] is incorporated to handle the DSA sequence. Specifically, given a DSA sequence , where , , and denote the number of channels, frame, height, and width, respectively. After DRM compression, the initial channel number is changed to 1. The compressed input is then fed into the 2D model, where property F replaces C as the input channel:
(12) |
Methods | Metrics | ||||||
---|---|---|---|---|---|---|---|
Dice | AUC | Acc | Sen | Spe | IOU | VC↓ | |
UNet | 0.7633 | 0.9797 | 0.9637 | 0.7601 | 0.9818 | 0.6198 | 79.60 |
Res-UNet | 0.765 | 0.9787 | 0.9644 | 0.7593 | 0.9826 | 0.6217 | 73.05 |
UNet++ | 0.7643 | 0.9808 | 0.9645 | 0.7537 | 0.9832 | 0.6211 | 68.68 |
CS2Net | 0.769 | 0.9817 | 0.9648 | 0.7663 | 0.9826 | 0.6272 | 68.40 |
TSI-Net | 0.7861 | 0.9816 | 0.9664 | 0.7971 | 0.9815 | 0.6492 | 42.70 |
where, is dimensionality squeeze operation and is a 2D model.
Fig. 3 shows the segmentation results. We perform minimizing operations on the entire sequence for each pixel along the sequence dimension to obtain minimum intensity projection (MIP) images. It can be seen that all models have comparable segmentation ability for large blood vessels, while for small blood vessels, there is a significant reduction of red pixels in the segmentation results of TSI-Net, which indicates that our network is more powerful for fine vessel segmentation.
The qualitative segmentation results for the DIAS dataset are presented in Table I. Overall, TSI-Net demonstrates the best performance on this dataset, exhibiting an approximately 2% improvement in Dice coefficient, which is outperforming to state-of-the-art methods.
Notably, TSI-Net exhibited a noticeable 3% improvement in the Sen score and a significant reduction in the VC compared to other methods, while only experiencing a marginal 0.1% decrease in Auc. These enhancements imply an improved ability of the model to accurately segment fine vessels and ensure improved vessel connectivity with fewer disruptions. Precisely capturing vessel distribution in interventional procedures serves as a crucial benchmark, and both metrics hold particular significance in scenarios characterized by highly imbalanced background and target distributions.
However, in contrast to the other composite evaluation metrics, the Spe is relatively low. Sen and Spe represent the proportions of correctly predicted positive and negative pixels, respectively. During the extraction of weakly vascularized pixels, TSI-Net introduces some false positives, indicating incorrect predictions of non-vascular pixels. We believe that the model further improves its ability to extract weak blood vessels while naturally increasing the acceptance of false positives.
Method | Metrics | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Baseline | UCM | BCM | SDB | Dice | AUC | Acc | Sen | Spe | IOU | VC |
✓ | 0.7633 | 0.9797 | 0.9637 | 0.7601 | 0.9818 | 0.6198 | 79.60 | |||
✓ | ✓ | 0.7851 | 0.9849 | 0.9681 | 0.7549 | 0.9868 | 0.6478 | 64.32 | ||
✓ | ✓ | 0.7842 | 0.9854 | 0.9671 | 0.7729 | 0.9844 | 0.6468 | 54.49 | ||
✓ | ✓ | ✓ | 0.7861 | 0.9816 | 0.9664 | 0.7971 | 0.9815 | 0.6492 | 42.70 |
V-B Ablation Studies
To validate the effectiveness of the components, we conducted ablation experiments and evaluated the impact of each component on the results on the DIAS dataset. Using UNet as the baseline for the experiments, We incrementally incorporated Uni-directional ConvGRU Module (UCM) ,BCM and SDM to the baseline of the ablation study. All experiments were conducted using the same hyperparameter configuration. In our ablation study, all competitors were run on the same computing environment with the same data augmentation to ensure a fair comparison. Again since UNet cannot handle DSA sequences, the initial number of channels is set to 1 and compressed by DRM. The compressed input is then fed into the 2D model where ”F” replaces ”C” as the input channel. The results are shown in the Table II.
The inclusion of ConvGRU resulted in significant improvements in all model scores compared to the baseline, with bi-directional connectivity demonstrating superior performance over uni-directional connectivity across multiple metrics. When compared to baseline + BCM, the addition of SDB further enhanced the Sen score by 2.4%, achieving a maximum score of 0.7971. Auc and Spe exhibited only a marginal decrease of approximately 0.4%. Considering the degree of imbalance between IA vascular and background and practical medical application scenarios, the noted improvement in Sen and VC indicates that the model facilitates enhanced preservation of fine vascular information and vascular topology, ultimately enhancing the performance of sequence segmentation.
The ablation results were further visualized, as depicted in Fig 4, which includes an IA sequence featuring numerous fine vessels. Consequently, we zoomed in on specific image details to facilitate clearer visualization, focusing on the highlighted rectangular area. Within the zoomed subfigure, the thin vessels in the original image exhibit low contrast, appearing blurred and challenging to discern by the human eye. The expert has labeled the blood vessels in the GT, and based on the visualization of the segmentation results, it can be observed that Baseline, Baseline+UCM, and Baseline+BCM show incomplete extraction of the thin blood vessels in this area. However, the inclusion of Baseline+BCM+SDB resolves this issue by introducing fewer false-negative pixels, thereby visibly enhancing vascular connectivity.
VI Conclusion
We propose a spatio-temporal segmentation method for IA based on a timing sequence image segmentation network. In this network, the incorporation of BCM enables the network to consider both past and future information to better learn IA sequences that contain temporal information. Unlike previous segmentation networks, this network takes 2D+time DSA sequences as input and outputs 2D segmented images. Evaluation metrics show that TSI-Net achieves excellent overall performance compared to other state-of-the-art methods. In addition, the visualization of IA segmentation results with the addition of SDB demonstrated the power of TSI-Net in extracting fine vessels and improving vascular connectivity. In the future, we will continue to explore this work to improve network sensitivity while being able to achieve better overall correctness.
References
- [1] M. S. Phipps and C. A. Cronin, “Management of acute ischemic stroke,” Bmj, vol. 368, 2020.
- [2] M. Portegies, P. Koudstaal, and M. Ikram, “Cerebrovascular disease,” Handbook of clinical neurology, vol. 138, pp. 239–261, 2016.
- [3] K. M. Welch, Primer on cerebrovascular diseases. Gulf Professional Publishing, 1997.
- [4] M. R. Goni, N. I. R. Ruhaiyem, M. Mustapha, A. Achuthan, and C. M. N. C. M. Nassir, “Brain vessel segmentation using deep learning-a review,” IEEE Access, 2022.
- [5] A. Raman, M. Uprety, M. J. Calero, M. R. B. Villanueva, N. Joshaghani, N. Villa, O. Badla, R. Goit, S. E. Saddik, S. N. Dawood et al., “A systematic review comparing digital subtraction angiogram with magnetic resonance angiogram studies in demonstrating the angioarchitecture of cerebral arteriovenous malformations,” Cureus, vol. 14, no. 6, 2022.
- [6] R. Su, M. van der Sluijs, S. Cornelissen, P. J. van Doormaal, R. v. d. Broek, W. van Zwam, J. Hofmeijer, D. Ruijters, W. Niessen, A. van der Lugt et al., “Spatio-temporal u-net for cerebral artery and vein segmentation in digital subtraction angiography,” arXiv preprint arXiv:2208.02355, 2022.
- [7] W. Liu, T. Tian, L. Wang, W. Xu, H. Li, W. Zhao, X. Pan, H. Yang, F. Gao, Y. Deng et al., “Dias: A comprehensive benchmark for dsa-sequence intracranial artery segmentation,” arXiv preprint arXiv:2306.12153, 2023.
- [8] C. Meng, K. Sun, S. Guan, Q. Wang, R. Zong, and L. Liu, “Multiscale dense convolutional neural network for dsa cerebrovascular segmentation,” Neurocomputing, vol. 373, pp. 123–134, 2020.
- [9] R.-Q. Li, G.-B. Bian, X.-H. Zhou, X. Xie, Z.-L. Ni, and Z. Hou, “Cau-net: A novel convolutional neural network for coronary artery segmentation in digital substraction angiography,” in Neural Information Processing: 27th International Conference, ICONIP 2020, Bangkok, Thailand, November 23–27, 2020, Proceedings, Part I 27. Springer, 2020, pp. 185–196.
- [10] N. Ballas, L. Yao, C. Pal, and A. Courville, “Delving deeper into convolutional networks for learning video representations,” arXiv preprint arXiv:1511.06432, 2015.
- [11] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer, 2015, pp. 234–241.
- [12] Y. Liu, J. Shen, L. Yang, G. Bian, and H. Yu, “Resdo-unet: A deep residual network for accurate retinal vessel segmentation from fundus images,” Biomedical Signal Processing and Control, vol. 79, p. 104087, 2023.
- [13] T. Zhang, J. Li, Y. Zhao, N. Chen, H. Zhou, H. Xu, Z. Guan, C. Yang, L. Xue, R. Chen et al., “Mc-unet multi-module concatenation based on u-shape network for retinal blood vessels segmentation,” arXiv preprint arXiv:2204.03213, 2022.
- [14] C. Meng, K. Sun, S. Guan, Q. Wang, R. Zong, and L. Liu, “Multiscale dense convolutional neural network for dsa cerebrovascular segmentation,” Neurocomputing, vol. 373, pp. 123–134, 2020.
- [15] L. Xia, H. Zhang, Y. Wu, R. Song, Y. Ma, L. Mou, J. Liu, Y. Xie, M. Ma, and Y. Zhao, “3d vessel-like structure segmentation in medical images by an edge-reinforced network,” Medical Image Analysis, vol. 82, p. 102581, 2022.
- [16] K. Lee, L. Sunwoo, T. Kim, and K. J. Lee, “Spider u-net: Incorporating inter-slice connectivity using lstm for 3d blood vessel segmentation,” Applied Sciences, vol. 11, no. 5, p. 2014, 2021.
- [17] C. Liu, P. Gu, Z. Xiao et al., “Multiscale u-net with spatial positional attention for retinal vessel segmentation,” Journal of Healthcare Engineering, vol. 2022, 2022.
- [18] Y. Yin, Z. Han, M. Jian, G.-G. Wang, L. Chen, and R. Wang, “Amsunet: A neural network using atrous multi-scale convolution for medical image segmentation,” Computers in Biology and Medicine, p. 107120, 2023.
- [19] Y. Liu, H.-S. Kwak, and I.-S. Oh, “Cerebrovascular segmentation model based on spatial attention-guided 3d inception u-net with multi-directional mips,” Applied Sciences, vol. 12, no. 5, p. 2288, 2022.
- [20] M. Zhang, F. Yu, J. Zhao, L. Zhang, and Q. Li, “Befd: Boundary enhancement and feature denoising for vessel segmentation,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, October 4–8, 2020, Proceedings, Part V 23. Springer, 2020, pp. 775–785.
- [21] P. Yan, G. Li, Y. Xie, Z. Li, C. Wang, T. Chen, and L. Lin, “Semi-supervised video salient object detection using pseudo-labels,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 7284–7293.
- [22] X. Xiao, S. Lian, Z. Luo, and S. Li, “Weighted res-unet for high-quality retina vessel segmentation,” in 2018 9th international conference on information technology in medicine and education (ITME). IEEE, 2018, pp. 327–331.
- [23] Z. Zhou, M. M. Rahman Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: A nested u-net architecture for medical image segmentation,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4. Springer, 2018, pp. 3–11.
- [24] L. Mou, Y. Zhao, H. Fu, Y. Liu, J. Cheng, Y. Zheng, P. Su, J. Yang, L. Chen, A. F. Frangi et al., “Cs2-net: Deep learning segmentation of curvilinear structures in medical imaging,” Medical image analysis, vol. 67, p. 101874, 2021.