This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

SSHNN: Semi-Supervised Hybrid NAS Network for Echocardiographic Image Segmentation

Abstract

Accurate medical image segmentation especially for echocardiographic images with unmissable noise requires elaborate network design. Compared with manual design, Neural Architecture Search (NAS) realizes better segmentation results due to larger search space and automatic optimization, but most of the existing methods are weak in layer-wise feature aggregation and adopt a “strong encoder, weak decoder” structure, insufficient to handle global relationships and local details. To resolve these issues, we propose a novel semi-supervised hybrid NAS network for accurate medical image segmentation termed SSHNN. In SSHNN, we creatively use convolution operation in layer-wise feature fusion instead of normalized scalars to avoid losing details, making NAS a stronger encoder. Moreover, Transformers are introduced for the compensation of global context and U-shaped decoder is designed to efficiently connect global context with local features. Specifically, we implement a semi-supervised algorithm Mean-Teacher to overcome the limited volume problem of labeled medical image dataset. Extensive experiments on CAMUS echocardiography dataset demonstrate that SSHNN outperforms state-of-the-art approaches and realizes accurate segmentation. Code will be made publicly available.

Index Terms—  Medical image segmentation, Hybrid NAS, Semi-supervised learning, Transformer

1 Introduction

2D echocardiography is a critical medical imaging technique for clinical routine to measure the cardiac morphology and function, and further obtaining a diagnosis [1]. Due to noise, artifacts and low contrast caused by uneven illumination, organ boundaries in echocardiographic images are blurred [2]. Besides, structures in echocardiographic images appear differently at various scales, requiring the model effectively handles the multi-scale features for medical image segmentation.

Benefiting from receptive field, Convolutional Neural Networks (CNNs) represented extraordinary feature extraction ability and are frequently adopted [3, 4, 5]. Especially in medical area, the proposal of U-shaped encoder-decoder network has propelled semantic segmentation to the superior development [6, 7, 8, 9]. Moreover, when the Transformer [10] is introduced into computer vision and converted to vision Transformer (ViT)[11] to enhance the global receptive ability, the combination of U-shaped network with Transformer is focused, such as TransUNet [12], Swin Transformer [13], etc. The ever-proposed manual designed networks improve segmentation accuracy while becoming more complex.

To resolve the difficulty and complexity of network designing, Neural Architecture Search (NAS) is proposed, aiming to design automatically and accurately. Among them, NAS-UNet [14] replaces the CNN portion with searchable cells to reduce network parameters. In order to expand search space for better generalization, Hierarchical NAS (HNAS) also searches in layer-level, such as Auto-DeepLab [15], Dynamic routing [16], etc. Moreover, HCT-Net [17] adopts ViT to add context information. However, these methods employ normalized scalars for feature fusion in lay-wise optimizing, prone to local detail loss caused by insufficient parameters. Besides, despite having usable multi-scale features, decoder only depends on concatenation and convolution layers, contributing to “strong encoder, weak decoder”, incapable to capture sufficient context and local information [15]. Hence, our focus is enhance model segmentation ability through the integration of exemplary manual design and HNAS.

Refer to caption
Fig. 1: The network framework of SSHNN. (a) Overview of SSHNN. Dashed lines denote candidate connections. We only show several connections among all skip connections to keep clarity. (b) Transformer layer module. (c) Decoder cell module. (d) Encoder cell module. Left: Feature fusion details. Middle: Inner cell search space. Right: Partial channel connections.

In this paper, we propose a novel hybrid NAS with semi-supervised learning considering the size limitation of labeled medical image dataset, named SSHNN. In the design of HNAS, we replace simple scalars with convolutions for local feature aggregation, pursuing higher degree of flexibility. To overcome “strong encoder, weak decoder”, add U-shaped decoder after HNAS to recover multi-scale features into original size, and Transformer is utilized to compensate for global context. Experiments prove SSHNN outperforms state-of-the-art (SOTA) methods on public medical image dataset on echocardiography: CAMUS [1].

2 Methodologies

In this section, We elaborate SSHNN in four parts: HNAS design, decoder, semi-supervised learning and optimization strategy. The network framework is shown in Fig. 1.

2.1 HNAS design

Inner Cell: Cell search space can be represented by a directed acyclic graph, consisting of blocks and edges, denoting the mapping from input tensors to 1 output tensor, and candidate operations, respectively. For the ithi^{th} block of a cell in the lthl^{th} layer, define a tuple (Iil,Oil)(I_{i}^{l},O_{i}^{l}) to denote the mapping, where IilI_{i}^{l}\in\mathcal{I} denotes the input tensor, and Oil𝒪O_{i}^{l}\in\mathcal{O} denotes the candidate operation. The set of input tensors consists of previous cell’s output Hl1H^{l-1}, previous-previous cell’s output Hl2H^{l-2}, and previous blocks’ output in the current cell H1l,,Hi1l{H_{1}^{l},\dots,H_{i-1}^{l}}. The set of candidate operations 𝒪\mathcal{O} includes: depthwise-separable convolution, dilated convolution, average polling, max polling, skip connection and no connection.

To reduce memory cost, partial channel connections [18] are adopted, where 1/n1/n portion of dimensional features are sent to the cell and the rest features remain unchanged. Moreover, continuous relaxation [19] is reused as differentiable search space is pursued and then the stochastic gradient descent (SGD) can be applied. Thus, the output tensor of block:

Hil=HjlOkl𝒪exp{αjik}m=1|𝒪|exp{αjim}Okl(PartjiHjl)+(1Partji)Hjl\begin{aligned} H_{i}^{l}&=\sum_{H_{j}^{l}\in\mathcal{I}}\sum_{O_{k}^{l}\in\mathcal{O}}\frac{exp\{\alpha_{j\rightarrow i}^{k}\}}{\sum_{m=1}^{|\mathcal{O}|}exp\{\alpha_{j\rightarrow i}^{m}\}}\cdot O_{k}^{l}(Part_{j\rightarrow i}\circ H_{j}^{l})\\ &+(1-Part_{j\rightarrow i})\circ H_{j}^{l}\end{aligned}

(1)

where PartjiPart_{j\rightarrow i} is sampling mask for channel selection, and α\alpha are normalized scalars, called architecture parameters, measuring the weight of candidate operation. To ensure the differentiable back propagation, softmax is implemented. Finally, the output tensor Hl=Concat({Hil|i<=B})H^{l}=Concat(\{H_{i}^{l}|i<=B\}), where Concat()Concat(\cdot) is concatenation and BB is the number of blocks. Cell level search is denoted as Hl=Cell(Hl1,Hl2;α)H^{l}=Cell(H^{l-1},H^{l-2};\alpha).
Out Layer: Layer level search aims at finding the optimal network backbone within network search space for specific dataset, with the specific procedure is combine features from different resolutions for better feature extraction. In past researches [15, 20, 21], normal algorithm uses another architecture parameter β\beta for feature fusion by linear combination. However, the amount of β\beta is a not large, which is usually in the hundreds, but are responsible for complex feature fusion, especially after applying partial channel connection. By estimation in Sec. 3.4, linear combination cannot fully fuse multi-scale features since hundreds of channels represent different features, which needs more parameters to elaborate.

Thus, we replace normalized scalars with convolution kernels to realize higher degree of flexibility, shown in Fig. 1(d, LeftLeft). Furthermore, long and short skip connection is utilized due to an incredible ability in image segmentation [7, 21]. The whole representation of layer design:

Hls=Conv(Concat(Down(Cell(s2Hl1,sHl2;α)),Up(Cell(2sHl1,sHl2;α)),Cell(sHl1,sHl2;α),{sHlHs|l<l}))\begin{aligned} {}^{s}H^{l}=&\ Conv(Concat(Down(Cell(^{\frac{s}{2}}H^{l-1},^{s}H^{l-2};\alpha)),\\ &Up(Cell(^{2s}H^{l-1},^{s}H^{l-2};\alpha)),Cell(^{s}H^{l-1},^{s}H^{l-2};\alpha),\\ &\{^{s}H^{l^{\prime}}\in{{}^{s}H}|l^{\prime}<l\}))\end{aligned}

(2)

where Conv()Conv(\cdot) denotes the convolution, transforming fused features to the same channel counts of ss-resolution, Down()Down(\cdot) denotes downsampling, and Up()Up(\cdot) denotes upsampling.

2.2 Decoder

Rather than simply concatenating multi-scale features and then processing by convolution layers to revert features to the original image size after searching optimal network structure, where these layers act as decoder and HNAS acts as encoder essentially, we apply ViT to add global context and use U-shape decoder structure, shown in Fig. 1(a).

Define input image 𝐱H×W×C\mathbf{x}\in\mathbb{R}^{H\times W\times C}, where H×WH\times W denotes spatial resolution and CC denotes channel counts. First, perform tokenization [22]. Reshape 𝐱\mathbf{x} into {𝐱piP×P×C|i=1,,N}\{\mathbf{x}_{p}^{i}\in\mathbb{R}^{P\times P\times C}|i=1,\dots,N\} by P×PP\times P convolution (stride = PP), where N=HWP2N=\frac{HW}{P^{2}}. Second, patch embedding and transformer. Map the patches 𝐱p\mathbf{x}_{p} into a D-dimensional embedding space and add specific position embedding to keep position information, and then apply Transformers, including Multi-head Self-Attention (MSA) and Multi-layer Perceptron (MLP) blocks:

𝐳l\displaystyle\small\mathbf{z}_{l}^{*} =MSA(LN(𝐳l1))+𝐳l1\displaystyle=MSA(LN(\mathbf{z}_{l-1}))+\mathbf{z}_{l-1} (3)
𝐳l\displaystyle\mathbf{z}_{l} =MLP(LN(𝐳l))+𝐳l\displaystyle=MLP(LN(\mathbf{z}_{l}^{*}))+\mathbf{z}_{l}^{*} (4)

where LN()LN(\cdot) represents the layer normalization.

Thus, Transformer output is 𝐳lHWP2×D\mathbf{z}_{l}\in\mathbb{R}^{\frac{HW}{P^{2}}\times D} and then we reshape it into HP×WP×D\mathbb{R}^{\frac{H}{P}\times\frac{W}{P}\times D} for decoder. Herein, we need to combine the global features extracted from Transformers with the local features extracted from NAS. To avoid losing low-level or high-level details, use U-shape decoder is necessary.

We first use a 2D convolution to adjust the channels of encoded features from Transformer to have the same number of channels as Hls{}^{s}H^{l} for feature aggregation. Since we do not expect that transformers bring excessive parameters, HP\frac{H}{P} and WP\frac{W}{P} are small and then should be upsampled to ss-resolution spatial size. After reshaping and resamping, concatenate two parts of features, followed with a convolution layer to match the number of channels with the next resolution Hls2{}^{\frac{s}{2}}H^{l} channel counts, and then attain the output of current resolution.

Similarly, “Upsample-Concatenation-Convolution” (shown in Fig. 1(c)) is implemented again for the next resolution features until combining s=4s=4 resolution features. For segmentation purposes, finally a upsample layer and a convolution layer are used to recover the features to the full resolution and specific number of classes for predicting the dense output.

2.3 Semi-Supervised Learning

As medical image dataset always has a small amount of labeled images and a large amount of unlabeled data. Thus, we try to combine NAS with Mean Teacher method [23], to enhance the effectiveness and generalisation of the model.

For Mean Teacher, we have two models: StudentStudent f(θs)f(\theta_{s}) and TeacherTeacher f(θt)f(\theta_{t}), where θs\theta_{s} and θt\theta_{t} denote the network parameters of StudentStudent and TeacherTeacher, respectively. θt\theta_{t} are updated by the exponential moving average (EMA) of θs\theta_{s}:

θt,i=αθs,i1+(1α)θs,i\theta_{t,i}=\alpha\theta_{s,i-1}+(1-\alpha)\theta_{s,i} (5)

where α\alpha is a hyperparameter controlling the updating speed, ii is iteration times. Given labeled dataset 𝒟l\mathcal{D}_{l} and unlabeled dataset 𝒟u\mathcal{D}_{u}, define the consistency regularization to update θs\theta_{s}:

c=𝒟llMSE(S(plt),S(pls))m+𝒟ulMSE(S(put),S(pus))n\begin{aligned} \mathcal{L}_{c}=\frac{\sum_{\mathcal{D}_{l}}{l}_{MSE}(S(p_{l}^{t}),S(p_{l}^{s}))}{m}+\frac{\sum_{\mathcal{D}_{u}}{l}_{MSE}(S(p_{u}^{t}),S(p_{u}^{s}))}{n}\end{aligned}

(6)

where pltp_{l}^{t} and plsp_{l}^{s} are the outputs of TeacherTeacher and StudentStudent on labeled images, putp_{u}^{t} and pusp_{u}^{s} are the outputs on unlabeled images, lMSEl_{MSE} is the Mean-Square Error, and S()S(\cdot) is softmax function applied in channel dimension for scale controlling.

The supervised loss of StudentStudent is computed by the output of TeacherTeacher and the ground truth of labeled images:

s=1m𝒟llCE(plt,yl)\mathcal{L}_{s}=\frac{1}{m}\sum_{\mathcal{D}_{l}}l_{CE}(p_{l}^{t},y_{l}) (7)

where lCEl_{CE} is cross entropy loss function. Thus, the total loss used to train the θs\theta_{s} is the sum of the consistency regularization and the supervised loss: total=λ0s+λ1c\mathcal{L}_{total}=\lambda_{0}\mathcal{L}_{s}+\lambda_{1}\mathcal{L}_{c}, where λ0=1\lambda_{0}=1 and λ1\lambda_{1} follows exponential ramp-up function in [24].

2.4 Optimization

To construct a differentiable computation graph, we use continuous relaxation for architecture parameters α\alpha controlling the connection in inner cell and differential convolution layers for architecture parameters γ\gamma controlling the connection in outer layer, which makes gradient descent possible. In the training, we split the labeled and unlabeled training dataset into two disjoint sets respectively: 𝒟l,A\mathcal{D}_{l,A} and 𝒟l,B\mathcal{D}_{l,B}, 𝒟u,A\mathcal{D}_{u,A} and 𝒟u,B\mathcal{D}_{u,B}. The optimization in each epoch can be summarized as:

  1. 1.

    Update weight parameters ww of f(θs)f(\theta_{s}) by
    w𝒟l,A,𝒟u,A(w,α,γ)\nabla_{w}\mathcal{L}_{\mathcal{D}_{l,A},\mathcal{D}_{u,A}}(w,\alpha,\gamma) using 𝒟l,A\mathcal{D}_{l,A} and 𝒟u,A\mathcal{D}_{u,A}

  2. 2.

    Update weight parameters ww of f(θt)f({\theta_{t}}) by EMA

  3. 3.

    Update architecture parameters α\alpha and γ\gamma of f(θs)f(\theta_{s}) by α,γ𝒟l,B,𝒟u,B(w,α,γ)\nabla_{\alpha,\gamma}\mathcal{L}_{\mathcal{D}_{l,B},\mathcal{D}_{u,B}}(w,\alpha,\gamma) using 𝒟l,B\mathcal{D}_{l,B} and 𝒟u,B\mathcal{D}_{u,B}

  4. 4.

    Update architecture parameters α\alpha and γ\gamma of f(θt)f(\theta_{t}) by EMA

where \mathcal{L} is the total loss on the segmentation mini-batch.

Refer to caption
Fig. 2: The visual comparison of the proposed method with partial SOTA approaches on the CAMUS dataset.
Table 1: Quantitative comparison with SOTA methods based on manual network, pure NAS and hybrid NAS on CAMUS dataset. The best results are highlighted in bold.

Type Model Params Dice\uparrow IoU\uparrow Manual UNet [6] 7.8M 0.921±\pm0.002 0.859±\pm0.003 UNet++ [7] 9.1M 0.919±\pm0.006 0.855±\pm0.009 Deeplabv3+ [25] 43.5M 0.909±\pm0.006 0.839±\pm0.009 Transfuse [26] 26.3M 0.923±\pm0.004 0.861±\pm0.006 PSPNet [5] 49.0M 0.916±\pm0.005 0.850±\pm0.008 Pure NAS NasUNet [14] 0.8M 0.914±\pm0.004 0.845±\pm0.005 Auto-DeepLab [15] 44.4M 0.918±\pm0.003 0.851±\pm0.006 MixSearch [20] 10.4M 0.917±\pm0.004 0.852±\pm0.004 Hybrid NAS HCT-Net [17] 31.1M 0.919±\pm0.005 0.856±\pm0.005 SSHNN-S 13.6M 0.916±\pm0.004 0.850±\pm0.006 SSHNN-M 23.5M 0.925±\pm0.002 0.865±\pm0.003 SSHNN-L 38.8M 0.932±\pm0.002 0.873±\pm0.002 1 In SSHNN-S, -M, and -L, F=8F=8, and L=4L=4, 66, and 88, respectively. 2 For a fair comparison, all tests are conducted under semi-supervised learning.

3 Experimental Results

Table 2: Ablation studies on network design and semi-supervised learning. The evaluation metric is Dice and the best results under different fractions are highlighted in bold. Filter multiplier F=8F=8.

Method Convolution fusion U-shaped decoder Unlabeled image count NuN_{u}, labeled image count NlN_{l} = 1800 & Transformer 18000 5400 3600 1800 0 SSHNN-L 0.901±\pm0.001 0.907±\pm0.001 0.910±\pm0.003 0.904±\pm0.003 0.898±\pm0.005 0.910±\pm0.001 0.915±\pm0.001 0.917±\pm0.002 0.914±\pm0.003 0.910±\pm0.004 0.913±\pm0.001 0.917±\pm0.002 0.921±\pm0.002 0.917±\pm0.002 0.913±\pm0.004 0.921±\pm0.000 0.927±\pm0.002 0.932±\pm0.002 0.929±\pm0.003 0.925±\pm0.003 SSHNN-S 0.906±\pm0.002 0.911±\pm0.003 0.916±\pm0.004 0.912±\pm0.004 0.908±\pm0.005 SSHNN-M 0.917±\pm0.001 0.922±\pm0.001 0.925±\pm0.002 0.923±\pm0.003 0.920±\pm0.004

3.1 Dataset and Evaluation Metrics

CAMUS [1] dataset is used to evaluate the performance of SSHNN, which is an open large-scale dataset in 2D echocardiography and were collected from 500 patients. Noticed that one four-chamber and one two-chamber view sequences were collected from each patient, lasting around 20 pictures long without labels, except for the moment of end diastole (ED) and end systole (ES). Therefore, accessible labelled dataset has 2000 echocardiographic images used for supervised loss calculation, and unlabeled dataset has around 19000 echocardiographic images. Segmentation labels have four types by manual annotations are the left ventricle endocardium, the myocardium, the left atrium and background. In this paper, we adopt the Intersection over Union (IoU), Dice Coefficient (Dice) and Parameters (Params) as the evaluation metrics.

3.2 Implementation Details

We use images of size 256×256256\times 256 as the network input. Each cell has B=5B=5 blocks. For the partial channel connections, n=4n=4. Use filter multiplier FF and layers LL to control the complexity of network and FF is initialized as 8. Thus, in s=4s=4, there are B×F×s4=40B\times F\times\frac{s}{4}=40 filters. When reducing spatial size from s2ss\rightarrow 2s, the number of filters doubles. We use 4 transformers to extract global context. Experiments were carried out on a Nvidia RTX3090Ti GPU. For weights ww and architecture γ\gamma, applied optimizer is SGD with momentum 0.9, weight decay 0.0003 and initial value 0.01. For architecture α\alpha, Adam [27] is applied with learning rate 0.003 and weight decay 0.001. The total number of epochs is 40 and architecture is optimized after 10 epochs since unstable update of ww makes local optima. TeacherTeacher is updated synchronously with StudentStudent for stable training. Finally, use 10% labeled images in validation and the rest in training with unlabeled images.

3.3 Experimental results

To demonstrate the effectiveness of SSHNN, we compare it with SOTA approaches. Table 1 and Fig .2 respectively elaborate the numerical and visual results for each network on the CAMUS dataset. Mean and standard deviation values for each metric are obtained from cross-validating on the 10 folds of the dataset. From SSHNN series, obviously larger LL brings higher model capacity, leading to better segmentation performance at the cost of more parameters (lower speed). Compared with Transfuse, which has suboptimal Dice in Table 1, SSHNN-L gets 0.98% Dice improvement and 1.05% IoU enhancement. Similarly compared, SSHNN-M also has the better Dice with fewer parameters, verifying the superiority.

3.4 Ablation Study

Shown in Table 2, we conducted multiple sets of tests to evaluate the impact of convolution fusion, U-shape decoder & Transformer and the fraction of labeled images. First, SSHNN-L with convolution fusion has a 1.00% higher Dice than without convolution fusion when unlabeled image versus labeled image NuNl=10\frac{N_{u}}{N_{l}}=10. Moreover, use U-shaped decoder with transformer also increases the Dice by 1.33% compared with no changes at this fraction, verifying the effectiveness of them. Obviously, combine convolution and U-shaped decoder with Transformer works better as SSHNN-L with them remains 0.921 Dice when NuNl=10\frac{N_{u}}{N_{l}}=10, aligning with the objective of semi-supervision of learning with minimal supervision.

Table 3: Ablation experiments on filter multiplier FF.

Model FF Params Dice\uparrow IoU\uparrow SSHNN-L 6 23.6M 0.928±\pm0.002 0.868±\pm0.004 8 38.8M 0.932±\pm0.002 0.873±\pm0.002 10 58.2M 0.933±\pm0.001 0.876±\pm0.002

Note that when NuNl=2\frac{N_{u}}{N_{l}}=2, the model outperforms the others as unlabeled data provides extra information and does not mask the characteristics of labeled data. The results of SSHNN-S and SSHNN-M prove the robustness of our method again.

Last but not least, Table 3 discusses the effect of FF, showing that higher model capacity promotes segmentation ability with expensive memory cost, which is a trade-off problem.

4 Conclusion

In this paper, we propose a semi-supervised hybrid NAS network termed SSHNN for medical image segmentation. We adopt convolution fusion to fully utilize multi-scale features extracted by layer-wise searching of HNAS. Furthermore, U-shaped decoder and Transformers are introduced for decode part. Semi-supervised learning Mean Teacher is implemented to overcome limited labeled dataset volumn. Finally, experiments on the CAMUS dataset demonstrate SSHNN realizes a more superior performance than existing methods.

References

  • [1] Sarah Leclerc, Erik Smistad, Joao Pedrosa, Andreas Østvik, Frederic Cervenansky, Florian Espinosa, Torvald Espeland, Erik Andreas Rye Berg, Pierre-Marc Jodoin, Thomas Grenier, et al., “Deep learning for segmentation using an open large-scale dataset in 2d echocardiography,” IEEE transactions on medical imaging, vol. 38, no. 9, pp. 2198–2210, 2019.
  • [2] Honghe Li, Yonghuai Wang, Mingjun Qu, Peng Cao, Chaolu Feng, and Jinzhu Yang, “Echoefnet: Multi-task deep learning network for automatic calculation of left ventricular ejection fraction in 2d echocardiography,” Computers in Biology and Medicine, vol. 156, pp. 106705, 2023.
  • [3] Ross Girshick, “Fast r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1440–1448.
  • [4] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille, “Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 4, pp. 834–848, 2017.
  • [5] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia, “Pyramid scene parsing network,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 2881–2890.
  • [6] Olaf Ronneberger, Philipp Fischer, and Thomas Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2015, pp. 234–241.
  • [7] Zongwei Zhou, Md Mahfuzur Rahman Siddiquee, Nima Tajbakhsh, and Jianming Liang, “Unet++: Redesigning skip connections to exploit multiscale features in image segmentation,” IEEE transactions on medical imaging, vol. 39, no. 6, pp. 1856–1867, 2019.
  • [8] Xiao Xiao, Shen Lian, Zhiming Luo, and Shaozi Li, “Weighted res-unet for high-quality retina vessel segmentation,” in International Conference on Information Technology in Medicine and Education. IEEE, 2018, pp. 327–331.
  • [9] Steven Guan, Amir A Khan, Siddhartha Sikdar, and Parag V Chitnis, “Fully dense unet for 2-d sparse photoacoustic tomography artifact removal,” IEEE journal of biomedical and health informatics, vol. 24, no. 2, pp. 568–576, 2019.
  • [10] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  • [11] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  • [12] Jieneng Chen, Yongyi Lu, Qihang Yu, Xiangde Luo, Ehsan Adeli, Yan Wang, Le Lu, Alan L Yuille, and Yuyin Zhou, “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021.
  • [13] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10012–10022.
  • [14] Yu Weng, Tianbao Zhou, Yujie Li, and Xiaoyu Qiu, “Nas-unet: Neural architecture search for medical image segmentation,” IEEE access, vol. 7, pp. 44247–44257, 2019.
  • [15] Chenxi Liu, Liang-Chieh Chen, Florian Schroff, Hartwig Adam, Wei Hua, Alan L Yuille, and Li Fei-Fei, “Auto-deeplab: Hierarchical neural architecture search for semantic image segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 82–92.
  • [16] Yanwei Li, Lin Song, Yukang Chen, Zeming Li, Xiangyu Zhang, Xingang Wang, and Jian Sun, “Learning dynamic routing for semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 8553–8562.
  • [17] Zhihong Yu, Feifei Lee, and Qiu Chen, “Hct-net: hybrid cnn-transformer model based on a neural architecture search network for medical image segmentation,” Applied Intelligence, pp. 1–17, 2023.
  • [18] Yuhui Xu, Lingxi Xie, Xiaopeng Zhang, Xin Chen, Guo-Jun Qi, Qi Tian, and Hongkai Xiong, “Pc-darts: Partial channel connections for memory-efficient architecture search,” arXiv preprint arXiv:1907.05737, 2019.
  • [19] Hanxiao Liu, Karen Simonyan, and Yiming Yang, “Darts: Differentiable architecture search,” arXiv preprint arXiv:1806.09055, 2018.
  • [20] Luyan Liu, Zhiwei Wen, Songwei Liu, Hong-Yu Zhou, Hongwei Zhu, Weicheng Xie, Linlin Shen, Kai Ma, and Yefeng Zheng, “Mixsearch: Searching for domain generalized medical image segmentation architectures,” arXiv preprint arXiv:2102.13280, 2021.
  • [21] Xiong Zhang, Hongmin Xu, Hong Mo, Jianchao Tan, Cheng Yang, Lei Wang, and Wenqi Ren, “Dcnas: Densely connected neural architecture search for semantic image segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 13956–13967.
  • [22] Xingang Yan, Weiwen Jiang, Yiyu Shi, and Cheng Zhuo, “Ms-nas: Multi-scale neural architecture search for medical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2020, pp. 388–397.
  • [23] Antti Tarvainen and Harri Valpola, “Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results,” Advances in neural information processing systems, vol. 30, 2017.
  • [24] Samuli Laine and Timo Aila, “Temporal ensembling for semi-supervised learning,” arXiv preprint arXiv:1610.02242, 2016.
  • [25] Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 801–818.
  • [26] Yundong Zhang, Huiye Liu, and Qiang Hu, “Transfuse: Fusing transformers and cnns for medical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2021, pp. 14–24.
  • [27] Diederik P Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.