This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: 1 Shanghai AI Lab, Shanghai, China
2 Shanghai Jiao Tong University, Shanghai, China
3 University of Surrey, Guildford GU2 7XH, United Kingdom
4 University of Electronic Science and Technology of China
5 College of Electronic and Information Engineering, Tongji University

Exploring Vanilla U-Net for Lesion Segmentation from Whole-body FDG-PET/CT Scans

Jin Ye 11    Haoyu Wang 1122    Ziyan Huang 1122    Zhongying Deng 1133    Yanzhou Su 44    Can Tu 1122    Qian Wu 1155    Yuncheng Yang 1122    Meng Wei 11    Jingqi Niu 1122    Junjun He Corresponding Author1122
Abstract

Tumor lesion segmentation is one of the most important tasks in medical image analysis. In clinical practice, Fluorodeoxyglucose Pos-itron-Emission Tomography (FDG-PET) is a widely used technique to identify and quantify metabolically active tumors. However, since FDG-PET scans only provide metabolic information, healthy tissue or benign disease with irregular glucose consumption may be mistaken for cancer. To handle this challenge, PET is commonly combined with Computed Tomography (CT), with the CT used to obtain the anatomic structure of the patient. The combination of PET-based metabolic and CT-based anatomic information can contribute to better tumor segmentation results. In this paper, we explore the potential of U-Net for lesion segmentation in whole-body FDG-PET/CT scans from three aspects, including network architecture, data preprocessing, and data augmentation.The experimental results demonstrate that the vanilla U-Net with proper input shape can achieve satisfactory performance. Specifically, our method achieves first place in both preliminary and final leaderboards of the autoPET 2022 challenge. Our code is available at https://github.com/Yejin0111/autoPET2022_Blackbean.

Keywords:
autoPET FGD-PET CT Segmentation U-Net.

1 Introduction

Fluorodeoxyglucose Positron-Emission Tomography (FDG-PET) and Computed Tomography (CT) are two vital modalities in the diagnostic workup for various malignant solid tumor entities. FDG-PET is widely used in oncology to reflect the glucose consumption of tissues for therapy control because tumor lesions typically have increased glucose consumption [4]. FDG-PET is a sensitive but non-tumor-specific method for identifying malignancy. It can also be seen in healthy tissue or benign disease as inflammation or posttraumatic repair could be mistaken for cancer [6]. To overcome this problem, PET is commonly combined with CT to attach anatomic information to the metabolic data and improve diagnostic accuracy in the real clinical scenario. As a result, efficiently utilizing the paired PET and CT images can help accurately diagnose malignant tumors and avoid needless surgery.

In this context, the Automatic Lesion Segmentation in Whole-Body FDG-PET/CT Challenge (autoPET) is organized to offer the opportunity to participants to develop an automatic bi-modal approach for the three-dimensional segmentation of tumors lesions in FDG-PET and CT scans. The purpose of this challenge is to provide a large-scale training data set to promote the deep-learning-based automated tumor lesion segmentation on whole-body FDG-PET /CT data111https://autoPET.grand-challenge.org/. The challenge provides more than 1000 cases of paired PET/CT scans, while the previously released PET/CT datasets usually contain less than 400 scans.

Deep learning methods with Convolutional Neural Networks (CNNs) have become de facto standard in automated PET/CT lesion segmentation. However, the performance of tumor lesion segmentation in whole-body PET/CT remains far from satisfactory due to limited training data in previous works [4, 3]. Based on the large-scale training data provided in the autoPET Challenge, we provide some empirical analysis of the U-Net [5] for the lesion segmentation task, including model architecture, data pre-processing, and data augmentation.The experimental results demonstrate that the vanilla U-Net with proper input shape can achieve satisfactory performance. Specifically, a Dice score of 92.81%, on the preliminary test set.

2 Material

The challenge involves only one task, i.e., accurate and fast lesion segmentation.

2.1 Dataset

The autoPET dataset consists of patients with histologically proven malignant melanoma, lymphoma, or lung cancer as well as negative control patients who were examined by FDG-PET/CT in two large medical centers (University Hospital Tübingen, Germany & University Hospital of the LMU in Munich, Germany). All PET/CT data within this challenge have been acquired on advanced PET/CT scanners (Siemens Biograph mCT, mCT Flow and Biograph 64, GE Discovery 690) using standardized protocols following international guidelines. In detail, the dataset provides 1,014 paired FDG-PET and CT studies from 900 patients as the training data, and 200 studies as the testing data. Furthermore, there are 5 cases from the testing data that are provided for participants to evaluate their algorithms.

2.2 Metrics

A combination of three metrics selected by the organizer to evaluate the methods: (1) Foreground Dice score of segmented lesions (DSC); (2) Volume of false positive connected components that do not overlap with positives (False Positive Volume, FPV); (3) Volume of positive connected components in the ground truth that do not overlap with the estimated segmentation mask (False Negative Volume, FNV). The final ranking of submitted algorithms is computed by the numerical mean of the single rankings,

score=0.5×RankDSC+0.25×RankFPV+0.25×RankFNV.\mathrm{score}=0.5\times\mathrm{Rank_{DSC}}+0.25\times\mathrm{Rank_{FPV}}+0.25\times\mathrm{Rank_{FNV}}. (1)

3 Method

Refer to caption
Figure 1: Illustration of the vanilla U-Net we used. The number of feature channels is denoted on the top of the box. The arrows denote different operations.

3.1 Network Architecture

The architecture is illustrated in Figure 1. We adopt the vanilla U-Net, which has a common encoder-decoder architecture. We simply concatenate the PET/CT scan pairs as the input. For feature extraction, the vanilla U-Net transfer the input into feature maps with 32 channels at the first layer and doubles the number of feature channels at each down-sampling step. At each stage of the decoder, two feature maps are concatenated as the input, one from the skip connections, and the other from the upsampled feature map of the previous stage.

Table 1: Some detailed settings of the data augmentations
Data augmentation Parameters
Random Brightness Transform range=(0.75,1.25),σ=0.10range=(0.75,1.25),\sigma=0.10
Random Gamma Transform range=(0.70,1.50),p=0.30range=(0.70,1.50),p=0.30
Random Flip p=0.5p=0.5
Random Rotation range=(15,+15)range=(-15,+15)

3.2 Data Preprocessing

According to the official pre-processing pipeline of autoPET222https://autoPET.grand-challenge.org/Dataset/, the provided CT and PET scans are resampled to 1.500 ×\times 1.0182 ×\times 1.0182 mm. In addition, the PET scans are normalized and converted to Standardized Update Values (SUV). We follow this pipeline and develop our method based on the provided pre-processed NifTI files.

3.3 Data Augmentation

3.3.1 Random Brightness Transform

Brightness transform contains multiplicative and additive operations on the intensity, formulated as

Xout=m×(Xin)+𝒩(0,σ),X_{out}=m\times(X_{in})+\mathcal{N}(0,\sigma), (2)

where mm is a scale factor randomly sampled from a uniform distribution of a fixed range, with the range shown in Table 1.

3.3.2 Random Gamma Transform

Gamma transform is a technique used to compensate for the non-linear display characteristics of a device, formulated as

Xout=(Xin)1/γ.X_{out}=(X_{in})^{1/\gamma}. (3)

The original image is corrected by sending the signal through the inverse function f(x)=x1/γf(x)=x^{1/\gamma} based on a fixed gamma value. The gamma value is randomly sampled from a uniform distribution of a fixed range.

Since the CT and PET scans share similar anatomical structures, we use the same random flip and rotation as the spatial data augmentations. Referring to nnU-Net [2], the parameters of data augmentations are listed in Table 1.

Table 2: Implementation details in the training stage of our final submission.
Training setup Details
Spacing [1.5, 1.01821005, 1.01821005]
Crop size 192×192×192192\times 192\times 192
Batch size 2
Data augmentation “default_3D_augmentation_params” in nnUNet
Learning rate 0.0001
Optimizer SGD + Poly decay policy
Number of iterations 250,000
Table 3: Implementation details in the testing stage of our final submission.
Testing setup Details
Spacing [1.5, 1.01821005, 1.01821005]
Crop size 192×192×192192\times 192\times 192
Inference strategy Sliding window strategy with step size 0.5
Postprocessing None
Ensemble strategy Ensemble the results of 5 folds

4 Experiments

In the training stage, we keep the original spacing, and the crop size is 192×192×192192\times 192\times 192. The batch size is fixed to 2 and each model is trained for 250,000 iterations. We use SGD optimizer with momentum of 0.99, weight decay of 0.001. The learning rate is set as 0.0001 at the beginning and decayed based on the poly decay policy (1iter/250000)0.9(1-iter/250000)^{0.9} for every 250 iterations. The final setups of our training and testing stage are listed in Table 2.

4.1 Implementation Details

Our development environment is CUDA 11.2, Pytorch 1.11.0, and Tesla A100 80G. We implement our method based on the nnU-Net codebase [2].

In the predicting stage, we just follow the nnU-Net with no post-processing. Specifically, nnU-Net adopts the sliding-window strategy for inference due to the high resolution of volumetric medical images. To find the trade-off between accuracy and speed, we ablate the step size of the sliding-window strategy in the ablation section. The final setups of our testing stage are listed in Table 3.

Table 4: Results of different network architectures on the preliminary test set. * indicates UNETR is implemented by ourselves.
Network architecture stage feature channels params DSC FPV FNV
UNETR* - - 133.0M 0.51 6.28 1.58
Shallow U-Net 4 (16,32,64,128)(16,32,64,128) 1.4M 0.71 1.22 3.39
Vanilla U-Net 5 (32,64,128,256,512)(32,64,128,256,512) 22.6M 0.92 1.99 1.68
Deeper U-Net 6 (32,64,128,256,512,1024)(32,64,128,256,512,1024) 58.6M 0.52 1.57 1.95

4.2 Ablation

Different architectures. As shown in Table 4, we first evaluate the performance of different number of stages and feature channels in U-Net. The vanilla U-Net outperforms other architectures for DSC and FPV on the preliminary test set. Moreover, the CNN-based vanilla U-Net improves the transformer-based UNETR [1] by 42% in DSC. In addition, the params of the vanilla U-Net is 22.6M, which is much less than Deeper U-Net (58.6M) and UNETR (133.0M). Note that the parameters are concluded by torchinfo333https://github.com/TylerYep/torchinfo.

Table 5: DSC of different input shapes in the training stage on our validation set. “Ori spacing” indicates that we keep original spacing for all data in training. “1/2 spacing” indicates that we decrease the spacing to 1/2 for training cases.
      Input shape       DSC
      Ori spacing       1/2 spacing
      128×128×128128\times 128\times 128       0.62       0.59
      160×160×160160\times 160\times 160       0.69       0.65
      192×192×192192\times 192\times 192       0.75       0.68
      224×224×224224\times 224\times 224       0.74       0.72

Input shape and spacing in the training stage. We further ablate different crop sizes in the training stage, and show the DSC results in Table 5. For the original spacing, the cropped input size of 192×192×192192\times 192\times 192 achieves the best DSC 75%, while the crop size of 224×224×224224\times 224\times 224 performs better with 1/2 spacing. Table 5 also shows that keeping original spacing is the best way to train a better model.

Table 6: Results of different input shapes in the inference stage on our validation set. The inference time shows the total time of 103 cases. Input shape DSC FPV FNV Inference time (s) 128×128×128128\times 128\times 128 0.73 2.43 7.88 5528 160×160×160160\times 160\times 160 0.75 3.11 6.83 4248 192×192×192192\times 192\times 192 0.75 4.24 6.10 4672 224×224×224224\times 224\times 224 0.75 3.73 6.07 3878 256×256×256256\times 256\times 256 0.75 4.65 5.73 4757 288×288×288288\times 288\times 288 0.74 6.73 5.44 3288 320×320×320320\times 320\times 320 0.72 10.87 5.12 4169 Table 7: Results of different step sizes of the sliding window strategy in the inference stage on our validation set. The inference time shows the total time of 103 cases. Step size DSC FPV FNV Inference time (s) 0.5×input_shape0.5\times input\_shape 0.75 4.24 6.10 4672 0.6×input_shape0.6\times input\_shape 0.75 4.03 6.43 3166 0.7×input_shape0.7\times input\_shape 0.75 3.67 6.43 2961 0.8×input_shape0.8\times input\_shape 0.75 3.71 6.51 2933 0.9×input_shape0.9\times input\_shape 0.75 3.70 6.51 2937 1.0×input_shape1.0\times input\_shape 0.75 3.66 6.51 2915

Input shape and step size in the testing stage. To find a faster and more accurate way of inference, we ablate different input sizes and the step sizes of the sliding window strategy. Table 7 shows the results of different input shapes in the testing stage. The DSC changes little with different input sizes, while FNV and FPV are very sensitive. The smaller (larger) input size we use, the lower FPV (FNV) we get. For step sizes, the results are shown in Table 7. The DSC is stable at 75% in all step sizes. However, FNV and FPV are also sensitive to different step size.

Table 8: Final submissions. “Status” indicates whether our submission succeeded in the submission system.
      Step size       Crop size       Status
      0.5       192×192×192192\times 192\times 192       Failed
      0.5       224×224×224224\times 224\times 224       Failed
      0.6       192×192×192192\times 192\times 192       Success
      0.7       192×192×192192\times 192\times 192       Success

Table 9: Comparison to other teams on the preliminary test set. Teamname DSC FNV FPV Blackbean 0.93 0.79 1.07 SM 0.93 1.42 0.90 BDAV 0.93 1.57 0.78 PET_TEAM 0.92 0.53 0.89 UIH-FL 0.93 1.57 0.81 Table 10: Comparison to other teams on the final test set. Teamname DSC FNV FPV Blackbean 0.62 0.54 2.84 BDAV 0.62 0.75 3.61 FightTumor 0.60 0.47 5.10 UIH-FL 0.61 0.83 4.85 heiligerl 0.61 0.63 5.87

4.3 Final submissions

According to the ablation studies, we adopt the original spacing and set the input shape to 192×192×192192\times 192\times 192 in the training stage. We simply train 5 folds models using the same splits with nnUNet, and combine the results of all models in the inference stage (Also the same as nnUNet). We first evaluate our model with step size 0.5 and crop size 192×192×192192\times 192\times 192 on the preliminary test set, the results are shown in Table 10, which achieves the Mean Position 17.3 (Can be found in the autoPET homepage) and ranks top-1 on the leaderboard (Team name: “Blackbean”). In the final submissions, to take both accuracy and inference time into consideration, we submitted different strategies, which are listed in Table 8. In the final leaderboard, our vanilla U-Net also achieves first place among all competitors, and the results are shown in Table 10.

References

  • [1] Hatamizadeh, A., Tang, Y., Nath, V., Yang, D., Myronenko, A., Landman, B., Roth, H.R., Xu, D.: Unetr: Transformers for 3d medical image segmentation. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. pp. 574–584 (2022)
  • [2] Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnu-net: a self-configuring method for deep learning-based biomedical image segmentation. Nature Methods 18(2), 203–211 (2021)
  • [3] Li, H., Jiang, H., Li, S., Wang, M., Wang, Y.: Densex-net: An end-to-end model for lymphoma segmentation in whole-body pet/ct images. IEEE Access PP(99),  1–1 (2019)
  • [4] Nakajo, M., Jinnouchi, S., Fukukura, Y., Tanabe, H., Tateno, R., Nakajo, M.: The efficacy of whole-body fdg-pet or pet/ct for autoimmune pancreatitis and associated extrapancreatic autoimmune lesions. European Journal of Nuclear Medicine & Molecular Imaging 34(12), 2088–2095 (2007)
  • [5] Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. pp. 234–241. Springer International Publishing, Cham (2015)
  • [6] Rosenbaum, S.J., Lind, T., Antoch, G., Bockisch, A.: False-positive fdg pet uptakethe role of pet/ct. European Radiology 16(5), 1054–1065 (2005)