AATCT-IDS: A Benchmark Abdominal Adipose Tissue CT Image Dataset for Image Denoising, Semantic Segmentation, and Radiomics Evaluation
Abstract
Background and Objective: The metabolic syndrome induced by obesity is closely associated with cardiovascular disease, and the prevalence is increasing globally, year by year. Obesity is a risk marker for detecting this disease. However, current research on computer-aided detection of adipose distribution is hampered by the lack of open-source large abdominal adipose datasets.
Methods: In this study, a benchmark Abdominal Adipose Tissue CT Image Dataset (AATTCT-IDS) containing 300 subjects is prepared and published. AATTCT-IDS publics 13,732 raw CT slices, and the researchers individually annotate the subcutaneous and visceral adipose tissue regions of 3,213 of those slices that have the same slice distance to validate denoising methods, train semantic segmentation models, and study radiomics. For different tasks, this paper compares and analyzes the performance of various methods on AATTCT-IDS by combining the visualization results and evaluation data. Thus, verify the research potential of this data set in the above three types of tasks.
Results: In the comparative study of image denoising, algorithms using a smoothing strategy suppress mixed noise at the expense of image details and obtain better evaluation data. Methods such as BM3D preserve the original image structure better, although the evaluation data are slightly lower. The results show significant differences among them. In the comparative study of semantic segmentation of abdominal adipose tissue, the segmentation results of adipose tissue by each model show different structural characteristics. Among them, BiSeNet obtains segmentation results only slightly inferior to U-Net with the shortest training time and effectively separates small and isolated adipose tissue. In addition, the radiomics study based on AATTCT-IDS reveals three adipose distributions in the subject population.
Conclusion: AATTCT-IDS contains the ground truth of adipose tissue regions in abdominal CT slices. This open-source dataset can attract researchers to explore the multi-dimensional characteristics of abdominal adipose tissue and thus help physicians and patients in clinical practice. AATCT-IDS is freely published for non-commercial purpose at: https://figshare.com/articles/dataset/AATTCT-IDS/23807256.
keywords:
CT Image Dataset , Abdominal Adipose Tissue , Image Denoising , Semantic Segmentation , Radiomics1 Introduction
1.1 Research Background and Motivation
Metabolic Syndrome (MetS) is a cluster of metabolic disorders that are becoming increasingly prevalent due to the rising rates of obesity. Each individual component of MetS is a well-established risk factor that is strongly associated with cardiovascular disease (CVD). The obese population tends to exhibit multiple risk factors that increase the incidence and severity of CVD, including heart failure, coronary artery calcification, myocardial infarction, microvascular dysfunction, and cardiac dysfunction [1]. The global prevalence of MetS is estimated to be around 25%, and the number of deaths due to obesity has increased by 28.3% between 1990 and 2015 [2]. Abdominal obesity, which is linked with all components of MetS, has been shown to possess a higher predictive value in determining the susceptibility to CVD than other indicators [3, 4]. The Japanese Metabolic Syndrome Criteria Examination Committee and the International Diabetes Federation have proposed two sets of MetS criteria, both of which consider abdominal obesity as a mandatory risk marker [5].
In clinical practice, the assessment of obesity involves various methods, such as waist circumference (WC) and body mass index (BMI), in addition to basic weight measurement [6]. BMI is determined by dividing weight (kg) by height (m) squared, with specific thresholds set by the World Health Organization for adults and children [7, 8, 4]. Conversely, WC thresholds differ based on gender or ethnic group, necessitating the use of a logistic regression model to determine this value. This model connects the WC threshold to cardiovascular disease (CVD) risk, allowing simultaneous acquisition of WC thresholds for MetS risk factors [9]. However, these processes often rely on manual manipulation and are prone to significant errors. Additionally, the application of different thresholds to diverse populations hinders efficient detection. Moreover, current non-imaging biomarkers, including BMI and WC, do not have the ability to characterize or quantify adipose distribution accurately, especially visceral adipose, nor can any of the measured values accurately reflect abdominal obesity alone. Therefore, there is an urgent need to incorporate computerized mechanical aids to achieve a non-invasive diagnosis, optimize the diagnostic process, and improve diagnostic results.
Medical imaging techniques, such as computed tomography (CT), provide a visual representation of the interior of the human body in a non-invasive manner, thus avoiding the controversy caused by metabolic differences across races when WC reflects subcutaneous and visceral fat [10]. Recent studies have shown that combining computer vision techniques and medical imaging can help improve the efficiency and accuracy of clinical diagnosis by radiologists and enable value-added tasks [11, 12]. Segmentation of medical images is a common and vital task in computer-aided diagnostic processes. It involves separating homogeneous regions within the images. One of the most effective approaches for this task is semantic segmentation using deep learning algorithms. These algorithms excel at rapidly and accurately extracting regions of interest (ROI), such as tissues, organs, and lesions, by classifying images at the pixel level [13, 14]. Notably, training with two-dimensional slices of abdominal CT images allows the creation of models capable of segmenting three-dimensional body components in abdominal CT scans. These models achieve accuracy levels that either match or exceed those achieved through manual segmentation by experts [15, 16]. Clinically, it is widely accepted that the visceral adipose tissue area in image slices of specific anatomical locations in the abdomen is highly correlated with overall visceral adipose tissue volume and whole-body adipose tissue volume [17, 18]. However, in the study of radiomics, mining higher-order data from digital images is crucial, and the conventional approach is to extract high-level features containing underlying physiological information from ROI [19]. These high-level features, which are imperceptible to the naked eye, can be employed to develop models for quantifying disease risk and characterizing pathology when integrated with features outside of radiomics [11, 12]. These models can assist radiologists in image analysis and serve as a foundation for supporting diagnostic decisions. The major obstacle for these applications is the lack of adequate large-scale datasets with usable labels [16].
This paper presents a benchmark Abdominal Adipose Tissue CT Image Dataset (AATCT-IDS) for image denoising, semantic segmentation, and radiomics evaluation. AATTCT-IDS is a publicly available dataset containing 13,777 images derived from 300 CT scans. According to a certain slice distance, the researchers extract 3,213 slices from the raw CT slices of AATTCT-IDS for manual annotation to identify the subcutaneous and visceral adipose tissue regions and extract a set of multidimensional features. In this study, the images in the dataset were segmented and clustered based on deep learning and machine learning algorithms, and the performance of the different algorithms was evaluated using the corresponding metrics. Furthermore, clinically common mixed noise species are introduced in the images, and the images are restored using various denoising techniques and evaluated. AATCT-IDS is available at the URL: https://figshare.com/articles/dataset/AATTCT-IDS/23807256.
The main contributions of this paper are as follows:
-
•
Multiple experts involved in the development of the abdominal adipose tissue CT image dataset.
-
•
Demonstrate that AATCT-IDS can be used to distinguish the performance of different algorithms in processing CT image denoising, clustering, and semantic segmentation tasks.
-
•
This abdominal adipose tissue CT image dataset is published as open source for non-commercial purposes.
1.2 Related Work
In the field of digital image processing for body composition segmentation, a study [20] employs a dataset consisting of abdominal MRI scans of 31 diabetic patients. This dataset contains 16 consecutive axial slices from the 4th to 5th lumbar vertebrae (L4-L5) for each patient. Nonetheless, the study is limited by utilizing annotations generated semi-automatically by a single radiologist as the reference standard. Another study [21] constructs a novel dataset to evaluate digital image processing and machine learning algorithms. This dataset comprises 168 single-slice abdominal CT scan images obtained from participants in the Baltimore Longitudinal Study of Aging, focusing specifically on the L4-L5 regions of the body. The images have been semi-manually annotated by the investigators. However, due to the limited sample size used for training, the model’s ability to generalize in identifying and removing food residues experiences a loss in specificity.
Some studies combining classical machine learning with deep learning employ proprietary datasets. For instance, the Second XiangYa Hospital of Central South University provides abdominal CT scan images of 20 patients for the segmentation and classification of abdominal fat [22]. These images are manually annotated by radiologists and uniformly resized to pixels. Another study [23] presented a dataset from abdominal MRI images of 65 patients with diabetes and hypercholesterolemia. The researchers apply Gaussian normalization to the initial data and augment it by performing rotations in four directions. Additionally, a study aiming to quantify adipose tissue employs abdominal MRI images from 75 subjects from the First Hospital of Jilin University [24]. To mitigate the risk of overfitting, the researchers augmented it with polar and affine transformations.
Adipose tissue segmentation based on deep learning has a broader demand for data. A cross-modal segmentation study [25] uses abdominal CT and abdominal MRI images of 34 National Institutes of Health patients. The study employs patient-level normalization to preprocess the gray values of CT and MRI images so that the histograms have a specific distribution. In addition, the research team individually rectified the MRI images to remove bias fields. To expand the amount of data, methods such as image translation, reflection, and scaling are applied for data augmentation. In another study [26], researchers manually annotate CT images of 40 ovarian cancer patients at the Health Science Center of University of Oklahoma for training a model. Similarly, a different study [27] describes a dataset consisting of enhanced CT images from 43 lung cancer patients at the Chinese Academy of Medical Sciences Cancer Hospital, Shenzhen Hospital. In order to improve the ability of the model to segment soft tissue, the CT value of the images is limited to the range of . However, the majority of images in the dataset exhibit small visceral fat regions, significantly hampering the model’s ability to accurately segment visceral adipose in tested images. Additionally, the limited number of patients in this dataset poses a challenge in demonstrating the model’s generalization ability. In another study [28], a dataset comprising abdominal MRI images of 45 diabetic patients is used to compare the performance of U-Net and V-Net in segmenting body composition. Researchers employ a lean tissue filter to assist in marking adipose tissue and remove the top 1% of the highest gray values in the histogram during data preprocessing. A study [29] uses the abdominal MRI images of 38 Rhineland subjects to build a dataset for training models. To ensure an accurate evaluation, two experts manually generate the labels without relying on semi-automatic support. Finally, in a study [30] conducted by researchers, they curated a dataset comprising 50 patients, with over half of them undergoing screening for oncological indications. The primary purpose of this dataset was to train a semantic segmentation model. However, it’s worth noting that the semantic segmentation in this study does not encompass the identification of visceral adipose tissue.
Table 1 presents a comprehensive overview of the pertinent dataset details. However, all the aforementioned datasets suffer from a common limitation - a small number of subjects. This becomes particularly challenging due to the significant interindividual variability and the inherent directional characteristics of medical images. Consequently, conventional data augmentation techniques prove inadequate in effectively tackling this concern. On the other hand, the datasets mentioned are non-public, and some are based on scans of patients with non-obesity-related diseases. Therefore, for the advancement of abdominal adipose segmentation research, there is a pressing need for an open-source dataset comprising an ample number of subjects acquired through routine obesity screenings.
Aim | Equipment | Amount | Subjects | Type | Method | Reference |
---|---|---|---|---|---|---|
Digital image processing | MRI | 31 patients | Diabetic | Private | Semiautomatic | [20] |
CT | 168 images | Healthy | Private | Semiautomatic | [21] | |
Machine learning | CT | 20 patients | Sick | Private | Manual | [22] |
MRI | 65 patients | Hypercholesterolemic | Private | Semiautomatic | [23] | |
MRI | 75 patients | Sick | Private | Semiautomatic | [24] | |
Deep learning | CT & MRI | 34 patients | Sick | Private | Semiautomatic | [25] |
CT | 40 patients | Ovarian cancer | Private | Manual | [26] | |
CT | 43 patients | Lung cancer | Private | Manual | [27] | |
MRI | 45 patients | Diabetic | Private | Semiautomatic | [28] | |
MRI | 38 patients | Healthy | Private | Manual | [29] | |
CT | 50 patients | Cancer (50%) | Private | Manual | [30] |
1.3 Structure of Paper
This section expounds on the background and motivation for preparing the dataset and the current datasets involved in related research fields. Section 2 will introduce the preparation process of AATCT-IDS in detail and evaluate the performance of the dataset from multiple perspectives. The evaluation results will be presented in Section 3. The content of Section 4 will systematically analyze the previous evaluation results. A summary and outlook will be given at the end of the article.
2 Material and Method
2.1 Dataset Preparation
AATCT-IDS contains 13,732 abdominal CT images from 300 subjects. The details of the abdominal adipose tissue CT image dataset are described below:
-
1.
(1) Data Source:
Stage 1: Shengjing Hospital of China Medical University provides 330 samples saved in DICOM format. Each sample corresponds to the CT scan sequence of a subject, and consecutive slices of L2-L3 sites are selected from the samples by experienced radiologists with a spacing of 1.25 mm between adjacent slices.
Stage 2: Under the guidance of radiologists, biomedical researchers at Northeastern University remove substandard data, such as images with visual defects like artifacts (Shown in Figure 2 - (a)), and ultimately retain 300 samples to prepare the dataset. -
2.
(2) Preparation Process:
Step 1: Only image data from DICOM files are extracted for open source to uphold subject privacy and facilitate data integration into diverse deep learning models for training purposes. Given the substantial variation in CT values across fat, muscle, skin, and bone tissues, and considering the acceptable segmentation accuracy achievable with uint8 data type, CT value within the range of is mapped to grayscale values spanning the range of . Save resultant images in JPEG format for compatibility and accessibility.
Step 2: Given that training with sparse annotations can build a model for segmenting dense volume [31] and adjacent slices of abdominal CT are highly similar, images are extracted with a slice spacing of 5 mm from the complete CT slices, and the extracted 3,213 images are combined into sub-datasets for manual labeling of adipose tissue regions.
Step 3: Resize all images to pixels by cropping.
Step 4: Seven biomedical researchers from Northeastern University and two Shengjing Hospital of China Medical University radiologists manually label subcutaneous and visceral adipose tissue without using semi-automatic support (See Section 2.2 for details).
Step 5: Three other senior radiologists from Shengjing Hospital of China Medical University further calibrate the labels after unifying the standard. Figure 1 illustrates the preparation process of AATCT-IDS. -
3.
(3) Scanning device: 256 slice spiral CT (Revolution CT, GE Healthcare).
-
4.
(4) Contrast media: Ioversol (320mgI/ml).
-
5.
(5) Annotation software: Labelme.
-
6.
(6) Data format: Both the complete sequence and extracted sub-dataset are saved in JPEG format. Annotations are stored in two forms, the annotations of subcutaneous adipose tissue and visceral adipose tissue are separated and stored as index images in PNG format, and the manually annotated adipose tissue boundary is recorded in JSON format, which is convenient for subsequent expansion or conversion to other formats (See Table 2 for details).
-
7.
(7) The Ethical Committee of Northeastern University approved the preparation and study of AATCT-IDS.

Aim | Type | Amount | Format |
---|---|---|---|
Segmentation | CT (Partial) | 3,213 | JPEG |
Label (Subcutaneous adipose) | 3,213 | PNG (Index) | |
Label (Visceral adipose) | 3,213 | PNG (Index) | |
Label (Adipose tissue) | 3,213 | JSON | |
Denoising & Radiomics | CT (Total) | 13,732 | JPEG |
2.2 Dataset Description
2.2.1 CT Scan
Compared to MRI, radiologists can more easily label adipose tissue in CT images [25]. In addition, CT scans are more readily available and have shorter acquisition times. Thus AATCT-IDS focuses on collecting CT images. This study establishes a set of criteria for volunteer recruitment to ensure a strong association between the data and obesity examination while minimizing confounding variables. The samples in this dataset are derived from a population with randomized BMI, ensuring a balanced representation of both genders. Additionally, the study participants are exclusively individuals who are either healthy or have obesity-related diseases. During the scanning procedure, subjects are instructed to maintain their arms above their heads and follow breath-holding instructions to mitigate the occurrence of artifacts.
2.2.2 Subcutaneous Adipose Tissue
CT employs Hounsfield units (HU) to quantify the extent of X-ray beam attenuation within tissues, thereby reflecting tissue density. Adipose tissue has a lower density than muscle, organs, and bone, with CT values typically in the , so adipose appears lower lightness on grayscale images. However, the presence of edema or fatty infiltration can influence radiation attenuation. Edema increases radiation attenuation within subcutaneous adipose tissue, whereas adipose infiltration reduces radiation attenuation in skeletal muscle. As a consequence, the boundaries between subcutaneous adipose and neighboring skeletal muscle become blurred, hindering the image segmentation process. In addition, a small percentage of extremely lean or obese subjects may also show indistinguishable adipose tissue boundaries in CT images. Such images are excluded during dataset preparation (Shown in Figure 2 - (b)(c)(d)).
2.2.3 Visceral Adipose Tissue
All images in this dataset originate from the L2-L3 segment of the subject, encompassing part of the small intestine’s structure (Shown in Figure 2 - (e)). Therefore, the presence of the intestinal cavity can often be observed in the images of the dataset. Notably, the intestinal cavity and visceral adipose tissue share close pixel characteristics, and both show low brightness, so the intestinal cavity can easily be misclassified as visceral adipose tissue [32]. In addition, coexistence with intestinal capillaries tissue and various viscera allows for a complex spatial distribution of visceral adipose tissue. As a result, visceral adipose tissue is isolated in numerous narrow spaces with complex and irregular boundaries. The above details are paid attention to during the dataset preparation process.

2.3 Methods of Denoising
The reconstruction of CT images relies on physical measurements and is sensitive to contrast ratio, so the introduction of noise cannot be avoided [33]. Denoising while preserving factual medical information is a prerequisite for precision medicine. In this paper, varying degrees of single or mixed types of noise are inserted into all CT images. Subsequently, multiple denoising means are used to restore the image separately, aiming to verify that AATCT-IDS can distinguish the denoising performance of different algorithms.
2.3.1 Add Noise
Noise in CT images generally follows Poisson and Gaussian distributions, and hardware or software failures can introduce small amounts of noise characterized as Speckle or Pulse types [34]. Poisson noise arises from the inherent uncertainty of photons, while Gaussian noise stems from sensor noise during the acquisition process [35]. In order to simulate the realistic scanning situation, single or mixed types of noise (Gaussian, Poisson, Speckle, Pulse) with a different mean (, ) and variance (, ) are introduced into the image.
2.3.2 Denoising
Image denoising requires restoring the original image as much as possible without destroying the actual image structure, such as boundaries, texture details, and contrast ratio. Denoising algorithms commonly operate in either the spatial or the transform domains, and in some instances, both domains are involved in the denoising process [36]. In this paper, denoising algorithms based on spatial domain include Median filter, Mean filter, Gaussian filter, and Non-local mean filter (NLM), while denoising algorithms based on transform domain include Wiener filter, Wavelet transform (WT) and Block Matching 3D (BM3D).
The classic smoothing filter is the most basic and popular image denoising method, which can remove some of the low noise with a good time complexity. The Mean filter assigns the arithmetic mean value to the central pixel within a sliding window, effectively reducing Gaussian and Pulse noise. The Median filter is a typical nonlinear filter. During the filtering process, the pixels of the picture will be replaced one by one with the median value of adjacent pixels. In addition, compared with other smoothing filters, it has less damage to edge details. The Gaussian filter is a low-pass filter that blurs the image while reducing noise. The Wiener filter reduces the mean square error of the image by analyzing the statistical characteristics of the image and noise. It exhibits sensitivity to additive noise, making it particularly effective in suppressing Gaussian white noise. The size of the filter window has a direct impact on the denoising effect. Therefore, the smoothing and Wiener filter in this paper applies different filter windows (size, ).
The advantage of the NLM algorithm is that it exploits the redundant information in the image [37]. The replaced gray value in NLM also comes from the weighted average of other pixels. However, unlike the conventional local smoothing filter, the pixels participating in the averaging process are not adjacent but exhibit structural similarity, and the similarity measurement between two pixels is determined based on the Gaussian-weighted Euclidean distance. Furthermore, the filter parameter () of this algorithm is associated with the attenuation of the exponential function. Excessive values of can lead to image blurring, while overly small values result in weakened denoising effects. In this paper, the filter parameters of NLM take the same value () as the standard deviation of the noise.
Following a WT, the image is decomposed into components with different frequencies in the wavelet domain. Based on this, the noise can be removed by employing threshold processing at the scale corresponding to each component. The wavelet coefficients associated with noise are eliminated, while those related to structural information are amplified [38]. In this paper, the decomposition and reconstruction of polluted CT images involve multiple wavelet basis (Daubechies, Dmeyer).
In the denoising process of BM3D [39], the algorithm combines the blocks with the minor differences from the reference blocks within a certain distance into a 3D matrix. Subsequently, it sets the smaller coefficients to zero in 3D space by hard thresholding and merges the inverse transformed 3D results to the initial position. It should be noted that a high time complexity accompanies the BM3D algorithm.
2.4 Methods of Segmentation
Training segmentation models is a pre-work of radiomics. Unlike traditional image segmentation algorithms that rely on unsupervised learning, semantic segmentation generates pixel-level predictions on visual inputs and is usually based on deep learning. In this paper, three annotations corresponding to the CT image sub-datasets are used for model training and validation, respectively. All models (FCN, U-Net, PSPNet, DeepLabV3, DenseASPP, BiSeNet, ICNet, DANet) are trained in the same device with the same set of parameters.
By replacing the fully connected layers of traditional CNN, FCN achieves semantic segmentation for the first time in deep learning [40]. The transposed convolution at the end of the FCN uses upsampling to maintain the spatial information of the image, so it is suitable for the input of arbitrary size. In addition, the skip architecture that combines global and local information improves the local prediction based on the global prediction. Therefore, the FCN-8s with relatively accurate segmentation effects are the training object in this paper.
U-Net [41] and PSPNet [42] optimize the structure of FCN in different strategies. U-Net pioneered a U-shaped symmetric structure (encoder-decoder), which achieves more thorough feature fusion in a spliced manner. Medical imaging exhibits fixed structure, complex gradients, and small scale. U-Net is widely used in medical image segmentation because it can be trained with smaller datasets while fusing low-resolution information with high-resolution information. PSPNet adds a module called pyramid pooling to the model, which increases the receptive field by paralleling global pooling layers for four scales so that the model can understand the context.
DeepLabV3 [43] introduces dilated convolutions in the cascade module, which solves the low-resolution problem caused by consecutive pooling and downsampling. In addition, the improvement of the ASPP module also allows the model to capture sufficient context information. This improvement is achieved by employing stacked or parallel ASPP modules consisting of batch normalization layers and dilated convolutions with varying dilated ratios. Another approach to enhance the ASPP module is demonstrated in DenseASPP [44]. The non-dense sampling of dilated convolutions may lead to losing much information. Therefore, this model designs a tighter structure splicing dilated convolution for obtaining a larger receptive field.
Both BiSeNet [45] and ICNet [46] focus on lightweight real-time semantic segmentation tasks. The remarkable feature of the BiSeNet model is the combination of spatial and semantic paths. The shallower networks and more channels of the former are suitable for preserving spatial features, while the deeper networks and fewer channels of the latter are suitable for extracting context. By fusing high-level and low-level features, BiSeNet considers both pixel prediction speed and prediction accuracy. ICNet is a variant of PSPNet. The compressed image is fed into this model from different branches than the original image. The low-resolution image is responsible for providing simple semantic prediction, and the integration of medium-resolution and high-resolution features by the cascade framework can refine the semantic predictions step by step. The performance of ICNet is manifested at less prediction cost.
The novelty of DANet [47] resides in its integration of attention modules (channel attention module, position attention module) and conventional dilated convolution FCN. This module can obtain the corresponding feature dependencies and update the position feature and channel maps according to the weights. Thus, the model learns better features.
2.5 Methods of Radiomics
In order to verify the potential of AATCT-IDS being used to study radiomics, features in multiple dimensions, including texture, area, and perimeter, are extracted from the image, and unsupervised clustering at the patient level is performed based on the features. In addition, the clustering results are analyzed by evaluation indicators.
-Means is a standard prototype-based clustering algorithm. After initializing the set into clusters, it will alternately iterate the centroid and clusters until the cluster center changes minimally after iteration. The shape of the clusters can judge the clustering effect. Usually, the spherical cluster symbolizes a better iteration result. DBSCAN is a standard density clustering algorithm that can find the largest set of density-connected samples. By recursively incorporating unvisited samples that meet the specified criteria and are in the neighborhood of visited samples into clusters, DBSCAN enables the expansion of clusters. Therefore, it does not necessitate a preset quantity of clusters and can uncover clusters of diverse geometries. The AGglomerative NESting (AGNES) algorithm employs a bottom-up aggregation strategy. Within the computation procedure of AGNES, each sample is initially treated as an independent cluster, and the algorithm continuously merges the two closest clusters while updating the distance matrix until the preset number of clusters is achieved.
2.6 Evaluation Metric
This study uses visualization and evaluation metrics to examine denoising, segmentation, and clustering methods. The evaluation of denoising methods encompasses indicators such as Structural Similarity (SSIM), Normalized Root Mean Squared Error (NRMSE), and Peak Signal-to-Noise Ratio (PSNR). The evaluation of segmentation methods involves Mean Intersection over Union (MIoU), Mean Pixel Accuracy (MPA), and Dice coefficient. Furthermore, clustering methods are evaluated based on Davies Bouldin Index (DBI) and Dunn Validity Index (DVI).
2.6.1 Denoising Evaluation Metric
SSIM measures the similarity between images involving three factors (brightness, contrast, structural information), in which brightness and contrast do not affect structural information. SSIM is defined as:
(1) |
Given reference image and denoised image . In Equation 1, , , and correspond to the dynamic range of the gray value, mean, and variance, respectively, and and are constants. According to the definition of the equation, the value of SSIM is in the range and positively correlates with the similarity.
The mean square error (MSE) after the squared root can be normalized to NRMSE. MSE can objectively reflect the difference between pixels, and RMSE reduces the impact of outliers on this basis. PSNR also base on MSE, which can reflect whether the image is distorted by calculating the ratio between practical information and noise. MSE, NRMSE, and PSNR can all be defined mathematically:
(2) |
(3) |
(4) |
The and in Equation 2 represent the length and width of the image, respectively. and correspond to the gray value at coordinates in image and image , respectively. And the in Equation 3 represents the mean value of the reference image. The max in Equation 4 is the maximum gray value of the image. Moreover, NRMSE negatively correlates with the denoising effect, and PSNR positively correlates with the denoising effect.
2.6.2 Segmentation Evaluation Metric
MIoU computes the average value of the ratio between the intersection and union of labels and predictions for each class present in the given image. On the other hand, MPA calculates the proportion of pixels in the predicted results that match the label to the total number of predictions. In addition, the Dice coefficient is an assessment tool for measuring the similarity between sets of pixels and exhibits a positive correlation with the intersection over union. These above metrics can be precisely defined using mathematical formulas:
(5) |
(6) |
(7) |
The in Equation 5 and Equation 6 represent the number of pixel types in the image except for the background, and the TP, FP, TN, and FN are defined in Equation 3.
Ground truth | Predict mask | |
---|---|---|
Positive | Negative | |
Positive | TP | FN |
Negative | FP | TN |
2.6.3 Clustering Evaluation Metric
No clinical information involving the privacy of subjects is publicly available in AATCT-IDS, so this paper uses an internal evaluation metric to evaluate the clustering results. DBI reflects the intra-cluster and inter-cluster similarity by calculating the ratio of intra-cluster distance to inter-cluster distance of any two clusters. Therefore, the smaller the metric, the better the clustering performance is characterized. DVI calculates the ratio of the minimum distance (inter-cluster) to the maximum distance (intra-cluster) among all sample points of any two clusters. Both can be defined mathematically:
(8) |
(9) |
The aim of Equation 8 is to cluster the samples into clusters, where represents the average Euclidean distance from the sample to the center in the cluster, and represents the Euclidean distance between the center of the cluster and cluster. Furthermore, in Equation 9, and represent the Euclidean distance between two samples belonging to different clusters and belonging to the same cluster, respectively.
3 Result and Analysis
3.1 Results and Analysis of Denoising Methods
Various denoising methods participate in the comparative study of this paper, and these methods are used to restore the AATCT-IDS data that are contaminated by noise. Figure 3 exhibits the visualization outcomes obtain from different combinations of denoising algorithms and noise, all of which correspond to the same original image. Due to space limitations, this figure presents the denoising results for some parameters. Specifically, identical parameters are applied to all types of noise (, ), with a filter window size of 3 for both the smoothing filter and the Wiener filter and Daubechies as the wavelet basis for the WT. Detailed multi-parameter results are presented in Table 4. Where Gaussian-1 and Gaussian-2 denote Gaussian noise with parameters , and , , respectively. Wavelet-1 and Wavelet-2 indicate that the WT with Daubechies and Dmeyer as wavelet functions, respectively. Also, the sizes in this table mean the filter window size.
This study uses images contaminated with low-intensity Poisson noise, so the relevant results all present nice visual effects. The data in Table 4 further indicate that the Wiener filter and the NLM algorithm are more specific to this noise. However, the initially complex visceral adipose boundary becomes smooth after NLM processing, which confirms that the noise removal by NLM sacrifices image details. In addition, the median filter with a larger filter window obtains better data results in removing Gaussian noise, but its corresponding image is too blurred. In particular, for some images of thin subjects, the partial outer boundaries of the subcutaneous adipose area are destroyed, which is also reflected in other methods using filter windows.
Mixed types of noise exhibit a more complex distribution, posing a greater challenge for algorithms to suppress the noise effectively. Although the smoothing strategy in this study yields a higher evaluation index for this kind of noise, the resulting images generally fail to maintain adequate sharpness. Conversely, denoising methods based on WT show advantages in restoring AATCT-IDS data contaminated by multiple noises. This method balances the Signal-to-Noise Ratio and image details, achieving comparable objective data to the smoothing strategy while effectively preserving the original image grayscale. In addition, BM3D also preserves most of the image structure, especially when dealing with contamination caused by pulse noise. The excellent performance of BM3D comes at the cost of increased time complexity, and its restoration of AATCT-IDS data consumes much computing time.

Noises | Methods | Size | Size | Methods | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|
SSIM | NRMSE | PSNR | SSIM | NRMSE | PSNR | SSIM | NRMSE | PSNR | |||
Poisson | Mean | 0.936 | 0.164 | 27.7 | 0.869 | 0.253 | 23.9 | NLM | 0.964 | 0.074 | 34.6 |
Gaussian | 0.954 | 0.133 | 29.5 | 0.927 | 0.176 | 27.1 | BM3D | 0.887 | 0.107 | 31.4 | |
Wiener | 0.965 | 0.074 | 34.6 | 0.932 | 0.093 | 32.6 | Wavelet-1 | 0.893 | 0.109 | 31.3 | |
Median | 0.951 | 0.100 | 32.0 | 0.898 | 0.186 | 26.7 | Wavelet-2 | 0.872 | 0.105 | 31.6 | |
Gaussian-1 | Mean | 0.229 | 0.300 | 22.6 | 0.200 | 0.340 | 21.4 | NLM | 0.210 | 0.308 | 22.2 |
Gaussian | 0.237 | 0.290 | 22.9 | 0.232 | 0.295 | 22.6 | BM3D | 0.206 | 0.311 | 22.0 | |
Wiener | 0.230 | 0.278 | 23.1 | 0.233 | 0.259 | 23.7 | Wavelet-1 | 0.210 | 0.305 | 22.3 | |
Median | 0.298 | 0.215 | 25.4 | 0.384 | 0.227 | 24.9 | Wavelet-2 | 0.211 | 0.303 | 22.4 | |
Gaussian-2 | Mean | 0.179 | 0.453 | 18.9 | 0.159 | 0.464 | 18.7 | NLM | 0.135 | 0.603 | 16.4 |
Gaussian | 0.183 | 0.455 | 18.9 | 0.185 | 0.443 | 19.1 | BM3D | 0.138 | 0.569 | 17.1 | |
Wiener | 0.164 | 0.478 | 18.4 | 0.167 | 0.441 | 19.1 | Wavelet-1 | 0.137 | 0.571 | 16.9 | |
Median | 0.204 | 0.337 | 21.4 | 0.270 | 0.291 | 22.7 | Wavelet-2 | 0.138 | 0.571 | 16.9 | |
Poisson | Mean | 0.227 | 0.297 | 22.6 | 0.194 | 0.341 | 21.4 | NLM | 0.207 | 0.314 | 22.1 |
Gaussian-1 | Gaussian | 0.234 | 0.288 | 22.8 | 0.230 | 0.297 | 22.6 | BM3D | 0.206 | 0.317 | 22.0 |
Wiener | 0.226 | 0.284 | 23.0 | 0.230 | 0.266 | 23.5 | Wavelet-1 | 0.207 | 0.310 | 22.2 | |
Median | 0.296 | 0.219 | 25.2 | 0.382 | 0.231 | 24.7 | Wavelet-2 | 0.207 | 0.308 | 22.2 | |
Poisson | Mean | 0.222 | 0.305 | 22.3 | 0.192 | 0.346 | 21.2 | NLM | 0.198 | 0.334 | 21.5 |
+Gaussian-1 | Gaussian | 0.229 | 0.296 | 22.6 | 0.226 | 0.304 | 22.4 | BM3D | 0.198 | 0.330 | 21.6 |
+Speckle | Wiener | 0.216 | 0.302 | 22.4 | 0.220 | 0.283 | 23.0 | Wavelet-1 | 0.197 | 0.329 | 21.7 |
Median | 0.293 | 0.230 | 24.8 | 0.383 | 0.242 | 24.3 | Wavelet-2 | 0.198 | 0.327 | 21.7 | |
Poisson | Mean | 0.212 | 0.343 | 21.3 | 0.183 | 0.370 | 20.7 | NLM | 0.185 | 0.456 | 18.8 |
+Gaussian-1 | Gaussian | 0.218 | 0.341 | 21.4 | 0.216 | 0.336 | 21.5 | BM3D | 0.187 | 0.437 | 19.8 |
+Pulse | Wiener | 0.202 | 0.426 | 19.4 | 0.201 | 0.363 | 20.8 | Wavelet-1 | 0.182 | 0.435 | 19.2 |
Median | 0.291 | 0.224 | 25.0 | 0.378 | 0.235 | 24.6 | Wavelet-2 | 0.183 | 0.435 | 19.3 |
3.2 Results and Analysis of Segmentation Methods
In this study, the training parameters of all models are unified to obtain data with comparative value. As the segmentation outcomes are solely employed for comparative analyses, these parameters are not tuned to be optimal states. Also, each training consisted of 150 epochs. Figure 4 shows the segmentation results of different models for the same abdominal CT slice, where the reference image covered with a mask is marked with the word ”label.” A more objective evaluation of the outcomes is presented in Table 5. Considering that certain clinical studies only examine specific types of adipose tissue, this study conducted separate training and verification using images of subcutaneous adipose, visceral adipose, and overall abdominal adipose.
As the fundamental model, FCN exhibits limitations in effectively separating the narrow visceral adipose tissue between organs. Nonetheless, the appeal of FCN lies in its lower training cost. Both DenseASPP and DeepLabV3 adopt the strategy of optimizing the ASPP module and show approximate segmentation performance on AATCT-IDS. The difference is that DenseASPP further reduces the training time. The visualization results and evaluation metrics of U-Net surpass those of other models, which confirms its wide application in medical image segmentation. It can separate small and isolated adipose tissues around the kidney while avoiding tiny structures such as small intestinal capillaries. BiSeNet stands out as a lightweight real-time semantic segmentation model, delivering impressive evaluation metrics second only to U-Net for AATCT-IDS segmentation. What sets BiSeNet apart is its ability to achieve these results while consuming the least training time, making it a noteworthy advantage. However, ICNet, another lightweight real-time semantic segmentation model, exhibits a considerable performance gap on this dataset, particularly in coping with the segmentation of subcutaneous adipose tissue. Additionally, given the substantial background area within abdominal CT images, the MPA metrics of each model tend to have approximate values.

Method | Type | MIoU | MPA | Dice | Training time |
---|---|---|---|---|---|
FCN8s | Abdominal | 0.828 | 0.960 | 0.888 | 5h5min |
Subcutaneous | 0.889 | 0.984 | 0.908 | 5h8min | |
Visceral | 0.831 | 0.962 | 0.884 | 5h6min | |
PSPNet | Abdominal | 0.855 | 0.968 | 0.903 | 14h6min |
Subcutaneous | 0.910 | 0.987 | 0.924 | 14h19min | |
Visceral | 0.847 | 0.969 | 0.897 | 14h22min | |
DeepLabV3 | Abdominal | 0.849 | 0.971 | 0.900 | 15h29min |
Subcutaneous | 0.900 | 0.990 | 0.918 | 15h43min | |
Visceral | 0.853 | 0.972 | 0.901 | 15h42min | |
DenseASPP | Abdominal | 0.845 | 0.966 | 0.897 | 8h42min |
Subcutaneous | 0.904 | 0.989 | 0.921 | 8h46min | |
Visceral | 0.844 | 0.964 | 0.895 | 8h53min | |
DANet | Abdominal | 0.898 | 0.978 | 0.944 | 14h56min |
Subcutaneous | 0.950 | 0.991 | 0.974 | 15h1min | |
Visceral | 0.895 | 0.979 | 0.942 | 15h10min | |
BiSeNet | Abdominal | 0.901 | 0.978 | 0.947 | 2h37min |
Subcutaneous | 0.957 | 0.991 | 0.978 | 2h36min | |
Visceral | 0.900 | 0.976 | 0.945 | 2h43min | |
ICNet | Abdominal | 0.790 | 0.953 | 0.867 | 4h42min |
Subcutaneous | 0.865 | 0.978 | 0.890 | 4h45min | |
Visceral | 0.800 | 0.955 | 0.860 | 4h44min | |
U-Net | Abdominal | 0.907 | 0.980 | 0.952 | 10h3min |
Subcutaneous | 0.960 | 0.991 | 0.980 | 10h8min | |
Visceral | 0.905 | 0.979 | 0.949 | 10h5min |
3.3 Results and Analysis of Clustering Methods
Since clustering high-dimensional features is difficult to visualize the results and the evaluation process is complicated, this study uses t-SNE to reduce the multi-dimensional features to two-dimensional space for clustering. In addition, the multi-dimensional features under consideration pertain to the shape or texture features associated with adipose distribution. The outcomes of clustering using -Means on this set of two-dimensional features and the associated evaluation results are presented in Figure 5, where each dot is related to a subject, and dots of the same color are classified as the same cluster. The table on the right side of this figure evaluates the closeness of the different clusters. Moreover, Figure 6 and Figure 7 demonstrate the clustering and related evaluation results of DBSCAN and AGNES, respectively.
The -Means method in this study performed ten operations, each with a different number of clusters and centroids, and the optimal result is shown in Figure 5. Similarly, the AGNES method undergoes multiple rounds of operations and has the same experimental results as the -Means method, although it defines different orders for the clusters. Both -Means and AGNES clearly classify this set of two-dimensional features into three relatively discrete clusters, which is highly similar to the clustering results of DBSCAN. Therefore, this study divides the subjects into three subsets according to the clustering results of AGNES and conducts a correlation analysis on the shape features and texture features of these three subsets, respectively. The outcomes of this analysis disclosed three distinct patterns of adipose distribution within the dataset population, thereby showcasing the potential of utilizing AATCT-IDS for radiomics research.



3.4 Discussion
Different from conventional natural image segmentation tasks, clinical diagnosis has higher requirements for image prediction accuracy, which is a prerequisite for achieving precision medicine. Some of the segmentation models involved in the comparative study in this paper have obtained good evaluation data. However, many tiny structures in the visualization results of these models still need to make accurate predictions. Therefore, there remains scope for enhancing the accuracy of adipose tissue segmentation in CT images. In addition, abdominal CT images can yield various categories of radiomics features, but only some of the shape and texture features are covered in this paper. Multi-dimensional features containing multiple categories may have a closer relationship with clinical diagnosis.
4 Conclusion and Future Works
This paper details an abdominal adipose tissue CT image dataset from 300 subjects, AATTCT-IDS. The dataset publics 13,732 raw CT slices, and the researchers individually annotate the subcutaneous and visceral adipose tissue regions of 3,213 of those slices that have the same slice distance. Additionally, all files in the dataset are stored on a subject-by-subject basis.
AATTCT-IDS has the function of examining denoising, segmentation, and radiomics methods. This paper conducts a comparative study on these three types of tasks. In the image denoising study, the CT images in AATTCT-IDS are contaminated with noises of different intensities and distributions. The seven denoising methods in the image restoration step demonstrate different suppression characteristics to these noises. The research delves into image segmentation, employing deep learning to assess the efficacy of eight semantic segmentation models across three clinical medical segmentation tasks. This paper focuses on analyzing the structural characteristics of the adipose tissue segmented by each model while discussing the ability of each model to segment abdominal adipose in combination with the actual training time. In addition, in the radiomics study based on AATTCT-IDS, this paper divides subjects into three subsets by referring to the clustering results of shape features and texture features after dimensionality reduction by various unsupervised methods. Correlation analysis of features performed on these subsets further validates the three fat distributions in AATTCT-IDS.
The comparative analysis of the datasets indicates potential areas for enhancement in the segmentation model. We are committed to exploring a diverse range of approaches for AATTCT-IDS, aiming to unlock a broader scope of possibilities in precision medicine. Furthermore, we intend to integrate clinical diagnostic results with multi-dimensional features spanning various categories, in order to enhance and deepen the study of radiomics.
Acknowledgments
This work is supported by the “National Natural Science Foundation of China” (No. 82220108007). We thank B.A. Qi Qiu, from Foreign Studies College in Northeastern University, China, for her professional English proofreading in this paper. We also thank Miss. Zixian Li and Mr. Guoxian Li for their important discussion in this work.
Declaration of Competing Interest
The authors declare that they have no conflict of interest.
Data Availability Statement
Ethics is proved by China Medical Universiy, China: No. 2022PS433K. The datasets presented in this study can be found in online repositories. The names of the repository and accession number can be found below: https://figshare.com/articles/dataset/AATTCT-IDS/23807256.
References
- [1] J. D. Tune, A. G. Goodwill, D. J. Sassoon, K. J. Mather, Cardiovascular consequences of metabolic syndrome, Translational Research 183 (2017) 57–70.
- [2] M. G. Saklayen, The global epidemic of the metabolic syndrome, Current hypertension reports 20 (2) (2018) 1–8.
- [3] R. Kumari, S. Kumar, R. Kant, An update on metabolic syndrome: Metabolic risk markers and adipokines in the development of metabolic syndrome, Diabetes & Metabolic Syndrome: Clinical Research & Reviews 13 (4) (2019) 2409–2417.
- [4] Y. Wang, M. A. Beydoun, The obesity epidemic in the united states—gender, age, socioeconomic, racial/ethnic, and geographic characteristics: a systematic review and meta-regression analysis, Epidemiologic reviews 29 (1) (2007) 6–28.
- [5] K. Yamagishi, H. Iso, The criteria for metabolic syndrome and the national health screening and education system in japan, Epidemiology and health 39 (2017) 2017003.
- [6] P. M. Graffy, P. J. Pickhardt, Quantification of hepatic and visceral fat by ct and mr imaging: relevance to the obesity epidemic, metabolic syndrome and nafld, The British journal of radiology 89 (1062) (2016) 20151024.
- [7] Y. Wang, M. A. Beydoun, L. Liang, B. Caballero, S. K. Kumanyika, Will all americans become overweight or obese? estimating the progression and cost of the us obesity epidemic, Obesity 16 (10) (2008) 2323–2330.
- [8] N. I. Parikh, M. J. Pencina, T. J. Wang, K. J. Lanier, C. S. Fox, R. B. D’Agostino, R. S. Vasan, Increasing trends in incidence of overweight and obesity over 5 decades, The American journal of medicine 120 (3) (2007) 242–250.
- [9] S. Zhu, S. B. Heymsfield, H. Toyoshima, Z. Wang, A. Pietrobelli, S. Heshka, Race-ethnicity–specific waist circumference cutoffs for identifying cardiovascular disease risk factors, The American journal of clinical nutrition 81 (2) (2005) 409–415.
- [10] R. H. Eckel, S. M. Grundy, P. Z. Zimmet, The metabolic syndrome, The Lancet 365 (9468) (2005) 1415–1428.
- [11] F. Pesapane, M. Codari, F. Sardanelli, Artificial intelligence in medical imaging: threat or opportunity? radiologists again at the forefront of innovation in medicine, European Radiology Experimental 2 (1) (2018) 35.
- [12] S. K. Zhou, H. Greenspan, C. Davatzikos, J. S. Duncan, B. Van Ginneken, A. Madabhushi, J. L. Prince, D. Rueckert, R. M. Summers, A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises, Proceedings of the IEEE 109 (5) (2021) 820–838.
- [13] M. H. Hesamian, W. Jia, X. He, P. Kennedy, Deep learning techniques for medical image segmentation: Achievements and challenges, Journal of Digital Imaging 32 (4) (2019) 582–596.
- [14] D. Shen, G. Wu, H.-I. Suk, Deep learning in medical image analysis, Annual Review of Biomedical Engineering 19 (1) (2017) 221–248.
- [15] A. D. Weston, P. Korfiatis, T. L. Kline, K. A. Philbrick, P. Kostandy, T. Sakinis, M. Sugimoto, N. Takahashi, B. J. Erickson, Automated abdominal segmentation of ct scans for body composition analysis using deep learning, Radiology 290 (3) (2019) 669–679.
- [16] G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, J. A. van der Laak, B. van Ginneken, C. I. Sánchez, A survey on deep learning in medical image analysis, Medical Image Analysis 42 (2017) 60–88.
- [17] W. Shen, M. Punyanitya, Z. Wang, D. Gallagher, M.-P. St.-Onge, J. Albu, S. B. Heymsfield, S. Heshka, Total body skeletal muscle and adipose tissue volumes: estimation from a single abdominal cross-sectional image, Journal of Applied Physiology 97 (6) (2004) 2333–2338.
- [18] W. Shen, M. Punyanitya, Z. Wang, D. Gallagher, M.-P. St-Onge, J. Albu, S. B. Heymsfield, S. Heshka, Visceral adipose tissue: relations between single-slice areas and total volume, The American Journal of Clinical Nutrition 80 (2) (2004) 271–278.
- [19] R. J. Gillies, P. E. Kinahan, H. Hricak, Radiomics: Images are more than pictures, they are data, Radiology 278 (2) (2016) 563–577.
- [20] J. Kullberg, H. Ahlström, L. Johansson, H. Frimmel, Automated and reproducible segmentation of visceral and subcutaneous adipose tissue from abdominal mri, International Journal of Obesity 31 (12) (2007) 1806–1817.
- [21] S. Makrogiannis, G. Caturegli, C. Davatzikos, L. Ferrucci, Computer-aided assessment of regional abdominal fat with food residue removal in ct, Academic Radiology 20 (11) (2013) 1413–1421.
- [22] Z. Wang, Y. Meng, F. Weng, Y. Chen, F. Lu, X. Liu, M. Hou, J. Zhang, An effective cnn method for fully automated segmenting subcutaneous and visceral adipose tissue on ct scans, Annals of Biomedical Engineering 48 (1) (2020) 312–328.
- [23] B. S. Devi, D. S. Misbha, Development of a fully convolutional network for the segmentation of adipose tissues on abdominal mri, in: A. P. Pandian, X. Fernando, W. Haoxiang (Eds.), Computer Networks, Big Data and IoT, Lecture Notes on Data Engineering and Communications Technologies, Springer Nature, Singapore, 2022, pp. 739–751.
- [24] N. Shen, X. Li, S. Zheng, L. Zhang, Y. Fu, X. Liu, M. Li, J. Li, S. Guo, H. Zhang, Automated and accurate quantification of subcutaneous and visceral adipose tissue from magnetic resonance imaging based on machine learning, Magnetic Resonance Imaging 64 (2019) 28–36.
- [25] S. Masoudi, S. M. Anwar, S. A. Harmon, P. L. Choyke, B. Turkbey, U. Bagci, Adipose tissue segmentation in unlabeled abdomen mri using cross modality domain adaptation, in: 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), 2020, pp. 1624–1628.
- [26] Y. Wang, Y. Qiu, T. Thai, K. Moore, H. Liu, B. Zheng, A two-step convolutional neural network based computer-aided detection scheme for automatically segmenting adipose tissue volume depicting on ct images, Computer Methods and Programs in Biomedicine 144 (2017) 97–104.
- [27] H. Shen, P. He, Y. Ren, Z. Huang, S. Li, G. Wang, M. Cong, D. Luo, D. Shao, E. Y.-P. Lee, R. Cui, L. Huo, J. Qin, J. Liu, Z. Hu, Z. Liu, N. Zhang, A deep learning model based on the attention mechanism for automatic segmentation of abdominal muscle and fat for body composition assessment, Quantitative Imaging in Medicine and Surgery 13 (3) (2023) 1384–1398.
- [28] T. Langner, A. Hedström, K. Mörwald, D. Weghuber, A. Forslund, P. Bergsten, H. Ahlström, J. Kullberg, Fully convolutional networks for automated segmentation of abdominal adipose tissue depots in multicenter water–fat mri, Magnetic Resonance in Medicine 81 (4) (2019) 2736–2745.
- [29] S. Estrada, R. Lu, S. Conjeti, X. Orozco-Ruiz, J. Panos-Willuhn, M. M. B. Breteler, M. Reuter, Fatsegnet: A fully automated deep learning pipeline for adipose tissue segmentation on abdominal dixon mri, Magnetic Resonance in Medicine 83 (4) (2020) 1471–1483.
- [30] S. Koitka, L. Kroll, E. Malamutmann, A. Oezcelik, F. Nensa, Fully automated body composition analysis in routine ct imaging using 3d semantic segmentation convolutional neural networks, European Radiology 31 (4) (2021) 1795–1804.
- [31] Ö. Çiçek, A. Abdulkadir, S. S. Lienkamp, T. Brox, O. Ronneberger, 3d u-net: Learning dense volumetric segmentation from sparse annotation, in: S. Ourselin, L. Joskowicz, M. R. Sabuncu, G. Unal, W. Wells (Eds.), Medical Image Computing and Computer-Assisted Intervention – MICCAI 2016, Lecture Notes in Computer Science, Springer International Publishing, Cham, 2016, pp. 424–432.
- [32] M. R. Baggerman, D. P. van Dijk, B. Winkens, R. M. Schnabel, R. J. van Gassel, M. E. Bol, F. C. Bakers, S. W. Olde Damink, M. C. van de Poll, Edema in critically ill patients leads to overestimation of skeletal muscle mass measurements using computed tomography scans, Nutrition 89 (2021) 111238.
- [33] M. Diwakar, M. Kumar, A review on ct image noise and its denoising, Biomedical Signal Processing and Control 42 (2018) 73–88.
- [34] A. Ravishankar, S. Anusha, H. K. Akshatha, A. Raj, S. Jahnavi, J. Madhura, A survey on noise reduction techniques in medical images, in: 2017 International conference of Electronics, Communication and Aerospace Technology (ICECA), Vol. 1, 2017, pp. 385–389.
- [35] D. Thanh, P. Surya, et al., A review on ct and x-ray images denoising methods, Informatica 43 (2).
- [36] R. Kaur, M. Juneja, A. K. Mandal, A comprehensive review of denoising techniques for abdominal ct images, Multimedia Tools and Applications 77 (17) (2018) 22735–22770.
- [37] A. Buades, B. Coll, J. M. Morel, A review of image denoising algorithms, with a new one, Multiscale Modeling & Simulation 4 (2) (2005) 490–530.
- [38] S. Mallat, A theory for multiresolution signal decomposition: the wavelet representation, IEEE Transactions on Pattern Analysis and Machine Intelligence 11 (7) (1989) 674–693.
- [39] K. Dabov, A. Foi, V. Katkovnik, K. Egiazarian, Image denoising by sparse 3-d transform-domain collaborative filtering, IEEE Transactions on Image Processing 16 (8) (2007) 2080–2095.
- [40] J. Long, E. Shelhamer, T. Darrell, Fully convolutional networks for semantic segmentation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3431–3440.
- [41] O. Ronneberger, P. Fischer, T. Brox, U-net: Convolutional networks for biomedical image segmentation, in: N. Navab, J. Hornegger, W. M. Wells, A. F. Frangi (Eds.), Lecture Notes in Computer Science, Lecture Notes in Computer Science, Springer International Publishing, Cham, 2015, pp. 234–241.
- [42] H. Zhao, J. Shi, X. Qi, X. Wang, J. Jia, Pyramid scene parsing network, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2881–2890.
- [43] L.-C. Chen, G. Papandreou, F. Schroff, H. Adam, Rethinking atrous convolution for semantic image segmentation (2017). arXiv:1706.05587.
- [44] M. Yang, K. Yu, C. Zhang, Z. Li, K. Yang, Denseaspp for semantic segmentation in street scenes, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 3684–3692.
- [45] C. Yu, J. Wang, C. Peng, C. Gao, G. Yu, N. Sang, Bisenet: Bilateral segmentation network for real-time semantic segmentation, in: Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 325–341.
- [46] H. Zhao, X. Qi, X. Shen, J. Shi, J. Jia, Icnet for real-time semantic segmentation on high-resolution images, in: Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 405–420.
- [47] J. Fu, J. Liu, H. Tian, Y. Li, Y. Bao, Z. Fang, H. Lu, Dual attention network for scene segmentation, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 3146–3154.