Synergistic Learning of Lung Lobe Segmentation and Hierarchical Multi-Instance Classification for Automated Severity Assessment of COVID-19 in CT Images
Abstract
Understanding chest CT imaging of the coronavirus disease 2019 (COVID-19) will help detect infections early and assess the disease progression. Especially, automated severity assessment of COVID-19 in CT images plays an essential role in identifying cases that are in great need of intensive clinical care. However, it is often challenging to accurately assess the severity of this disease in CT images, due to variable infection regions in the lungs, similar imaging biomarkers, and large inter-case variations. To this end, we propose a synergistic learning framework for automated severity assessment of COVID-19 in 3D CT images, by jointly performing lung lobe segmentation and multi-instance classification. Considering that only a few infection regions in a CT image are related to the severity assessment, we first represent each input image by a bag that contains a set of 2D image patches (with each cropped from a specific slice). A multi-task multi-instance deep network (called M2UNet) is then developed to assess the severity of COVID-19 patients and also segment the lung lobe simultaneously. Our M2UNet consists of a patch-level encoder, a segmentation sub-network for lung lobe segmentation, and a classification sub-network for severity assessment (with a unique hierarchical multi-instance learning strategy). Here, the context information provided by segmentation can be implicitly employed to improve the performance of severity assessment. Extensive experiments were performed on a real COVID-19 CT image dataset consisting of 666 chest CT images, with results suggesting the effectiveness of our proposed method compared to several state-of-the-art methods.
Index Terms:
COVID-19, CT, Severity Assessment, Lung lobe Segmentation, Multi-Task Learning, Multi-Instance LearningI Introduction
THE coronavirus disease 2019 (COVID-19) is spreading fast worldwide since the end of . Until April 5, about million patients are confirmed with this infectious disease, reported by [1]. This raises a Public Health Emergency of International Concern (PHEIC) of WHO. In the field of medical image analysis, many imaging-based artificial intelligence methods have been developed to help fight against this disease, including automated diagnosis [2, 3, 4], segmentation [5, 6, 7], and follow-up and prognosis [8].


Previous imaging-based studies mainly focus on identifying COVID-19 patients from non-COVID-19 subjects. As the golden standard for COVID-19 is the Reverse Transcription-Polymerase Chain Reaction (RT-PCR) test, the effectiveness of those imaging-base applications is limited.
Moreover, approximately of patients with COVID-19 have only mild to moderate symptoms [9], while the remaining patients have severe symptoms. Based on previous studies [10, 5], the imaging-based characters of COVID-19 patients are distinct to related diseases, e.g., viral pneumonia. Therefore, the severity assessment of the disease is of high clinical value, which helps effectively allocate medical resources such as ventilator. Among various radiological examinations, chest CT imaging plays an essential role in fighting this infectious disease by helping early identify lung infections and assess the severity of the disease. Previous studies show that computed tomography (CT) has the ability to provide valuable information in the screening and diagnosis [10]. In this work, CT help the clinicians to evaluate the condition of the patients in advance, by which necessary measures or treatments could better proceed, especially for severe patients in time.
However, automatically assess the severity of COVID-19 in CT images is a very challenging task. First, infections caused by COVID-19 often occur in small regions of the lungs and are difficult to identify in CT images, as shown in Fig. 1. Second, imaging biomarkers of COVID-19 patients caused by an infection are similar in some severe and non-severe cases are similar, including ground-glass opacities (GGO), mosaic sign, air bronchogram, and interlobular septal thickening (Fig. 1). In addition, there are large inter-case variations in CT images of COVID-19 patients (Fig. 2), because these images are usually acquired by multiple imaging centers with different scanners and different scanning parameters.
Several recent methods have been proposed for the diagnosis of COVID-19 [2, 3, 4], with only some specifically designed for severity assessment of the disease. In several studies [5, 6, 7], segmentation of lung or lung lobe is used as a prerequisite procedure for diagnosis purposes.
However, most of these methods treat the lung lobe segmentation and disease diagnosis as two separate tasks, ignoring their underlying association. Note that the segmentation of lung lobe can provide rich information regarding spatial locations and tissue types in CT images. Therefore, it is intuitively reasonable to jointly perform lung lobe segmentation and severity assessment/prediction, where the context information provided by segmentation results can be used to improve the prediction performance. The joint learning scheme is obviously faster than the two-stage framework, since detecting and cropping the lung field are not needed. Besides, as demonstrated by related works of class activation maps (CAMs) [11], the classification task raises high signal responses in lung lobe area, where the infection patterns of lung lobe in disease progression could also provide useful guidance for the segmentation of lung lobes.
Moreover, most of the previous works are based on 2D image slices [2, 3, 4]. However, the annotation of 2D CT slices is a heavy workload for radiologists. It is interesting to directly employ 3D CT images for automated severity assessment of COVID-19, which is desired for real-world clinical applications.
To this end, in this paper, we propose a synergistic learning framework for automated severity assessment of COVID-19 in 3D CT images, by jointly performing severity assessment of COVID-19 patients and lung lobe segmentation. The proposed method extends the previous conference work [12] to deal with 3D images, and makes substantial methodological improvements for the task of COVID-19 severity assessment. Specifically, considering that only a few slices in CT images are related to severity assessment, each input CT image is represented by a bag of 2D image patches, each of which is randomly cropped from a specific slice. Furthermore, each slice is represented by a bag of infection regions represented by intermediate embedding features. With each bag as input, a multi-task multi-instance deep neural network (called M2UNet) is developed, including 1) a shared patch-level encoder, 2) a segmentation sub-network for lung lobe segmentation, and 3) a classification sub-network for severity assessment of COVID-19 patients (i.e., severe or non-severe) using a hierarchical multi-instance learning strategy. Here, the segmentation results are used to provide context information of input CT images to boost the performance of severity assessment. Extensive experiments have been performed on a real-world COVID-19 dataset with chest CT images, with the results demonstrating the effectiveness of the proposed method compared to several state-of-the-art methods.
The contributions of this work are three-fold:
-
•
A multi-task multi-instance learning framework is proposed to jointly assess the severity of the COVID-19 patients and segment lung lobes in chest CT images, where the segmentation task provides context information to aid the task of severity assessment in chest CT image.
-
•
A unique hierarchical multi-instance learning strategy is developed to predict the severity of patients in a weakly supervised manner.
-
•
We evaluate the proposed method on a real clinical dataset with 666 3D CT images of COVID-19 patients, achieving promising results in severity assessment compared to several state-of-the-art methods.
The rest of the paper is organized as follows. In Section 2, we introduce the related works for the segmentation and diagnosis of CT images of COVID-19 patients, as well as related studies on deep multi-instance learning. Then, we introduce the proposed method in Section 3. In Section 4, we present the materials, experimental setup, and experimental results. Finally, we conclude this paper and present several future research directions in Section 5.
II Related Work
In this section, we briefly review the most relevant studies from the following three aspects: 1) lung segmentation of CT images with COVID-19, 2) automated diagnosis of COVID-19 patients, and 3) deep multi-instance learning.
II-A Lung Segmentation of CT Images with COVID-19
Segmentation of lung or lung lobe has been used as a common pre-requisite procedure for automatic diagnosis of COVID-19 based on chest CT images. Several deep learning methods have been proposed for the segmentation of lung in CT images with COVID-19. For instance, U-Net et al .[13] has been widely used for segmentation of both lung regions and lung lesions in COVID-19 applications [14, 15, 16, 7]. Qi et al . [7] use U-Net to delineate the lesions in the lung and extract radiometric features of COVID-19 patients with the initial seeds given by a radiologist for predicting hospital stay. Also, several variants of U-Net have been applied to the diagnosis or severity assessment of COVID-19. Jin et al . [6] design a two-stage pipeline to screen COVID-19 in CT images, and they utilize U-Net++ [17] to detect the whole lung region and to separate lesions from lung regions. Besides, V-Net [18] is also used in various segmentation applications. Shan et al . [5] integrates human-in-the-loop strategy into the training process of VB-Net (a variant of V-Net). The human-aided strategy is an intuitive way to address the issue of lacking manual labels during segmentation in CT images.
II-B Automated Diagnosis of COVID-19
Both X-rays [19] and CT images [10] can provide effective information for the computer-assisted diagnosis of COVID-19. Compared with X-rays, chest CT imaging contains hundreds of slices, which is clearer and more precise but has to take more time for specialists to diagnose. Therefore, there is a great demand to use CT images for automated diagnosis of COVID-19. In general, the existing methods for COVID-19 diagnosis based on CT images can be roughly divided into two categories: 1) classification; 2) severity assessment. In the former category, many studies have been conducted to determine whether patients are infected with COVID-19 disease. For example, Chen et al . [2] exploits a UNet++ based segmentation model to segment COVID-19 related lesions in chest CT images of 51 COVID-19 patients and 55 patients with other diseases, and finally determine the label (COVID-19 or non-COVID-19) of each image based on the segmented lesions. Ying et al . [3] propose a CT diagnosis system, namely DeepPneumonia, which is based on the ResNet50 model to identify patients with COVID-19 from bacteria pneumonia patients and healthy people. In the second category, Tang et al .[4] proposed to first adopt VB-Net to separate the lung into anatomical sub-regions, and then use these sub-regions to compute quantitative features for training a random forest (RF) model for COVID-19 severity assessment (with labels of being non-severe or severe).

II-C Deep Multi-instance Learning
The scenario of multi-instance learning (MIL) [20, 21, 22] or learning from weakly annotated data [23] arises when only a general statement of the category is given, but multiple instances can be observed. MIL aims to learn a model that can predict the label of a bag accurately, and many recent studies have focused on implementing MIL via deep neural networks. For instance, Oquab et al . [23] train a deep model with multiple image patches of multiple scales as input, and aggregate the prediction results of multiple inputs by using a max-pooling operation. Besides, many studies [24, 25, 26] propose to formulate image classification as a MIL problem so as to address the weakly supervised problem. Moreover, MIL is particularly suitable for problems with only a limited number (e.g., tens or hundreds) of training samples in various medical image-based applications, such as computer-assisted disease diagnosis [27, 28, 29, 30]. For instance, Yan et al .[29] propose a two-stage deep MIL method to find discriminative local anatomies, where the first-stage convolutional neural network (CNN) is learned in a MIL fashion to locate discriminative image patches and the second-stage CNN is boosted using those selected patches. More recently, a landmark-based deep MIL framework [30] is developed to learn both local and global representations of MRI for automated brain disease diagnosis, leading to a new direction for handling limited training samples in the domain of medical image analysis. Since there are only a limited number of cases at hand, it is desirable to employ the multi-instance learning strategy for severity assessment of COVID-19 patients in chest CT images.
III Proposed Method
III-A Framework
The framework of the proposed method is illustrated in Fig. 3, where the input is the raw 3D CT image and the output is the lung segmentation and severity assessment of COVID-19 patients (i.e., severe or non-severe). Specifically, each 3D CT image is processed via several image pre-processing steps. Then, a set of 2D image patches is randomly cropped from the processed image to construct an instance bag, and each bag represents a specific input CT image. This bag is regarded as the input of the proposed multi-task multi-instance U-Net (M2UNet). The M2UNet is designed to learn two tasks jointly, i.e., severity assessment of a COVID-19 patient and segmentation of the lung lobe.
As shown in Fig. 3, in M2UNet, an encoding module is first used for patch-level feature extraction of 2D patches in each input bag, followed by two sub-networks for joint severity assessment and lung lobe segmentation. Specifically, in the classification sub-network, these extracted patch-level features are fed into a feature embedding module and an image-level feature learning module to capture the local-to-global volume representation of the input CT image. With the learned volume features, a classification layer is finally used to assess the severity of each COVID-19 patient (i.e., severe or non-severe). In the segmentation sub-network, those patch-level features are fed into a decoding module to perform lung lobe segmentation for each patch in the input bag. Since these two sub-networks are trained jointly with a shared patch-level encoder, the context information provided by the segmentation results can be implicitly employed to improve the performance of severity assessment.
III-B Data Preparation
To eliminate the effect of the background noise in each raw 3D CT image, we crop each scan to only keep the region containing the human body, by using a threshold-based processing method. Specifically, we first binarize the image using the threshold of zero, through which the human tissues and the gas regions will be separated. Then, the human body region is cropped according to the binary mask. Each body region image has a size of at least for the axial plane in this work.
While image resampling is commonly used in many deep learning methods for segmentation and classification [31, 32], we do not resample the raw CT images in order to preserve their original data distributions. Since our method is clinical-oriented with inconsistent imaging qualities, CT images used in this study are not as clean as those in benchmark datasets [33]. For example, the physical spacing of our data has large variation, e.g., from to between slices, because of the use of different CT scanners and scanning parameters. Using a common interpolation method (e.g., tri-linear interpolation) to resample a CT image into , one will introduce heavy artifacts to the image. Besides, only a few infection regions in each CT image are related to severity assessment. To this end, we employ the weakly supervised multi-instance learning (MIL) strategy for handing these inconsistent CT images. Specifically, for each pre-processed CT image, we randomly crop a set of 2D patches sampled from 2D slices (with each patch from a specific slice) in each image to construct an instance bag, and each bag is used to represent a specific CT image and treated as the input of the subsequent M2UNet. In this way, the inter-slice/patch relationships can be implicitly captured by our M2UNet. In addition, this MIL strategy represents each 3D image through a set of 2D image patches rather than sequential slices. This can partially alleviate the problem of data inconsistency, so our method has high practical value in real-world applications.
III-C Network Architecture
As shown in Fig. 3, using each bag (consisting of a set of 2D image patches) as the input data, the proposed M2UNet first employs an encoding module for patch-level feature extraction. Based on these features, the classification and segmentation sub-networks are then used to jointly perform two tasks, respectively, i.e., 1) severity assessment of the patients, and 2) segmentation of lung lobes in each patch. Specifically, the classification sub-network uses a unique hierarchical MIL strategy to extract the local-to-global representation of each input image, with an embedding-level MIL module, an image-level MIL module, and a classification module. The segmentation sub-network contains a decoding module to segment lung lobes of 2D image patches in each bag.
The detailed network architecture is listed in Table I. The combination of the encoder and decoder is U-Net like, with four down-sampling blocks in the encoder and four up-sampling blocks in the decoder. The outputs of the same level blocks in the encoder and decoder are concatenated and fed into the next block of the decoder. Limited by computational resources, all the convolutional layers in the encoder and decoder have the same number (i.e., ) of kernels, except the last block in the encoder. The last block of encoder outputs dimensional features to help build a more robust classification for severity assessment. The decoder outputs the corresponding segmentation mask of five types of lung lobes for each image patch.
Block Name | Num. | Layers | Parameter Setting | Input | # |
Encoding block 1 | 2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | 2D image patches | |
Pool 1 | 1 | max-pooling | K: {}, STR:2 | Encoding block 1 | |
Encoding block 2 | 2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Pool 1 | |
Pool 2 | 1 | max-pooling | K: {}, STR:2 | Encoding block 2 | |
Encoding block 3 | 2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Pool 2 | |
Pool 3 | 1 | max-pooling | K: {}, STR:2 | Encoding block 3 | |
Encoding block 4 | 2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Pool 3 | |
Pool 4 | 1 | max-pooling | K: {}, STR:2 | Encoding block 4 | |
Encoding block 5 | 1 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Pool 4 | |
1 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | |||
Embedding-Level MIL | 1 | GCP | Num. Concepts: 256 | Encoding block 5 | |
1 | conv | K: {}, PAD:0, STR:1 | |||
Image-Level MIL | 1 | GCP | Num. Concepts: 128 | Embedding-Level MIL | |
1 | conv | K: {}, PAD:0, STR:1 | |||
Clssifier | 1 | conv | K: {}, PAD:0, STR:1 | Image-Level MIL | |
Decoding block 5 | 1 | {up-sample, conv, batchnorm, ReLU, concat} | K: {}, PAD:1, STR:1 | Encoding block 5 | |
2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Encoding block 4 | ||
Decoding block 4 | 1 | {up-sample, conv, batchnorm, ReLU, concat} | K: {}, PAD:1, STR:1 | Decoding block 5 | |
2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Encoding block 3 | ||
Decoding block 3 | 1 | {up-sample, conv, batchnorm, ReLU, concat} | K: {}, PAD:1, STR:1 | Decoding block 4 | |
2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Encoding block 2 | ||
Decoding block 2 | 1 | {up-sample, conv, batchnorm, ReLU, concat} | K: {}, PAD:1, STR:1 | Decoding block 3 | |
2 | {conv, batchnorm, ReLU} | K: {}, PAD:1, STR:1 | Encoding block 1 | ||
Decoding block 1 | 1 | conv | K: {}, PAD:0, STR:1 | Decoding block 2 |

III-D Hierarchical Multi-Instance Learning
While infection regions of the lung, related to COVID-19 (e.g., nodule and GGO) are usually located in regions of the CT image, the category of each CT image is labeled at the entire image level, rather than the region-level. That is, many regions are actually unrelated to the classification task for severity assessment.
Multi-instance learning (MIL) provides a useful tool to solve such a weakly supervised problem. Conventional MIL represents a 2D image as a bag, and each bag consists of multiple regions of the input image (i.e., instances). Their overall prediction is made at the bag-level by roughly two kinds of methods, i.e., the embedding-level methods and the instance-level methods. The former learns the relationship among instances by projecting them into a new embedding space. The latter directly generates the bag-level predictions by performing voting on the instance predictions. However, both methods are inapplicable for the classification of 3D images, as 3D images contain multiple 2D image slices, and the class labels are also related to some local regions of the slices.
In this work, based on the previous study on pathological images [12], we propose a hierarchical MIL strategy in the classification sub-network of our M2UNet to perform severity assessment of COVID-19 in 3D CT images, as shown in Fig. 3. As mentioned in Section III-B, we represent each input 3D volumetric image as a bag consisting of a set of 2D image patches, and these patches are regarded as the instances in the MIL problem settings. Formally, we first construct a bag with 2D patches cropped from the regional slices to represent each input CT image. Denote the -th and the -th 3D CT image as and , respectively, where and . Here, indicates the -th instance of the -th image. Then, these 2D patches (size: ) are fed into the encoding module for patch/instance-level feature extraction. These instance-level features are further fed into an embedding-level MIL module, which will be introduced later. After obtaining the instance-level features, the bag/image-level feature are then generated by our proposed global contrast pooling (GCP) layer in the image-level MIL module.
As illustrated in Fig. 4, the proposed GCP layer aims to make the instance features closer to the relevant concepts, and also push those irrelevant instance features and concepts away from each other. In this work, the term ‘concept’ denotes the to-be-learned feature of GCP layer that is discriminative for severity assessment. Theoretically, the concept is a normalized weight to map features in instance feature space to an ordered embedding space. Specifically, in the GCP layer, we assume the bag-level feature is represented by the relationship between instance features and concepts. Here, these concepts are learned to reveal the data structure in a global perspective. The bag-level feature is then denoted as a dimensional feature vector, with each dimension denoting the maximum similarity between one concept and all instance features. We use the cosine function to measure such relationships. Thus, the bag feature and the similarity can be written as
(1) |
(2) |
where () is the maximum similarity between the instance features of the -th bag and the -th image-level concept . denotes the commonly used regularization term used in deep networks. With Eqs. (1)-(2), one can observe that the proposed GCP layer can automatically learn the concepts that are related to those discriminative instances, thus reducing the influence of those irrelevant instances. Note such a GCP layer can be also used in other weakly supervised problems, where only a small portion of regions in an image are related to the task at hand (such as MRI-based brain disorder diagnosis [30]).
We further use an embedding-level MIL module (with a GCP layer) to learn embedding-level representations, by regarding each image patch as a bag and the intermediate patch-level features produced by the encoder as instances. In this way, the relationships among small regions in each patch can be modeled by our method. Based on the embedding-level features, an image-level MIL module (with a GCP layer) is further used to generate the volume features. Based on the volume features, we use a fully-connected layer followed by a cross-entropy loss to predict the severity score (i.e., severe or non-severe) of each input CT image. The final loss function in the proposed hierarchical MIL network for severity assessment can be formulated as
(3) |
where denotes the mapping function of the fully-connected layer, and denotes the severity type confirmed by clinicians.
III-E Multi-task Learning for Joint Lung Segmentation and Severity Assessment
The segmentation task is supervised by the aggregation of cross-entropy loss and Dice loss as follows
(4) |
where and denote the predicted and ground-truth segmentation masks for the -th patch in the -th category. In this work, we segment categories, including five parts of lung lobes and the background. It is worth noting that most of the cases in our dataset do not have ground-truth segmentation masks. For these cases, we simply avoid calculating the segmentation loss for them.
Finally, the losses in Eqs. 3 and 4 are trained simultaneously in a multi-task learning manner, and the overall loss of the proposed method is written as
(5) |
where is the trade-off parameter used to balance the contributions of these two tasks. In this work, is empirically set to .
Method | Accuracy | Precision | Recall | F1 Score | AUC |
---|---|---|---|---|---|
ResNet50+Max [34] | 0.924 0.497 | 0.856 0.154 | 0.793 0.106 | 0.816 0.120 | 0.803 0.090 |
Gated Att. MIL [35] | 0.955 0.015 | 0.879 0.054 | 0.946 0.019 | 0.906 0.037 | 0.973 0.024 |
Tang et al . [4]* | 0.875 | - | 0.933 | - | 0.910 |
Yang et al . [36] | - | - | 0.750 | - | 0.892 |
Cls. Only (Ours) | 0.969 0.023 | 0.928 0.073 | 0.958 0.031 | 0.938 0.045 | 0.980 0.013 |
M2UNet (Ours) | 0.985 0.005 | 0.975 0.022 | 0.952 0.011 | 0.963 0.011 | 0.991 0.010 |
Method | # (MB) | DSC | SEN | PPV |
---|---|---|---|---|
U-Net | 0.776 0.050 | 0.759 0.037 | 0.834 0.033 | |
U-Net++ | 0.784 0.035 | 0.773 0.038 | 0.821 0.018 | |
Seg. Only (Ours) | 0.759 0.055 | 0.756 0.064 | 0.785 0.045 | |
M2UNet (Ours) | 0.785 0.058 | 0.783 0.059 | 0.799 0.051 |
III-F Implementation
The proposed method is implemented based on the open-source deep learning library Pytorch. The training of the network is accelerated by four NVidia Tesla V100 GPUs (each with 32 GB memory). For feasible learning of the lung region images, we clamp the intensities of the image into , which indicates that we use the width of and the level of for the pulmonary window. Then, the data is normalized to the value of , as used by other deep learning methods. The dataset is highly imbalanced as the number of severe patients is much fewer than the non-severe patient. The ratio of the severe patient is less than in our dataset. Therefore, we augmented the data by directly duplicated the severe cases in the training set. This can be done because the proposed method uses a random cropping strategy to construct the inputs. This makes the duplicated cases not the same to each other for the training of the network. We also use the random cropping strategy in the testing stage, by assuming that the data distribution is already well learned in training. Other cropping strategies, e.g., center cropping, may not be suitable here, as the center of pulmonary is dominated by the trachea and other tissues.
In both training and testing stage, we randomly crop image patches from each input 3D CT image to construct the image-level bag (i.e., with the bag size of ). And we use the output of the encoder to construct the embedding-level bag that contains feature maps. We train M2UNet using the learning rate of with a decay strategy of ‘Poly’ (with the power of ). The network is optimized by a standard Stochastic Gradient Descent (SGD) algorithm with epochs. And the weights are decayed by a rate of with the momentum of .
IV Experiments
In this section, we first introduce the materials, competing methods, and experimental setup. We then present the experimental results achieved by our method and several state-of-the-art methods. We finally investigate the influence of two major strategies used in our method. More results on parameter analysis can be found in the Supplementary Materials.
IV-A Materials
The real COVID-19 dataset contains a total of 3D chest CT scans acquired from patients who are confirmed with COVID-19 (i.e., RT-PCR Test Positive). These CT images are collected from seven hospitals with a variety of CT scanners, including Philips (Ingenuity CT iDOSE4), GE (Bright speed S), Siemens (Somatom perspective), Hitachi (ECLOS), and Anke (ANATOM 16HD). The images are of large variation in terms of the image size of , and the spatial resolution of , and . Obviously, diagnosis based on these images is a very challenging task. The severity of the patient is confirmed by clinicians, following the guideline of 2019-nCoV (trail version 7) issued by the China National Health Commission. The severity of the patient is categorized into four types, i.e., mild, moderate, severe, and critical. In clinical practice, patients are often divided into two groups with different treatment regimens, i.e., severe and non-severe. The segmentation of 152 out of 666 images were delineated by an AI-based software and confirmed by experienced radiologists. In this work, we employ this partitioning strategy. That is, mild and moderate are treated as non-severe, while severe and critical are regarded as severe. Therefore, the task of severe assessment is formulated into a binary classification problem. Therefore, the dataset is partitioned into severe and non-severe patients.
IV-B Competing Methods
We first compare the proposed M2UNet with four state-of-the-art methods in [4, 36, 34, 35] for severity assessment of COVID-19 patients. The first two methods [4, 36] are both based on hand-crafted features of CT images, while the last two [34, 35] are deep learning-based methods that can learn imaging features automatically from data. Specifically, Tang et al . [4] first segment the lung, lung lobe and lesions in CT images. Then, the quantitative features of COVID-19 patients, e.g., the infection volume and the ratio of the whole lung, are calculated based on the segmentation results. The prediction is done by a random forest method. Yang et al . [36] propose to aggregate infection scores calculated on 20 lung regions for severity assessment. Besides, the ResNet50+Max method [34] is also compared for patch-wise classification. ResNet50+Max is a non-MIL method, which has a ResNet-50 network architecture and performs image-level classification through max-voting. In addition, we apply the Gated Att. MIL method proposed in [35] on our dataset, which is a one-stage MIL method with an attention mechanism. For fair comparison, this method shares the same multi-instance pool as our M2UNet.
We further compare our method with two state-of-the-art methods for lung lobe segmentation, including 1) UNet [13], and 2) UNet++ [17]. The parameter settings for these five competing methods are the same as those in their respective papers.
To evaluate the influence of the proposed multi-task learning and hierarchical MIL strategies used in M2UNet, we further compare M2UNet with its two variants: 1) M2UNet with only the classification sub-network (denoted as Clas. Only), 2) M2UNet with only the segmentation sub-network (denoted as Seg. Only).

Method | Accuracy | Precision | Recall | F1 Score | AUC |
---|---|---|---|---|---|
ResNet50+Max [34] | 0.924 0.497 | 0.856 0.154 | 0.793 0.106 | 0.816 0.120 | 0.803 0.090 |
Gated Att. MIL [35] | 0.955 0.015 | 0.879 0.054 | 0.946 0.019 | 0.906 0.037 | 0.973 0.024 |
Cls. Only (Ours) | 0.969 0.023 | 0.928 0.073 | 0.958 0.031 | 0.938 0.045 | 0.980 0.013 |
Method | Accuracy | Precision | Recall | F1 Score | AUC |
---|---|---|---|---|---|
Cls. Only | 0.969 0.023 | 0.928 0.073 | 0.958 0.031 | 0.938 0.045 | 0.980 0.013 |
M2UNet | 0.985 0.005 | 0.975 0.022 | 0.952 0.011 | 0.963 0.011 | 0.991 0.010 |
IV-C Experimental Setup
A five-fold cross-validation (CV) strategy is used in the experiments for performance evaluation. Specifically, the whole dataset is first randomly partitioned into five subsets (with approximately equal sample size of subjects). We treat one subset as the testing set (), while the remaining four subsets are combined to construct the training set () and validation set (). The validation set here is used for selecting the hyper-parameters. This process is iterated until each subsets serve as a testing set once. The final results are reported on the test set.
Two tasks are included in the proposed method, i.e., classification of severity assessment, and 2) segmentation of the lung lobe. For performance evaluation, two sets of metrics are used in these two tasks, with the details given below.
IV-C1 Metrics for Classification
We use five commonly used metrics to evaluate the classification performance achieved by different methods in the severity assessment task, i.e., Accuracy, Precision, Recall, F1 Score, and the area under the receiver operating characteristic curve (AUC).
(6) |
(7) |
(8) |
where TP, TN, FP and FN denote true positive, true negative, false positive and false negative, respectively.
IV-C2 Metrics for Segmentation
We use three metrics, i.e., Dice Similarity Coefficient (DSC), Positive Predict Value (PPV) and Sensitivity (SEN), to evaluate the segmentation performance of different methods, with the definitions given below.
(9) |
(10) |
where and represent the ground-truth and predicted segmentation maps for each scan, respectively.

IV-D Comparison with State-of-the-art Methods
IV-D1 Results of Severity Assessment
We first report the performance of six different methods in the task of severity assessment for COVID-19 patients, with the results shown in Table II. Note that the results from the competing methods are directly obtained from the respective papers. As can be seen, four deep learning-based methods usually outperform two hand-crafted feature-based methods in most cases. For some specific metrics, the method in [4] achieves the Recall of , which is significantly better than the non-MIL method ResNet50+Max. The conventional MIL-based method in [35] gets a performance improvement in terms of accuracy by . Three MIL methods (i.e., [35], Cls. Only, and M2UNet) yield satisfying performance, and the proposed M2UNet achieves the best results (e.g.., the accuracy of and F1 Score of ). However, the proposed method with multiple instances in multi-instance learning achieves the accuracy of and F1 Score of . The receiver operating characteristic (ROC) curves of six competing methods are illustrated in Fig. 6. Note that this ROC curve is plotted based on the results on one fold testing data, which is slightly different from the average performance on five-folds in Table II. Table II and Fig. 6 clearly suggest that our M2UNet generates the overall best performance in the task of severity assessment of COVID-19 patients based on chest CT images.
IV-D2 Results of Lung Lobe Segmentation
We then report the results of lung lobe segmentation achieved by four different methods in Table III. Comparing Seg. Only and the conventional U-Net, the former dramatically reduces the parameter from to . As a consequence, the performance in terms of DSC and PPV is also decreased by and , respectively. By using multi-task learning, M2Net improves the performance, from to in terms of DSC, which also outperform the performance of conventional U-Net, with a decreasing of parameters, from to . The proposed M2UNet also achieves a slightly higher performance compared with U-Net++.
The visualization of segmentation results achieved by three different methods on two subjects is shown in Fig. 5. From this figure, we can see that M2UNet generates the overall best segmentation masks, while U-Net and U-Net++ usually yield under-segmentation results on these cases. These results further show the advantage of our M2UNet.
Bag Size | Accurate | Precision | Recall | F1 Score |
---|---|---|---|---|
0.933 0.045 | 0.845 0.072 | 0.936 0.043 | 0.878 0.060 | |
0.940 0.030 | 0.844 0.069 | 0.948 0.037 | 0.882 0.060 | |
0.965 0.032 | 0.914 0.061 | 0.959 0.042 | 0.931 0.048 | |
0.971 0.019 | 0.929 0.044 | 0.946 0.036 | 0.935 0.029 | |
0.969 0.023 | 0.928 0.073 | 0.958 0.031 | 0.938 0.045 |
IV-E Ablation Study
We further evaluate the influence of two major strategies used in our M2UNet, i.e., 1) the hierarchical MIL strategy for classification, and 2) the multi-task learning strategy for joint severity assessment and lung lobe segmentation.
IV-E1 Influence of Hierarchical MIL Strategy
To evaluate the effectiveness of the hierarchical MIL strategy, we compare the variant of the proposed M2UNet (i.e., Cls. Only without the segmentation subnetwork) with a non-MIL method (i.e., ResNet50+Max) and a one-stage MIL method (i.e., Gated Att. MIL [35]). The classification results of these three methods in the task of severity assessment are reported in Table IV. As shown in Table IV, two MIL methods (i.e., Gated Att. MIL and Cls. Only) can generate more accurate decisions under the weakly supervised setting, compared with the non-MIL method ResNet50+Max. Besides, our hierarchical MIL strategy can further boost the classification performance compared to the conventional one-stage MIL strategy. For instance, our Cls. Only method achieves an F1 Score of , which is higher than that (i.e., ) yielded by Gated Att. MIL with a one-stage MIL strategy. These results suggest the effectiveness of the proposed hierarchical MIL strategy.
IV-E2 Influence of Multi-task Learning Strategy
Our M2UNet can jointly learn the segmentation task and the classification task in a multi-task learning manner. Here, we also investigate the influence of such a multi-task learning paradigm, by comparing M2UNet with its two single-task variants, i.e., ‘Cls. Only’ for classification and ‘Seg. Only’ for segmentation. The performance comparison in two tasks for severity assessment and lung lobe segmentation are reported in Table V and Table VII, respectively. Table V suggests that, compared with Cls. Only, the multi-task learning paradigm used in M2UNet helps to improve the classification accuracy by , while increasing the precision score by over and the F1 Score by . Notably, the F1 and precision of the Cls. Only method are already higher than , which are hard to be improved. This is more valuable in this classification scheme, as the F1 score is more representative in evaluating such an imbalanced classification task.
As can be observed from Table VII, although M2UNet is not specifically designed for lung lobe segmentation, it still improves the overall segmentation performance in terms of three metrics, compared with its single-task variant (i.e., Seg. Only). This implies that the proposed multi-task learning strategy is useful in boosting the learning performance of both tasks of severity assessment and lung lobe segmentation.
Method | DSC | SEN | PPV |
---|---|---|---|
Seg. Only | 0.759 0.055 | 0.756 0.064 | 0.785 0.045 |
M2UNet | 0.785 0.058 | 0.783 0.059 | 0.799 0.051 |
IV-F Influence of Bag Size
We further investigate the performance of our method using different size of bags, and the results are shown in Table VI. Specifically, we vary the bag size within . As shown in the table, the performance of M2UNet for classification changes along with the bag size. The proposed method achieves the best performance with the bag size of . Another observation is that, the table suggests the performance of the proposed AI-based severity assessment model is not sensitive with the bag size when larger than , indicating that at least patches are required for the proposed method to achieve an acceptable result.
V Conclusion and Future Work
In this paper, we propose a synergistic learning framework for automated severity assessment and lung segmentation of COVID-19 in 3D chest CT images. In this framework, we first represent each input image by a bag to deal with the challenging problem that the severity is related to local infected regions in the CT image. We further develop a multi-task multi-instance deep network (called M2UNet) to assess the severity of COVID-19 patients and segment the lung lobe simultaneously, where the context information provided by segmentation can be employed to boost the performance of severity assessment. A hierarchical multi-instance learning strategy is also proposed in M2UNet for severity assessment. Experimental results on a real COVID-19 CT image dataset demonstrate that our method achieves promising results in severity assessment of COVID-19 patients, compared with several state-of-the-art methods.
In the current work, the severity assessment of COVID-19 only relies on one time-point data, without considering longitudinal imaging biomarkers. It is interesting to perform a longitudinal study to investigate the progression of the disease, which is one of our future work. Since annotations for lung lobe in 3D CT images are usually tedious and error-prone, only a small number of subjects in our dataset have ground-truth segmentation. Therefore, it is highly desired to develop automated or even semi-automated image annotation methods, which will also be studied in the future.
References
- [1] Worldometers, “COVID-19 CORONAVIRUS PANDEMIC,” https://www.worldometers.info/coronavirus/, 2020, [Online].
- [2] J. Chen, L. Wu, J. Zhang, L. Zhang, D. Gong, Y. Zhao, S. Hu, Y. Wang, X. Hu, B. Zheng et al., “Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography: A prospective study,” medRxiv, 2020.
- [3] Y. Song, S. Zheng, L. Li, X. Zhang, X. Zhang, Z. Huang, J. Chen, H. Zhao, Y. Jie, R. Wang et al., “Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images,” medRxiv, 2020.
- [4] Z. Tang, W. Zhao, X. Xie, Z. Zhong, F. Shi, J. Liu, and D. Shen, “Severity assessment of coronavirus disease 2019 (COVID-19) using quantitative features from chest CT images,” arXiv preprint arXiv:2003.11988, 2020.
- [5] F. Shan, Y. Gao, J. Wang, W. Shi, N. Shi, M. Han, Z. Xue, D. Shen, and Y. Shi, “Lung infection quantification of COVID-19 in CT images with deep learning,” arXiv preprint arXiv:2003.04655, 2020.
- [6] S. Jin, B. Wang, H. Xu, C. Luo, L. Wei, W. Zhao, X. Hou, W. Ma, Z. Xu, Z. Zheng et al., “Ai-assisted CT imaging analysis for COVID-19 screening: Building and deploying a medical AI system in four weeks,” medRxiv, 2020.
- [7] X. Qi, Z. Jiang, Q. Yu, C. Shao, H. Zhang, H. Yue, B. Ma, Y. Wang, C. Liu, X. Meng et al., “Machine learning-based CT radiomics model for predicting hospital stay in patients with pneumonia associated with SARS-CoV-2 infection: A multicenter study,” medRxiv, 2020.
- [8] F. Shi, J. Wang, J. Shi, Z. Wu, Q. Wang, Z. Tang, K. He, Y. Shi, and D. Shen, “Review of artificial intelligence techniques in imaging data acquisition, segmentation and diagnosis for COVID-19,” CoRR, vol. abs/2004.02731, 2020. [Online]. Available: https://arxiv.org/abs/2004.02731
- [9] R. Verity, L. C. Okell, I. Dorigatti, P. Winskill, C. Whittaker, N. Imai, G. Cuomo-Dannenburg, H. Thompson, P. G. Walker, H. Fu et al., “Estimates of the severity of coronavirus disease 2019: a model-based analysis,” The Lancet Infectious Diseases, 2020.
- [10] X. Xie, Z. Zhong, W. Zhao, C. Zheng, F. Wang, and J. Liu, “Chest ct for typical 2019-ncov pneumonia: relationship to negative rt-pcr testing,” Radiology, p. 200343, 2020.
- [11] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, “Learning deep features for discriminative localization,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2921–2929.
- [12] K. He, J. Huo, Y. Shi, Y. Gao, and D. Shen, “Midcn: A multiple instance deep convolutional network for image classification,” in Pacific Rim International Conference on Artificial Intelligence. Springer, 2019, pp. 230–243.
- [13] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2015, pp. 234–241.
- [14] C. Zheng, X. Deng, Q. Fu, Q. Zhou, J. Feng, H. Ma, W. Liu, and X. Wang, “Deep learning-based detection for COVID-19 from chest CT using weak label,” medRxiv, 2020.
- [15] Y. Cao, Z. Xu, J. Feng, C. Jin, X. Han, H. Wu, and H. Shi, “Longitudinal assessment of COVID-19 using a deep learning–based quantitative CT pipeline: Illustration of two cases,” Radiology: Cardiothoracic Imaging, vol. 2, no. 2, p. e200082, 2020.
- [16] L. Huang, R. Han, T. Ai, P. Yu, H. Kang, Q. Tao, and L. Xia, “Serial quantitative chest CT assessment of COVID-19: Deep-learning approach,” Radiology: Cardiothoracic Imaging, vol. 2, no. 2, p. e200075, 2020.
- [17] Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: A nested u-net architecture for medical image segmentation,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. Springer, 2018, pp. 3–11.
- [18] F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 2016 Fourth International Conference on 3D Vision (3DV). IEEE, 2016, pp. 565–571.
- [19] H. Y. F. Wong, H. Y. S. Lam, A. H.-T. Fong, S. T. Leung, T. W.-Y. Chin, C. S. Y. Lo, M. M.-S. Lui, J. C. Y. Lee, K. W.-H. Chiu, T. Chung et al., “Frequency and distribution of chest radiographic findings in covid-19 positive patients,” Radiology, p. 201160, 2020.
- [20] T. G. Dietterich, R. H. Lathrop, and T. Lozano-Pérez, “Solving the multiple instance problem with axis-parallel rectangles,” Artificial Intelligence, vol. 89, no. 1-2, pp. 31–71, 1997.
- [21] K. Xu, J. Ba, R. Kiros, K. Cho, A. Courville, R. Salakhudinov, R. Zemel, and Y. Bengio, “Show, attend and tell: Neural image caption generation with visual attention,” in International Conference on Machine Learning, 2015, pp. 2048–2057.
- [22] M. Liu, J. Zhang, E. Adeli, and D. Shen, “Joint classification and regression via deep multi-task multi-channel learning for Alzheimer’s disease diagnosis,” IEEE Transactions on Biomedical Engineering, vol. 66, no. 5, pp. 1195–1206, 2018.
- [23] M. Oquab, L. Bottou, I. Laptev, J. Sivic et al., “Weakly supervised object recognition with convolutional neural networks,” in NIPS, 2014, pp. 1545–5963.
- [24] J. Feng and Z.-H. Zhou, “Deep miml network,” in Thirty-First AAAI Conference on Artificial Intelligence, 2017.
- [25] M. Sun, T. X. Han, M.-C. Liu, and A. Khodayari-Rostamabad, “Multiple instance learning convolutional neural networks for object recognition,” in 2016 23rd International Conference on Pattern Recognition (ICPR). IEEE, 2016, pp. 3270–3275.
- [26] J. Wu, Y. Yu, C. Huang, and K. Yu, “Deep multiple instance learning for image classification and auto-annotation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3460–3469.
- [27] T. Tong, R. Wolz, Q. Gao, R. Guerrero, J. V. Hajnal, and D. Rueckert, “Multiple instance learning for classification of dementia in brain MRI,” Medical Image Analysis, vol. 18, no. 5, pp. 808–818, 2014.
- [28] Y. Xu, J.-Y. Zhu, I. Eric, C. Chang, M. Lai, and Z. Tu, “Weakly supervised histopathology cancer image segmentation and classification,” Medical Image Analysis, vol. 18, no. 3, pp. 591–604, 2014.
- [29] Z. Yan, Y. Zhan, Z. Peng, S. Liao, Y. Shinagawa, S. Zhang, D. N. Metaxas, and X. S. Zhou, “Multi-instance deep learning: Discover discriminative local anatomies for bodypart recognition,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1332–1343, 2016.
- [30] M. Liu, J. Zhang, E. Adeli, and D. Shen, “Landmark-based deep multi-instance learning for brain disease diagnosis,” Medical Image Analysis, vol. 43, pp. 157–168, 2018.
- [31] K. He, X. Cao, Y. Shi, D. Nie, Y. Gao, and D. Shen, “Pelvic organ segmentation using distinctive curve guided fully convolutional networks,” IEEE Transactions on Medical Imaging, vol. 38, no. 2, pp. 585–595, 2018.
- [32] C. Lian, M. Liu, J. Zhang, and D. Shen, “Hierarchical fully convolutional network for joint atrophy localization and Alzheimer’s disease diagnosis using structural MRI,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.
- [33] S. Bakas, M. Reyes, A. Jakab, S. Bauer, M. Rempfler, A. Crimi, R. T. Shinohara, C. Berger, S. M. Ha, M. Rozycki et al., “Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge,” arXiv preprint arXiv:1811.02629, 2018.
- [34] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4, inception-resnet and the impact of residual connections on learning,” in Thirty-first AAAI Conference on Artificial Intelligence, 2017.
- [35] M. Ilse, J. M. Tomczak, and M. Welling, “Attention-based deep multiple instance learning,” CoRR, vol. abs/1802.04712, 2018. [Online]. Available: http://arxiv.org/abs/1802.04712
- [36] R. Yang, X. Li, H. Liu, Y. Zhen, X. Zhang, Q. Xiong, Y. Luo, C. Gao, and W. Zeng, “Chest CT severity score: An imaging tool for assessing severe COVID-19,” Radiology: Cardiothoracic Imaging, vol. 2, no. 2, p. e200047, 2020.