Next Article in Journal
Controlling a Mouse Pointer with a Single-Channel EEG Sensor
Next Article in Special Issue
A CNN Deep Local and Global ASD Classification Approach with Continuous Wavelet Transform Using Task-Based FMRI
Previous Article in Journal
Quantifying Uncertainty in Pulsed Thermographic Inspection by Analysing the Thermal Diffusivity Measurements of Metals and Composites
Previous Article in Special Issue
An Automated CAD System for Accurate Grading of Uveitis Using Optical Coherence Tomography Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Precise Segmentation of COVID-19 Infected Lung from CT Images Based on Adaptive First-Order Appearance Model with Morphological/Anatomical Constraints

1
BioImaging Laboratory, Department of Bioengineering, University of Louisville, Louisville, KY 40292, USA
2
College of Computer and Information Science, Princess Nourah Bint Abdulrahman University, Riyadh 11564, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(16), 5482; https://doi.org/10.3390/s21165482
Submission received: 11 June 2021 / Revised: 8 August 2021 / Accepted: 10 August 2021 / Published: 14 August 2021
(This article belongs to the Special Issue Computer Aided Diagnosis Sensors)

Abstract

:
A new segmentation technique is introduced for delineating the lung region in 3D computed tomography (CT) images. To accurately model the distribution of Hounsfield scale values within both chest and lung regions, a new probabilistic model is developed that depends on a linear combination of Gaussian (LCG). Moreover, we modified the conventional expectation-maximization (EM) algorithm to be run in a sequential way to estimate both the dominant Gaussian components (one for the lung region and one for the chest region) and the subdominant Gaussian components, which are used to refine the final estimated joint density. To estimate the marginal density from the mixed density, a modified k-means clustering approach is employed to classify the Gaussian subdominant components to determine which components belong properly to a lung and which components belong to a chest. The initial segmentation, based on the LCG-model, is then refined by the imposition of 3D morphological constraints based on a 3D Markov–Gibbs random field (MGRF) with analytically estimated potentials. The proposed approach was tested on CT data from 32 coronavirus disease 2019 (COVID-19) patients. Segmentation quality was quantitatively evaluated using four metrics: Dice similarity coefficient (DSC), overlap coefficient, 95th-percentile bidirectional Hausdorff distance (BHD), and absolute lung volume difference (ALVD), and it achieved 95.67 ± 1.83 % , 91.76 ± 3.29 % , 4.86 ± 5.01 , and 2.93 ± 2.39 , respectively. The reported results showed the capability of the proposed approach to accurately segment healthy lung tissues in addition to pathological lung tissues caused by COVID-19, outperforming four current, state-of-the-art deep learning-based lung segmentation approaches.

1. Introduction

Pulmonary diseases are serious public heath threats that may happen after having inflammation or fluid accumulation in the lung, causing a respiratory failure, such as coronavirus disease 2019 (COVID-19). The primary reason for COVID-19 death is acute respiratory distress syndrome (ARDS) [1]. According to Gupta et al. [2], 83.9 % of the COVID-19 patients in their study needed a mechanical ventilation support, of whom 87.95 % had ARDS. Therefore, detection and diagnosis of COVID-19 grades is vital to prioritize patient’s need for ventilator support. The accuracy attainable by computer-aided diagnostic (CAD) system using lung imaging data for COVID-19 depends on how accurate the segmentation is. Accurate lung segmentation is a challenging task as different pathologies affect the appearance of the lung, and if the infected regions are missed during the segmentation, it will affect the entire task. Therefore, this paper focuses on developing an automatic system to detect and segment the lungs in chest computed tomography (CT), which is one of the popular noninvasive clinical modalities used by physicians to diagnose lung pathologies.
In the last few years, many preliminary studies have been conducted to detect and segment lung as well as pathological lesions. Some of these studies [3,4,5,6,7,8,9] proposed threshold-based approaches for lung segmentation, which performed well on normal CT scans but failed in pathological cases, especially severe cases, whereas lungs in the normal CT scan can be discriminated easily from background due to huge differences in attenuation [10]. Therefore, to overcome this problem, more recent studies employed texture, shapes, deep learning, or hybrid techniques to accurately segment normal and different lung pathologies. These studies are briefly discussed below.
In [11,12,13,14], authors considered texture analysis, shape analysis, or both of them in their system to discriminate between objects. A recent study by Oulefki et al. [15] proposed a system to automatically segment COVID-19 lung infected region by applying a multilevel entropy-based threshold approach, namely a modified Kapur method. Their system achieved a sensitivity, specificity, Dice similarity coefficient (DSC), and precision of 73.3 % , 99.4 % , 71.4 % , and  73.9 % , respectively. Another study by Korfiatis et al. [16] employed k-means clustering to partition CT voxels into four classes: lung, muscle, fat, and bone based on intensity values. After that, the initial lung region was extracted by applying a filling operation. Finally, a support vector machine (SVM) was used to determine the final border of the lung based on intensity and wavelet-based descriptors. In [17], authors proposed a segmentation system by eliminating unwanted regions and segmenting lung initially using a threshold approach. Moreover, a 3D gray-level co-occurrence matrix (GLCM) was constructed for a window of size 15 × 15 × 15 centered on each voxel. Then, predefined features were extracted from the GLCM, and a new image was constructed, being the product of the entropy and the inverse difference moment of the GLCM. Subsequently, the abnormal regions were identified from the constructed image using a threshold approach. Finally, the later and initial segmentation were merged together to determine the final segmentation. Dehmeshki et al. [18] used a genetic algorithm (GA) to construct a system to identify spherical nodules within CT images. First, the lung was segmented using adaptive thresholding. Then, the authors utilized a geometric feature, namely, volumetric shape index (VSI), for the segmented lung as a weighted factor in the fitness function of GA. VSI of a spherical object is 1, while that of a cylindrical object is 0.75 , so the values of fitness function for nodules were higher than for blood vessels. Convergence criteria of GA to select the shape as a nodule was a threshold-based. The detection rate of their system was approximately 90 % with a 14.6 false positive per scan. Moreover, Nakagomi [19] presented a min-cut graph segmentation algorithm based on multiple shapes and prior information of neighbors structure to detect and segment lung infected by pleural effusion. In [20], authors presented a lung segmentation system for different lung pathologies. Their system first determined two seed points within both lungs using a thresholding approach, then a fuzzy connectedness (FC) algorithm was used to extract the lung. Furthermore, multiple refinement stages based on machine learning classification and neighboring anatomy-guided learning mechanisms were included in their system to detect pathological regions during FC segmentation. A recent study by Houssein et al. [21] developed a segmentation system that employed a heuristic method, called manta ray foraging optimization (MRFO), based on an opposition-based learning (OBL), using Otsu’s method as a fitness function, to get the best threshold values using COVID-19 CT images. More information about texture- and shape-based lung segmentation can be found in [22].
Recently, deep learning approaches have been employed to segment normal as well as pathological lung caused by COVID-19. For example, Saood et al. [23] investigated two deep learning approaches to semantically segment infected/non-infected lung using CT images. These included SegNet [24] and U-Net [25] networks. The author employed these networks for binary and multi-class classification. They conducted multiple experiments with different hyperparameters. The best reported results for the binary (multi-class) classification gave accuracy of 95.4 ± 2.9 % ( 90.7 ± 6 % ) and 94.9 ± 4.3 % ( 90.8 ± 6.5 % ) using SegNet and U-Net, respectively. A similar study [26] proposed a segmentation system using a convolution neural network (CNN). Their network employed feature variation block to enhance the efficiency of feature representation as well as progressive atrous spatial pyramid pooling to deal with appearance and shape differences caused by sophisticated infection. The DSC (sensitivity, specificity) of their system was 72.6 % ( 75.1 % , 72.6 % ) and 98.7 % ( 98.6 % , 99 % ) for COVID-19 infections and normal lung CT images, respectively. A multi-task deep learning-based system was implemented by Amyar et al. [27]. This study included reconstruction for better feature representation; segmentation to extract lesion regions; and classification to categorize the scan into normal, COVID-19, and other diseases. Their system employed encoder-decoder architecture based on U-Net network which used a common encoder for the three tasks. The best reported DSC of their segmentation task was 88 % . Recent study by Fan et al. [28] developed a binary and multi-class segmentation system using CT chest images, called Inf-Net. This system was mainly based on a deep learning. Moreover, to compensate the limited number of labeled images, they included a random sampling-based semi-supervised learning, namely, Semi-Inf-Net. Their system employed edge attention as well as reverse attention to improve the feature representation by modeling lung boundaries. In addition, high-level features were exploited by their network and combined by a parallel partial decoder. The performance of their infection segmentation system achieved a DSC of 68.2 % and 73.9 % , sensitivity of 69.2 % and 72.5 % , and specificity of 94.3 % and 96 % using Inf-Net and Semi-Inf-Net, respectively. A similar study [29] proposed an automatic deep learning-based multi-class segmentation system of COVID-19 using CT chest images. The latter exploited aggregated residual transformations in addition to soft attention mechanism to better represent the features and to increase the system’s ability to distinguish between different COVID-19 lesions. The reported DSC (accuracy, precision) of their system was 94 % ( 89 % , 95 % ) and 83 % ( 79 % , 82 % ) with and without data augmentation, respectively. Another study [30] proposed a semi-supervised deep learning-based segmentation system, called FSS-2019-nCov, to detect lesion infection in COVID-19 patients. The latter was based on encoder–decoder architecture with Res2Net [31] encoder backbone. In the proposed encoder–decoder architecture, the authors used a context enrichment module, namely, smoothed atrous convolution block and the multi-scale pyramid pooling block, to overcome any debilitation occurred in the represented knowledge generated in the encoder phase. This system was consisted of three modules: conditioner path, adaptive interaction module, and segmentation path. Conditioner path was responsible to learn feature maps from support sets which contain CT slice and its ground-truth. Subsequently, these feature maps were transmitted to the segmentation path using adaptive interaction module which was responsible for detecting lesion in the CT slice. The performance of their system achieved a DSC of 79.8 % , sensitivity of 80.3 % , and specificity of 98.6 % . In [32], the authors employed V-Net [33] to segment lung in COVID-19 CT images that was refined by a shape deformation module. A similar study by Li et al. [34] employed U-Net to segment lung on CT images. Then, they proposed a deep learning network, called COVNet, with a ResNet-50 [35] backbone to detect COVID-19 lesions. A recent study [36] developed a deep learning-based segmentation system, called LungINFseg, to detect COVID-19 lesions in CT images. This system was built on the basis of encoder–decoder architecture. The authors employed a 2D discrete wavelet transform (DWT) with four Haar filters and a receptive field aware (RFA) module in the encoder phase, which were able to change the size of receptive field, to capture more relevant features related to infected regions. Their system achieved a DSC and intersection over union (IoU) score of 80.34 % and 68.77 % , respectively. Other studies have also employed deep learning as a segmentation system with varying accuracy as reported in [36,37,38,39,40,41,42,43].
Segmentation techniques for CT data using deep learning method consider the current, state-of-the-art approaches. However, they have some drawbacks in practical applications, such as the need for huge databases to learn the different pathology of the lung regions which makes the training of such network is very high computational [44]. Moreover, segmentation approaches based on deformable models, which optimize a trade-off between smoothness of the deformable boundary and homogeneity of the region inside the boundary, suffer from high computational complexity and limited capabilities when the desired boundary has concavities or encompasses a region that is naturally inhomogeneous, such as infected lung regions. To overcome the aforementioned limitations, we are proposing an unsupervised lung segmentation approach that is based on modeling the first-order appearance model of CT data by using a probabilistic model based on a linear combination of Gaussian (LCG) that estimates dominant components, corresponding to lung and chest regions, as well as subdominant components. Subsequently, these subdominant components are clustered to one of the dominant components for marginal density estimation. This model can capture the variability in the Hounsfield distributions that may come from changing the screening protocols and severity of lung infections. Finally, we refine the lung segmentation by applying 3D morphological constraints based on the Markov–Gibbs random field (MGRF) model with analytical parameter estimations.

2. Methods

A fully automated segmentation framework is presented to extract both healthy lung tissues as well as pathological lung tissues that may be caused by COVID-19. The major steps of the framework, depicted in Figure 1, are as follows: (i) preprocessing 3D chest CT scans to identify background voxels; (ii) modeling the gray-level distribution of the CT data as a Gaussian mixture model with parameters estimated using a novel, sequential, expectation-maximization (EM)-based approach; (iii) preliminary segmentation of the lung region based on the use of a Bayes classifier; and (iv) refining the segmentation using a three-dimensional, rotation- and translation-invariant MGRF to impose morphological constraint. Below, we will describe the details of each step.

2.1. First-Order Visual Appearance Model

The ultimate goal is accurate labeling of voxels as belonging to lung tissue or background, where accuracy is defined as close agreement with “ground-truth” lung region delineated by a radiologist. The main challenge in modeling the distribution of the radiodensities (in Hounsfield units) of lung and chest tissues, i.e., the relative frequency histogram of CT voxel values, is dependent upon slice thickness and the severity of lung infection as shown in Figure 2.
To address this challenge, we will assume that the first-order visual appearance model of the CT data ( H ) can be modeled with linear combination of Gaussian distributions with K 2 components [45]. The first two components, called the dominant modes, corresponding to the lung region ( k = 1 ) and the chest region exterior to the lungs ( k = 2 ). The remaining Gaussian components k = 3 , , K are called subdominant modes. Thus, the proposed probabilistic model is
p ( h ) = w 1 φ ( h ; θ 1 ) + w 2 φ ( h ; θ 2 ) + k = 3 K w k τ ( h ; θ k ) ,
where the w k > 0 are mixing weights, and  φ is a Gaussian density with parameters θ k = ( μ k , σ k ) . In order for p ( h ) to be a density function, the weights must satisfy the constraint
k = 1 K w k = 1 .
Given the number K of Gaussian components, the  3 K parameters of Equation (1), including mixing weights W and means and variances Θ , are estimated by maximizing the log-likelihood of the empirical data
L ( W , Θ ) = h = 0 H n ( h ) log p ( h ; W , Θ ) ,
where n ( h ) is the histogram of the CT data, whose voxel values range from 0 to H. The corresponding relative frequency histogram is denoted f ( h ) = n ( h ) / N , N being the total number of voxels. To maximize the likelihood in Equation (3), we employ an iterative block relaxation process as follows.
Let τ indicate an iteration such that ( W [ τ ] , Θ [ τ ] ) are the parameter estimates on that iteration, and 
p [ τ ] ( h ) = p ( h ; W [ τ ] , Θ [ τ ] ) = k = 1 K w k [ τ ] τ ( h ; θ k [ τ ] )
is the proposed probabilistic model for the CT data. The conditional weights are estimated as follows:
π [ τ ] ( k | h ) = w k [ τ ] τ ( h ; θ k [ τ ] ) p [ τ ] ( h ) ;
This conditional probability specifies the relative contributions of voxel value h to each component at step τ . Using these variables, Equation (3) can be written in the equivalent form:
L ( W [ τ ] , Θ [ τ ] ) = h = 0 H n ( h ) log k = 1 K π [ τ ] ( k | h ) τ ( h ; θ k [ τ ] )
From given starting values at τ = 0 , the block relaxation scheme converges to a local maximum of the likelihood function in Equation (6) through iteration of the following two steps:
  • E-step [ τ + 1 ] : estimate W [ τ + 1 ] , Θ [ τ + 1 ] , which maximize L ( W , Θ ) under the fixed conditional weights of Equation (5) at step τ .
  • M-step [ τ + 1 ] : recalculate weights, which maximize L holding parameters W [ τ + 1 ] and Θ [ τ + 1 ] fixed.
The process is repeated until the changes of all the parameters become small.
The E-step maximizes the likelihood function of Equation (6) subject to the constraints Equation (2). The solution for the weights is
w k [ τ + 1 ] = h = 0 H f ( h ) π [ τ ] ( k | h )
Then, parameters of each Gaussian component are found using the ordinary (unconstrained) maximum likelihood estimates:
μ k [ τ + 1 ] = 1 w k [ τ + 1 ] h = 0 H h · f ( h ) π [ τ ] ( k | h ) ( σ k [ τ + 1 ] ) 2 = 1 w k [ τ + 1 ] h = 0 H h μ k [ τ + 1 ] 2 · f ( h ) π [ τ ] ( k | h )
We will follow Algorithm 1 to illustrate the steps for estimating the parameters of the proposed probabilistic model. The final estimated density will consist of the two dominant Gaussian components and K 2 subdominant Gaussian components. Jensen–Shannon divergence (JSD) [46] is employed in this algorithm to measure the similarity between empirical density and mixed density for use as convergence criteria to determine the number of Gaussian components. The latter is a symmetric version of a Kullback–Leibler divergence [47].
Algorithm 1: Estimation of the proposed probabilistic model parameters
Sensors 21 05482 i001
Estimation of the marginal density: The K 2 subdominant components of the final estimated model p ( h ) need to be partitioned among the two dominant modes. Each subordinate component is associated with one dominant component in order to minimize the expected misclassification rate. This is accomplished using the proposed Algorithm 2.
Algorithm 2: The proposed clustering algorithm
Sensors 21 05482 i002

2.2. MGRF-Based Morphological Constraints

To get a consistent segmentation of the lung region, we applied rotation invariant spatial constraints by using a generic Markov–Gibbs model of region maps [45]. The model, which incorporates voxel–voxel interaction effects as shown in Figure 3, has, in general, an arbitrary interaction structure and corresponding Gibbs potentials. For simplicity, we restrict the interaction neighborhood system ( N ) to the nearest 9 neighbors in the above CT slice and 9 neighbors in the below CT-slice.
To model the interactions between the CT voxels, we will assume all the interactions as the same within each region. To estimate this interaction in analytical way, let V : X × X { V eq , V ne } denote a bi-valued Gibbs potential describing pairwise interactions, where
V ( x , χ ) = V eq x = χ V ne x χ
Then, the Gibbs probability distribution (GPD) of region maps on the 3D lattice R is as follows [45]:
P ( m ) exp ( i , j , z ) R ( ξ , η , ζ ) N V ( m i , j , z , m i + ξ , j + η , z + ζ )
By modifying the derivation scheme in [45] to fit our model, the following first approximation of the maximum likelihood estimator (MLE) of the potential values for a given map m is obtained:
V eq = X 2 X 1 f ( m ) 1 X V ne = X 2 X 1 f ( m ) 1 + 1 X
where f ( m ) and f ( m ) denote the relative frequency of the equal and non-equal pairs of the labels in all the equivalent voxels pairs ( ( i , j , z ) , ( i + ξ , j + η , z + ζ ) ) : ( i , j , z ) R ; ( i + ξ , j + η , z + ζ ) R ; ( ξ , η , ζ ) N , respectively.

2.3. Joint MGRF Model and Lung Segmentation Algorithm

In order to integrate the first-order appearance model with the spatial probabilistic model that describes the morphological/anatomical constrains, we will assume that the CT data ( g ) consisting of visual appearance model and its spatial map ( m , data labels) follow the following two-level MGRF model:
P ( g , m ) = P ( m ) P ( g | m )
Here, P ( m ) is an unconditional distribution of maps that is modeled by MGRF probabilistic model that is demonstrated in Equation (10). P ( g | m ) is a conditional distribution of gray levels for a given labeling. The Bayesian maximum a posteriori estimation (MAP) of the labeling, given the image g , m * = argmax m L ( g , m ) maximizes the log-likelihood,
L ( g , m ) = log P ( g | m ) + log P ( m ) .
In order to summarize the proposed segmentation system, the basic steps are demonstrated in Algorithm 3.
Algorithm 3: Lung Extraction Algorithm
input: A test 3D CT image.
    output: Final 3D lung segmentation.
  • 1st Order Density Estimation: Estimate the marginal density function for lung ( p l ( h ) ) and marginal density function for chest ( p c ( h ) ).
  • Initial Segmentation/Labeling: Use Bayes classifier to delineate the initial lung region by using the marginal estimated densities.
  • Estimation of Gibbs Potentials: Applying Equation (11) on the initial segmentation to estimate the Gibbs potentials.
  • Refine Segmentation: Use iterative conditional mode (ICM) algorithm [45] to find the map that maximize the likelihood of joint MGRF model shown in Equation (13).

3. Evaluation Metrics

This section describes the metrics used to gauge the performance of our proposed system: Dice similarity coefficient (DSC), overlap coefficient, and absolute lung volume difference (ALVD). Each of these quantifies in some way either the agreement or dissimilarity between the segmentation algorithm result and the corresponding ground-truth segmentation. More detailed explanation is presented in Section 3.1, Section 3.2 and Section 3.3, respectively. Furthermore, a fourth metric, the 95th-percentile bidirectional Hausdorff distance (BHD) (Section 3.4), is employed to quantify the accuracy of the boundary of the segmented region relative to ground-truth.

3.1. Dice Similarity Coefficient (DSC)

Dice similarity coefficient (DSC) is one of the most common similarity metric to measure the similarity between two different areas. This metric is used to evaluate the result of the proposed system by estimating the similarity between the black-white segmented lung (L) and the ground-truth (G), i.e., the percentage of common region (i.e., the green part) in both images as shown in Figure 4a. The range of this metric is between 0 and 1, as 0 and 1 mean dissimilar and similar, respectively. It is computed as follows:
D S C = 2 × n L G n L + n G
where n L G is the cardinality of white pixels in the intersection between the segmented lung (L) and the ground-truth (G), while n L and n G are the cardinality of the white pixels in the segmentation (L) and the ground-truth (G), respectively.

3.2. Overlap Coefficient

Overlap coefficient is used in our assessment pipeline to measure the similarity between the predicted object and its ground-truth by computing the overlap percentage between them, see Figure 4c. The overlap coefficient of identical objects gives 1, while it gives 0 for heterogeneous one. The latter is estimated as follows:
o v e r l a p = n L G n L G
where n L G is the cardinality of white pixels in the union between the segmented lung (L) and the ground-truth (G).

3.3. Absolute Lung Volume Difference (ALVD)

Another metric used to assess our work is absolute lung volume difference (ALVD). ALVD computes the similarity between two images by measuring the differences between the ground-truth (G) and the black-white segmented lung (L) (Figure 4d). The ALVD of similar objects gives 0. This metric is defined as
A L V D = n G n L n G
where n G n L is the absolute difference between the cardinality of white pixels in the ground-truth (G) and segmentation (L).

3.4. Bidirectional Hausdorff Distance (BHD)

This section describes the last metric called bidirectional Hausdorff distance (BHD), which is used to evaluate our proposed system in addition to the previous three metrics. BHD is the bidirectional estimation of Hausdorff distance (HD) between the black-white segmented lung (L) and the ground-truth (G), and vice versa. HD is the maximum Euclidean distance between the points in the border of the black-white segmented lung (L) and its closest point in the border of the ground-truth (G), as visualized in Figure 4b, which is computed as follows [48,49]:
H D ( L , G ) = max l L { min g G { d ( l , g ) } }
where l and g are sets of the points border in the L and G, respectively, and d ( g , l ) is the Euclidean distance between the two points.
As, B H D ( L , G ) is estimated as
B H D ( L , G ) = m a x { HD ( L , G ) , HD ( G , L ) }
In this paper, the 95th-percentile BHD is used to evaluate our proposed system. Instead of getting the maximum Euclidean distance between L and G, 95th-percentile of all computed distances is selected to overcome the outliers.

4. Experimental Results

The segmentation framework described above was applied to the problem of segmenting lung with pathological tissue in COVID-19 patients. The proposed segmentation system is evaluated and tested on 32 CT chest volume with different severity of COVID-19 infections, selected from 249 CT volume in COVID-19 [50]. Four of them had healthy/mild COVID-19 infections whose image size ranges from 512 × 512 × 51 to 512 × 512 × 125 , while 17 patients of size 512 × 512 × 36 607 who had moderate infections as well as 11 CT chest volume of size 512 × 512 × 44 577 had severe COVID-19 infections. To compare our framework with other approaches that depend on a training dataset, we select another 34 3D CT chest volume (i.e., 3713 images in total) from the same dataset to use them as a training. Table 1 summarized the dataset characteristics used in our experimental results. The data are graded according to the radiology protocol in [51]. To obtain more accurate segmentation, we included morphological/anatomical constraints based on the use of a rotation invariant MGRF model.
To demonstrate step by step how our proposed approach works, Figure 5b shows empirical density for the 3D CT chest volume (Figure 5a), and the two Gaussian mixtures approximating its dominant modes are presented in Figure 5c. Furthermore, Figure 5c demonstrates the JSD between the empirical density and the two estimated dominant Gaussian components. Figure 5d shows the changes of JSD and the best-estimated number of Gaussian components that are demonstrated in Figure 5e. Figure 5f demonstrated the classification of the subdominant Gaussian components based on the use of the proposed clustering algorithm (Algorithm 2). Figure 5g,h demonstrates the final marginal densities for lung and chest as well as the final estimated mixed density. Figure 6 and Figure 7 demonstrate the ability of the proposed probabilistic model to handle the variability in the empirical density that it may occur due to the severity of infections or the variability of the scanning protocol.
To highlight the promise of including the rotation invariant MGRF-based morphological/anatomical constraints with the adaptive first-order appearance model, the system’s performance is evaluated before and after inclusion, as demonstrated in Table 2. As shown in the table, the proposed segmentation is enhanced after including MGRF-based morphological/anatomical constraints, particularly in severe cases where the DSC is significantly increased from 82.15 ± 13.13 % to 95.15 ± 1.91 % . To more prove the attainable enhancement of the system, Figure 8 presents three examples of the proposed system before and after the inclusion for healthy/mild, moderate, and severe COVID-19 infected lung. As shown in the figure, the developed system outperforms the proposed appearance model alone (i.e., LCG) for three examples, whereas the proposed system shows its ability to segment a severe COVID-19 infection with 94.55 % DSC compared to the proposed appearance model which gives 76.49 % DSC. Moreover, Figure 9 presents the proposed segmentation for a severe lung COVID-19 infection at different cross-sections (i.e., 2D axial, coronal, and saggital) to visually show the efficiency of the proposed system. Overall, the proposed system achieves a DSC, overlap, BHD, and ALVD of 95.76 ± 1.83 % , 91.76 ± 3.29 % , 4.06 ± 5.01 , and 2.93 ± 2.39 , respectively. Finally, to prove the robustness of the proposed segmentation approach, deep learning approaches are adopted as a comparison: DeepLabv3+ [52] using ResNet-50 network as a backbone, Inf-Net [28] with backbone ResNet-50 network, U-Net [25], and 3D U-Net [53]. The results are reported in Table 3. As demonstrated in the table, the 3D U-Net approach gives a worst performance of 66.08 ± 35.99 % DSC, 58.30 ± 34.81 % overlap, 44.44 ± 48.60 BHD, and 66.77 ± 133.01 ALVD, while the proposed segmentation approach gives the best performance compared to these deep learning approaches. Moreover, to visually demonstrate the capability of the proposed system, three different examples of healthy/mild, moderate, and severe lung COVID-19 infections are segmented using these approaches, as presented in Figure 10. As demonstrated in the figure, the proposed approach segments the three examples better than the other four approaches. Moreover, the U-Net approach has a DSC close to the proposed approach. However, there are some parts that segment incorrectly as demonstrated in the figures, e.g., classifying part of trachea or chest as lung. Therefore, the proposed system is much better due to its segmentation being closer to the ground-truth. Therefore, it is highly recommended to use our approach to segment the lung infected by COVID-19 as it shows better performance than the state-of-the-art deep learning approaches. In addition, it is unsupervised technique, thus it will not suffer from the underfitting and overfitting problems.

5. Discussion and Conclusions

Experiments demonstrate that the proposed framework is promising and achieved high accuracy, with identification of the first-order appearance model followed by 3D morphological constraints based on analytical estimation MGRF parameters producing good results when segmenting the COVID-19 infected lung region from CT images. Quantitative metrics of accuracy including the DSC, overlap coefficient, 95th-percentile BHD, and the ALVD metrics all show consistent performance on our sample data set of 32 subjects, outperforming current, state-of-the-art deep learning-based lung segmentation approaches. The results herein demonstrate the ability of the developed system to segment lung on a CT image, whose DSC is improved from 89.59 ± 9.76 % to 95.67 ± 1.83 % when 3D morphological MGRF-based constraints are included in the system pipeline. However, the accuracy of the proposed segmentation system will get affected if the lung is significantly damaged or filled with water, or the appearance of the lung is closed to the chest. Thus, separation based on appearance model will be very challenging task. Therefore, we plan to add some shape model approach in our system to overcome these problems. Moreover, a future extension of this work would integrate the proposed segmentation approach into a computer-aided diagnostic system to assess pulmonary function and risk of mortality in COVID-19 patients, which is the ultimate goal of our research group. Furthermore, the morphological constraints could be made to support large-scale inhomogeneity of the kind seen in severe lung infection. This will be accomplished by expanding the neighborhood system to include larger cliques so that the MGRF model incorporates higher order interaction effects.

Author Contributions

Methodology, A.S. (Ahmed Sharafeldeen), M.E., N.S.A., A.S. (Ahmed Soliman), and A.E.-B.; Software, A.S. (Ahmed Sharafeldeen) and A.E.-B.; Validation, A.S. (Ahmed Sharafeldeen) and A.E.-B.; Formal analysis, A.S. (Ahmed Sharafeldeen), M.E., N.S.A., A.S. (Ahmed Soliman), and A.E.-B.; investigation, A.E.-B.; Writing—original draft preparation, A.S. (Ahmed Sharafeldeen), M.E., N.S.A., A.S. (Ahmed Soliman), and A.E.-B.; Writing—review and editing, A.S. (Ahmed Sharafeldeen), M.E., N.S.A., A.S. (Ahmed Soliman), and A.E.-B.; Visualization, A.S. (Ahmed Sharafeldeen) and A.E.-B.; Supervision, A.E.-B.; Project administration, N.S.A. and A.E.-B. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number (PNU-DRI-Targeted-20-001).

Institutional Review Board Statement

We used public data.

Informed Consent Statement

We used public data.

Data Availability Statement

The data presented in this study are openly available in The Cancer Imaging Archive (TCIA) at https://doi.org/10.7937/TCIA.2020.GQRY-NC81 accessed date 5 June 2021, reference number [50].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Machhi, J.; Herskovitz, J.; Senan, A.M.; Dutta, D.; Nath, B.; Oleynikov, M.D.; Blomberg, W.R.; Meigs, D.D.; Hasan, M.; Patel, M.; et al. The Natural History, Pathobiology, and Clinical Manifestations of SARS-CoV-2 Infections. J. Neuroimmune Pharmacol. 2020, 15, 359–386. [Google Scholar] [CrossRef] [PubMed]
  2. Gupta, S.; Hayek, S.S.; Wang, W.; Chan, L.; Mathews, K.S.; Melamed, M.L.; Brenner, S.K.; Leonberg-Yoo, A.; Schenck, E.J.; Radbel, J.; et al. Factors Associated With Death in Critically Ill Patients With Coronavirus Disease 2019 in the US. JAMA Intern. Med. 2020, 180, 1436. [Google Scholar] [CrossRef]
  3. Vijayaraj, J. Various Segmentation Techniques for Lung Cancer Detection using CT Images: A Review. Turk. J. Comput. Math. Educ. 2021, 12, 918–928. [Google Scholar] [CrossRef]
  4. Silveira, M.; Nascimento, J.; Marques, J. Automatic segmentation of the lungs using robust level sets. In Proceedings of the 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Lyon, France, 22–26 August 2007. [Google Scholar] [CrossRef]
  5. Sun, X.; Zhang, H.; Duan, H. 3D Computerized Segmentation of Lung Volume With Computed Tomography. Acad. Radiol. 2006, 13, 670–677. [Google Scholar] [CrossRef] [PubMed]
  6. Leader, J.K.; Zheng, B.; Rogers, R.M.; Sciurba, F.C.; Perez, A.; Chapman, B.E.; Patel, S.; Fuhrman, C.R.; Gur, D. Automated lung segmentation in X-ray computed tomography. Acad. Radiol. 2003, 10, 1224–1236. [Google Scholar] [CrossRef]
  7. Hu, S.; Hoffman, E.; Reinhardt, J. Automatic lung segmentation for accurate quantitation of volumetric X-ray CT images. IEEE Trans. Med. Imaging 2001, 20, 490–498. [Google Scholar] [CrossRef]
  8. Brown, M.S.; Goldin, J.G.; McNitt-Gray, M.F.; Greaser, L.E.; Sapra, A.; Li, K.T.; Sayre, J.W.; Martin, K.; Aberle, D.R. Knowledge-based segmentation of thoracic computed tomography images for assessment of split lung function. Med. Phys. 2000, 27, 592–598. [Google Scholar] [CrossRef]
  9. Brown, M.; McNitt-Gray, M.; Mankovich, N.; Goldin, J.; Hiller, J.; Wilson, L.; Aberie, D. Method for segmenting chest CT image data using an anatomical model: Preliminary results. IEEE Trans. Med. Imaging 1997, 16, 828–839. [Google Scholar] [CrossRef]
  10. Van Rikxoort, E.M.; de Hoop, B.; Viergever, M.A.; Prokop, M.; van Ginneken, B. Automatic lung segmentation from thoracic computed tomography scans using a hybrid approach with error detection. Med. Phys. 2009, 36, 2934–2947. [Google Scholar] [CrossRef]
  11. Tan, J.H.; Acharya, U.R. Active spline model: A shape based model—Interactive segmentation. Digit. Signal Process. 2014, 35, 64–74. [Google Scholar] [CrossRef] [Green Version]
  12. Gill, G.; Toews, M.; Beichel, R.R. Robust Initialization of Active Shape Models for Lung Segmentation in CT Scans: A Feature-Based Atlas Approach. Int. J. Biomed. Imaging 2014, 2014, 1–7. [Google Scholar] [CrossRef]
  13. Lassen, B.; van Rikxoort, E.M.; Schmidt, M.; Kerkstra, S.; van Ginneken, B.; Kuhnigk, J.M. Automatic Segmentation of the Pulmonary Lobes From Chest CT Scans Based on Fissures, Vessels, and Bronchi. IEEE Trans. Med Imaging 2013, 32, 210–222. [Google Scholar] [CrossRef]
  14. Birkbeck, N.; Kohlberger, T.; Zhang, J.; Sofka, M.; Kaftan, J.; Comaniciu, D.; Zhou, S.K. Lung Segmentation from CT with Severe Pathologies Using Anatomical Constraints. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2014; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 804–811. [Google Scholar] [CrossRef] [Green Version]
  15. Oulefki, A.; Agaian, S.; Trongtirakul, T.; Laouar, A.K. Automatic COVID-19 lung infected region segmentation and measurement using CT-scans images. Pattern Recognit. 2021, 114, 107747. [Google Scholar] [CrossRef]
  16. Korfiatis, P.; Kalogeropoulou, C.; Karahaliou, A.; Kazantzi, A.; Skiadopoulos, S.; Costaridou, L. Texture classification-based segmentation of lung affected by interstitial pneumonia in high-resolution CT. Med. Phys. 2008, 35, 5290–5302. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, J.; Li, F.; Li, Q. Automated segmentation of lungs with severe interstitial lung disease in CT. Med. Phys. 2009, 36, 4592–4599. [Google Scholar] [CrossRef] [Green Version]
  18. Dehmeshki, J.; Ye, X.; Lin, X.; Valdivieso, M.; Amin, H. Automated detection of lung nodules in CT images using shape-based genetic algorithm. Comput. Med Imaging Graph. 2007, 31, 408–417. [Google Scholar] [CrossRef] [PubMed]
  19. Nakagomi, K.; Shimizu, A.; Kobatake, H.; Yakami, M.; Fujimoto, K.; Togashi, K. Multi-shape graph cuts with neighbor prior constraints and its application to lung segmentation from a chest CT volume. Med. Image Anal. 2013, 17, 62–77. [Google Scholar] [CrossRef] [PubMed]
  20. Mansoor, A.; Bagci, U.; Xu, Z.; Foster, B.; Olivier, K.N.; Elinoff, J.M.; Suffredini, A.F.; Udupa, J.K.; Mollura, D.J. A Generic Approach to Pathological Lung Segmentation. IEEE Trans. Med. Imaging 2014, 33, 2293–2310. [Google Scholar] [CrossRef] [Green Version]
  21. Houssein, E.H.; Emam, M.M.; Ali, A.A. Improved manta ray foraging optimization for multi-level thresholding using COVID-19 CT images. Neural Comput. Appl. 2021. [Google Scholar] [CrossRef]
  22. El-Baz, A.; Beache, G.M.; Gimel’farb, G.; Suzuki, K.; Okada, K.; Elnakib, A.; Soliman, A.; Abdollahi, B. Computer-Aided Diagnosis Systems for Lung Cancer: Challenges and Methodologies. Int. J. Biomed. Imaging 2013, 2013, 942353. [Google Scholar] [CrossRef] [Green Version]
  23. Saood, A.; Hatem, I. COVID-19 lung CT image segmentation using deep learning methods: U-Net versus SegNet. BMC Med. Imaging 2021, 21. [Google Scholar] [CrossRef]
  24. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  26. Yan, Q.; Wang, B.; Gong, D.; Luo, C.; Zhao, W.; Shen, J.; Shi, Q.; Jin, S.; Zhang, L.; You, Z. COVID-19 Chest CT Image Segmentation—A Deep Convolutional Neural Network Solution. arXiv 2020, arXiv:2004.10987. [Google Scholar]
  27. Amyar, A.; Modzelewski, R.; Li, H.; Ruan, S. Multi-task deep learning based CT imaging analysis for COVID-19 pneumonia: Classification and segmentation. Comput. Biol. Med. 2020, 126, 104037. [Google Scholar] [CrossRef]
  28. Fan, D.P.; Zhou, T.; Ji, G.P.; Zhou, Y.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Inf-Net: Automatic COVID-19 Lung Infection Segmentation From CT Images. IEEE Trans. Med Imaging 2020, 39, 2626–2637. [Google Scholar] [CrossRef]
  29. Chen, X.; Yao, L.; Zhang, Y. Residual Attention U-Net for Automated Multi-Class Segmentation of COVID-19 Chest CT Images. arXiv 2020, arXiv:2004.05645. [Google Scholar]
  30. Abdel-Basset, M.; Chang, V.; Hawash, H.; Chakrabortty, R.K.; Ryan, M. FSS-2019-nCov: A deep learning architecture for semi-supervised few-shot segmentation of COVID-19 infection. Knowl. Based Syst. 2021, 212, 106647. [Google Scholar] [CrossRef] [PubMed]
  31. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2Net: A New Multi-Scale Backbone Architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Zhao, C.; Xu, Y.; He, Z.; Tang, J.; Zhang, Y.; Han, J.; Shi, Y.; Zhou, W. Lung segmentation and automatic detection of COVID-19 using radiomic features from chest CT images. Pattern Recognit. 2021, 119, 108071. [Google Scholar] [CrossRef]
  33. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016. [Google Scholar] [CrossRef] [Green Version]
  34. Li, L.; Qin, L.; Xu, Z.; Yin, Y.; Wang, X.; Kong, B.; Bai, J.; Lu, Y.; Fang, Z.; Song, Q.; et al. Using Artificial Intelligence to Detect COVID-19 and Community-acquired Pneumonia Based on Pulmonary CT: Evaluation of the Diagnostic Accuracy. Radiology 2020, 296, E65–E71. [Google Scholar] [CrossRef]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef] [Green Version]
  36. Singh, V.K.; Abdel-Nasser, M.; Pandey, N.; Puig, D. LungINFseg: Segmenting COVID-19 Infected Regions in Lung CT Images Based on a Receptive-Field-Aware Deep Learning Framework. Diagnostics 2021, 11, 158. [Google Scholar] [CrossRef] [PubMed]
  37. Shan, F.; Gao, Y.; Wang, J.; Shi, W.; Shi, N.; Han, M.; Xue, Z.; Shen, D.; Shi, Y. Abnormal lung quantification in chest CT images of COVID-19 patients with deep learning and its application to severity prediction. Med. Phys. 2021, 48, 1633–1645. [Google Scholar] [CrossRef] [PubMed]
  38. Gerard, S.E.; Herrmann, J.; Xin, Y.; Martin, K.T.; Rezoagli, E.; Ippolito, D.; Bellani, G.; Cereda, M.; Guo, J.; Hoffman, E.A.; et al. CT image segmentation for inflamed and fibrotic lungs using a multi-resolution convolutional neural network. Sci. Rep. 2021, 11. [Google Scholar] [CrossRef]
  39. Pan, F.; Li, L.; Liu, B.; Ye, T.; Li, L.; Liu, D.; Ding, Z.; Chen, G.; Liang, B.; Yang, L.; et al. A novel deep learning-based quantification of serial chest computed tomography in Coronavirus Disease 2019 (COVID-19). Sci. Rep. 2021, 11, 1–11. [Google Scholar] [CrossRef]
  40. Ma, J.; Wang, Y.; An, X.; Ge, C.; Yu, Z.; Chen, J.; Zhu, Q.; Dong, G.; He, J.; He, Z.; et al. Toward data-efficient learning: A benchmark for COVID-19 CT lung and infection segmentation. Med. Phys. 2021, 48, 1197–1210. [Google Scholar] [CrossRef]
  41. Elharrouss, O.; Subramanian, N.; Al-Maadeed, S. An encoder-decoder-based method for COVID-19 lung infection segmentation. arXiv 2020, arXiv:2007.00861. [Google Scholar]
  42. Müller, D.; Rey, I.S.; Kramer, F. Automated Chest CT Image Segmentation of COVID-19 Lung Infection based on 3D U-Net. arXiv 2020, arXiv:2007.04774. [Google Scholar]
  43. Tilborghs, S.; Dirks, I.; Fidon, L.; Willems, S.; Eelbode, T.; Bertels, J.; Ilsen, B.; Brys, A.; Dubbeldam, A.; Buls, N.; et al. Comparative study of deep learning methods for the automatic segmentation of lung, lesion and lesion type in CT scans of COVID-19 patients. arXiv 2020, arXiv:2007.15546. [Google Scholar]
  44. Tajbakhsh, N.; Jeyaseelan, L.; Li, Q.; Chiang, J.N.; Wu, Z.; Ding, X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med. Image Anal. 2020, 63, 101693. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. El-Baz, A.S.; Gimel’farb, G.L.; Suri, J.S. Stochastic Modeling for Medical Image Analysis; OCLC: 1086143882; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  46. Fuglede, B.; Topsoe, F. Jensen-Shannon divergence and Hilbert space embedding. In Proceedings of the International Symposium onInformation Theory 2004, Chicago, IL, USA, 27 June–2 July 2004. [Google Scholar] [CrossRef]
  47. MacKay, D.J.C. Information Theory, Inference, and Learning Algorithms; Cambridge University Press: New York, NY, USA, 2003; p. 34. [Google Scholar]
  48. Gerig, G.; Jomier, M.; Chakos, M. Valmet: A New Validation Tool for Assessing and Improving 3D Object Segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2001; Springer: Berlin/Heidelberg, Germany, 2001; pp. 516–523. [Google Scholar] [CrossRef] [Green Version]
  49. Soliman, A.; Khalifa, F.; Alansary, A.; Gimel’farb, G.; El-Baz, A. Performance Evaluation of an Automatic MGRF-Based Lung Segmentation Approach; AIP: College Park, MA, USA, 2013. [Google Scholar] [CrossRef] [Green Version]
  50. An, P.; Xu, S.; Harmon, S.A.; Turkbey, E.B.; Sanford, T.H.; Amalou, A.; Kassin, M.; Varble, N.; Blain, M.; Anderson, V.; et al. CT Images in COVID-19. Available online: https://wiki.cancerimagingarchive.net/display/Public/CT+Images+in+COVID-19 (accessed on 5 June 2021).
  51. Kasper, J.; Decker, J.; Wiesenreiter, K.; Römmele, C.; Ebigbo, A.; Braun, G.; Häckel, T.; Schwarz, F.; Wehler, M.; Messmann, H.; et al. Typical Imaging Patterns in COVID-19 Infections of the Lung on Plain Chest Radiographs to Aid Early Triage. In RöFo-Fortschritte auf dem Gebiet der RöNtgenstrahlen und der Bildgebenden Verfahren; Georg Thieme Verlag KG: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  52. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–4 September 2018. [Google Scholar]
  53. Çiçek., Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Schematic illustration of the pipeline of the proposed segmentation system using CT images.
Figure 1. Schematic illustration of the pipeline of the proposed segmentation system using CT images.
Sensors 21 05482 g001
Figure 2. An illustrative example of variability of CT appearance (distribution of radiodensities) for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infections.
Figure 2. An illustrative example of variability of CT appearance (distribution of radiodensities) for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infections.
Sensors 21 05482 g002
Figure 3. Illustration of the 3D MGRF-Based morphological constraints on the anatomical segmentation. The middle column shows the selected slice, and its upper and lower slices; the left column shows the selected pixel and its neighbors at the upper slice while the right column shows the selected pixel and its neighbors at the lower slice.
Figure 3. Illustration of the 3D MGRF-Based morphological constraints on the anatomical segmentation. The middle column shows the selected slice, and its upper and lower slices; the left column shows the selected pixel and its neighbors at the upper slice while the right column shows the selected pixel and its neighbors at the lower slice.
Sensors 21 05482 g003
Figure 4. Illustration of the evaluation metrics: (a) DSC, (b) HD, (c) overlap coefficient, and (d) ALVD. Note that TP, TN, FP, and FN are true positive (correct lung pixel), true negative (correct background pixel), false positive (incorrect lung pixel), and false negative (incorrect background pixel).
Figure 4. Illustration of the evaluation metrics: (a) DSC, (b) HD, (c) overlap coefficient, and (d) ALVD. Note that TP, TN, FP, and FN are true positive (correct lung pixel), true negative (correct background pixel), false positive (incorrect lung pixel), and false negative (incorrect background pixel).
Sensors 21 05482 g004
Figure 5. An illustrative example of the proposed system: (a) CT chest volume, (b) empirical density, (c) two dominants Gaussian components, (d) JSD between empirical density and mixed density, (e) two dominant and K 2 subdominant Gaussian components, (f) proposed cluster algorithm, (g) marginal density for lung and chest, and (h) final mixed density for all components. Note that JSD stands for Jensen–Shannon divergence.
Figure 5. An illustrative example of the proposed system: (a) CT chest volume, (b) empirical density, (c) two dominants Gaussian components, (d) JSD between empirical density and mixed density, (e) two dominant and K 2 subdominant Gaussian components, (f) proposed cluster algorithm, (g) marginal density for lung and chest, and (h) final mixed density for all components. Note that JSD stands for Jensen–Shannon divergence.
Sensors 21 05482 g005
Figure 6. An illustrative example of the proposed appearance model estimated from Thick-Section CT appearance model for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that first, second, and third rows represent empirical, marginal, and mixed densities, respectively.
Figure 6. An illustrative example of the proposed appearance model estimated from Thick-Section CT appearance model for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that first, second, and third rows represent empirical, marginal, and mixed densities, respectively.
Sensors 21 05482 g006
Figure 7. An illustrative example of the proposed appearance model estimated from Thin-Section CT appearance model for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that first, second, and third rows represent empirical, marginal, and mixed densities, respectively.
Figure 7. An illustrative example of the proposed appearance model estimated from Thin-Section CT appearance model for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that first, second, and third rows represent empirical, marginal, and mixed densities, respectively.
Sensors 21 05482 g007
Figure 8. An illustrative example of the proposed segmentation for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that red border (green border or yellow region) refers to ground-truth (segmentation).
Figure 8. An illustrative example of the proposed segmentation for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that red border (green border or yellow region) refers to ground-truth (segmentation).
Sensors 21 05482 g008
Figure 9. An illustrative example of the proposed segmentation (second and third rows) for a severe COVID-19 infected lung at (a) 2D axial, (b) coronal, and (c) sagittal cross sections of an original image (first row). Note that red border (green border or yellow region) refers to ground-truth (segmentation).
Figure 9. An illustrative example of the proposed segmentation (second and third rows) for a severe COVID-19 infected lung at (a) 2D axial, (b) coronal, and (c) sagittal cross sections of an original image (first row). Note that red border (green border or yellow region) refers to ground-truth (segmentation).
Sensors 21 05482 g009
Figure 10. An illustrative example of the proposed segmentation compared to other deep learning approaches for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that red border (green border or yellow region) refers to ground-truth (segmentation).
Figure 10. An illustrative example of the proposed segmentation compared to other deep learning approaches for (a) healthy/mild, (b) moderate, and (c) severe COVID-19 infected lung. Note that red border (green border or yellow region) refers to ground-truth (segmentation).
Sensors 21 05482 g010
Table 1. Dataset characteristics.
Table 1. Dataset characteristics.
ClassResolution#Slices#PatientsTotal
TrainingHealthy/Mild 512 × 512 43 54 234
Moderate35–39720
Severe 46 321 12
TestingHealthy/Mild 51 125 432
Moderate 36 607 17
Severe 44 577 11
Table 2. Quantitative evaluation of the proposed segmentation system before and after applying rotation invariant Markov–Gibbs random field (MGRF). Note that LCG, DSC, BHD, and ALVD stand for linear combination of Gaussian, Dice similarity coefficient, 95th-percentile bidirectional Hausdorff distance, and absolute lung volume difference, respectively.
Table 2. Quantitative evaluation of the proposed segmentation system before and after applying rotation invariant Markov–Gibbs random field (MGRF). Note that LCG, DSC, BHD, and ALVD stand for linear combination of Gaussian, Dice similarity coefficient, 95th-percentile bidirectional Hausdorff distance, and absolute lung volume difference, respectively.
DSCOverlapBHDALVD
LCG-modelHealthy/Mild 96.37 ± 0.47 % 92.99 ± 0.88 % 11.32 ± 4.59 3.58 ± 2.06
Moderate 92.57 ± 4.64 % 86.47 ± 7.61 % 9.59 ± 5.30 7.31 ± 7.39
Severe 82.51 ± 13.13 % 71.98 ± 17.26 % 13.77 ± 7.20 23.16 ± 19.29
Overall 89.59 ± 9.76 % 82.31 ± 13.72 % 11.24 ± 6.08 12.29 ± 14.63
Final SystemHealthy/Mild 97.53 ± 0.56 % 95.18 ± 1.06 % 2.41 ± 1.12 1.72 ± 1.09
Moderate 95.54 ± 1.91 % 91.53 ± 3.41 % 4.25 ± 3.73 3.48 ± 2.59
Severe 95.19 ± 1.66 % 90.87 ± 3.01 % 6.70 ± 6.97 2.52 ± 2.32
Overall 95.76 ± 1.83 % 91.76 ± 3.29 % 4.86 ± 5.01 2.93 ± 2.39
Table 3. Quantitative evaluation of the proposed segmentation system compared with other deep learning approaches. Note that DSC, BHD, and ALVD stand for Dice similarity coefficient, 95th-percentile bidirectional Hausdorff distance, and absolute lung volume difference, respectively.
Table 3. Quantitative evaluation of the proposed segmentation system compared with other deep learning approaches. Note that DSC, BHD, and ALVD stand for Dice similarity coefficient, 95th-percentile bidirectional Hausdorff distance, and absolute lung volume difference, respectively.
DSCOverlapBHDALVD
DeepLabv3+ [52]Healthy/Mild 80.32 ± 32.97 % 74.93 ± 37.81 % 23.58 ± 40.93 116.36 ± 220.88
Moderate 95.15 ± 1.52 % 90.79 ± 2.75 % 8.85 ± 21.37 7.32 ± 3.68
Severe 93.80 ± 2.47 % 88.41 ± 4.29 % 27.01 ± 46.72 8.78 ± 4.24
Overall 92.88 ± 11.49 % 88.07 ± 13.22 % 16.79 ± 34.17 21.41 ± 77.87
U-Net [25]Healthy/Mild 93.47 ± 7.46 % 88.39 ± 12.36 % 25.68 ± 15.60 5.46 ± 6.00
Moderate 95.09 ± 2.02 % 90.71 ± 3.62 % 19.04 ± 35.58 3.61 ± 3.36
Severe 91.68 ± 5.07 % 84.99 ± 8.30 % 44.81 ± 73.04 9.41 ± 7.24
Overall 93.77 ± 4.30 % 88.56 ± 7.16 % 28.49 ± 50.08 5.76 ± 5.80
Inf-Net [28]Healthy/Mild 89.77 ± 10.41 % 82.56 ± 15.81 % 17.41 ± 27.32 14.41 ± 9.32
Moderate 92.73 ± 1.64 % 86.49 ± 2.85 % 13.40 ± 25.32 13.43 ± 3.29
Severe 90.20 ± 4.56 % 82.42 ± 7.06 % 36.03 ± 38.77 14.91 ± 5.65
Overall 91.54 ± 4.52 % 84.67 ± 6.99 % 21.44 ± 31.40 14.00 ± 4.96
3D U-Net [53]Healthy/Mild 78.11 ± 35.98 % 72.80 ± 39.39 % 29.41 ± 41.91 140.67 ± 276.19
Moderate 68.46 ± 35.82 % 60.58 ± 33.85 % 42.04 ± 51.53 33.91 ± 35.35
Severe 55.72 ± 37.95 % 46.95 ± 35.31 % 55.25 ± 50.43 93.15 ± 158.26
Overall 66.08 ± 35.99 % 58.30 ± 34.81 % 44.44 ± 48.60 66.77 ± 133.01
Our SystemHealthy/Mild 97.53 ± 0.56 % 95.18 ± 1.06 % 2.41 ± 1.12 1.72 ± 1.09
Moderate 95.54 ± 1.91 % 91.53 ± 3.41 % 4.25 ± 3.73 3.48 ± 2.59
Severe 95.19 ± 1.66 % 90.87 ± 3.01 % 6.70 ± 6.97 2.52 ± 2.32
Overall 95.67 ± 1.83 % 91.76 ± 3.29 % 4.86 ± 5.01 2.93 ± 2.39
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sharafeldeen, A.; Elsharkawy, M.; Alghamdi, N.S.; Soliman, A.; El-Baz, A. Precise Segmentation of COVID-19 Infected Lung from CT Images Based on Adaptive First-Order Appearance Model with Morphological/Anatomical Constraints. Sensors 2021, 21, 5482. https://doi.org/10.3390/s21165482

AMA Style

Sharafeldeen A, Elsharkawy M, Alghamdi NS, Soliman A, El-Baz A. Precise Segmentation of COVID-19 Infected Lung from CT Images Based on Adaptive First-Order Appearance Model with Morphological/Anatomical Constraints. Sensors. 2021; 21(16):5482. https://doi.org/10.3390/s21165482

Chicago/Turabian Style

Sharafeldeen, Ahmed, Mohamed Elsharkawy, Norah Saleh Alghamdi, Ahmed Soliman, and Ayman El-Baz. 2021. "Precise Segmentation of COVID-19 Infected Lung from CT Images Based on Adaptive First-Order Appearance Model with Morphological/Anatomical Constraints" Sensors 21, no. 16: 5482. https://doi.org/10.3390/s21165482

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop