Next Article in Journal
Computer-Aided Detection (CADe) System with Optical Coherent Tomography for Melanin Morphology Quantification in Melasma Patients
Next Article in Special Issue
CSGBBNet: An Explainable Deep Learning Framework for COVID-19 Detection
Previous Article in Journal
Quantitative SARS-CoV-2 Spike Antibody Response in COVID-19 Patients Using Three Fully Automated Immunoassays and a Surrogate Virus Neutralization Test
Previous Article in Special Issue
Transfer Learning for the Detection and Diagnosis of Types of Pneumonia including Pneumonia Induced by COVID-19 from Chest X-ray Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Residual-Shuffle Network with Spatial Pyramid Pooling Module for COVID-19 Screening

by
Mohd Asyraf Zulkifley
1,*,†,
Siti Raihanah Abdani
2,†,
Nuraisyah Hani Zulkifley
3 and
Mohamad Ibrani Shahrimin
2
1
Department of Electrical, Electronic and Systems Engineering, Faculty of Engineering and Built Environment, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia
2
Faculty of Humanities, Management and Science, Universiti Putra Malaysia Bintulu Campus, Bintulu 97008, Sarawak, Malaysia
3
Community Health Department, Faculty of Medicine and Health Sciences, Universiti Putra Malaysia, Serdang 43400, Selangor, Malaysia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2021, 11(8), 1497; https://doi.org/10.3390/diagnostics11081497
Submission received: 19 June 2021 / Revised: 14 August 2021 / Accepted: 16 August 2021 / Published: 19 August 2021

Abstract

:
Since the start of the COVID-19 pandemic at the end of 2019, more than 170 million patients have been infected with the virus that has resulted in more than 3.8 million deaths all over the world. This disease is easily spreadable from one person to another even with minimal contact, even more for the latest mutations that are more deadly than its predecessor. Hence, COVID-19 needs to be diagnosed as early as possible to minimize the risk of spreading among the community. However, the laboratory results on the approved diagnosis method by the World Health Organization, the reverse transcription-polymerase chain reaction test, takes around a day to be processed, where a longer period is observed in the developing countries. Therefore, a fast screening method that is based on existing facilities should be developed to complement this diagnosis test, so that a suspected patient can be isolated in a quarantine center. In line with this motivation, deep learning techniques were explored to provide an automated COVID-19 screening system based on X-ray imaging. This imaging modality is chosen because of its low-cost procedures that are widely available even in many small clinics. A new convolutional neural network (CNN) model is proposed instead of utilizing pre-trained networks of the existing models. The proposed network, Residual-Shuffle-Net, comprises four stacks of the residual-shuffle unit followed by a spatial pyramid pooling (SPP) unit. The architecture of the residual-shuffle unit follows an hourglass design with reduced convolution filter size in the middle layer, where a shuffle operation is performed right after the split branches have been concatenated back. Shuffle operation forces the network to learn multiple sets of features relationship across various channels instead of a set of global features. The SPP unit, which is placed at the end of the network, allows the model to learn multi-scale features that are crucial to distinguish between the COVID-19 and other types of pneumonia cases. The proposed network is benchmarked with 12 other state-of-the-art CNN models that have been designed and tuned specially for COVID-19 detection. The experimental results show that the Residual-Shuffle-Net produced the best performance in terms of accuracy and specificity metrics with 0.97390 and 0.98695, respectively. The model is also considered as a lightweight model with slightly more than 2 million parameters, which makes it suitable for mobile-based applications. For future work, an attention mechanism can be integrated to target certain regions of interest in the X-ray images that are deemed to be more informative for COVID-19 diagnosis.

1. Introduction

Even after a year since the start of the Coronavirus disease 2019 (COVID-19) pandemic, there are still many countries that struggle with the increased number of COVID-19 cases everyday [1]. Moreover, there are several new COVID-19 mutations that are more prevalent among the younger peoples that have a higher infection rate [2]. Luckily, several of the developed vaccines like AstraZeneca, Pfizer, and Moderna have been proven to work well in reducing the number of infected cases even for these new COVID-19 variants [3]. Thus, the screening task of the COVID-19 cases is becoming more important so that high-risk patients can be identified immediately, which is a crucial step in breaking the infection chain of the virus. These patients were then to be quarantined in dedicated centers, whereby patients with severe symptoms need to be admitted to the hospitals for further intensive treatment. According to the World Health Organization, the suggested diagnosis method to detect this severe acute respiratory syndrome coronavirus 2, also known as SARS-CoV-2, which is the cause for the COVID-19 disease is through a reverse transcription-polymerase chain reaction (RT-PCR) test [4]. However, the cost of the RT-PCR test can be considered as expensive for most of the developing countries, whereby their economies have also been affected heavily by this pandemic. Hence, an effective screening method that can detect the disease immediately such as a rapid antigen test has been proposed to screen the patient with a high likelihood to be positive of COVID-19 [5]. In general, the rapid antigen test is slightly less accurate with 90% sensitivity and hence the results still need to be confirmed with the RT-PCR test [6]. Nevertheless, the advantage of fast screening outweighs the danger of spreading the viruses caused by the late notification of SARS-CoV-2 infection.
According to the current market price, a rapid antigen test costs around 30 USD, and it is still a pricey cost in the case of mass testing, especially in the industrial settings [7]. Therefore, an automated X-ray based screening method has been proposed to offer a low-cost screening alternative, while still producing fast screening results. The cost of an X-ray procedure is around one-fourth of the rapid antigen test and the machine is widely available all over the world [8]. Therefore, no additional specific equipment is needed, whereby small clinics are also known to have an X-ray machine. However, this imaging procedure needs to be complement with an automated screening method since not all health practitioners are well versed in the COVID-19 detection [9]. An X-ray-based screening requires a lot of prior experience, where the general health practitioners need to familiarize themselves with the COVID-19 prognosis as seen from the chest X-ray image. In line with this argument, this paper proposes an automated screening algorithm to detect the likelihood of COVID-19 cases based on X-ray images using an advanced machine learning technique. The proposed system will only focus on full-frontal chest X-ray images; as such, any side or sliced frontal X-ray images as shown in Figure 1 will be removed from the dataset. This step is taken because of two-fold reasons, which are to limit the variation of the input images and to reduce inaccurate learning of the model due to noisy data in the COVID-19 class.
In this work, a lightweight convolutional neural network (CNN) is proposed using the Residual-Shuffle network concept, in which the main branch will be split into two sub-networks that will later be concatenated and shuffled together to better learn the features from various prior layers. A residual skip connection [10] is also added to reduce the possibility of a zero-gradient diminishing issue, where the full architecture of the proposed network consists of four stacks of the Residual-Shuffle unit. At the bottom layer, three parallel branches of a spatial pyramid pooling (SPP) unit [11] are also added to improve the network capability in handling multi-scale detection. The added unit will be useful for the detection of COVID-19 cases with various severity levels, whereby the size of the air pocket in the X-ray images will be different as the disease becomes more severe. The design of this SPP unit is in contrast to the simplified SPP network as used in [12], in which their multi-scale feature maps are directly obtained through repeated down-pooling without performing any convolution operation. Overall, the proposed architecture utilizes slightly more than 2 million parameters, which is a lot smaller than the upper threshold of a lightweight network as defined in [8]. Due to its lightweight nature, it can be implemented on various types of mobile platforms and still achieve acceptable processing speed. Therefore, the main novelties of the proposed Residual-Shuffle-Net are its lightweight and accurate CNN model with just ∼2 million parameters, and the residual-shuffle unit that allows better feature learning across several groups of channels. Moreover, the network has also been integrated with three parallel branches of an SPP unit to extract multi-scale features, which are crucial in distinguishing COVID-19 cases and other types of pneumonia cases. It is worth noting that the proposed method aims to be an early screening method, after which the patients still need to be diagnosed by qualified medical practitioners. Therefore, this method is more applicable to developing countries, whereby the cost of mass diagnosis is considerably high compared to the average wage rate. Besides that, due to the lack of good laboratory facilities in developing countries, the results of the RT-PCR test will usually come out a few days after the samples are taken, which makes the proposed fast screening method a good complimentary test.
This paper is organized into five sections, where a review of the CNN-based system in screening COVID-19 disease is summarized in Section 2. Section 3 discusses in detail the full architecture of the proposed Residual-Shuffle Network (Residual-Shuffle-Net). The source code for the Residual-Shuffle-Net can be found at https://github.com/asyrafzulkifley/Residual-Shuffle-Net/blob/main/Residual_Shuffle_Net%20Model (accessed on 1 August 2021). Section 4 provides details of the database used for the validation tests and the outputs of the experiments, whereby comprehensive discussions on their performance are elaborated with respect to the state-of-the-art benchmarked methods. Section 5 summarizes the limitations of the proposed work, while a concise section on conclusions and future works is proposed at the end of this paper.

2. Convolutional Neural Networks for COVID-19 Detection

Convolutional neural networks have been successfully applied in many applications such as video analytic [13], intelligent remote sensing [14], two-dimensional signal processing [15], biomedical diagnosis [16], and many more. The technology relies on a set of optimal features to represent a dedicated application, whereby the features are trained using large numbers of data, contrary to the handcrafted features in the standard machine learning approach. As an example of the standard machine learning approach, the work in [17] has extracted a set of handcrafted COVID-19 features as the input to a naive Bayes classifier. They have utilized textures and several morphological features to represent the possibility of COVID-19 cases. Moreover, there are two popular imaging modalities have been explored in COVID-19 detection, which are X-ray images and computed tomography (CT) scans. According to Sverzellati et al. [18], X-ray is the better imaging modality compared to the CT scan in screening the possibility of COVID-19 cases, coupled with its low-cost procedures and wide availability of the X-ray machine. However, researchers in [19] found out that the segmentation of lung condition using CT scans produced a clearer mask for identifying the COVID-19 symptoms. In general, researchers have focused on either using a pre-trained model or developing a new dedicated model for the task of COVID-19 detection. This issue has been explored initially by Pham et al. [20], in which they have run several tests on the existing CNN models to verify the network’s effectiveness for the COVID-19 detection. They concluded that using pre-trained models of AlexNet, GoogleNet, and SqueezeNet were enough to achieve more than 98% accuracies on a three-class classification problem. However, there is an issue of class imbalance where they have used only 55 images of COVID-19 X-ray images compared to more than 1000 images for the other two classes.
Hence, several researchers have also explored basic pre-trained CNN such as the works by Pandit et al. [21] and Panwar et al. [22], who have retrained the VGG-16 architecture for a two-class problem to identify either COVID-19 or normal cases. Kikkisetti et al. [23] have also used the same VGG-16 architecture with minimal model modification, just by changing the last layer connection to a set of four nodes. Again, the transfer learning approach is used due to a limited number of datasets, whereby only 108 COVID-19 samples were used during training and testing phases. Apostolopoulos and Mpesiana [24] have experimented on five existing models for two and three-class classification problems, which are VGG-19 [25], MobileNet V2 [26], Inception [27], Xception [28], and Inception ResNet V2 [29]. They found out that the simplest VGG-19 model produced the highest accuracy. These results were obtained from limited numbers of training data with just 224 images of COVID-19 X-ray, whereby the other deeper models might experience under-fitting issues. Similarly, Narin et al. [9] have tested different sets of five CNN models, which are ResNet-50, ResNet-101, ResNet-152 [10], Inception V3 [30], and Inception ResNet V2 [29], and concluded that the simplest model ResNet-50 produces the highest accuracy. Again, only 341 X-ray images of COVID-19 patients were used during their experiments, which might point to the same underlying issue. Therefore, Loey et al. [31] have suggested the usage of a generative adversarial network (GAN) to create a set of synthetic images so that the training data for the COVID-19 class can be increased. A variety of GAN models have also been successfully applied to other applications such as traffic sign recognition [32], in which the synthetic training dataset has managed to improve their object detection accuracy. The work in [33] improvised on the GAN design, in which a conditional deep convolutional GAN is used to create the synthetic image separately for each class. Through this approach, more accurate and dedicated samples for the COVID-19 class were generated, since the other class data are more than enough for CNN training. In contrast, Ucar and Korkmaz [34] utilized a lightweight CNN model, SqueezeNet [35], that uses 2 million parameters to screen the COVID-19 cases. They have also optimized the hyper-parameters setup using Bayesian optimization. This optimization approach was also applied in [36], where an optimal set of hyper-parameters can boost the performance of the machine learning network.
On the other hand, Khan et al. [37] have modified the Xception network by changing the bottom layer with a flatten operator instead of a regular global average pooling operator. Even though Xception uses a separable convolution as its building block, the network runs comparatively slow due to its large number of parameters. Hence, Panahi et al. [38] have devised a fast network, FCOD, which requires just a total of 85,321 parameters. Their lightweight model has utilized a separable convolution format with a low number of convolution filter sets. Like the network by Khan et al., the work by Abdani et al. [8] has modified the bottom layer of SqueezeNet to integrate a spatial pyramid pooling unit for a three-class classification problem. They have argued that the air pockets in the X-ray images varied in size, especially between different severity levels of the COVID-19 cases. Hence, the CNN network must be able to capture these various scales of unique features, so that a robust COVID-19 identification system can be produced. Contrary to the parallel down-pooling unit in the SPP approach, Mahmud et al. [39] have introduced a network with several parallel atrous convolutions with different dilation rates. Thus, the captured receptive field size became bigger when the dilation rate was increased. They have also utilized a two-stage training, whereby the first stage focuses on the general classification of various pneumonia cases, and the second stage only focuses on differentiating between a COVID-19 case or not. The network introduced by Gilanie et al. [40] is unique in the sense that all convolution kernels are in the size of 5 × 5, which makes it less optimal for Tensorflow application. Their straightforward network consists of eight layers of convolution operator without applying any batch normalization technique. The CoroDet, which was introduced in [41], utilized a stack of seven down-sampling and up-sampling modules to produce a lightweight COVID-19 classifier, which has been extensively tested for two, three, and four-class classification problems. They have utilized a Leaky ReLU activation function, whereby the down-sampling operation is done through maximum down-pooling, while the up-sampling operation is done through feature map resize. These repeated down and up-sampling procedures were operated at small-size feature maps, which leads to a lot of information loss. Apart from COVID-19 detection, CNN has also been used to predict the severity level of the COVID-19 infection. In [42], a pre-trained VGG-16 architecture is used to predict the lung condition of the COVID-19 patients based on chest X-ray imaging. The transfer learning approach is used to initialize the network parameters, in which the last layer is modified to be a regressor node. Two different networks were trained for two different regression goals, which are to identify the geographic extent and degree of opacity of the lung. Similar to the previous work goals, Wong et al. [43] use a more complex network architecture of COVID-Net S to determine the severity level of the patient’s lung conditions. They have used the stratified Monte Carlo cross-validation method to further improve sampling strategy by grouping the chest X-ray images according to the age, sex, geographical location, imaging view, and imaging position attributes. They have also augmented their training dataset with image manipulation methods so that the dataset variability will be increased through translations, rotations, horizontal flips, zooms, intensity shifts, cutout, and Gaussian noise addition.

3. Residual-Shuffle Network with Spatial Pyramid Pooling Module

The Residual-Shuffle Network (Residual-Shuffle-Net) is a specialized lightweight CNN model built for COVID-19 detection for X-ray-based imaging systems. The network core component combines the residual skip connection with a compact shuffle unit, whereby a set of group convolutions is applied to let the network to learn multiple local features, instead of a set of global features. The design of this network is derived from the basic module of the ShuffleNet V2 [44], whereby the main branch is split into two branches at the beginning of the network and only one of the branches is treated with a convolution operator, while the other branch acts as a feed-forward unit. Generally, the Residual-Shuffle-Net consists of three modules of network flows, which are bottom, middle, and top modules. Let us define the Residual-Shuffle-Net, RS as the following with M representing the module and I representing the input image with the size of 256 × 256 pixels:
RS = M 1 . M 2 . M 3 . I
The bottom module M 1 consists of an entry network that will extract rough features using two layers of convolution ( C ) and maximum down-pooling ( P ) operators. Each of the convolution operators applied in the Residual-Shuffle-Net will be followed by a batch normalization operator and a leaky rectified linear unit (Leaky ReLU) activation function. A small number of filter sizes is used for both convolution operations in the bottom network with just 8 and 16 channels to keep the number of parameters minimal during the early part of the network:
M 1 = C 1 . P 1 . C 2 . P 2
The middle module M 2 comprises of the core Residual-Shuffle unit ( RSU ). There will be a stack of four sequential RSU , where P is added between them to down-scale the feature map size except for the fourth unit:
M 2 = ( i = 0 3 RSU i . P i ) . RSU 4
The initial feature map size to this module is 64 × 64 pixels and the output feature map size at the end of this module is 8 × 8 pixels. Therefore, the range of the feature map size for this module is considerably small in order to maintain the lightweight nature of the proposed model. There will be three convolution operators in each RSU , where its full building block is shown in Figure 2. Residual-Shuffle-Net utilizes a bottleneck design for the convolution operations; as such, the middle convolution operator will have a smaller number of the filter set compared to the first and third convolution operators. The residual skip connection input layer comes from the output of the first convolution, which will be added to the output of the third convolution. Group and shuffle operations are performed during the second convolution operation, in which a simple channel split procedure is used to divide the feature maps into two equal groups, where it will be combined back using a concatenate operator. The addition of group and shuffle operations force the network to learn from various sets of feature branches, instead of one set of general features [45]. The design of sequential RSU allows the network to learn multiple local features instead of global features in each convolution layer.
The final module M 3 utilizes a spatial pyramid pooling ( SPP ) unit, followed by a composite function of global average pooling and dense connection. Since this study focuses on a three-class problem, a SoftMax activation function with three output classes is used to come out with the likelihood of an image belong to each respective class. The SPP unit is added with the goal of improving the network’s capability in extracting multi-scale features of the diseases from X-ray images. In this paper, the dataset used to validate the network performance consists of diseases from various severity levels of COVID-19. Thus, the size of the air pocket seen in the X-ray images varies in size, and some of them are difficult to identify and differentiate using naked eyes, especially for the case between COVID-19 and other types of pneumonia cases. In line with the previous reasoning, three parallel branches of down-pooling operators are employed to extract the multi-scale features, which will be resized and concatenated back at the end of the unit. Since the input feature map size is 8 × 8 pixels, a set of 2 by 2, 4 by 4, and 6 by 6 down-pooling kernels has been utilized to produce equally spaced down-pooling scales as shown in Figure 3.
Then, a standard global average pooling G operator is used to sample the best multi-scale features before the classification process is done using a dense feedforward layer. The full design of the Residual-Shuffle-Net architecture is given in Table 1, where each layer information is detailed out in terms of the filter size, kernel size, and stride step for all bottom, middle, and top modules. Let D represent the dense connection layer with three output classes, where a SoftMax activation function is applied to complete the Residual-Shuffle-Net classifier with a total of 2,090,491 parameters.
M 3 = SPP . G . D

4. Experiments and Discussion

4.1. Dataset

The proposed model is validated by using a three-class problem of X-ray image classification, which consists of COVID-19, other types of pneumonia, and normal cases. Due to the limited number of X-ray images for each category of bacterial and viral pneumonia, this study has chosen a three-class classification problem. Therefore, the class of other types of pneumonia contains both the virus and bacteria-caused pneumonia, except for COVID-19 cases for a fair deep learning comparison. The images for all three classes of X-ray radiography dataset were downloaded from two publicly available databases, which are Medical Imaging Databank of the Valencia Region (BIMCV) [46] and Radiological Society of North America (RSNA) [47]. The BIMCV dataset is the sole provider for the COVID-19 cases, while the RSNA dataset is the provider for the other types of pneumonia and normal cases. The dataset was skimmed through to select a set of quality images such that X-ray images with unrelated patterns or conditions will be removed. The main reason for this removal is to limit the possibility of the deep learning model in learning unrelated features to the disease such as color variation, background objects, side view images, and many more. If these images were not removed, these noisy patterns will be captured by the CNN model as parts of the disease patterns, which will produce an unfair comparison to the good X-ray imaging. In this case, the overall quality of the RSNA dataset is better than the BIMCV dataset. Besides that, even if the BIMCV dataset provides the radiological findings, this paper has omitted them as the respective findings are not available for the RSNA dataset. For the COVID-19 cases, all positive cases have also been confirmed with the RT-PCR test. The original mean age of the COVID-19 patient is 63 years old with a relatively fair distribution between the gender, with 46% of them being male and 54% of them being female. The dataset was captured from various types of X-ray machines, which were initially saved in Digital Imaging and Communications in Medicine (DICOM) format. According to the report in Vaya et al. [46], the majority of X-ray images were captured using fixed X-ray machines that include Konica Minolta 0862 342, GMM Accord DR 255, and Siemens FD-X X-ray machines. In the end, there are 1341 X-ray images for COVID-19 cases, 1341 images for other types of pneumonia cases, and 1341 images for normal cases, which sums up to 4023 X-ray images in total. All images are then saved in Portable Network Graphics (PNG) format with a standard resolution of 1024 × 1024 pixels, which is bigger than the input requirements of all benchmarked CNN models. Some samples of the X-ray images used in this paper are shown in Figure 4.

4.2. Evaluation Metrics

There are six evaluation metrics used to evaluate the performance of the proposed Residual-Shuffle-Net and its benchmarked methods. The selected metrics are accuracy ( A C C ¯ ) , sensitivity ( S E N ¯ ) , specificity ( S P E ¯ ) , precision ( P R E ¯ ) , F1-Score, and number of parameters. ( A C C ¯ ) concerns more on the true detection rate either for the positive or negative cases, while ( S E N ¯ ) and ( S P E ¯ ) concern more true positive rates and true negative rates, respectively. On the other hand, ( P R E ¯ ) measures the ratio of the correctly detected case and the total samples that have been predicted as positive, while F1-Score measures the harmonic mean between ( S E N ¯ ) and ( P R E ¯ ) . All five of these metrics rely on basic units of true positive ( T P o s ) , true negative ( T N e g ) , false positive ( F P o s ) , and false negative ( F N e g ) . The ( F P o s ) and the ( F N e g ) are the cases when the predicted class does not match the labeled ground truth, whereby the prediction should have been positive and negative detection, respectively. On contrary, the ( T P o s ) and the ( T N e g ) are the cases when the predicted class exactly matches the ground truth label for both of the cases. Finally, the number of parameters represents the total number of trainable and non-trainable parameters utilized by the respective CNN model. The evaluation metrics are calculated as follows:
A C C ¯ = T P o s + T N e g T P o s + T N e g + F P o s + F N e g
S E N ¯ = T P o s T P o s + F N e g
S P E ¯ = T N e g T N e g + F P o s
P R E ¯ = T P o s T P o s + F P o s
F 1 S c o r e = 2 T P o s 2 T P o s + F P o s + F N e g

4.3. Experimental Setup

There are 12 state-of-the-art CNN models from recent COVID-19 works that have been selected to be the performance benchmark for the proposed Residual-Shuffle-Net. All 12 models have utilized CNN classifiers for their COVID-19 detection system based on input from X-ray images, which are Hussain et al. [41], Abdani et al. [12], Khan et al. [37], Panahi et al. [38], Pandit et al. [21], Ozturk et al. [48], Mahmud et al. [39], Loey et al. [31], Ucar et al. [34], Panwar et al. [22], Narin et al. [9], and Gilanie et al. [40]. Five of the methods have used existing popular models, in which the models have been properly defined by the original authors, while the other seven methods have been selected because of their networks’ details were fully explained in their paper. All benchmarked models and the proposed Residual-Shuffle-Net have been coded on the Python platform using the Keras-Tensorflow library for a fair comparison, whereby their hyper-parameter settings have been tuned for the maximum classification performance. This is because the majority of the benchmarked papers have tested less than 600 images of COVID-19 cases. The main criterion used to judge for optimized hyperparameter settings for all methods is error convergence for training and validation loss function. The cutoff threshold value for both of the errors is 0.1, as such all models have been trained to produce errors less than the pre-set threshold value. Hence, a set of optimized settings as shown in Table 2 has been found using grid search methodology, whereby all the models achieved error convergence in their training and validation datasets. The performance metrics were also coded and analyzed using the Numpy library from Python software. One hot encoded labeling with SoftMax activation function is standardized as the last dense layer for all models. The Adam optimizer with a fixed learning rate is used to update the parameter values during the training phase, which has been set up to optimize the cross-entropy loss function with an accuracy performance metric. No simple or complex data augmentation was utilized, except for the image resizing operation to fit the input requirement of each model. In the Residual-Shuffle-Net case, the input image is resized to the resolution of 256 × 256 pixels. Batch size selection will depend on the model size, whereby the maximum possible batch size is used to train each of the models using a single Nvidia RTX 2080 Ti graphics card. Our Intel i9-9900K machine with a 3.60 GHz clock rate can afford to process Residual-Shuffle-Net with a batch size of 64 images. The proposed Residual-Shuffle-Net uses a total of 2,090,491 parameters, whereby 2,087,275 of them are trainable parameters, while 3216 of them are non-trainable parameters.
A five-fold cross-validation scheme is used to divide the dataset into general training and testing pots, so that the sampling bias can be reduced, whereas the over-fitting issue on the selected samples can be minimized. Then, the testing pot is further divided into two equal classes between validation and test dataset, whereby the final dataset is divided according to the ratio of 8:1:1 between training, validation, and testing phases. Therefore, for one validation fold, the number of X-ray images used for the training, validation, and testing are 3217, 403, and 403 images, respectively.

4.4. Discussion on the Residual-Shuffle-Net and Its Benchmarked Models Performance

In general, all the tested methods have been trained until convergence for both accuracy and loss functions as shown in Figure 5 and Figure 6, respectively. During the training phase, the error for all CNN models converged towards zero value, while the accuracy for 10 out of the 12 models converged towards maximum accuracy of 1.0. Besides that, the other two models by Panwar et al. [22] and Hussain et al. [41] have converged to 0.95 accuracies after 80 epochs of training update. The convergence pattern assumption is also supported by the validation loss curves as shown in Figure 7 that proved the Residual-Shuffle-Net and all the benchmark methods have been trained until convergence. The trend for both training and validation losses for all methods has converged towards zero error. Although there are six performance metrics were calculated in this study, two of them carry more weightage in determining the best detection method for COVID-19 screening. The goal of the screening stage is to detect as many true positive cases as possible that will be confirmed later by the RT-PCR test. Therefore, the A C C ¯ and P R E ¯ metrics were prioritized in ranking the benchmark methods, where both of them measure a certain ratio of positive cases over the total number of cases. However, it is worth noting that the false negative metric still plays an important role in COVID-19 screening. A screening method will be rendered useless if none of the cases were screened at the early stage, which will directly increase the cost of healthcare to the government. Table 3 shows the performance of the Residual-Shuffle-Net and its benchmark methods using all six performance metrics, which were ranked using A C C ¯ and P R E ¯ . As a whole, Residual-Shuffle-Net performed the best in five out of the six evaluation metrics, except for the total number of parameters. It achieved the highest A C C ¯ and P R E ¯ with 0.97390 and 0.97403, respectively, while maintaining a relatively lightweight model with just 2 million parameters.
It is interesting to note that the second and the third-best CNN models, which are the methods by Gilanie et al. [40] and Abdani et al. [12], respectively, are both a specialized model designed for COVID-19 detection. The method by Gilanie et al. achieved an A C C ¯ of 0.96868, while the method by Abdani et al. achieved an A C C ¯ of 0.96395. Both of them can also be regarded as lightweight CNN models with total usage of parameters of less than 10 million. The uniqueness of the method by Gilanie et al. is in the selection of convolution kernel size, whereby 5 × 5 kernels were used throughout their network. A bigger kernel size can better capture the unique features on the X-ray images but comes with the main weakness of slower processing speed. On the other hand, Abdani et al. achieved good detection performance by relying on the simplified spatial pyramid pooling module that was able to capture multi-scale features of the X-ray images, which was crucial in distinguishing the cases of COVID-19 and other types of pneumonia. Their three parallel down-pooling branches did not consist of convolution operation, where the features maps were directly flattened for dense connections, which allows their network to maintain a small size of the total number of parameters.
The best pre-trained model performance among the benchmark models is returned by Khan et al. [37] method through the usage of Xception-71 architecture. They have slightly modified the top layer of the network to include a global average pooling operator instead of the original flatten operation. Their method managed to record an A C C ¯ of 0.96247 and a S P E ¯ of 0.98123 but requires a large model size of 88 million parameters, which is more than 42 times total number of parameters compared to the proposed Residual-Shuffle-Net. Even with a big-sized model, their model has utilized separable convolution schemes to reduce the demand of memory usage but its three-layer convolution unit still uses a large filter size of 728 channels. Surprisingly, a simple VGG-16 model, which was employed by Pandit et al. [21] and Panwar et al. [22] delivered the next best evaluation performance among the pre-trained models. Their architecture used 13 layers of the convolutional operation without utilizing any residual or feedforward branches. However, Panwar et al. modified the top layer of the network, by using a global average pooling operator, which resulted in much smaller model size, from the original 33 million parameters down to 14 million parameters. The three-layer dense connections in the original network used by Pandit et al. require a large number of parameters because there is a connection on each of the 4096 nodes, which also produced worse A C C ¯ of 0.94858 than the work by Panwar et al.
The overall worst performing method is recorded by the model designed by Hussain et al. [41]. Their model is relatively lightweight, in which they utilized repeated down-sampling and up-sampling processes on a small feature map size. The up-sampling operations that were applied to the small feature map size did not increase their network capability in extracting meaningful features from the X-ray images, which resulted in low A C C ¯ and F1-Score values of 0.78685 and 0.77785, respectively. However, the lowest F1-Score of 0.72718 was returned by the method of Narin et al. [9] that uses a pre-trained model of ResNet-50 architecture. Their low F1-score indicates that ResNet-50 produced a low ratio of true positive detection compared to the number of false detection, which makes it not suitable for the screening task. However, the method by Narin et al. produced a much higher A C C ¯ and S P E ¯ with 0.83765 and 0.88078, respectively, when compared to the method by Hussain et al. Contrary to that, a smaller version of ResNet-18 used by Loey et al. [31] managed to produce a better A C C ¯ of 0.94058 compared to the larger ResNet versions. Similar findings were also concluded in the work of Apostolopoilos and Mpesiana [24], where their simplest model among VGG-19, MobileNet V2, Inception, Xception-41, and Inception ResNet V2 produced the best COVID-19 detection. The reasoning behind these findings can be pointed towards small differences in COVID-19 features and other types of pneumonia cases, where the addition of compact multi-scale approach used in Residual-Shuffle-Net and method by Abdani et al. is more crucial to the classification performance rather than a deep residual connection network.
Figure 8 shows the confusion matrix of the proposed Residual-Shuffle-Net in identifying the three classes of normal, COVID-19, and other types of pneumonia cases. This confusion matrix reports the performance of each class with respect to the other classes, instead of average performance information as shown in Table 3. The matrix provides the exact number of true positive, true negative, false positive, and false negative detections with regard to each of two other classes. The total number of samples on each class is uniform so that a fair comparison can be made to identify the weakness of the Residual-Shuffle-Net. The best true positive cases among the classes were recorded by the COVID-19 cases with 1329 out of 1341 X-ray images were correctly identified. Only eight cases of COVID-19 X-ray images were wrongly identified to be the other types of pneumonia cases. The main weakness of the proposed Residual-Shuffle-Net can be traced to false detection in the case of normal patients, where 51 normal cases were screened as other types of pneumonia cases. Similarly, 28 cases of other types of pneumonia were wrongly identified to be normal cases. One of the contributing factors behind this weakness is the quality of the X-ray images, whereby the datasets for normal and other types of pneumonia cases were captured in a more uniform setup. Contrary to that, X-ray images for COVID-19 cases were captured by using various machines with a different setup that leads to more variety in the imaging quality. Hence, it is easier to distinguish the COVID-19 cases compared to the other classes. However, the true positive detection for all classes remains high with the lowest case of 1282 true detections for the other types of pneumonia class, which is still a high accuracy with 95.6% true detection. For completion of the results, Figure 9 shows the receiver operating characteristic (ROC) curves for the proposed Residual-Shuffle-Net. The area under the curve (AUC) value for each validation fold is also provided, whereby the highest AUC of 0.9981 is achieved by the first fold, while the lowest AUC of 0.9963 is achieved by the second fold. Generally, the performance difference between the folds is very minimal with an AUC variance of 4.85 × 10 7 .

5. Limitations

There are few limitations of the proposed work, mainly due to the hardware limitation, emergence of new variants of concerns, and clinical test requirements. The core of the proposed method relies on the deep convolutional network that requires heavy computational power, especially during the training process. Therefore, an efficient computational platform is crucial during the training phase, while a lesser intensive computation is needed during the testing or screening phase. Moreover, the proposed deep network cannot distinguish the various types of the COVID-19 mutation, especially with regard to the recent variants of concern such as delta and lambda variants. This is because most of the existing X-ray images were taken from the early variants of COVID-19, whereby images of the newer mutations are still being added continuously to the dataset. Besides this, the proposed method still requires confirmation diagnosis from the medical practitioners. This approach is still in the early phase of development, whereby more clinical testings need to be performed before it is suitable for mass usage.

6. Conclusions

In conclusion, this study managed to prove the effectiveness of the Residual-Shuffle-Net in detecting COVID-19 cases based on the X-ray imaging input. The main novelty of the proposed network lies in its lightweight residual-shuffle unit that combines the split and shuffle unit with a residual skip connection. This architecture allows the network to better learn the distinguishing features between the COVID-19 cases and other class categories. In addition, the network is also embedded with a spatial pyramid pooling unit that enables it to extract multi-scale features, which is important for detecting COVID-19 cases of various severity levels. The Residual-Shuffle-Net returned the best performance for five performance metrics, which are A C C ¯ , S E N ¯ , S P E ¯ , P R E ¯ , and F1-Score with 0.97390, 0.97390, 0.98695, 0.97403, and 0.97387, respectively. Although the method by Panahi et al. uses the lowest total number of parameters, Residual-Shuffle-Net is still considered as a lightweight model with just 2,090,491 parameters. The classification performance can be further improved by considering an attention mechanism that allows the network to focus on selected regions of interest, rather than treating the whole image as equal. Besides this, a separable convolution approach can also be implemented to reduce memory usage.

Author Contributions

Conceptualization, M.A.Z., S.R.A. and N.H.Z.; software, M.A.Z. and S.R.A.; formal analysis, M.A.Z., S.R.A. and N.H.Z.; writing—original draft preparation, M.A.Z., S.R.A. and N.H.Z.; writing—review and editing, M.A.Z., S.R.A., N.H.Z. and M.I.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Universiti Kebangsaan Malaysia with Grant No. GUP-2019-008 and Ministry of Higher Education Malaysia with Grant No. FRGS/1/2019/ICT02/ UKM/02/1.

Institutional Review Board Statement

The original COVID-19 radiography images were collected in accordance with the Declaration of Helsinki, and the ethics protocol was approved by the ethical board of Arnau de Vilanova Hospital in Valencia Region with the project code of CElm: 12/2020, which was approved in December 2020.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SPPSpatial Pyramid Pooling
CNNConvolutional Neural Networks
COVID-19Coronavirus Disease 2019
RT-PCRReverse Transcription Polymerase Chain Reaction
CTComputed Tomography
GANGenerative Adversarial Network
Residual-Shuffle-NetResidual-Shuffle Network
Leaky ReLULeaky Rectified Linear Unit
BIMCVMedical Imaging Databank of the Valencia Region
RSNARadiological Society of North America
DICOMDigital Imaging and Communications in Medicine
PNGPortable Network Graphics
ROCReceiver Operating Characteristic

References

  1. Lancet, T. COVID-19: The worst may be yet to come. Lancet 2020, 396, 71. [Google Scholar] [CrossRef]
  2. da Silva, J.C.; Félix, V.B.; Leão, S.A.B.F.; Trindade Filho, E.M.; Scorza, F.A. New Brazilian Variant Of The Sars-Cov-2 (P1) of Covid-19 in Alagoas State. Braz. J. Infect. Dis. 2021, 25, 101588. [Google Scholar] [CrossRef]
  3. Vallee, A.; Chan-Hew-Wai, A.; Bonan, B.; Lesprit, P.; Parquin, F.; Catherinot, É.; Choucair, J.; Billard, D.; Amiel-Taieb, C.; Camps, È.; et al. Oxford—AstraZeneca COVID-19 vaccine: Need of a reasoned and effective vaccine campaign. Public Health 2021, 196, 135–137. [Google Scholar] [CrossRef] [PubMed]
  4. McGrath, J.; Kenny, C.; Smyth, H.; McGinty, T.; Sheehan, G.; Gaine, S.; McCullagh, B.; MacMahon, P.; Egan, J.; Cotter, A. A multidisciplinary evaluation of suspected, non-confirmed cases of COVID-19 including chest CT, as compared to World Health Organization recommendations. Clin. Radiol. 2021, 76, 384–390. [Google Scholar] [CrossRef] [PubMed]
  5. Bruzzone, B.; De Pace, V.; Caligiuri, P.; Ricucci, V.; Guarona, G.; Pennati, B.M.; Boccotti, S.; Orsi, A.; Domnich, A.; Da Rin, G.; et al. Comparative diagnostic performance of rapid antigen detection tests for COVID-19 in a hospital setting. Int. J. Infect. Dis. 2021, 107, 215–218. [Google Scholar] [CrossRef] [PubMed]
  6. Bouassa, R.S.M.; Veyer, D.; Péré, H.; Bélec, L. Analytical performances of the point-of-care SIENNA™ COVID-19 Antigen Rapid Test for the detection of SARS-CoV-2 nucleocapsid protein in nasopharyngeal swabs: A prospective evaluation during the COVID-19 s wave in France. Int. J. Infect. Dis. 2021, 106, 8–12. [Google Scholar] [CrossRef]
  7. Book COVID—19 Drive—Thru, Clinic and Home Screening Test Services Online. 2021. Available online: https://www.doctoroncall.com.my/medicine/coronavirus-covid-19-test-kit (accessed on 21 June 2021).
  8. Abdani, S.R.; Zulkifley, M.A.; Zulkifley, N.H. A Lightweight Deep Learning Model for Covid-19 Detection. In Proceedings of the IEEE Symposium on Industrial Electronics & Applications (ISIEA), Shah Alam, Malaysia, 17–18 July 2020. [Google Scholar]
  9. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (covid-19) using X-ray images and deep convolutional neural networks. arXiv 2020, arXiv:2003.10849. [Google Scholar]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. CoRR 2015. Available online: http://xxx.lanl.gov/abs/1512.03385 (accessed on 6 June 2021).
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. In Proceedings of the Computer Vision—ECCV, Zurich, Switzerland, 5–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 346–361. [Google Scholar]
  12. Abdani, S.R.; Zulkifley, M.A. DenseNet with Spatial Pyramid Pooling for Industrial Oil Palm Plantation Detection. In Proceedings of the 2019 International Conference on Mechatronics, Robotics and Systems Engineering, Bali, Indonesia, 4–6 December 2019. [Google Scholar]
  13. Zulkifley, M.A.; Trigoni, N. Multiple-Model Fully Convolutional Neural Networks for Single Object Tracking on Thermal Infrared Video. IEEE Access 2018, 6, 42790–42799. [Google Scholar] [CrossRef]
  14. Sumbul, G.; Demir, B. A Novel Multi-Attention Driven System for Multi-Label Remote Sensing Image Classification. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 5726–5729. [Google Scholar] [CrossRef] [Green Version]
  15. Shao, G.; Chen, Y.; Wei, Y. Deep Fusion for Radar Jamming Signal Classification Based on CNN. IEEE Access 2020, 8, 117236–117244. [Google Scholar] [CrossRef]
  16. Zulkifley, M.A.; Abdani, S.R.; Zulkifley, N.H. Pterygium-Net: A deep learning approach to pterygium detection and localization. Multimed. Tools Appl. 2019, 78, 34563–34584. [Google Scholar] [CrossRef]
  17. Lal, H.; Nguyen, T.; Li, H.; Abbasi, A.A.; Lone, K.J.; Zhao, Z.; Zaib, M.; Chen, A.; Duong, T.Q. Machine-learning classification of texture features of portable chest X-ray accurately classifies COVID-19 lung infection. Biomed. Eng. Online 2020, 19, 88. [Google Scholar]
  18. Sverzellati, N.; Ryerson, C.J.; Milanese, G.; Renzoni, E.A.; Volpi, A.; Spagnolo, P.; Bonella, F.; Comelli, I.; Affanni, P.; Veronesi, L.; et al. Chest X-ray or CT for COVID-19 pneumonia? Comparative study in a simulated triage setting. Eur. Respir. J. 2021. [Google Scholar] [CrossRef]
  19. Romanov, A.; Bach, M.; Yang, S.; Franzeck, F.C.; Sommer, G.; Anastasopoulos, C.; Bremerich, J.; Stieltjes, B.; Weikert, T.; Sauter, A.W. Automated CT Lung Density Analysis of Viral Pneumonia and Healthy Lungs Using Deep Learning-Based Segmentation, Histograms and HU Thresholds. Diagnostics 2021, 11, 738. [Google Scholar] [CrossRef]
  20. Pham, T.D. Classification of COVID-19 chest X-rays with deep learning: New models or fine tuning? Health Inf. Sci. Syst. 2021, 9, 1–11. [Google Scholar] [CrossRef]
  21. Pandit, M.K.; Banday, S.A.; Naaz, R.; Chishti, M.A. Automatic detection of COVID-19 from chest radiographs using deep learning. Radiography 2021, 27, 483–489. [Google Scholar] [CrossRef] [PubMed]
  22. Panwar, H.; Gupta, P.; Siddiqui, M.K.; Morales-Menendez, R.; Singh, V. Application of deep learning for fast detection of COVID-19 in X-rays using nCOVnet. Chaos Solitons Fractals 2020, 138, 109944. [Google Scholar] [CrossRef] [PubMed]
  23. Kikkisetti, S.; Zhu, J.; Shen, B.; Li, H.; Duong, T.Q. Deep-learning convolutional neural networks with transfer learning accurately classify COVID-19 lung infection on portable chest radiographs. PeerJ 2020, 8, e10309. [Google Scholar] [CrossRef] [PubMed]
  24. Apostolopoulos, I.D.; Mpesiana, T.A. Covid-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition; Technical Report; University of Oxford: Oxford, UK, 2014. [Google Scholar]
  26. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  27. Szegedy, C.; Wei, L.; Yangqing, J.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  28. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
  29. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  30. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  31. Loey, M.; Smarandache, F.; Khalifa, M.E.N. Within the Lack of Chest COVID-19 X-ray Dataset: A Novel Detection Model Based on GAN and Deep Transfer Learning. Symmetry 2020, 12, 651. [Google Scholar] [CrossRef] [Green Version]
  32. Dewi, C.; Chen, R.C.; Liu, Y.T.; Jiang, X.; Hartomo, K.D. Yolo V4 for Advanced Traffic Sign Recognition With Synthetic Training Data Generated by Various GAN. IEEE Access 2021, 9, 97228–97242. [Google Scholar] [CrossRef]
  33. Zulkifley, M.A.; Abdani, S.R.; Zulkifley, N.H. COVID-19 Screening Using a Lightweight Convolutional Neural Network with Generative Adversarial Network Data Augmentation. Symmetry 2020, 12, 1530. [Google Scholar] [CrossRef]
  34. Ucar, F.; Korkmaz, D. COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from X-ray images. Med. Hypotheses 2020, 140, 109761. [Google Scholar] [CrossRef] [PubMed]
  35. Iandola, F.N.; Moskewicz, M.W.; Ashraf, K.; Han, S.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50× fewer parameters and <1 MB model size. CoRR 2016. Available online: http://xxx.lanl.gov/abs/1602.07360 (accessed on 6 June 2021).
  36. Muad, A.M.; Zaki, S.K.M.; Jasim, S.A. Optimizing Hopfield Neural Network for Super-Resolution Mapping. J. Kejuruter. 2020, 32, 91–97. [Google Scholar] [CrossRef]
  37. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Comput. Methods Programs Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  38. Panahi, A.H.; Rafiei, A.; Rezaee, A. FCOD: Fast COVID-19 Detector based on deep learning techniques. Inform. Med. Unlocked 2021, 22, 100506. [Google Scholar] [CrossRef] [PubMed]
  39. Mahmud, T.; Rahman, M.A.; Fattah, S.A. CovXNet: A multi-dilation convolutional neural network for automatic COVID-19 and other pneumonia detection from chest X-ray images with transferable multi-receptive feature optimization. Comput. Biol. Med. 2020, 122, 103869. [Google Scholar] [CrossRef]
  40. Gilanie, G.; Bajwa, U.I.; Waraich, M.M.; Asghar, M.; Kousar, R.; Kashif, A.; Aslam, R.S.; Qasim, M.M.; Rafique, H. Coronavirus (COVID-19) detection from chest radiology images using convolutional neural networks. Biomed. Signal Process. Control 2021, 66, 102490. [Google Scholar] [CrossRef]
  41. Hussain, E.; Hasan, M.; Rahman, M.A.; Lee, I.; Tamanna, T.; Parvez, M.Z. CoroDet: A deep learning based classification for COVID-19 detection using chest X-ray images. Chaos Solitons Fractals 2021, 142, 110495. [Google Scholar] [CrossRef]
  42. Zhu, J.; Shen, B.; Abbasi, A.; Hoshmand-Kochi, M.; Li, H.; Duong, T.Q. Deep transfer learning artificial intelligence accurately stages COVID-19 lung disease severity on portable chest radiographs. PLoS ONE 2020, 15, e0236621. [Google Scholar] [CrossRef]
  43. Wong, A.; Lin, Z.Q.; Wang, L.; Chung, A.G.; Shen, B.; Abbasi, A.; Hoshmand-Kochi, M.; Duong, T.Q. Towards computer-aided severity assessment via deep neural networks for geographic and opacity extent scoring of SARS-CoV-2 chest X-rays. Sci. Rep. 2021, 11, 9315. [Google Scholar] [CrossRef] [PubMed]
  44. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet V2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  45. Abdani, S.R.; Zulkifley, M.A.; Zulkifley, N.H. Group and Shuffle Convolutional Neural Networks with Pyramid Pooling Module for Automated Pterygium Segmentation. Diagnostics 2021, 11, 1104. [Google Scholar] [CrossRef]
  46. Vaya, M.d.l.I.; Saborit, J.M.; Montell, J.A.; Pertusa, A.; Bustos, A.; Cazorla, M.; Galant, J.; Barber, X.; Orozco-Beltrán, D.; Garcia, F.; et al. BIMCV COVID-19+: A large annotated dataset of RX and CT images from COVID-19 patients. arXiv 2020, arXiv:2006.01174. [Google Scholar]
  47. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
  48. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Samples of removed X-ray images. The first two samples are removed because they were captured from the side view, while the third sample is removed because of the incomplete information on the frontal chest X-ray image.
Figure 1. Samples of removed X-ray images. The first two samples are removed because they were captured from the side view, while the third sample is removed because of the incomplete information on the frontal chest X-ray image.
Diagnostics 11 01497 g001
Figure 2. Architecture of the Residual-Shuffle unit.
Figure 2. Architecture of the Residual-Shuffle unit.
Diagnostics 11 01497 g002
Figure 3. Architecture of the spatial pyramid pooling unit.
Figure 3. Architecture of the spatial pyramid pooling unit.
Diagnostics 11 01497 g003
Figure 4. Samples of X-ray images for each category of COVID-19, normal and other types of pneumonia cases.
Figure 4. Samples of X-ray images for each category of COVID-19, normal and other types of pneumonia cases.
Diagnostics 11 01497 g004
Figure 5. Training accuracy for Residual-Shuffle-Net and all of its benchmark methods.
Figure 5. Training accuracy for Residual-Shuffle-Net and all of its benchmark methods.
Diagnostics 11 01497 g005
Figure 6. Training loss for Residual-Shuffle-Net and all of its benchmark methods.
Figure 6. Training loss for Residual-Shuffle-Net and all of its benchmark methods.
Diagnostics 11 01497 g006
Figure 7. Validation loss for the Residual-Shuffle-Net and all of its benchmark methods.
Figure 7. Validation loss for the Residual-Shuffle-Net and all of its benchmark methods.
Diagnostics 11 01497 g007
Figure 8. Confusion matrix performance of the Residual-Shuffle-Net in identifying the three classes of COVID-19, normal and other types of pneumonia cases using frontal chest X-ray images.
Figure 8. Confusion matrix performance of the Residual-Shuffle-Net in identifying the three classes of COVID-19, normal and other types of pneumonia cases using frontal chest X-ray images.
Diagnostics 11 01497 g008
Figure 9. Receiver operating characteristic (ROC) curves for the Residual-Shuffle-Net with its respective area under the curve values.
Figure 9. Receiver operating characteristic (ROC) curves for the Residual-Shuffle-Net with its respective area under the curve values.
Diagnostics 11 01497 g009
Table 1. Overall architecture of the Residual-Shuffle-Net.
Table 1. Overall architecture of the Residual-Shuffle-Net.
No.LayerOutput SizeFilter SizeKernel SizeStride
1Convolution256 × 25683 × 31 × 1
2Max Pooling128 × 128-2 × 22 × 2
3Convolution128 × 128163 × 31 × 1
4Max Pooling64 × 64-2 × 22 × 2
5Residual-Shuffle (1)64 × 6432-16-323 × 3 − 1 × 1 − 3 × 31 × 1
6Max Pooling32 × 32-2 × 22 × 2
7Residual-Shuffle (2)32 × 3264-32-643 × 3 − 1 × 1 − 3 × 31 × 1
8Max Pooling16 × 16-2 × 22 × 2
9Residual-Shuffle (3)16 × 16128-64-1283 × 3 − 1 × 1 − 3 × 31 × 1
10Max Pooling8 × 8-2 × 22 × 2
11Residual-Shuffle (4)8 × 8256-128-2563 × 3 − 1 × 1 − 3 × 31 × 1
12Spatial Pyramid Pooling8 × 8128-128-1282 × 2 − 4 × 4 − 6 × 61 × 1
13Global Average Pooling1 × 1-8 × 81 × 1
14Dense + SoftMax-31 × 11 × 1
Table 2. Hyper-parameter settings for the Residual-Shuffle-Net.
Table 2. Hyper-parameter settings for the Residual-Shuffle-Net.
CriteriaHyper-Parameter Setting
Batch size64
Training epoch80
Backpropagation methodAdam optimizer
Input image size256 × 256 pixels
Optimizer learning rate0.0001
Optimizer momentums β 1 = 0.9, β 2 = 0.999
Loss functioncategorical cross-entropy
Labeling formatOne-hot encoded
Table 3. Performance results of the Residual-Shuffle-Net and the benchmark methods.
Table 3. Performance results of the Residual-Shuffle-Net and the benchmark methods.
Method ACC ¯ SEN ¯ SPE ¯ PRE ¯ F 1 Score Total ParametersTrainable Parameters
Hussain et al. [41]0.786950.786950.893470.819980.777853,798,0833,798,083
Narin et al. [9]0.837650.761550.880780.83760.7271823,567,29923,514,179
Ozturk et al. [48]0.846290.846290.923180.925720.819111,167,3631,164,143
Ucar et al. [34]0.928410.928410.964200.931300.928661,078,2111,078,211
Mahmud et al. [39]0.936860.936860.968430.940250.936091,338,2911,164,143
Loey et al. [31]0.940580.940580.970290.943870.9406711,192,00311,182,275
Panahi et al. [38]0.944060.944060.972030.946740.9438585,32183,849
Pandit et al. [21]0.948580.948580.974290.948590.9483433,609,53933,609,539
Panwar et al. [22].0.957000.957000.9785000.957460.9569214,747,71514,747,715
Khan et al. [37]0.962470.962470.981230.963340.9624788,054,13987,958,763
Abdani et al. [8]0.963950.963950.981970.964540.963881,862,331859,883
Gilanie et al. [40]0.968680.968680.984340.968840.968632,339,2672,339,267
Residual-Shuffle-Net0.973900.973900.986950.974030.973872,090,4912,087,275
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zulkifley, M.A.; Abdani, S.R.; Zulkifley, N.H.; Shahrimin, M.I. Residual-Shuffle Network with Spatial Pyramid Pooling Module for COVID-19 Screening. Diagnostics 2021, 11, 1497. https://doi.org/10.3390/diagnostics11081497

AMA Style

Zulkifley MA, Abdani SR, Zulkifley NH, Shahrimin MI. Residual-Shuffle Network with Spatial Pyramid Pooling Module for COVID-19 Screening. Diagnostics. 2021; 11(8):1497. https://doi.org/10.3390/diagnostics11081497

Chicago/Turabian Style

Zulkifley, Mohd Asyraf, Siti Raihanah Abdani, Nuraisyah Hani Zulkifley, and Mohamad Ibrani Shahrimin. 2021. "Residual-Shuffle Network with Spatial Pyramid Pooling Module for COVID-19 Screening" Diagnostics 11, no. 8: 1497. https://doi.org/10.3390/diagnostics11081497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop