CLASSIFIED COVID-19 BY DENSENET121-BASED DEEP TRANSFER LEARNING FROM CT-SCAN IMAGES
Walat R. Ibrahima, a*, Mayyadah R. Mahmood a
a,* Faculty of Science, University of Zakho, Zakho, Kurdistan Region, Iraq – (walat.ebrahim@stud.uoz.edu.krd, mayyadah.mahmood@uoz.edu.krd )
Received: 21 June, 2023 / Accepted: 22 Oct., 2023 / Published: 19 Dec, 2023. https://doi.org/10.25271/sjuoz.2023.11.4.1166
SARS-CoV-2 or COVID-19 is the worldwide epidemic caused by Coronavirus illness in 2019 caused severe acute respiratory syndrome (SARS) and was found for the first time in Wuhan, which is located in China. In March 2020, the World Health Organization (WHO) gave the disease the name COVID-19 and predicted that it would spread over the world as an epidemic (Cantarero et al. 2021;De Anda-Suarez et al. 2022). COVID-19 negatively affects the health of people, especially the elderly, people with existing medical issues, and people whose immune systems are already impaired. By the middle of July 2020, more than 13 million individuals had contracted COVID-19, and more than 570,000 people had already died from the epidemic (Sakib et al. 2020). COVID-19's most pervasive clinical characteristics are a dry cough, fever, and tiredness (Yan et al. 2021). patients often have the most severe symptoms of pneumonia after they have been infected with the coronavirus (Rahman, Khandakar, and Member 2021). Despite this, it is possible that some COVID-19 cases may exhibit no symptoms at all. In severe circumstances, intensive care unit treatment or oxygen support therapy may be required (Munusamy et al. 2021).
"Reverse Transcriptase Polymerase Chain Reaction (RT-PCR)" real-time testing was utilized as a diagnostic method to halt the COVID-19 illness from spreading after the genome of SARS-CoV-2 was sequenced (Mondal et al. 2022).
The present pandemic state does not support the low-sensitivity RT-PCR test because RT-PCR takes considerable time and has a significant number of false-negative results (Ibrahim and Mahmood 2023). A test result that is falsely negative can result in an illness spreading extensively (Tabik et al. 2020). Sometimes, infected individuals may not be identified right away and may not receive the proper care (Shah et al. 2021). Consequently, diagnostic imaging, especially chest X-rays and chest computerized tomographic (CT) (Dai et al. 2020), is frequently used in combination with other tests to aid in earlier COVID-19 screening and treatment (Tai et al. 2021).
In light of this, lesions with ground-glass opacities (GGO) and pulmonary fibrosis are COVID-19 lung imaging features (Dong et al. 2021). CT scans are extremely sensitive, so they help assess the severity of the illness; however, their main limitation is the time required for diagnosis; even radiology specialists require approximately 21.5 minutes to parse each case's test results (Wu et al. 2021). For these reasons, automated diagnosis technologies, such as models of deep learning (DL), have been created to screen for COVID-19 in CT images recently (He et al. 2020). According to several recent studies, CNNs are a particular kind of DL technique (Hussain et al. 2021), having a high capability for COVID-19 identification (Shi et al. 2021). With an acceptable performance prediction using COVID-19-infected patients' chest CT scan images (Mohammed et al. 2020). These techniques can be applied by fine-tuning the weights or knowledge gained from the pre-trained CNN model on the large dataset and adapting it to perform well on specific tasks with limited data and achieve better results compared to training from scratch (Shukla et al. 2020). As in the dataset of COVID-19 for the diagnosis of COVID-19 from CT-scan chest images (Sarker et al. 2020). A densely connected convolutional network (DenseNet), a specialized CNN technique, has the advantages of mitigating the problem of vanishing gradients, improving feature reuse, and decreasing parameter usage. Despite these advantages, it has some limitations, such as computational complexity and the substantial amount of GPU memory; furthermore, DenseNet-121 is prone to overfitting (Albelwi 2022).
Therefore, the main focus of this study is on using pre-trained DenseNet-121 models of DL to automatically identify and diagnose COVID-19 using chest CT-scan data.
The following is a list of the important contributions that this article has made:
1. The DenseNet121-based Deep Transfer Learning (DTL) model was modified by the exclusion of the last layer (i.e., the top) and the addition of three dense layers, including Dense128, Dense64, and Dense2, to the model to enhance the model and achieve high accuracy.
2. The model is used as an approach for automatically detecting and diagnosing COVID-19 patients on lung CT-scan images.
3. To ensure a precise and effective evaluation of the model, the image datasets go through a pre-processing stage.
4. Employing a modified model on two CT-scan image datasets offers a faster alternative to the widely used RT-PCR testing kit.
The remaining parts of this article have been organized into the following sections and order: The relevant works are listed in section 2. The suggested DTL model and data description are described in Section 3. In section 4 the findings and their analysis are discussed. Finally, section 5 presents the points of conclusion.
Various study projects have been carried out extensively and fast in order to build AI strategies for combating the COVID-19 worldwide pandemic. Some relevant research on using a chest CT scan to diagnose and classify COVID-19 is reviewed.
Wang, Liu, and Dou (2020) suggested a new joint learning framework to use different datasets to detect COVID-19 disease. The network has been split into two parts. The first part has a lightweight structure and four separate layers of convolution. The second part has more densely packed learning blocks that connect. The COVID-CT dataset and SARS-CoV-2 were used in this work. The work results show that the SARS-CoV-2 dataset outperformed the COVID-CT dataset with an accuracy of 90%, 95% precision, 85% recall, and a 90% F1-score.
Silva et al. (2020) presented a model called Efficient CovidNet applied to lung CT-scan radiography to identify COVID-19 patients, which uses voting techniques and cross-dataset analytics. Experimental results obtained from SARS-COV-2 CT-Scan datasets show a sensitivity of 98.80%, a precision of 99.2%, and an accuracy of 98.99%.
Seum et al. (2020) compared 12 pre-trained CNNs models in classifying COVID-19 patients by utilizing an available dataset (the SARS-COV-2 CT-Scan). This work was carried out in two phases: one without segmentation and another with segmentation. The study results showed in the first phase that the performance of the DenseNet169 model is higher than other models in terms of F1-scores, and accuracy with 88.60% and 89.31%, respectively. In the second phase, DenseNet201 and ResNet18 both obtained an accuracy of 89.92%, implying that using classification after segmentation improves the efficiency of a classification model.
Ramzy, Sherin, and Karma (2021) suggested (COV-CAF), which is a computer-aided framework to identify COVID-19 cases through 3D CT-scan pictures. The researchers combined both DL and traditional techniques in COV-CAF, which has two stages: one for preparatory work and the second one for extracting and classifying. The first stage is focused on 3D CT volumes and proposes an effective cut selection method for selecting useful CT slices, and the second stage uses a recently developed CNN called Norm-VGG16 to extract features automatically. They employed two datasets, the SARS-CoV-2 CT-scan and MosMedData. The findings obtained on the SARS-COV-2 CT scan were 97.59% accuracy and 98.41% sensitivity; whereas, it gained an overall accuracy of 97.76% and an average sensitivity of 96.73% on the MosMedData dataset.
Biswas et al. (2021) conducted a study to use effective transfer learning (TL) algorithms to construct an accurate COVID-19 prediction model on SARS-CoV-2 CT datasets. Initially, they utilized three kinds of (DL) models separately: ResNet50, Xception, and VGG-16, and then combined these three models to increase the model's prediction accuracy. The suggested model has a high F1-score of 99% and an accuracy of 98.79%.
Jaiswal et al. (2021) developed a DTL model to detect COVID-19 infection. It is a DenseNet201 – based model, which was trained before with a massive dataset known as the ImageNet dataset. The provided method produced results with an accuracy of 0.9625, 96.2% recall, 96.2% precision, and a specificity of 96.21%.
Hasan et al. (2021), using CNN, suggested a method to find COVID-19 persons via images from lung scanning. They use the newest updated CNN model (DenseNet-121). The findings showed a satisfactory level of performance for the COVID-19 prediction, with an accuracy of 92% and a recall of 95%.
Horry et al. (2020) built a mechanism for COVID-19 detection based on the DL by utilizing the VGG19 algorithm on various kinds of medical photographs (CT, x-ray, ultrasound), achieving a precision of 84% for CT scans, 86% for x-rays, and 100% for ultrasounds, and this system was able to identify COVID-19 in 100% of the ultrasound images.
Islam and Matin (2020) employed LeNet-5 CNN networks after a simple CNN model to predict COVID-19 from thoracic images. For training purposes, the obtained results were an accuracy of 86.06%, an F1-score of 87%, 89% recall, 85% precision, and a 0.86 ROC.
Pham (2020) showed results from 16 pre-trained CNN models of COVID-19 classification. According to this study, TL led to higher classification rates than data augmentation. ResNet-18 had the best sensitivity (98%), MobileNet-v2 earned the maximum accuracy (95%), and F1-score (96%), and DenseNet-201 had the highest degree of specificity (96%). This research utilized various networks, including ShuffleNet, ResNet-18, DenseNet-201, and MobileNet-v2.
Garain et al. (2021), for COVID-19 screening, created a three-layer DCSNN using CT images. The proposed method obtained a 99% F1 score for the potential-based model. For chest CT images, the suggested SNN-based model outperforms CNN models, but it takes longer to train. Furthermore, in comparison to older DL models, the current method is more efficient.
Chen et al. (2021) conducted a study on automatically identifying positive COVID-19 patients using chest CT scan radiography, a prototype network that has been pre-trained by a momentum-contrasting learning strategy was applied. They obtained values of 87% of accuracy, 93.2% of AUC, 88.5% of precision, and recall 87.4%.
Huang and Liao (2022) proposed LightEfficientNetV2 model and fine-tunes seven CNN techniques involving EfficientNet-B0, InceptionV3, Xception, MobileNetV2, DenseNet121, EfficientNetV2, and ResNet50V2 to diagnose COVID-19 cases on three different datasets (NIH Chest X-rays, SARS-CoV-2, and COVID-CT). The proposed LightEfficientNetV2 model obtained an accuracy of 99.47% on the SARS-CoV-2 dataset and 88.67% on COVID-CT.
Kathamuthu et al. (2023) used several deep transfer learning-based CNN approaches including InceptionV3, Resnet50, Xception, VGG16, VGG19, and Densenet121 to screen COVID-19 in chest CT images and achieved an accuracy of 98% with VGG16 model on SARS-CoV-2 CT-scan dataset.
In light of the limited number of COVID-19 radiography pictures in our dataset, it is difficult to develop a DL model in this situation because DL approaches need a huge dataset to start from scratch. Transfer learning (TL) could be utilized as an effective solution to this issue because TL techniques use the knowledge gained in one field to carry out tasks in other fields. For instance, when initializing DL models with weights obtained from the ImageNet (Fei-Fei, Deng, and Li 2010), as the ImageNet is a massive dataset comprised of 14 million images from various resources, by that time it has learned several features for diverse images, so these newly acquired features can then be applied to another dataset. For this reason, the weights gained from pre-trained DenseNet-121 on the ImageNet dataset were employed for the SARS-COV2 and the COVID19-CT datasets in this study.
The study system's processes start with two distinct
datasets that go through data pre-processing techniques before being split into
training, validation, and test sets. Hence, the data are ready to fit into a
modified model. Figure 1 shows the general study diagram.
Figure 2: Non-COVID (left) and COVID-19 (right).
Figure 3: CT-scan images of different sizes.
Figure 4: The differences in contrast CT scan images.
Dataset |
SARS-COV-2 Ct-Scan |
COVID19-CT |
# Patients |
120 |
387 |
# CT-scan images |
2481 |
746 |
Train sets images |
1983 |
596 |
Valid sets images |
249 |
75 |
Test sets images |
249 |
75 |
The term used to describe all the changes made to initial data before feeding it into machine learning (ML) or DL algorithms named pre-processing. All images are resized from different resolutions to 224 x 224 resolution using the OpenCV library interpolation function INTER_AREA. Additionally, the pixels' values might range from 0 to 255. If a DL model is applied to the image as-is, computing high numeric values may become more challenging. We can minimize this by adjusting the values to be
between 0 and 1, also known as normalization, which is a significant stage to guarantee that each input parameter (in this case, a pixel) has a uniform data distribution. This makes convergence faster while training the network.
In data analysis, data augmentation is utilized to raise the number of training data sets either by producing drastically altered copies of already existing data or by developing synthetic data from earlier existing data. This is accomplished by one of these two methods. This technique enhances the generalization performance and classification accuracy of the model while also serving as a regularise to reduce overfitting (Tang, Yuan, and Zhu 2020). Table 2 lists the parameters used for image augmentation in the current study.
In this study, a DenseNet121-based Deep Transfer Learning model-COVID-19 is modified and called (DDTL-COV) to classify COVID-19-infected patients from chest CT-scan images. The DDTL-COV was trained using TL, with the pre-trained.
DenseNet121 weights on the ImageNet dataset used as the initial weights. Following this, the model was fine-tuned on the COVID-19 dataset. Figure 5 shows the architecture of the DDTL-COV model, which employs DenseNet121 for accurately classifying COVID-19 patients.
The main parts of the DDTL-COV model are the feature extraction and classification parts.
In the feature extraction part, the input layer of the CNN is defined by the shape of the input image (224,224,3) followed by a convolutional layer of 3 kernels with size (3 x 3) with padding 'same'. The DDTL-COV model was trained using TL with the pre-trained DenseNet121 weights on the ImageNet dataset.
values |
Functions |
|
Rotation-range |
360 |
The degree range of rotations randomly |
Width-shift-range |
0.2 |
The range of shifting horizontally is randomly selected. |
Height-shift-range |
0.2 |
The range of shifting vertically is randomly selected. |
True |
The input randomly flips in the horizontal direction. |
|
True |
The input randomly flips in the vertical direction. |
|
Zoom range |
0.2 |
The randomized zooming range. |
DenseNet is a logical extension of ResNet and boosts performance by merging the feature maps at each layer with those of the layer that came before it inside a dense block. This permits the layers of the network that come after them to directly use the features of the layers that came before them, which promotes the reuse of features across the network as shown in Figure 6, which facilitates solving the vanishing gradient issue and also minimizes the parameters' number.
Depending on the network version, the DenseNet network is structured into four blocks of layers, each of which has a particular number of layers. For instance, the DenseNe-169 has 4 blocks with 6, 12, 32, and 32 layers, respectively, and DenseNe-121 was used in this study; it is shown in Error! Reference source not found. 7 as having four blocks with a total of 6, 12, 24, and 16 layers, respectively.
Layers for Batch Normalization, Relu functions, consonant layers, and Pooling layers are put between these blocks.
The DenseNet121 architecture implements direct interconnections between all antecedent and subsequent layers in furtherance of connectivity.
The feature concatenation's mathematical formulation is as follows:
(1)
Here, is a transition that is not linear. It consists of BN and ReLU (Yu et al. 2020). And mathematically can be represented as in Eq (2)
(2)
And followed by a convolution layer of (3 x 3) windows, is the feature map string for layers 0 to l-1 that has been concatenated into one tensor for simplicity of use, for down sampling. In the model architecture, transition layers made up of a BN layer are used to build and separate dense blocks, a 2×2 average pooling layer at the end, and a (1×1) convolution layer between them.
The growth rate, represented by the hyperparameter k in DenseNet121, indicates how the dense design achieves superior outcomes. The architecture of DenseNet121 evolves by observing feature maps as the network's global state. Therefore, despite having a reduced growth rate, it performs efficiently. Every succeeding layer of the network is able to access all of the feature maps that were generated by the layer that came before it. Additionally, each layer adds k feature maps to the overall state, resulting in a total input feature map count across the lth layer known as FM:
(3)
Here, k0 is the symbol for the channels on the input layer. To improve the efficiency of computation, an (1x1) convolution layer is added before every (3x3) convolution, decreasing the number of inputs for the feature map, which is typically more than the number of output feature maps k. The bottleneck layer, which is a (1x1) convolution layer, is inserted and creates 4k feature maps.
In the classification part, the fully connected layer at the end pre-trained DenseNet121model (i.e., the top) is excluded (include_top = False), an additional three dense layers including (Dense128_Relu_BN_Dropout 0.3), (Dense64_Relu_BN_Dropout 0.25), and last dense (2) is added to the model after GobalAveragePooling2D layer and BN, hence, Dense () indicates the layer's neurons, BN represents the Batch Normalization layer, Relu is the activation function, Dropout () is used to regularize CNNs in addition, it overcomes the over-fitting issue. For binary classification, the last Dense layer along the sigmoid is utilized as the output layer since the sigmoid is an activation function that turns non-normalized outputs into binary 0s and 1s, where (0 = COVID-19, 1= Non_COVID).
The sigmoid function can be expressed using the formula below:
The history of the model is presented in Table 3.
Layer(type) |
Output shape |
Parameter # |
|||
Input layer |
(None,224,224,3) |
0 |
|||
Conv2D |
(None,224,224,3) |
84 |
|||
Densenet121 (Functional) |
(None, None, None, 1024) |
7037504 |
|||
GlobalAveragePooling2d |
(None, 1024) |
0 |
|||
Batch normalization |
(None, 1024) |
4096 |
|||
Dense |
(None, 128) |
131200 |
|||
Batch normalization_1 |
(None, 128) |
512 |
|||
Dropout |
(None, 128) |
0 |
|||
Dense 1 |
(None, 64) |
8256 |
|||
Batch normalization_2 |
(None, 64) |
256 |
|||
Dropout 1 |
(None, 64) |
0 |
|||
Dense 2 |
(None, 2) |
130 |
|||
Total parameters: 7,182,038 Trainable parameters: 7,095,958 Non-trainable parameters: 86,080
|
The model completed training around 20 minutes with 75 epochs and a batch size of 32. The binary cross-entropy is utilized as a loss function that is capable of being lowered using the backpropagation approach while the model is being trained. The following equation demonstrates the loss function:
Here, the number of training samples is represented by n, the number of classes by k, the model parameter is θ, the actual level of the ith training sample is , and the output for the ith training sample is at the kth node.
The model is optimized by using the Adam optimizer, which is a subset of the stochastic gradient descent approach that is based on adaptive estimation of the first moment estimate , second-moment estimate , with α = 0.003 as the initial learning rate. updating the model weight θ with Adam's optimizer is illustrated in the below equation:
(6)
Where
(7)
(8)
Here, represents the weights of the gradient concerning the loss function, the weight at time t−1 is indicated by θt−1, the weight at time t is represented by θt, β1, and β2 are hyperparameters, the default value of β1 = 0.9 and β2 = 0.999 were used, With the ReduceLROnPlateau callback, the learning rate is decreased by a factor of 0.70 until it reaches a lower limit (min lr = 0.0001) when the validation loss remains constant for about 5 epochs (patience = 5).
This section demonstrates the efficacy of the suggested model in identifying if a person has COVID-19 illness or not by using the datasets SARS-CoV-2 CT-scan and COVID19-CT. Initially, we analyzed SARS-CoV-2 CT results then COVID19-CT and compared them. Additionally, each dataset's results were then compared with those from earlier studies and research using the same datasets. The pretraining model that is used in this study is DenseNet-121.
In this study, a Windows 10 Pro 64-bit operating system has been used with the Anaconda compiler supported by Python 3.7 and the TensorFlow library and Keras library, which both support GPU hardware acceleration on some COVID-19 CT scans for DL and CNN purposes, probably for achieving perfect infection detection. And, as for hardware, the following computer and key specifications were used: I5-10400F (10th Generation Intel Core i5 Desktop Processor with 6 cores and 12 threads), GeForce RTX 2070 with 8GB of vRAM, and 16 GB of DDR4 at 2666 MHz (RAM/Memory).
Three sets are created from the dataset for the performance analysis. 80% of the data used to train the model is in the training
set. 10% of the data makes up the validation set, which is used to fine-tune the model's hyperparameters and guard against overfitting. The remaining 10% of the data makes up the testing set, which is used to assess the model's ultimate performance. The loss of training and validation decreases gradually through each epoch, indicating that the model is learning and generalizing well. After epoch 55, the validation loss stops decreasing, which means that the model has reached its optimal performance. It is crucial to remember that the ideal number of epochs may change based on the model's complexity and the quantity of the dataset. The training and validation accuracy of the DDTL-COV model across various epochs are arranged in Figure 8, and the loss analyses of the model over the same number of epochs is illustrated in Figure 9.
Figure 10 shows a confusion matrix that is used to summarize the results of samples of each class that were successfully and mistakenly categorized. Based on the confusion matrix, the performance of the model was evaluated using four metrics: precision, recall, F1 score, and accuracy. The DDTL-COV model only produced one false positive case and no false negative cases, indicating a high score of accuracy of 99.6% and an F1-score of 100%. So, it will be of significance in the medical research field, as shown in Table 4.
Class\metrics |
precision |
recall |
F1-score |
Class 0\COVID-19 |
100% |
99% |
100% |
Class 1\non-COVID |
99% |
100% |
100% |
Accuracy overall 100% |
|||
Macro avg |
100% |
100% |
100% |
Weighted avg |
100% |
100% |
100% |
In this research, several innovative approaches developed for COVID-19 identification using the SARS-CoV-2 dataset are compared, as shown in Table 5. Some studies yielded fewer results, such as Panwar et al. (2020), which used the VGG model to obtain an accuracy of 95% and a recall of 94.4%. Wang, Liu, and Dou (2020) used the CNN model and achieved 95% accuracy and a 94.4% of F1-scor. As for Ramzy, Sherin, and Karma (2021), with VGG16, accuracy is 97.59%, recall is 98.41%, and so on. On the other hand, some other studies introduced various DL models and got superior results, as Silva et al. (2020) were capable to obtain an accuracy of 98.99%, a recall of 98.80%, and a precision of 99.20% through the utilization of Efficient CovidNet, and Biswas et al. (2021), by merging three previously trained CNNs of models (Xception, ResNet50, and VGG-16), were able to get an F1-score of 99% and an accuracy of 98.79%. The overall performance of the DDTL-COV model is superior to all other studies.
Table 5: A comparison of the DDTL-COV model with the modern techniques using the SARS-CoV-2 CT dataset (Acc: accuracy, Pre: precision, Rec: recall, F1-S: F1-score).
Ref. |
Method |
Acc. % |
Pre. % |
% |
% |
(Wang, Liu, and Dou 2020) |
CNN |
90 |
95 |
85 |
90 |
(Panwar et al. 2020) |
VGG |
95 |
/ |
94.4 |
/ |
(Silva et al. 2020) |
Efficient CovidNet |
98.99 |
99.20 |
98.80 |
/ |
(Seum et al. 2020) |
12 pre-training CNNs |
89.92 |
99.50 |
86.80 |
89.67
|
(Ramzy, Sherin, and Karma 2021) |
VGG16 |
97.59 |
/ |
98.41 |
/ |
(Biswas et al. 2021) |
Combine 3 pre-trained CNNs models (VGG-16, ResNet50, Xception) |
98.79 |
/ |
/ |
99 |
(Jaiswal et al. 2021) |
DenseNet201 |
96.21 |
96.25 |
96.2 |
96.2 |
(Hasan et al. 2021) |
DenseNet-121 |
92 |
/ |
95 |
/ |
(Huang and Liao 2022) |
LightEfficientNetV2 |
99.47 |
99.19 |
99.25 |
99.04 |
(Kathamuthu et al. 2023) |
VGG16 |
98 |
97.99 |
97.99 |
97.99 |
This work |
DDTL_COV (DenseNet-121) |
99.6 |
100 |
100 |
100 |
As in a previous experiment, the COVID19 CT dataset is partitioned randomly into three distinct sets of training, validation, and testing at 80, 10, and 10 percent, respectively. The model is trained over 75 epochs at a speed of 5 seconds per epoch, and its convergence is at a noticeably good speed. Figures 11 and 12 illustrate the accuracy of training and validation, as well as the training loss, and the validation loss of the DDTL-COV model on the COVID19-CT dataset, respectively.
Class \ Metrics |
Precision |
Recall |
F1-score |
Class 0\COVID-19 |
81% |
97% |
88% |
Class 1\non-COVID |
97% |
84% |
90% |
Accuracy overall 89% |
|||
Macro avg |
89% |
91% |
89% |
Weighted avg |
91% |
89% |
89% |
The overall performance of the COVID19-CT dataset in the manner of accuracy in comparison to that of the SARS-CoV-2 dataset is less as shown in Table 8. This is explained by the cross-source variability of the dataset's CT scans. The non-COVID-19 CT images were obtained from various sources and display a variety of results, making it challenging to differentiate between COVID-19 and other findings associated with lung disorders due to the possibility of visual manifestations overlapping. Another explanation is that the COVID19-CT dataset's CT images exhibit a significant contrast variation, varying spatial resolution, and other visual traits that can impair the model's capacity to extract more discriminative and generalizable features.
Method |
Acc. % |
Pre. % |
Rec. % |
F1-S % |
|
(Horry et al. 2020) |
VGG19 |
/ |
84 |
/ |
78
|
(Islam and Matin 2020) |
CNN + LeNet-5 |
86.06 |
85 |
89 |
87 |
(Pham 2020) |
Sixteen pre-trained CNNs |
95 % (MobileNet-v2) |
/ |
98 (ResNet-18)
|
96 (MobileNet-v2) |
(Garain et al. 2021) |
Three-layer DCSNN |
/ |
/ |
/ |
99 |
(Chen et al. 2021) |
prototypical network |
93.2 |
88.5 |
87.4 |
/ |
(Huang and Liao 2022) |
LightEfficientNetV2 |
88.67 |
87.28 |
87.43 |
87.55 |
This work |
DDTL_COV (DenseNet-121) |
89 |
89 |
91 |
89 |
Accuracy |
|
SARS-CoV-2 CT |
99.6% |
COVID19- CT |
89% |
This study suggests the DDTL-COV model, a deep transfer learning model based on DenseNet121, using weights from the ImageNet dataset to categorize patients on CT scans as either COVID or non-COVID. The SARS-CoV-2 CT scan and COVID19-CT datasets were utilized to assess the model, with image enhancement in pre-processing using the OpenCV library interpolation function INTER_AREA and data augmentation techniques. The suggested model achieves outstanding performance in the SARS-CoV-2 CT-scan dataset, with an accuracy of 99.6%. However, on the COVID19-CT dataset, it performed with an accuracy rate of 89%. The limitation of this model, its performance is not good on the COVID19-CT dataset.
The future study will design a structure combining DL models that have already been pre-trained. Additionally, the technique of k-fold cross-validation will be used as a partitioning strategy for the dataset
Albelwi, Saleh A. 2022. “Deep Architecture Based on DenseNet-121 Model for Weather Image Recognition.” International Journal of Advanced Computer Science and Applications 13(10): 559–65.
De Anda-Suarez, Juan et al. 2022. “A Novel Metaheuristic Framework Based on the Generalized Boltzmann Distribution for COVID-19 Spread Characterization.” IEEE Access 10: 7326–40.
Biswas, Shreya et al. 2021. “Prediction of Covid-19 from Chest Ct Images Using an Ensemble of Deep Learning Models.” Applied Sciences (Switzerland) 11(15).
Cantarero, Ruben et al. 2021. “COVID19-Routes: A Safe Pedestrian Navigation Service.” IEEE Access 9: 93433–49.
Chen, Xiaocong et al. 2021. “Momentum Contrastive Learning for Few-Shot COVID-19 Diagnosis from Chest CT Images.” Pattern Recognition 113: 107826. https://doi.org/10.1016/j.patcog.2021.107826.
Dai, Hui et al. 2020. “High-Resolution Chest CT Features and Clinical Characteristics of Patients Infected with COVID-19 in Jiangsu, China.” International Journal of Infectious Diseases 95: 106–12. https://doi.org/10.1016/j.ijid.2020.04.003.
Dong, Di et al. 2021. “The Role of Imaging in the Detection and Management of COVID-19: A Review.” IEEE Reviews in Biomedical Engineering 14: 16–29.
Fei-Fei, L., J. Deng, and K. Li. 2010. “ImageNet: Constructing a Large-Scale Image Database.” Journal of Vision 9(8): 1037–1037.
Garain, Avishek et al. 2021. “Detection of COVID-19 from CT Scan Images: A Spiking Neural Network-Based Approach.” Neural Computing and Applications 1. https://doi.org/10.1007/s00521-021-05910-1.
Hasan, Najmul, Yukun Bao, Ashadullah Shawon, and Yanmei Huang. 2021. “DenseNet Convolutional Neural Networks Application for Predicting COVID-19 Using CT Image.” SN Computer Science 2(5): 1–11. https://doi.org/10.1007/s42979-021-00782-7.
He, Xuehai et al. 2020. “Sample-Efficient Deep Learning for COVID-19 Diagnosis Based on CT Scans.” IEEE Transactions on Medical Imaging XX(Xx): 10.
Horry, Michael J. et al. 2020. “COVID-19 Detection through Transfer Learning Using Multimodal Imaging Data.” IEEE Access 8: 149808–24.
Huang, Mei Ling, and Yu Chieh Liao. 2022. “A Lightweight CNN-Based Network on COVID-19 Detection Using X-Ray and CT Images.” Computers in Biology and Medicine 146(March): 105604. https://doi.org/10.1016/j.compbiomed.2022.105604.
Hussain, Emtiaz et al. 2021. “CoroDet: A Deep Learning Based Classification for COVID-19 Detection Using Chest X-Ray Images.” Chaos, Solitons and Fractals 142: 110495. https://doi.org/10.1016/j.chaos.2020.110495.
Ibrahim, Walat Ramadhan, and Mayyadah Ramiz Mahmood. 2023. “COVID-19 Detection Based on Convolution Neural Networks from CT-Scan Images : A Review.” 29(3): 1668–77.
Islam, Md Rakibul, and Abdul Matin. 2020. “Detection of COVID 19 from CT Image by the Novel LeNet-5 CNN Architecture.” ICCIT 2020 - 23rd International Conference on Computer and Information Technology, Proceedings: 19–21.
Jaiswal, Aayush et al. 2021. “Classification of the COVID-19 Infected Patients Using DenseNet201 Based Deep Transfer Learning.” Journal of Biomolecular Structure and Dynamics 39(15): 5682–89. https://doi.org/10.1080/07391102.2020.1788642.
Kathamuthu, Nirmala Devi et al. 2023. “A Deep Transfer Learning-Based Convolution Neural Network Model for COVID-19 Detection Using Computed Tomography Scan Images for Medical Applications.” Advances in Engineering Software 175(October 2022): 103317. https://doi.org/10.1016/j.advengsoft.2022.103317.
Mohammed, Mazin Abed et al. 2020. “Benchmarking Methodology for Selection of Optimal COVID-19 Diagnostic Model Based on Entropy and TOPSIS Methods.” IEEE Access 8: 99115–31.
Mondal, Arnab Kumar, Arnab Bhattacharjee, Parag Singla, and A. P. Prathosh. 2022. “XViTCOS: Explainable Vision Transformer Based COVID-19 Screening Using Radiography.” IEEE Journal of Translational Engineering in Health and Medicine 10(November 2021): 1–10.
Munusamy, Hemalatha et al. 2021. “FractalCovNet Architecture for COVID-19 Chest X-Ray Image Classification and CT-Scan Image Segmentation.” Biocybernetics and Biomedical Engineering 41(3): 1025–38. https://doi.org/10.1016/j.bbe.2021.06.011.
Panwar, Harsh et al. 2020. “A Deep Learning and Grad-CAM Based Color Visualization Approach for Fast Detection of COVID-19 Cases Using Chest X-Ray and CT-Scan Images.” Chaos, Solitons and Fractals 140: 110190. https://doi.org/10.1016/j.chaos.2020.110190.
Pham, Tuan D. 2020. “A Comprehensive Study on Classification of COVID-19 on Computed Tomography with Pretrained Convolutional Neural Networks.” Scientific Reports 10(1): 1–8. https://doi.org/10.1038/s41598-020-74164-z.
Rahman, Tawsifur, Amith Khandakar, and Senior Member. 2021. “Development and Validation of an Early Scoring System for Prediction of Disease Severity in COVID-19 Using Complete Blood Count Parameters.” IEEE Access 9: 120422–41.
Ramzy, Mohamed, Ibrahim Sherin, and M Youssef Karma. 2021. “Abnormality Detection and Intelligent Severity Assessment of Human Chest Computed Tomography Scans Using Deep Learning : A Case Study on SARS ‑ COV ‑ 2 Assessment.” Journal of Ambient Intelligence and Humanized Computing (0123456789). https://doi.org/10.1007/s12652-021-03282-x.
Sakib, Sadman et al. 2020. “DL-CRC: Deep Learning-Based Chest Radiograph Classification for Covid-19 Detection: A Novel Approach.” IEEE Access 8(July): 171575–89.
Sarker, Laboni et al. 2020. “COVID-DenseNet: A Deep Learning Architecture to Detect COVID-19 from Chest Radiology Images.” Preprints (May). https://github.com/mmiemon/COVID-DenseNet%0Ahttp://proxy.library.stonybrook.edu/login?url=https://search.proquest.com/docview/2414040424?accountid=14172%0Ahttps://www.preprints.org/manuscript/202005.0151/v1/download%0Ahttp://linksource.ebsco.com/linking.a.
Seum, Ashek, Amir Hossain Raj, Shadman Sakib, and Tonmoy Hossain. 2020. “A Comparative Study of CNN Transfer Learning Classification Algorithms with Segmentation for COVID-19 Detection from CT Scan Images.” Proceedings of 2020 11th International Conference on Electrical and Computer Engineering, ICECE 2020: 234–37.
Shah, Vruddhi et al. 2021. “Diagnosis of COVID-19 Using CT Scan Images and Deep Learning Techniques.” Emergency Radiology 28(3): 497–505.
Shi, Feng et al. 2021. “Review of Artificial Intelligence Techniques in Imaging Data Acquisition, Segmentation, and Diagnosis for COVID-19.” IEEE Reviews in Biomedical Engineering 14: 4–15.
Shukla, Prashant Kumar et al. 2020. “Efficient Prediction of Drug–Drug Interaction Using Deep Learning Models.” IET Systems Biology 14(4): 211–16.
Silva, Pedro et al. 2020. “COVID-19 Detection in CT Images with Deep Learning: A Voting-Based Scheme and Cross-Datasets Analysis.” Informatics in Medicine Unlocked 20: 100427. https://doi.org/10.1016/j.imu.2020.100427.
Soares, Eduardo, and Plamen Angelov. 2020. “A Large Dataset of Real Patients CT Scans for COVID-19 Identification.” Harv. Dataverse 1: 1–8.
Tabik, S. et al. 2020. “COVIDGR Dataset and COVID-SDNet Methodology for Predicting COVID-19 Based on Chest X-Ray Images.” IEEE Journal of Biomedical and Health Informatics 24(12): 3595–3605.
Tai, Yonghang et al. 2021. “Intelligent Intraoperative Haptic-AR Navigation for COVID-19 Lung Biopsy Using Deep Hybrid Model.” IEEE Transactions on Industrial Informatics 17(9): 6519–27.
Tang, Shengnan, Shouqi Yuan, and Yong Zhu. 2020. “Data Preprocessing Techniques in Convolutional Neural Network Based on Fault Diagnosis towards Rotating Machinery.” IEEE Access 8: 149487–96.
Wang, Zhao, Quande Liu, and Qi Dou. 2020. “Contrastive Cross-Site Learning with Redesigned Net for COVID-19 CT Classification.” IEEE Journal of Biomedical and Health Informatics 24(10): 2806–13.
Wu, Yu Huan et al. 2021. “JCS: An Explainable COVID-19 Diagnosis System by Joint Classification and Segmentation.” IEEE Transactions on Image Processing 30: 3113–26.
Yan, Qingsen et al. 2021. “COVID-19 Chest CT Image Segmentation Network by Multi-Scale Fusion and Enhancement Operations.” IEEE Transactions on Big Data 7(1): 13–24.
Yu, Yongbin et al. 2020. “RMAF: Relu-Memristor-Like Activation Function for Deep Learning.” IEEE Access 8: 72727–41.