THE EFFECT OF FEATURE EXTRACTION ON COVID-19 CLASSIFICATION
Rebin Abdulkareem Hamaamin1, Shakhawan Hares Wady 2, Ali Wahab Kareem Sangawi 3
1Computer Science , College of Sciences, Charmo University, Chamchamal, Sulaimani, KRG, Iraq
Email: - rebin.abdulkarim@charmouniversity.org
2 Business Administration, College of Business, Charmo University, Chamchamal, Sulaimani, KRG, Iraq
Email: - shakhawan.hares@charmouniversity.org
3 General Science, College of Education, Charmo University, Chamchamal, Sulaimani, KRG, Iraq
Email: - ali.kareem@charmouniversity.org
Received: 15 Sep., 2023/ Accepted: 13 Mar., 2024 / Published: 3 June., 2024. https://doi.org/10.25271/sjuoz.2024.12.2.1204
ABSTRACT:
X-ray imaging stands as a prominent technique for diagnosing COVID-19, and it also serves as a crucial tool in the medical field for the analysis of various diseases. Numerous approaches are available to facilitate this analysis. Among these techniques, one involves the utilization of a Feature Extractor, which effectively captures pertinent characteristics from X-ray images. In a recent study, a comprehensive examination was conducted using 25 distinct feature extractors on X-ray images specific to COVID-19 cases. These images were categorized into two classes: COVID-19-positive and non-COVID-19. To enable a thorough evaluation, a sequence of machine learning classifiers was employed on these categorized images. The outcomes derived from this experimentation gauged the magnitude of impact that each individual feature exerted on COVID-19-related imagery. This assessment aimed to determine the efficacy levels of various feature extractors in terms of detection capability. Consequently, a distinction emerged between the more effective and less effective feature extractors, shedding light on their varying degrees of contribution to the detection process. Moreover, the comparative performance of different classifiers became evident, revealing the classifiers that exhibited superior performance when measured against their counterparts.
KEYWORDS: - Feature extraction, Machine learning, Effect of feature, X-Ray, Covid-19.
1. INTRODUCTION
The COVID-19 virus has a significant impact on individuals around the globe. The early diagnosis of this infectious disease is critical to prevent its global and local spread. Generally, scientists have tested numerous ways and methods to detect people and analyze the virus. Interestingly, one of the methods used for COVID-19 diagnosis is X-Rays that recognize whether the person is infected or not. Moreover, the researchers attempted to use methods and technologies that yielded quicker and more accurate results [1][2]. A series of feature extractors can be used to extract features from an image. A texture feature is one kind of feature. The texture is an important low-level feature in images, it can be used to characterize the contents of an image or a region. In addition to color features, color features alone cannot identify the image because distinct images can have similar histograms [3].
The essential factor here is to identify the relevant features and then choose the suitable classification method for these features. Once a set of features can be extracted from the images, it sends those features to machine learning for Classification. Classification is the process of categorizing a set of data into classes. It can be done on both structured and unstructured data. Determining the class of incoming data points is the first step in the procedure. The classes are also known as the target, label, or categories. Several machine learning classifiers are used to classify data, such as Support Vector Machine (SVM), k-nearest neighbors (KNN), Ensemble, Decision Tree (DT), and Naïve Bayes (NB) [4].
To review the previous studies, the first part of the research discusses the effect of feature extractors on images. The second part highlights the research study on the diagnosis of COVID-19, which include [5] classification of histopathology images and identification of malignant areas using each feature extractor Gray-Level Co-Occurrence Matrix (GLCM), Local Binary Pattern (LBP), Local Binary Gray Level Co-occurrence Matrix (LBGLCM), gray-level run-length matrix (GLRLM) and Segmentation-based Fractal Texture Analysis (SFTA), Using SVM, KNN, Linear Discriminant Analysis (LDA) and Boosted Tree classifiers. The feature extractor is GLCM, GLRM, LBP, CLBP, and LTP and has been tested on the following diseases Brain, Breast, Colon, Skin, and Thyroid [6]. Also, in [7], A technique has been developed by combining both HOG and LBP feature extractors to identify plant leaves by SVM classifier. Taha J. Alhindi et al. in [8] state that work has been done on medical imaging using LBP, HOG, and a pre-trained deep network for histopathology images. For this purpose, support vector machines, decision trees, and artificial neural networks were used to identify and classify images through the obtained features. Another texture feature used to detect breast cancer in women is
called Monogenic Binary Coding (MBC) [9]. Cigdem Turan and Kin-Man Lam in [10], some feature extractors were used in facial-expression recognition. All the features used are local descriptors. Seyyid Ahmed Medj in [11], states that a set of feature extractors has been developed to extract parts from images for both binary and multiclass. According to Samy Bakheet and Ayoub Al-Hamadi In [12] work has been done on the GLCM feature extractor to detect and diagnose COVID-19 patients based on X-ray images, with it, the latent-dynamic conditional random fields (LDCRFs) classifier. Also, in [13], feature extractors extract features from images. in [14], Violent event has been worked on with two feature extractors for detecting a violent event.
The final section of this research discusses some papers that have worked on COVID-19 in general to diagnose the disease. Author in [15] proposed a new testing methodology to determine whether a patient has been infected by the COVID-19 virus using the SDD300 model. The in-depth feature plus SVM-based procedure was proposed in [16] for identifying coronavirus-infected patients by applying CXR images. SVM was utilized for classification rather than DL-based classifiers, which require an extensive database for training and validation. A. Helwan et al. in [17] introduced a transfer learning approach to diagnose patients who tested positive for COVID-19 and distinguish them from healthy patients using ResNet-18, ResNet-50, and DenseNet-201. For this purpose, 2617 chest CT images of non-COVID-19 and COVID-19 have been experimented with.
M.
Alruwaili et al. in [18] proposed an improved Inception-ResNetV2 DL model for
accurately diagnosing chest CXR images. A Grad-CAM technique was also computed
to improve the visibility of infected lung parts in CXR scans. D. Ji et al. in
[19] presented a COVID-19 detection approach based on image modal feature
fusion. Small-sample enhancement preprocessing, including spinning,
translation, and randomized transformation, was initially conducted using this
methodology. Five classic pertaining models, including VGG19, ResNet152,
Xception, DenseNet201, and InceptionResnetV2, were used to extract CXR images'
features. Pramod Gaur et al. in [20] presented an innovative methodology for
preprocessing CT images and identifying COVID-19 positive and negative. The
suggested approach used the principle of empiric wavelet transformation for
preprocessing, with the optimal elements of the image's red, green, and blue
channels being learned on the presented approach. Deep and transfer learning
procedures as recommended by [21] to differentiate COVID-19 cases by assessing
CXR images.
This paper uses a series of feature extractors to extract features from X-ray images of COVID-19. All the parts are sent to different classifiers. Then they are measured by several metrics to see which feature has higher results and has less impact on COVID-19. The following is the flow of this paper. Section 2 highlights the methods used to extract features on the COVID-19 image and the classifiers used with the evaluation metrics. In section 3, All results are presented with different classifiers for each classifier, a set of metrics used to measure the data. The conclusion is described in the last section.
Material and Methodology
The process by which features are extracted from images in this research is visually outlined in Figure (1). This diagram succinctly portrays the procedural steps and components utilized to carry out the feature extraction from the images under investigation. It serves as a graphical representation of the methodology, aiding in conveying a clear understanding of the sequential and conceptual aspects of how image features are harnessed for analysis within the context of this study. In this system, a standard database is used to test several feature extractors to distinguish between COVID-19 and non-COVID-19 by several classifiers to take the effect of each feature extractor to see which good results are measured by several metrics.
Figure 1: System overview.
1.1 Dataset
The COVID-19 radiographic database, a publicly available database, is used in this paper. The database consists of 21165 CXR images, of which 10,192 belong to normal, 3616 correspond to COVID-19 positive, and 6012 belong to lung opacity (non-COVID lung infection), and 1345 are labeled as viral pneumonia cases. The data for this paper includes 7232 CXR images, 3616 of which with a positive COVID-19 diagnosis and 3616 negative cases randomly selected to create the balanced database. All images in the database have been preprocessed before being featured as follows: Figure (2) visually represents the sequential process involved in the manipulation of images. The initial step involves converting the images into grayscale, resulting in representations composed solely of varying shades of gray. In the subsequent step, all images are uniformly resized to dimensions of 512 x 512 pixels. Finally, the region of interest (ROI) is meticulously extracted from each image. This ROI encapsulates specific, critical features relevant to the analysis. The figure provides a graphical overview of these operations, effectively illustrating the transformation process, from grayscale conversion and resizing to the selective extraction of the region of interest.
1.2. Feature extraction
Feature extraction is the technique of extracting numbers from data using unique processing that allows you to operate on the original data without destroying it. It produces better results than working with raw data, and it makes them faster. In this study, a total of twenty-five distinct feature extractors were employed to extract features from the data. Each of these extractors possessed a unique set of features, employing varied techniques, as comprehensively outlined in Table (1).
Table 1: The list of feature extractors.
No. Features |
Feature Extractors |
N |
8 |
Gray Level Run Length Matrix (GLRLM) |
1 |
22 |
Gray Level Co-occurrence Matrix (GLCM) |
2 |
32 |
Weber Local Descriptor (WLD) |
3 |
37 |
Local Gradient Increasing Pattern (LGIP) |
4 |
48 |
Gabor Wavelet (GW) |
5 |
56 |
Local Directional Pattern (LDiP) |
6 |
56 |
Local Directional Number Pattern (LDNP) |
7 |
72 |
Local Directional Texture Pattern (LDTP) |
8 |
72 |
Wavelet Transform (WT) |
9 |
81 |
Histogram of Oriented Gradients (HOG) |
10 |
168 |
Pyramid of Histogram of Oriented Gradients (PHOG) |
11 |
256 |
Local Monotonic Pattern (LMP) |
12 |
256 |
Local Phase Quantization (LPQ) |
13 |
256 |
Local Transitional Pattern (LTrP) |
14 |
256 |
Median Binary Pattern (MBP) |
15 |
256 |
Gradient Directional Pattern (GDP) |
16 |
272 |
Local Arc Pattern (LAP) |
17 |
512 |
Local Frequency Descriptor (LFD) |
18 |
512 |
Median Ternary Pattern (MTP) |
19 |
512 |
Local Ternary Pattern (LTeP) |
20 |
512 |
Gradient Local Ternary Pattern (GLTP) |
21 |
800 |
Median Robust Extended Local Binary Pattern (MRELBP) |
22 |
1062 |
Binary Pattern of Phase Congruency (BPPC) |
23 |
2048 |
Improved Weber Binary Coding (IWBC) |
24 |
3072 |
Monogenic Binary Coding (MBC) |
25 |
1.3. Classification and Evaluation
After the features are extracted from the images, they are sent to the machine learning classifier. Basically, classifiers are algorithms that automatically sort or categorize data into one or more classes known as ‘Classification’. Classes are referred to as targets, labels, and categories. This research used a set of classifiers to separate the data into two types: COVID-19 and non-COVID-19. The classifiers used in this research are the KNN: which is a supervised machine learning algorithm method that can be used to solve both classification and regression problems and is simple to implement [22]. The second classifier is Ensemble, a method for creating a variety of base classifiers [23]. The third classifier is DT: This model works by parsing our data into a DT. The fourth classifier is NB: a set of straightforward "probabilistic classifiers" based on Bayes' theorem and strong (naive) independence [22]. Furthermore, the last classifier SVM, a supervised machine learning technique that can be used to solve classification and regression problems. In SVM used, different kernels for classifying data, the SVM linear kernel (SVML), SVM Radial Basis Function kernel (SVMRBF), and SVM polynomial kernel (SVMP) [24][25]. All three types of kernels were tested in the research. The final part of the study is to tether data with several different metrics to see how each feature receptor affects the image of COVID-19 [26].
2. EXPERIMENTAL RESULTS AND DISCUSSIONS
Numerous texture feature descriptors as presented in Table (1), were generated for assessing the CXR images of COVID-19, which is either positive or negative, and the result were given as second opinion to the clinicians to make exact decisions. Besides, to classify the CXR images, five well-known ML classifiers were utilized, namely, SVM (linear, RBF, and polynomial), KNN, Ensemble, DT, and NB classifiers. From the experiments performed on COVID-19 radiography dataset, the analysis of each descriptor was estimated in terms of a series of indicators (F-Score, Matthew's correlation coefficient (MCC) specificity, precision, sensitivity, and Accuracy) with Ensemble (Table 2), KNN (Table 3), NB (Table 4), DT (Table 5), SVML (Table 6), SVMRBF (Table 7), and SVMP (Table 8) classifiers respectively.
As depicted in the initial experiment detailed in Table (2), the MRELBP feature descriptor emerged as the frontrunner. It exhibited the highest accuracy rate of 95.09% across all metrics, in tandem with the Ensemble classifier. In contrast, the classification accuracy stemming from the GLRLM method lagged behind, registering a comparatively lower score of 78.15%.
Table 2: All the feature extractors used in the Ensemble Classifier
Measurement Metrics |
|
|||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- Score |
Feature Extractor |
N |
|
86.76 |
87.13 |
86.27 |
87.25 |
73.53 |
43.35 |
WT |
1 |
|
84.46 |
85.54 |
82.95 |
85.98 |
68.96 |
42.11 |
GLCM |
2 |
|
78.15 |
78.25 |
78.01 |
78.30 |
56.33 |
39.07 |
GLRLM |
3 |
|
91.65 |
91.64 |
91.69 |
91.62 |
83.31 |
45.83 |
HOG |
4 |
|
84.61 |
86.68 |
81.78 |
87.43 |
69.33 |
42.08 |
GW |
5 |
|
92.92 |
92.81 |
93.06 |
92.78 |
85.85 |
46.47 |
BPPC |
6 |
|
90.17 |
90.46 |
89.83 |
90.51 |
80.36 |
45.07 |
GDP |
7 |
|
90.92 |
91.56 |
90.15 |
91.69 |
81.86 |
45.43 |
GLTP |
8 |
|
90.19 |
90.62 |
89.70 |
90.69 |
80.42 |
45.08 |
LAP |
9 |
|
89.05 |
89.84 |
88.06 |
90.03 |
78.12 |
44.47 |
LDNP |
10 |
|
84.32 |
84.73 |
83.75 |
84.90 |
68.66 |
42.12 |
LDTP |
11 |
|
89.26 |
89.21 |
89.32 |
89.20 |
78.52 |
44.63 |
LFD |
12 |
|
92.08 |
92.11 |
92.06 |
92.10 |
84.17 |
46.04 |
LMP |
13 |
|
92.47 |
92.60 |
92.32 |
92.61 |
84.94 |
46.23 |
LPQ |
14 |
|
90.10 |
89.76 |
90.53 |
89.67 |
80.20 |
45.07 |
LTrP |
15 |
|
89.23 |
89.52 |
88.87 |
89.59 |
78.47 |
44.60 |
LDiP |
16 |
|
81.19 |
81.47 |
80.77 |
81.60 |
62.39 |
40.56 |
WLD |
17 |
|
91.87 |
92.50 |
91.13 |
92.60 |
83.76 |
45.91 |
MBC |
18 |
|
92.76 |
92.87 |
92.64 |
92.88 |
85.53 |
46.38 |
IWBC |
19 |
|
90.96 |
91.20 |
90.69 |
91.23 |
81.94 |
45.47 |
PHOG |
20 |
|
90.62 |
90.61 |
90.65 |
90.59 |
81.25 |
45.31 |
LGIP |
21 |
|
95.09 |
95.17 |
95.01 |
95.17 |
90.19 |
47.54 |
MRELBP |
22 |
|
89.81 |
90.08 |
89.47 |
90.14 |
79.62 |
44.89 |
MBP |
23 |
|
82.19 |
83.51 |
80.24 |
84.15 |
64.45 |
40.92 |
MTP |
24 |
|
89.44 |
90.13 |
88.59 |
90.29 |
78.90 |
44.68 |
LTeP |
25 |
Upon the utilization of the KNN classifier, distinct categories of features exhibited disparate responses in terms of their projected performance. The findings, as elucidated in Table (3), indicate that the MRELBP method showcased notable efficacy, achieving an accuracy rate of 91.59% in the classification of chest X-ray (CXR) images when juxtaposed with the array of alternative techniques implemented.
Table 3: All the feature extractors used in the KNN Classifier
Measurement Metrics |
|
|||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- score |
Feature extractor |
N |
|
85.59 |
84.98 |
86.49 |
84.70 |
71.21 |
42.86 |
WT |
1 |
|
84.00 |
83.86 |
84.22 |
83.79 |
68.02 |
42.02 |
GLCM |
2 |
|
75.30 |
74.20 |
77.58 |
73.02 |
50.66 |
37.93 |
GLRLM |
3 |
|
89.74 |
88.10 |
91.89 |
87.58 |
79.56 |
44.98 |
HOG |
4 |
|
82.25 |
81.69 |
83.15 |
81.34 |
64.53 |
41.21 |
GW |
5 |
|
91.20 |
90.30 |
92.34 |
90.07 |
82.44 |
45.65 |
BPPC |
6 |
|
88.24 |
86.64 |
90.43 |
86.04 |
76.56 |
44.25 |
GDP |
7 |
|
85.21 |
82.88 |
88.77 |
81.66 |
70.62 |
42.86 |
GLTP |
8 |
|
86.96 |
85.47 |
89.09 |
84.83 |
74.00 |
43.62 |
LAP |
9 |
|
84.98 |
83.19 |
87.69 |
82.27 |
70.07 |
42.69 |
LDNP |
10 |
|
79.71 |
80.13 |
79.06 |
80.36 |
59.44 |
39.80 |
LDTP |
11 |
|
85.51 |
83.39 |
88.73 |
82.30 |
71.18 |
42.99 |
LFD |
12 |
|
88.58 |
85.52 |
92.89 |
84.26 |
77.45 |
44.53 |
LMP |
13 |
|
90.16 |
88.34 |
92.53 |
87.79 |
80.41 |
45.19 |
LPQ |
14 |
|
85.74 |
82.17 |
91.31 |
80.17 |
71.95 |
43.25 |
LTrP |
15 |
|
85.74 |
83.93 |
88.44 |
83.04 |
71.61 |
43.06 |
LDiP |
16 |
|
77.75 |
76.51 |
80.15 |
75.35 |
55.58 |
39.14 |
WLD |
17 |
|
90.70 |
89.04 |
92.84 |
88.56 |
81.48 |
45.45 |
MBC |
18 |
|
88.82 |
84.94 |
94.41 |
83.24 |
78.16 |
44.71 |
IWBC |
19 |
|
87.06 |
84.17 |
91.30 |
82.82 |
74.40 |
43.80 |
PHOG |
20 |
|
87.35 |
84.76 |
91.09 |
83.61 |
74.92 |
43.91 |
LGIP |
21 |
|
91.59 |
90.39 |
93.10 |
90.08 |
83.23 |
45.86 |
MRELBP |
22 |
|
85.73 |
82.70 |
90.36 |
81.09 |
71.76 |
43.18 |
MBP |
23 |
|
77.32 |
77.48 |
77.04 |
77.61 |
54.65 |
38.63 |
MTP |
24 |
|
81.06 |
79.43 |
83.86 |
78.26 |
62.23 |
40.79 |
LTeP |
25 |
Another classifier used in the research is NB. It is a probabilistic machine learning algorithm based on the Bayes theorem. This classifier has a problem that when a feature that is output contains a zero result, this classifier does not output a result, and the result is non. As depicted in Table (4), the feature structures encompassing MRELBP, MBP, MTP, and LTep have yielded non-results, reflecting zero outcomes in their respective feature sets. In contrast, among the other feature extractors explored in this study, the highest achievement was observed with the PHOG feature extractor, achieving a noteworthy accuracy score of 80.80% in the accuracy metric specifically for COVID-19 detection. Notably, when these results are juxtaposed with the preceding ones, it becomes apparent that they fall short in comparison to the outcomes generated by the diverse classifier implementations.
Table 4: All the feature extractors used in the NB Classifier
Measurement Metrics |
|
|||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- score |
Feature extractor |
N |
|
75.80 |
76.47 |
74.56 |
77.03 |
51.62 |
37.75 |
WT |
1 |
|
76.18 |
77.93 |
73.04 |
79.32 |
52.47 |
37.70 |
GLCM |
2 |
|
67.32 |
63.92 |
79.57 |
55.08 |
35.75 |
35.45 |
GLRLM |
3 |
|
77.95 |
75.19 |
83.47 |
72.43 |
56.26 |
39.56 |
HOG |
4 |
|
72.14 |
80.09 |
58.95 |
85.33 |
45.92 |
33.96 |
GW |
5 |
|
77.39 |
73.01 |
86.93 |
67.84 |
55.81 |
39.68 |
BPPC |
6 |
|
76.39 |
74.52 |
80.21 |
72.57 |
52.95 |
38.63 |
GDP |
7 |
|
70.80 |
71.52 |
69.14 |
72.45 |
41.62 |
35.15 |
GLTP |
8 |
|
75.78 |
71.34 |
86.21 |
65.35 |
52.73 |
39.04 |
LAP |
9 |
|
74.58 |
71.98 |
80.48 |
68.67 |
49.51 |
38.00 |
LDNP |
10 |
|
65.06 |
84.66 |
36.79 |
93.32 |
36.51 |
25.65 |
LDTP |
11 |
|
72.91 |
68.75 |
84.02 |
61.80 |
47.01 |
37.81 |
LFD |
12 |
|
76.92 |
73.81 |
83.47 |
70.36 |
54.32 |
39.17 |
LMP |
13 |
|
79.45 |
76.45 |
85.13 |
73.76 |
59.28 |
40.28 |
LPQ |
14 |
|
71.63 |
68.11 |
81.40 |
61.85 |
44.13 |
37.08 |
LTrP |
15 |
|
72.79 |
69.94 |
79.94 |
65.64 |
46.07 |
37.30 |
LDiP |
16 |
|
70.55 |
77.29 |
58.19 |
82.90 |
42.41 |
33.20 |
WLD |
17 |
|
78.09 |
74.60 |
85.19 |
71.00 |
56.76 |
39.77 |
MBC |
18 |
|
77.27 |
74.02 |
84.04 |
70.50 |
55.05 |
39.36 |
IWBC |
19 |
|
80.80 |
77.92 |
85.98 |
75.62 |
61.94 |
40.87 |
PHOG |
20 |
|
73.06 |
69.78 |
81.37 |
64.74 |
46.77 |
37.57 |
LGIP |
21 |
|
- |
- |
- |
- |
- |
- |
MRELBP |
22 |
|
- |
- |
- |
- |
- |
- |
MBP |
23 |
|
- |
- |
- |
- |
- |
- |
MTP |
24 |
|
- |
- |
- |
- |
- |
- |
LTeP |
25 |
DT is another classifier implemented in this paper. This classifier works as a tree on the nodes and then extends the branches down. The only reason these classifiers were used in these feature extractions was to examine how they performed compared to other classifiers. The results obtained from this classifier are weaker than the KNN and ensemble classifiers, presented in Table (5), parallel to the assortment of classifiers, the optimal feature extractor remains MRELBP. Remarkably, among the alternative feature extraction approaches, Access achieves the peak outcome of 85.50%, standing out as the most elevated result among the other feature extractors.
Table 5: All the feature extractors used in the DT Classifier
Measurement Metrics |
|
|||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- score |
Feature extractor |
N |
|
79.92 |
79.61 |
80.44 |
79.39 |
59.85 |
40.01 |
WT |
1 |
|
80.18 |
80.09 |
80.35 |
80.01 |
60.38 |
40.11 |
GLCM |
2 |
|
72.49 |
72.20 |
73.20 |
71.78 |
44.99 |
36.35 |
GLRLM |
3 |
|
82.72 |
82.20 |
83.54 |
81.89 |
65.45 |
41.43 |
HOG |
4 |
|
76.58 |
76.50 |
76.79 |
76.38 |
53.19 |
38.32 |
GW |
5 |
|
83.24 |
83.21 |
83.28 |
83.20 |
66.49 |
41.62 |
BPPC |
6 |
|
82.00 |
81.22 |
83.28 |
80.72 |
64.05 |
41.12 |
GDP |
7 |
|
81.88 |
81.54 |
82.45 |
81.31 |
63.79 |
41.00 |
GLTP |
8 |
|
81.23 |
81.05 |
81.55 |
80.91 |
62.48 |
40.65 |
LAP |
9 |
|
80.03 |
79.96 |
80.15 |
79.92 |
60.08 |
40.03 |
LDNP |
10 |
|
76.75 |
77.02 |
76.32 |
77.18 |
53.54 |
38.33 |
LDTP |
11 |
|
80.50 |
80.20 |
81.01 |
79.99 |
61.01 |
40.30 |
LFD |
12 |
|
83.26 |
83.20 |
83.40 |
83.13 |
66.55 |
41.65 |
LMP |
13 |
|
83.69 |
83.24 |
84.40 |
82.99 |
67.41 |
41.91 |
LPQ |
14 |
|
80.77 |
80.48 |
81.27 |
80.26 |
61.55 |
40.44 |
LTrP |
15 |
|
81.92 |
81.23 |
83.08 |
80.76 |
63.88 |
41.07 |
LDiP |
16 |
|
72.79 |
72.47 |
73.53 |
72.05 |
45.59 |
36.50 |
WLD |
17 |
|
82.32 |
82.67 |
81.80 |
82.85 |
64.66 |
41.12 |
MBC |
18 |
|
83.09 |
82.64 |
83.80 |
82.38 |
66.20 |
41.61 |
IWBC |
19 |
|
80.13 |
79.61 |
81.02 |
79.24 |
60.28 |
40.16 |
PHOG |
20 |
|
82.90 |
83.15 |
82.53 |
83.26 |
65.80 |
41.42 |
LGIP |
21 |
|
85.50 |
85.69 |
85.26 |
85.74 |
71.03 |
42.74 |
MRELBP |
22 |
|
80.59 |
80.43 |
80.86 |
80.32 |
61.19 |
40.32 |
MBP |
23 |
|
75.08 |
74.80 |
75.75 |
74.41 |
50.21 |
37.64 |
MTP |
24 |
|
77.94 |
77.85 |
78.15 |
77.73 |
55.90 |
39.00 |
LTeP |
25 |
Tables (6 - 8) comprehensively present the performance outcomes of all SVM models. Drawing insights from the research findings across the three distinct kernels, a consistent pattern emerges: the features extracted through the utilization of MRELBP techniques consistently outshine alternative scenarios. This phenomenon is evident when employing the SVML, SVMRBF, and SVMP classifiers, where the corresponding accuracy rates for MRELBP features stand at 85.50%, 95.10%, and an impressive 95.87%, respectively. This consistency in the superiority of MRELBP-derived features across diverse SVM kernels underscores the robustness and effectiveness of this feature extraction approach. The exceptional accuracy rates obtained with SVMP highlight its particular aptitude in capturing the relevant patterns present in the data, contributing to accurate classification of COVID-19 images. Consequently, these findings emphasize the potential of MRELBP-based features in enhancing the performance and accuracy of machine learning models for disease
detection and classification.
Table 6: All the feature extractors used in the SVML Classifier
Measurement Metrics |
|
|||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- score |
Feature extractor |
N |
|
79.20 |
79.05 |
79.47 |
78.92 |
58.41 |
39.63 |
WT |
1 |
|
77.30 |
79.90 |
72.99 |
81.62 |
54.82 |
38.14 |
GLCM |
2 |
|
76.78 |
72.77 |
85.66 |
67.91 |
54.45 |
39.35 |
GLRLM |
3 |
|
88.86 |
87.96 |
90.06 |
87.66 |
77.75 |
44.50 |
HOG |
4 |
|
83.18 |
89.10 |
75.63 |
90.73 |
67.14 |
40.91 |
GW |
5 |
|
92.45 |
92.39 |
92.53 |
92.37 |
84.91 |
46.23 |
BPPC |
6 |
|
87.82 |
88.48 |
86.97 |
88.67 |
75.66 |
43.86 |
GDP |
7 |
|
88.96 |
90.44 |
87.12 |
90.79 |
77.97 |
44.38 |
GLTP |
8 |
|
86.82 |
86.74 |
86.93 |
86.71 |
73.64 |
43.42 |
LAP |
9 |
|
83.95 |
84.17 |
83.62 |
84.27 |
67.91 |
41.95 |
LDNP |
10 |
|
82.54 |
86.08 |
77.63 |
87.44 |
65.40 |
40.82 |
LDTP |
11 |
|
86.90 |
86.02 |
88.13 |
85.67 |
73.84 |
43.53 |
LFD |
12 |
|
89.72 |
89.16 |
90.44 |
88.99 |
79.45 |
44.90 |
LMP |
13 |
|
91.67 |
91.49 |
91.91 |
91.42 |
83.35 |
45.85 |
LPQ |
14 |
|
86.24 |
86.29 |
86.18 |
86.29 |
72.49 |
43.12 |
LTrP |
15 |
|
87.98 |
88.29 |
87.59 |
88.37 |
75.97 |
43.97 |
LDiP |
16 |
|
78.02 |
80.85 |
73.49 |
82.56 |
56.30 |
38.50 |
WLD |
17 |
|
92.51 |
93.19 |
91.74 |
93.28 |
85.04 |
46.23 |
MBC |
18 |
|
90.82 |
90.33 |
91.45 |
90.18 |
81.66 |
45.44 |
IWBC |
19 |
|
89.24 |
88.39 |
90.35 |
88.13 |
78.50 |
44.68 |
PHOG |
20 |
|
85.67 |
86.71 |
84.29 |
87.05 |
71.39 |
42.74 |
LGIP |
21 |
|
92.86 |
91.82 |
94.12 |
91.60 |
85.76 |
46.48 |
MRELBP |
22 |
|
87.16 |
88.33 |
85.66 |
88.66 |
74.36 |
43.49 |
MBP |
23 |
|
81.72 |
85.72 |
76.11 |
87.32 |
63.84 |
40.32 |
MTP |
24 |
|
85.20 |
87.74 |
81.84 |
88.56 |
70.57 |
42.34 |
LTeP |
25 |
Table 7: All the feature extractors used in the SVMRBF Classifier
Measurement Metrics |
|
|||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- score |
Feature extractor |
N |
|
88.81 |
89.95 |
87.41 |
90.21 |
77.67 |
44.33 |
WT |
1 |
|
86.11 |
88.05 |
83.58 |
88.64 |
72.33 |
42.88 |
GLCM |
2 |
|
79.11 |
80.19 |
77.36 |
80.86 |
58.28 |
39.37 |
GLRLM |
3 |
|
93.26 |
93.53 |
92.96 |
93.57 |
86.53 |
46.62 |
HOG |
4 |
|
85.48 |
90.74 |
79.03 |
91.92 |
71.56 |
42.24 |
GW |
5 |
|
94.57 |
94.37 |
94.80 |
94.34 |
89.15 |
47.29 |
BPPC |
6 |
|
91.73 |
91.87 |
91.56 |
91.89 |
83.47 |
45.86 |
GDP |
7 |
|
90.17 |
91.98 |
88.01 |
92.32 |
80.42 |
44.98 |
GLTP |
8 |
|
91.25 |
91.86 |
90.54 |
91.96 |
82.52 |
45.60 |
LAP |
9 |
|
90.84 |
90.77 |
90.93 |
90.75 |
81.69 |
45.42 |
LDNP |
10 |
|
80.94 |
77.77 |
86.69 |
75.19 |
62.31 |
40.99 |
LDTP |
11 |
|
90.73 |
90.74 |
90.75 |
90.72 |
81.48 |
45.37 |
LFD |
12 |
|
93.89 |
94.56 |
93.15 |
94.63 |
87.80 |
46.93 |
LMP |
13 |
|
94.46 |
94.54 |
94.38 |
94.54 |
88.93 |
47.23 |
LPQ |
14 |
|
91.12 |
91.06 |
91.20 |
91.04 |
82.25 |
45.57 |
LTrP |
15 |
|
90.30 |
91.26 |
89.14 |
91.45 |
80.63 |
45.09 |
LDiP |
16 |
|
83.13 |
85.00 |
80.50 |
85.75 |
66.36 |
41.34 |
WLD |
17 |
|
93.14 |
93.85 |
92.34 |
93.94 |
86.30 |
46.54 |
MBC |
18 |
|
93.33 |
92.88 |
93.86 |
92.79 |
86.67 |
46.68 |
IWBC |
19 |
|
92.54 |
93.00 |
92.02 |
93.07 |
85.10 |
46.25 |
PHOG |
20 |
|
92.10 |
92.50 |
91.65 |
92.56 |
84.22 |
46.03 |
LGIP |
21 |
|
95.10 |
95.78 |
94.36 |
95.84 |
90.21 |
47.53 |
MRELBP |
22 |
|
91.96 |
92.40 |
91.44 |
92.48 |
83.92 |
45.96 |
MBP |
23 |
|
75.66 |
71.49 |
85.41 |
65.91 |
52.33 |
38.91 |
MTP |
24 |
|
84.00 |
81.13 |
88.66 |
79.35 |
68.31 |
42.36 |
LTeP |
25 |
Table 8: All the feature extractors used in the SVMP Classifier
|
||||||||
|
Accuracy |
Sensitivity |
Precision |
Specificity |
MCC |
F- score |
Feature extractor |
N |
|
87.52 |
88.14 |
86.72 |
88.33 |
75.07 |
43.71 |
WT |
1 |
|
86.67 |
88.66 |
84.12 |
89.23 |
73.46 |
43.17 |
GLCM |
2 |
|
56.37 |
51.60 |
80.30 |
32.43 |
16.26 |
31.41 |
GLRLM |
3 |
|
93.40 |
92.93 |
93.96 |
92.85 |
86.82 |
46.72 |
HOG |
4 |
|
89.12 |
89.02 |
89.27 |
88.98 |
78.26 |
44.57 |
GW |
5 |
|
94.41 |
94.22 |
94.63 |
94.19 |
88.83 |
47.21 |
BPPC |
6 |
|
91.42 |
90.76 |
92.25 |
90.59 |
82.87 |
45.75 |
GDP |
7 |
|
90.94 |
90.77 |
91.16 |
90.72 |
81.89 |
45.48 |
GLTP |
8 |
|
91.09 |
90.11 |
92.31 |
89.86 |
82.21 |
45.60 |
LAP |
9 |
|
91.76 |
90.79 |
92.97 |
90.55 |
83.56 |
45.93 |
LDNP |
10 |
|
71.87 |
71.91 |
86.47 |
57.28 |
44.33 |
39.26 |
LDTP |
11 |
|
90.80 |
90.08 |
91.70 |
89.90 |
81.62 |
45.44 |
LFD |
12 |
|
94.18 |
93.70 |
94.73 |
93.62 |
88.36 |
47.11 |
LMP |
13 |
|
95.01 |
94.47 |
95.63 |
94.40 |
90.04 |
47.52 |
LPQ |
14 |
|
91.24 |
90.34 |
92.37 |
90.11 |
82.50 |
45.67 |
LTrP |
15 |
|
91.59 |
90.62 |
92.79 |
90.39 |
83.21 |
45.85 |
LDiP |
16 |
|
86.19 |
85.44 |
87.28 |
85.10 |
72.41 |
43.17 |
WLD |
17 |
|
93.64 |
93.60 |
93.71 |
93.58 |
87.30 |
46.83 |
MBC |
18 |
|
93.72 |
92.96 |
94.62 |
92.82 |
87.47 |
46.89 |
IWBC |
19 |
|
92.72 |
92.25 |
93.29 |
92.16 |
85.46 |
46.38 |
PHOG |
20 |
|
93.41 |
92.76 |
94.18 |
92.64 |
86.83 |
46.73 |
LGIP |
21 |
|
95.87 |
95.48 |
96.31 |
95.44 |
91.75 |
47.94 |
MRELBP |
22 |
|
91.99 |
91.30 |
92.84 |
91.15 |
84.00 |
46.03 |
MBP |
23 |
|
53.88 |
60.37 |
81.15 |
26.61 |
11.61 |
34.62 |
MTP |
24 |
|
88.96 |
88.85 |
89.11 |
88.80 |
77.92 |
44.49 |
LTeP |
25 |
Upon
extensive evaluation of various feature extractors with diverse classifiers,
noteworthy findings have emerged. Notably, the most effective feature extractor
across all classifiers, with the exception of the NB classifier, is determined
to be MRELBP. However, it is worth highlighting that MRELBP yields no results
within the NB classifier due to its lack of features, resulting in a count of
zero. Intriguingly, within the context of the NB classifier, the optimal
feature extractor is identified as PHOG. These outcomes, outlined in Figure 2,
distinctly outline the superiority of specific feature extractors within
distinct classifier contexts, thus contributing valuable insights to the
overall research findings.
Figure 3: effect of feature extraction on Covid-19 images in different classifier
In our analysis of various works, it has become evident that the features employed in this study surpass several other functions, particularly those highlighted in Table (9). However, the real potential lies in the synergy achieved through combining these features, leading to improved results. This underscores the importance of a holistic approach in leveraging the strengths of individual features for a more comprehensive and effective outcome.
Table 9: Comparison of proposed classification accuracy with existing state-of-the-art techniques
Previous study |
Methods/ classifier |
Accuracy (%) |
A. I. Khan et al[27] |
(CNN) CoroNet |
89.60 |
A. Zargari Khuzani et al[28] |
ResNet-18 |
92.49 |
L. Wang et al[1] |
COVID-Net |
93.30 |
M. Owais et al[29] |
Ensemble-Net |
94.72 |
F. Saiz and I. Barandiaran[15] |
SDD300 |
94.92 |
F. H. Ahmad and S. H. Wady[30] |
SVM (CT+GWT+LGIP) |
96.18 |
R. A. Hamaamin et al[31] |
HOG +LPQ |
97.15 |
R. A. Hamaamin et al[32] |
Resnet 50 |
98.05 |
Proposed |
MRELBP |
95.87 |
CONCLUSION
Due to the global prevalence of the COVID-19 pandemic, a multitude of researchers have been diligently striving to devise an intelligent approach for the swift identification of the disease. One prominent technique within this realm involves the utilization of feature extractors to distill pertinent information from images. In this pursuit, a particular study undertook the deployment of a series of feature extractors to scrutinize COVID-19 images, aiming to discern the optimal feature extraction methodology.
Through a meticulous evaluation of diverse feature extractors, this research endeavors to unearth the most efficacious one. Among the twenty-five feature extractors subjected to assessment, certain features emerged with exceptional efficacy in delineating the disease. Notably, the investigation pinpointed "MRELBP" as the preeminent feature extractor, showcasing a remarkable accuracy rate of 95.87% when employed in conjunction with the SVMP classifier. It is worth mentioning that this study exhibited a comprehensive approach, encompassing an array of classifiers to culminate in its findings. Moreover, a multifaceted set of metrics was employed to gauge the performance of each featured extractor.
The primary issue with the study lies in the time-consuming nature of obtaining results for certain features. Furthermore, some features are sourced from machine learning missions that were not undertaken
REFERENCES
Wei Wang, Yutao Li, Ji Li, Peng Zhang, Xin Wang, "Detecting COVID-19 in Chest X-Ray Images via MCFF-Net", Computational Intelligence and Neuroscience, vol. 2021, Article ID 3604900, 8 pages, 2021. https://doi.org/10.1155/2021/3604900.
Ebrahim Mohammed Senan, Ali Alzahrani, Mohammed Y. Alzahrani, Nizar Alsharif, Theyazn H. H. Aldhyani, "Automated Diagnosis of Chest X-Ray for Early Detection of COVID-19 Disease", Computational and Mathematical Methods in Medicine, vol. 2021, Article ID 6919483, 10 pages, 2021. https://doi.org/10.1155/2021/6919483.
J. Lian, M. Zhang, N. Jiang, W. Bi, and X. Dong, “Feature Extraction of Kidney Tissue Image Based on Ultrasound Image Segmentation,” J. Healthc. Eng., vol. 2021, 2021, doi: 10.1155/2021/9915697.
Nugrahaeni, Ratna Astuti and Kusprasapta Mutijarsa. “Comparative analysis of machine learning KNN, SVM, and random forests algorithm for facial expression classification.” 2016 International Seminar on Application for Technology of Information and Communication (ISemantic) (2016): 163-168.
Ş. Öztürk and B. Akdemir, “ScienceDirect ScienceDirect ScienceDirect Application of Feature Extraction and Classification Methods for Histopathological Image using GLCM , Application of Feature Extraction and Classification Methods for and GLCM , Histopathological Image using a SFTA,” Procedia Comput. Sci., vol. 132, no. Iccids, pp. 40–46, 2018, doi: 10.1016/j.procs.2018.05.057.
Vani Kumari, S., Usha Rani, K. (2020). Analysis on Various Feature Extraction Methods for Medical Image Classification. In: Jyothi, S., Mamatha, D., Satapathy, S., Raju, K., Favorskaya, M. (eds) Advances in Computational and Bio-Engineering. CBE 2019. Learning and Analytics in Intelligent Systems, vol 16. Springer, Cham. https://doi.org/10.1007/978-3-030-46943-6_3.
Keivani, M., Mazloum, J., Sedaghatfar, E., Tavakoli, M.B. (2020). Automated analysis of leaf shape, texture, and color features for plant classification. Traitement du Signal, Vol. 37, No. 1, pp. 17-28. https://doi.org/10.18280/ts.370103.
T. J. Alhindi, S. Kalra, K. H. Ng, A. Afrin, and H. R. Tizhoosh, “Comparing LBP , HOG and Deep Features for Classification of Histopathology Images,” 2018 Int. Jt. Conf. Neural Networks, pp. 1–7, 2018, doi: 10.1109/IJCNN.2018.8489329.
E. Mohammadi, E. Fatemizadeh, H. Sheikhzadeh, and S. Khoubani, “A Textural Approach for Recognizing Architectural Distortion In Mammograms,” pp. 136–140, 2013.
C. Turan and K. Lam, “Histogram-based Local Descriptors for Facial Expression Recognition ( FER ): A comprehensive Study,” J. Vis. Commun. Image Represent., 2018, doi: 10.1016/j.jvcir.2018.05.024.
S. A. Medjahed, “A Comparative Study of Feature Extraction Methods in Images Classification,” no. February, pp. 16–23, 2015, doi: 10.5815/ijigsp.2015.03.03.
S. Bakheet and A. Al-hamadi, “Automatic detection of COVID-19 using pruned GLCM-Based texture features and LDCRF classification,” Comput. Biol. Med., vol. 137, no. August, p. 104781, 2021, doi: 10.1016/j.compbiomed.2021.104781.
Saipullah, K.M., Kim, DH. A robust texture feature extraction using the localized angular phase. Multimed Tools Appl 59, 717–747 (2012). https://doi.org/10.1007/s11042-011-0766-5.
M. Aradhya and D. S. Guru, “Violent Video Event Detection Based on Integrated LBP and GLCM Texture Features Revue d ’ Intelligence Artificielle Violent Video Event Detection Based on Integrated LBP and GLCM Texture Features,” no. May, 2020, doi: 10.18280/ria.340208.
F. Saiz and I. Barandiaran, “COVID-19 Detection in Chest X-ray Images using a Deep Learning Approach,” Int. J. Interact. Multimed. Artif. Intell., vol. 6, no. 2, p. 4, 2020, doi: 10.9781/ijimai.2020.04.003.
A. K. Singh, A. Kumar, M. Mahmud, M. S. Kaiser, and A. Kishore, “COVID-19 Infection Detection from Chest X-Ray Images Using Hybrid Social Group Optimization and Support Vector Classifier,” Cognit. Comput., no. 0123456789, 2021, doi: 10.1007/s12559-021-09848-3.
A. Helwan, M. K. S. Ma’Aitah, H. Hamdan, D. U. Ozsahin, and O. Tuncyurek, “Radiologists versus Deep Convolutional Neural Networks: A Comparative Study for Diagnosing COVID-19,” Comput. Math. Methods Med., vol. 2021, 2021, doi: 10.1155/2021/5527271.
M. Alruwaili, A. Shehab, and S. Abd El-Ghany, “COVID-19 Diagnosis Using an Enhanced Inception-ResNetV2 Deep Learning Model in CXR Images,” J. Healthc. Eng., vol. 2021, no. Dl, 2021, doi: 10.1155/2021/6658058.
D. Ji, Z. Zhang, Y. Zhao, and Q. Zhao, “Research on Classification of COVID-19 Chest X-Ray Image Modal Feature Fusion Based on Deep Learning,” J. Healthc. Eng., vol. 2021, 2021, doi: 10.1155/2021/6799202.
P. Gaur, V. Malaviya, A. Gupta, G. Bhatia, R. B. Pachori, and D. Sharma, “COVID-19 disease identification from chest CT images using empirical wavelet transformation and transfer learning,” Biomed. Signal Process. Control, vol. 71, no. PA, p. 103076, 2022, doi: 10.1016/j.bspc.2021.103076.
Talal S. Qaid, Hussein Mazaar, Mohammad Yahya H. Al-Shamri, Mohammed S. Alqahtani, Abeer A. Raweh, Wafaa Alakwaa, "Hybrid Deep-Learning and Machine-Learning Models for Predicting COVID-19", Computational Intelligence and Neuroscience, vol. 2021, Article ID 9996737, 11 pages, 2021. https://doi.org/10.1155/2021/9996737
Aljabri, Malak, Amal A. Alahmadi, Rami Mustafa A. Mohammad, Menna Aboulnour, Dorieh M. Alomari, and Sultan H. Almotiri. 2022. "Classification of Firewall Log Data Using Multiclass Machine Learning Models" Electronics 11, no. 12: 1851. https://doi.org/10.3390/electronics11121851.
Dietterich, T.G. (2000). Ensemble Methods in Machine Learning. In: Multiple Classifier Systems. MCS 2000. Lecture Notes in Computer Science, vol 1857. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45014-9_1
S. A. Jafar Zaidi, I. Chatterjee, and S. Brahim Belhaouari, “COVID-19 Tweets Classification during Lockdown Period Using Machine Learning Classifiers,” Appl. Comput. Intell. Soft Comput., vol. 2022, 2022, doi: 10.1155/2022/1209172.
J. Zeng et al., “Prediction of peak particle velocity caused by blasting through the combinations of boosted-chaid and svm models with various kernels,” Appl. Sci., vol. 11, no. 8, 2021, doi: 10.3390/app11083705.
U. Özkaya, Ş. Öztürk, and M. Barstugan, “Coronavirus (COVID-19) Classification Using Deep Features Fusion and Ranking Technique,” Stud. Big Data, vol. 78, pp. 281–295, 2020, doi: 10.1007/978-3-030-55258-9_17.
A. I. Khan, J. L. Shah, and M. M. Bhat, “CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images,” Comput. Methods Programs Biomed., vol. 196, p. 105581, 2020, doi: 10.1016/j.cmpb.2020.105581.
A. Zargari Khuzani, M. Heidari, and S. A. Shariati, “COVID-Classifier: an automated machine learning model to assist in the diagnosis of COVID-19 infection in chest X-ray images,” Sci. Rep., vol. 11, no. 1, pp. 1–6, 2021, doi: 10.1038/s41598-021-88807-2.
M. Owais, H. S. Yoon, T. Mahmood, A. Haider, H. Sultan, and K. R. Park, “Light-weighted ensemble network with multilevel activation visualization for robust diagnosis of COVID19 pneumonia from large-scale chest radiographic database,” Appl. Soft Comput., vol. 108, no. April, p. 107490, 2021, doi: 10.1016/j.asoc.2021.107490.
F. H. Ahmad and S. H. Wady, “COVID-19 Infection Detection from Chest X-Ray Images Using Feature Fusion and Machine Learning,” Sci. J. Cihan Univ. – Sulaimaniya, vol. 5, no. 2, pp. 10–30, 2021.
R. A. Hamaamin, S. H. Wady, and A. W. Kareem, “Classification of COVID-19 on Chest X-Ray Images Through the Fusion of HOG and LPQ Feature Sets,” vol. 4, no. 2, pp. 135–143, 2022, doi: 10.24271/psr.51.
Hamaamin, R. A., Wady, S. H., & Sangawi, A. W. K. (2022). COVID-19 Classification based on Neutrosophic Set Transfer Learning Approach. UHD Journal of Science and Technology, 6(2), 11–18. https://doi.org/10.21928/uhdjst.v6n2y2022.pp11-18.