THE EFFECT OF FEATURE EXTRACTION ON COVID-19 CLASSIFICATION

Rebin Abdulkareem Hamaamin1, Shakhawan Hares Wady 2, Ali Wahab Kareem Sangawi 3

1Computer Science , College of Sciences, Charmo University, Chamchamal, Sulaimani, KRG, Iraq

Email: - rebin.abdulkarim@charmouniversity.org

2 Business Administration, College of Business, Charmo University, Chamchamal, Sulaimani, KRG, Iraq

Email: - shakhawan.hares@charmouniversity.org

3 General Science, College of Education, Charmo University, Chamchamal, Sulaimani, KRG, Iraq

Email: - ali.kareem@charmouniversity.org

 

Received: 15 Sep., 2023/ Accepted: 13 Mar., 2024 / Published: 3 June., 2024.             https://doi.org/10.25271/sjuoz.2024.12.2.1204

ABSTRACT:                                        

X-ray imaging stands as a prominent technique for diagnosing COVID-19, and it also serves as a crucial tool in the medical field for the analysis of various diseases. Numerous approaches are available to facilitate this analysis. Among these techniques, one involves the utilization of a Feature Extractor, which effectively captures pertinent characteristics from X-ray images. In a recent study, a comprehensive examination was conducted using 25 distinct feature extractors on X-ray images specific to COVID-19 cases. These images were categorized into two classes: COVID-19-positive and non-COVID-19. To enable a thorough evaluation, a sequence of machine learning classifiers was employed on these categorized images. The outcomes derived from this experimentation gauged the magnitude of impact that each individual feature exerted on COVID-19-related imagery. This assessment aimed to determine the efficacy levels of various feature extractors in terms of detection capability. Consequently, a distinction emerged between the more effective and less effective feature extractors, shedding light on their varying degrees of contribution to the detection process. Moreover, the comparative performance of different classifiers became evident, revealing the classifiers that exhibited superior performance when measured against their counterparts.

KEYWORDS: - Feature extraction, Machine learning, Effect of feature, X-Ray, Covid-19.

 


1.        INTRODUCTION


         The COVID-19 virus has a significant impact on individuals around the globe. The early diagnosis of this infectious disease is critical to prevent its global and local spread. Generally, scientists have tested numerous ways and methods to detect people and analyze the virus. Interestingly, one of the methods used for COVID-19 diagnosis is X-Rays that recognize whether the person is infected or not. Moreover, the researchers attempted to use methods and technologies that yielded quicker and more accurate results [1][2]. A series of feature extractors can be used to extract features from an image. A texture feature is one kind of feature. The texture is an important low-level feature in images, it can be used to characterize the contents of an image or a region. In addition to color features, color features alone cannot identify the image because distinct images can have similar histograms [3].

        The essential factor here is to identify the relevant features and then choose the suitable classification method for these features. Once a set of features can be extracted from the images, it sends those features to machine learning for Classification. Classification is the process of categorizing a set of data into classes. It can be done on both structured and unstructured data. Determining the class of incoming data points is the first step in the procedure. The classes are also known as the target, label, or categories. Several machine learning classifiers are used to classify data, such as Support Vector Machine (SVM), k-nearest neighbors (KNN), Ensemble, Decision Tree (DT), and Naïve Bayes (NB) [4].

        To review the previous studies, the first part of the research discusses the effect of feature extractors on images. The second part highlights the research study on the diagnosis of COVID-19, which include [5] classification of histopathology images and identification of malignant areas using each feature extractor Gray-Level Co-Occurrence Matrix (GLCM), Local Binary Pattern (LBP), Local Binary Gray Level Co-occurrence Matrix (LBGLCM), gray-level run-length matrix (GLRLM) and Segmentation-based Fractal Texture Analysis (SFTA), Using SVM, KNN, Linear Discriminant Analysis (LDA) and Boosted Tree classifiers. The feature extractor is GLCM, GLRM, LBP, CLBP, and LTP and has been tested on the following diseases Brain, Breast, Colon, Skin, and Thyroid [6]. Also, in [7], A technique has been developed by combining both HOG and LBP feature extractors to identify plant leaves by SVM classifier. Taha J. Alhindi et al. in [8] state that work has been done on medical imaging using LBP, HOG, and a pre-trained deep network for histopathology images. For this purpose, support vector machines, decision trees, and artificial neural networks were used to identify and classify images through the obtained features. Another texture feature used to detect breast cancer in women is


called Monogenic Binary Coding (MBC) [9]. Cigdem Turan and Kin-Man Lam in [10], some feature extractors were used in facial-expression recognition. All the features used are local descriptors. Seyyid Ahmed Medj in [11], states that a set of feature extractors has been developed to extract parts from images for both binary and multiclass. According to Samy Bakheet and Ayoub Al-Hamadi In [12] work has been done on the GLCM feature extractor to detect and diagnose COVID-19 patients based on X-ray images, with it, the latent-dynamic conditional random fields (LDCRFs) classifier. Also, in [13], feature extractors extract features from images. in [14], Violent event has been worked on with two feature extractors for detecting a violent event.

        The final section of this research discusses some papers that have worked on COVID-19 in general to diagnose the disease. Author in [15] proposed a new testing methodology to determine whether a patient has been infected by the COVID-19 virus using the SDD300 model. The in-depth feature plus SVM-based procedure was proposed in [16] for identifying coronavirus-infected patients by applying CXR images. SVM was utilized for classification rather than DL-based classifiers, which require an extensive database for training and validation. A. Helwan et al. in [17] introduced a transfer learning approach to diagnose patients who tested positive for COVID-19 and distinguish them from healthy patients using ResNet-18, ResNet-50, and DenseNet-201. For this purpose, 2617 chest CT images of non-COVID-19 and COVID-19 have been experimented with.


         M. Alruwaili et al. in [18] proposed an improved Inception-ResNetV2 DL model for accurately diagnosing chest CXR images. A Grad-CAM technique was also computed to improve the visibility of infected lung parts in CXR scans. D. Ji et al. in [19] presented a COVID-19 detection approach based on image modal feature fusion. Small-sample enhancement preprocessing, including spinning, translation, and randomized transformation, was initially conducted using this methodology. Five classic pertaining models, including VGG19, ResNet152, Xception, DenseNet201, and InceptionResnetV2, were used to extract CXR images' features. Pramod Gaur et al. in [20] presented an innovative methodology for preprocessing CT images and identifying COVID-19 positive and negative. The suggested approach used the principle of empiric wavelet transformation for preprocessing, with the optimal elements of the image's red, green, and blue channels being learned on the presented approach. Deep and transfer learning procedures as recommended by [21] to differentiate COVID-19 cases by assessing CXR images.

        This paper uses a series of feature extractors to extract features from X-ray images of COVID-19. All the parts are sent to different classifiers. Then they are measured by several metrics to see which feature has higher results and has less impact on COVID-19. The following is the flow of this paper. Section 2 highlights the methods used to extract features on the COVID-19 image and the classifiers used with the evaluation metrics. In section 3, All results are presented with different classifiers for each classifier, a set of metrics used to measure the data. The conclusion is described in the last section.

Material and Methodology

        The process by which features are extracted from images in this research is visually outlined in Figure (1). This diagram succinctly portrays the procedural steps and components utilized to carry out the feature extraction from the images under investigation. It serves as a graphical representation of the methodology, aiding in conveying a clear understanding of the sequential and conceptual aspects of how image features are harnessed for analysis within the context of this study. In this system, a standard database is used to test several feature extractors to distinguish between COVID-19 and non-COVID-19 by several classifiers to take the effect of each feature extractor to see which good results are measured by several metrics.


Figure 1: System overview.


1.1 Dataset

        The COVID-19 radiographic database, a publicly available database, is used in this paper. The database consists of 21165 CXR images, of which 10,192 belong to normal, 3616 correspond to COVID-19 positive, and 6012 belong to lung opacity (non-COVID lung infection), and 1345 are labeled as viral pneumonia cases. The data for this paper includes 7232 CXR images, 3616 of which with a positive COVID-19 diagnosis and 3616 negative cases randomly selected to create the balanced database. All images in the database have been preprocessed before being featured as follows: Figure (2) visually represents the sequential process involved in the manipulation of images. The initial step involves converting the images into grayscale, resulting in representations composed solely of varying shades of gray. In the subsequent step, all images are uniformly resized to dimensions of 512 x 512 pixels. Finally, the region of interest (ROI) is meticulously extracted from each image. This ROI encapsulates specific, critical features relevant to the analysis. The figure provides a graphical overview of these operations, effectively illustrating the transformation process, from grayscale conversion and resizing to the selective extraction of the region of interest.


 



Figure 2: Preprocessing overall framework


1.2.  Feature extraction

        Feature extraction is the technique of extracting numbers from data using unique processing that allows you to operate on the original data without destroying it. It produces better results than working with raw data, and it makes them faster. In this study, a total of twenty-five distinct feature extractors were employed to extract features from the data. Each of these extractors possessed a unique set of features, employing varied techniques, as comprehensively outlined in Table (1).


 


 

Table 1: The list of feature extractors.


No. Features

Feature Extractors

N

8

Gray Level Run Length Matrix (GLRLM)

1

22

Gray Level Co-occurrence Matrix (GLCM)

2

32

Weber Local Descriptor (WLD)

3

37

Local Gradient Increasing Pattern (LGIP)

4

48

Gabor Wavelet (GW)

5

56

Local Directional Pattern (LDiP)

6

56

Local Directional Number Pattern (LDNP)

7

72

Local Directional Texture Pattern (LDTP)

8

72

Wavelet Transform (WT)

9

81

Histogram of Oriented Gradients (HOG)

10

168

Pyramid of Histogram of Oriented Gradients (PHOG)

11

256

Local Monotonic Pattern (LMP)

12

256

Local Phase Quantization (LPQ)

13

256

Local Transitional Pattern (LTrP)

14

256

Median Binary Pattern (MBP)

15

256

Gradient Directional Pattern (GDP)

16

272

Local Arc Pattern (LAP)

17

512

Local Frequency Descriptor (LFD)

18

512

Median Ternary Pattern (MTP)

19

512

Local Ternary Pattern (LTeP)

20

512

Gradient Local Ternary Pattern (GLTP)

21

800

Median Robust Extended Local Binary Pattern (MRELBP)

22

1062

Binary Pattern of Phase Congruency (BPPC)

23

2048

Improved Weber Binary Coding (IWBC)

24

3072

Monogenic Binary Coding (MBC)

25


1.3. Classification and Evaluation

         After the features are extracted from the images, they are sent to the machine learning classifier. Basically, classifiers are algorithms that automatically sort or categorize data into one or more classes known as ‘Classification’. Classes are referred to as targets, labels, and categories. This research used a set of classifiers to separate the data into two types: COVID-19 and non-COVID-19. The classifiers used in this research are the KNN: which is a supervised machine learning algorithm method that can be used to solve both classification and regression problems and is simple to implement [22]. The second classifier is Ensemble, a method for creating a variety of base classifiers [23]. The third classifier is DT: This model works by parsing our data into a DT. The fourth classifier is NB: a set of straightforward "probabilistic classifiers" based on Bayes' theorem and strong (naive) independence [22]. Furthermore, the last classifier SVM, a supervised machine learning technique that can be used to solve classification and regression problems. In SVM used, different kernels for classifying data, the SVM linear kernel (SVML), SVM Radial Basis Function kernel (SVMRBF), and SVM polynomial kernel (SVMP) [24][25]. All three types of kernels were tested in the research. The final part of the study is to tether data with several different metrics to see how each feature receptor affects the image of COVID-19 [26].

2. EXPERIMENTAL RESULTS AND DISCUSSIONS

        Numerous texture feature descriptors as presented in Table (1), were generated for assessing the CXR images of COVID-19, which is either positive or negative, and the result were given as second opinion to the clinicians to make exact decisions. Besides, to classify the CXR images, five well-known ML classifiers were utilized, namely, SVM (linear, RBF, and polynomial), KNN, Ensemble, DT, and NB classifiers. From the experiments performed on COVID-19 radiography dataset, the analysis of each descriptor was estimated in terms of a series of indicators (F-Score, Matthew's correlation coefficient (MCC) specificity, precision, sensitivity, and Accuracy) with Ensemble (Table 2), KNN (Table 3), NB (Table 4), DT (Table 5), SVML (Table 6), SVMRBF (Table 7), and SVMP (Table 8) classifiers respectively.

        As depicted in the initial experiment detailed in Table (2), the MRELBP feature descriptor emerged as the frontrunner. It exhibited the highest accuracy rate of 95.09% across all metrics, in tandem with the Ensemble classifier. In contrast, the classification accuracy stemming from the GLRLM method lagged behind, registering a comparatively lower score of 78.15%.


Table 2: All the feature extractors used in the Ensemble Classifier


Measurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- Score

Feature Extractor

N

 

86.76

87.13

86.27

87.25

73.53

43.35

WT

1

 

84.46

85.54

82.95

85.98

68.96

42.11

GLCM

2

 

78.15

78.25

78.01

78.30

56.33

39.07

GLRLM

3

 

91.65

91.64

91.69

91.62

83.31

45.83

HOG

4

 

84.61

86.68

81.78

87.43

69.33

42.08

GW

5

 

92.92

92.81

93.06

92.78

85.85

46.47

BPPC

6

 

90.17

90.46

89.83

90.51

80.36

45.07

GDP

7

 

90.92

91.56

90.15

91.69

81.86

45.43

GLTP

8

 

90.19

90.62

89.70

90.69

80.42

45.08

LAP

9

 

89.05

89.84

88.06

90.03

78.12

44.47

LDNP

10

 

84.32

84.73

83.75

84.90

68.66

42.12

LDTP

11

 

89.26

89.21

89.32

89.20

78.52

44.63

LFD

12

 

92.08

92.11

92.06

92.10

84.17

46.04

LMP

13

 

92.47

92.60

92.32

92.61

84.94

46.23

LPQ

14

 

90.10

89.76

90.53

89.67

80.20

45.07

LTrP

15

 

89.23

89.52

88.87

89.59

78.47

44.60

LDiP

16

 

81.19

81.47

80.77

81.60

62.39

40.56

WLD

17

 

91.87

92.50

91.13

92.60

83.76

45.91

MBC

18

 

92.76

92.87

92.64

92.88

85.53

46.38

IWBC

19

 

90.96

91.20

90.69

91.23

81.94

45.47

PHOG

20

 

90.62

90.61

90.65

90.59

81.25

45.31

LGIP

21

 

95.09

95.17

95.01

95.17

90.19

47.54

MRELBP

22

 

89.81

90.08

89.47

90.14

79.62

44.89

MBP

23

 

82.19

83.51

80.24

84.15

64.45

40.92

MTP

24

 

89.44

90.13

88.59

90.29

78.90

44.68

LTeP

25


         Upon the utilization of the KNN classifier, distinct categories of features exhibited disparate responses in terms of their projected performance. The findings, as elucidated in Table (3), indicate that the MRELBP method showcased notable efficacy, achieving an accuracy rate of 91.59% in the classification of chest X-ray (CXR) images when juxtaposed with the array of alternative techniques implemented.


Table 3: All the feature extractors used in the KNN Classifier


Measurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- score

Feature extractor

N

 

85.59

84.98

86.49

84.70

71.21

42.86

WT

1

 

84.00

83.86

84.22

83.79

68.02

42.02

GLCM

2

 

75.30

74.20

77.58

73.02

50.66

37.93

GLRLM

3

 

89.74

88.10

91.89

87.58

79.56

44.98

HOG

4

 

82.25

81.69

83.15

81.34

64.53

41.21

GW

5

 

91.20

90.30

92.34

90.07

82.44

45.65

BPPC

6

 

88.24

86.64

90.43

86.04

76.56

44.25

GDP

7

 

85.21

82.88

88.77

81.66

70.62

42.86

GLTP

8

 

86.96

85.47

89.09

84.83

74.00

43.62

LAP

9

 

84.98

83.19

87.69

82.27

70.07

42.69

LDNP

10

 

79.71

80.13

79.06

80.36

59.44

39.80

LDTP

11

 

85.51

83.39

88.73

82.30

71.18

42.99

LFD

12

 

88.58

85.52

92.89

84.26

77.45

44.53

LMP

13

 

90.16

88.34

92.53

87.79

80.41

45.19

LPQ

14

 

85.74

82.17

91.31

80.17

71.95

43.25

LTrP

15

 

85.74

83.93

88.44

83.04

71.61

43.06

LDiP

16

 

77.75

76.51

80.15

75.35

55.58

39.14

WLD

17

 

90.70

89.04

92.84

88.56

81.48

45.45

MBC

18

 

88.82

84.94

94.41

83.24

78.16

44.71

IWBC

19

 

87.06

84.17

91.30

82.82

74.40

43.80

PHOG

20

 

87.35

84.76

91.09

83.61

74.92

43.91

LGIP

21

 

91.59

90.39

93.10

90.08

83.23

45.86

MRELBP

22

 

85.73

82.70

90.36

81.09

71.76

43.18

MBP

23

 

77.32

77.48

77.04

77.61

54.65

38.63

MTP

24

 

81.06

79.43

83.86

78.26

62.23

40.79

LTeP

25


        Another classifier used in the research is NB. It is a probabilistic machine learning algorithm based on the Bayes theorem. This classifier has a problem that when a feature that is output contains a zero result, this classifier does not output a result, and the result is non. As depicted in Table (4), the feature structures encompassing MRELBP, MBP, MTP, and LTep have yielded non-results, reflecting zero outcomes in their respective feature sets. In contrast, among the other feature extractors explored in this study, the highest achievement was observed with the PHOG feature extractor, achieving a noteworthy accuracy score of 80.80% in the accuracy metric specifically for COVID-19 detection. Notably, when these results are juxtaposed with the preceding ones, it becomes apparent that they fall short in comparison to the outcomes generated by the diverse classifier implementations.


Table 4: All the feature extractors used in the NB Classifier


Measurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- score

Feature extractor

N

 

75.80

76.47

74.56

77.03

51.62

37.75

WT

1

 

76.18

77.93

73.04

79.32

52.47

37.70

GLCM

2

 

67.32

63.92

79.57

55.08

35.75

35.45

GLRLM

3

 

77.95

75.19

83.47

72.43

56.26

39.56

HOG

4

 

72.14

80.09

58.95

85.33

45.92

33.96

GW

5

 

77.39

73.01

86.93

67.84

55.81

39.68

BPPC

6

 

76.39

74.52

80.21

72.57

52.95

38.63

GDP

7

 

70.80

71.52

69.14

72.45

41.62

35.15

GLTP

8

 

75.78

71.34

86.21

65.35

52.73

39.04

LAP

9

 

74.58

71.98

80.48

68.67

49.51

38.00

LDNP

10

 

65.06

84.66

36.79

93.32

36.51

25.65

LDTP

11

 

72.91

68.75

84.02

61.80

47.01

37.81

LFD

12

 

76.92

73.81

83.47

70.36

54.32

39.17

LMP

13

 

79.45

76.45

85.13

73.76

59.28

40.28

LPQ

14

 

71.63

68.11

81.40

61.85

44.13

37.08

LTrP

15

 

72.79

69.94

79.94

65.64

46.07

37.30

LDiP

16

 

70.55

77.29

58.19

82.90

42.41

33.20

WLD

17

 

78.09

74.60

85.19

71.00

56.76

39.77

MBC

18

 

77.27

74.02

84.04

70.50

55.05

39.36

IWBC

19

 

80.80

77.92

85.98

75.62

61.94

40.87

PHOG

20

 

73.06

69.78

81.37

64.74

46.77

37.57

LGIP

21

 

-

-

-

-

-

-

MRELBP

22

 

-

-

-

-

-

-

MBP

23

 

-

-

-

-

-

-

MTP

24

 

-

-

-

-

-

-

LTeP

25


 


         DT is another classifier implemented in this paper. This classifier works as a tree on the nodes and then extends the branches down. The only reason these classifiers were used in these feature extractions was to examine how they performed compared to other classifiers. The results obtained from this classifier are weaker than the KNN and ensemble classifiers, presented in Table (5), parallel to the assortment of classifiers, the optimal feature extractor remains MRELBP. Remarkably, among the alternative feature extraction approaches, Access achieves the peak outcome of 85.50%, standing out as the most elevated result among the other feature extractors.


Table 5: All the feature extractors used in the DT Classifier


Measurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- score

Feature extractor

N

 

79.92

79.61

80.44

79.39

59.85

40.01

WT

1

 

80.18

80.09

80.35

80.01

60.38

40.11

GLCM

2

 

72.49

72.20

73.20

71.78

44.99

36.35

GLRLM

3

 

82.72

82.20

83.54

81.89

65.45

41.43

HOG

4

 

76.58

76.50

76.79

76.38

53.19

38.32

GW

5

 

83.24

83.21

83.28

83.20

66.49

41.62

BPPC

6

 

82.00

81.22

83.28

80.72

64.05

41.12

GDP

7

 

81.88

81.54

82.45

81.31

63.79

41.00

GLTP

8

 

81.23

81.05

81.55

80.91

62.48

40.65

LAP

9

 

80.03

79.96

80.15

79.92

60.08

40.03

LDNP

10

 

76.75

77.02

76.32

77.18

53.54

38.33

LDTP

11

 

80.50

80.20

81.01

79.99

61.01

40.30

LFD

12

 

83.26

83.20

83.40

83.13

66.55

41.65

LMP

13

 

83.69

83.24

84.40

82.99

67.41

41.91

LPQ

14

 

80.77

80.48

81.27

80.26

61.55

40.44

LTrP

15

 

81.92

81.23

83.08

80.76

63.88

41.07

LDiP

16

 

72.79

72.47

73.53

72.05

45.59

36.50

WLD

17

 

82.32

82.67

81.80

82.85

64.66

41.12

MBC

18

 

83.09

82.64

83.80

82.38

66.20

41.61

IWBC

19

 

80.13

79.61

81.02

79.24

60.28

40.16

PHOG

20

 

82.90

83.15

82.53

83.26

65.80

41.42

LGIP

21

 

85.50

85.69

85.26

85.74

71.03

42.74

MRELBP

22

 

80.59

80.43

80.86

80.32

61.19

40.32

MBP

23

 

75.08

74.80

75.75

74.41

50.21

37.64

MTP

24

 

77.94

77.85

78.15

77.73

55.90

39.00

LTeP

25


 


         Tables (6 - 8) comprehensively present the performance outcomes of all SVM models. Drawing insights from the research findings across the three distinct kernels, a consistent pattern emerges: the features extracted through the utilization of MRELBP techniques consistently outshine alternative scenarios. This phenomenon is evident when employing the SVML, SVMRBF, and SVMP classifiers, where the corresponding accuracy rates for MRELBP features stand at 85.50%, 95.10%, and an impressive 95.87%, respectively. This consistency in the    superiority of MRELBP-derived features across diverse SVM kernels underscores the robustness and effectiveness of this feature extraction approach. The exceptional accuracy rates obtained with SVMP highlight its particular aptitude in capturing the relevant patterns present in the data, contributing to accurate classification of COVID-19 images. Consequently, these findings emphasize the potential of MRELBP-based features in enhancing the performance and accuracy of machine learning models for disease


detection and classification.

 

 

 

 


 

 

 

 

 

Table 6: All the feature extractors used in the SVML Classifier


Measurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- score

Feature extractor

N

 

79.20

79.05

79.47

78.92

58.41

39.63

WT

1

 

77.30

79.90

72.99

81.62

54.82

38.14

GLCM

2

 

76.78

72.77

85.66

67.91

54.45

39.35

GLRLM

3

 

88.86

87.96

90.06

87.66

77.75

44.50

HOG

4

 

83.18

89.10

75.63

90.73

67.14

40.91

GW

5

 

92.45

92.39

92.53

92.37

84.91

46.23

BPPC

6

 

87.82

88.48

86.97

88.67

75.66

43.86

GDP

7

 

88.96

90.44

87.12

90.79

77.97

44.38

GLTP

8

 

86.82

86.74

86.93

86.71

73.64

43.42

LAP

9

 

83.95

84.17

83.62

84.27

67.91

41.95

LDNP

10

 

82.54

86.08

77.63

87.44

65.40

40.82

LDTP

11

 

86.90

86.02

88.13

85.67

73.84

43.53

LFD

12

 

89.72

89.16

90.44

88.99

79.45

44.90

LMP

13

 

91.67

91.49

91.91

91.42

83.35

45.85

LPQ

14

 

86.24

86.29

86.18

86.29

72.49

43.12

LTrP

15

 

87.98

88.29

87.59

88.37

75.97

43.97

LDiP

16

 

78.02

80.85

73.49

82.56

56.30

38.50

WLD

17

 

92.51

93.19

91.74

93.28

85.04

46.23

MBC

18

 

90.82

90.33

91.45

90.18

81.66

45.44

IWBC

19

 

89.24

88.39

90.35

88.13

78.50

44.68

PHOG

20

 

85.67

86.71

84.29

87.05

71.39

42.74

LGIP

21

 

92.86

91.82

94.12

91.60

85.76

46.48

MRELBP

22

 

87.16

88.33

85.66

88.66

74.36

43.49

MBP

23

 

81.72

85.72

76.11

87.32

63.84

40.32

MTP

24

 

85.20

87.74

81.84

88.56

70.57

42.34

LTeP

25

 

Table 7: All the feature extractors used in the SVMRBF Classifier

Measurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- score

Feature extractor

N

 

88.81

89.95

87.41

90.21

77.67

44.33

WT

1

 

86.11

88.05

83.58

88.64

72.33

42.88

GLCM

2

 

79.11

80.19

77.36

80.86

58.28

39.37

GLRLM

3

 

93.26

93.53

92.96

93.57

86.53

46.62

HOG

4

 

85.48

90.74

79.03

91.92

71.56

42.24

GW

5

 

94.57

94.37

94.80

94.34

89.15

47.29

BPPC

6

 

91.73

91.87

91.56

91.89

83.47

45.86

GDP

7

 

90.17

91.98

88.01

92.32

80.42

44.98

GLTP

8

 

91.25

91.86

90.54

91.96

82.52

45.60

LAP

9

 

90.84

90.77

90.93

90.75

81.69

45.42

LDNP

10

 

80.94

77.77

86.69

75.19

62.31

40.99

LDTP

11

 

90.73

90.74

90.75

90.72

81.48

45.37

LFD

12

 

93.89

94.56

93.15

94.63

87.80

46.93

LMP

13

 

94.46

94.54

94.38

94.54

88.93

47.23

LPQ

14

 

91.12

91.06

91.20

91.04

82.25

45.57

LTrP

15

 

90.30

91.26

89.14

91.45

80.63

45.09

LDiP

16

 

83.13

85.00

80.50

85.75

66.36

41.34

WLD

17

 

93.14

93.85

92.34

93.94

86.30

46.54

MBC

18

 

93.33

92.88

93.86

92.79

86.67

46.68

IWBC

19

 

92.54

93.00

92.02

93.07

85.10

46.25

PHOG

20

 

92.10

92.50

91.65

92.56

84.22

46.03

LGIP

21

 

95.10

95.78

94.36

95.84

90.21

47.53

MRELBP

22

 

91.96

92.40

91.44

92.48

83.92

45.96

MBP

23

 

75.66

71.49

85.41

65.91

52.33

38.91

MTP

24

 

84.00

81.13

88.66

79.35

68.31

42.36

LTeP

25

 

Table 8: All the feature extractors used in the SVMP Classifier

 

easurement Metrics

 

 

Accuracy

Sensitivity

Precision

Specificity

MCC

F- score

Feature extractor

N

 

87.52

88.14

86.72

88.33

75.07

43.71

WT

1

 

86.67

88.66

84.12

89.23

73.46

43.17

GLCM

2

 

56.37

51.60

80.30

32.43

16.26

31.41

GLRLM

3

 

93.40

92.93

93.96

92.85

86.82

46.72

HOG

4

 

89.12

89.02

89.27

88.98

78.26

44.57

GW

5

 

94.41

94.22

94.63

94.19

88.83

47.21

BPPC

6

 

91.42

90.76

92.25

90.59

82.87

45.75

GDP

7

 

90.94

90.77

91.16

90.72

81.89

45.48

GLTP

8

 

91.09

90.11

92.31

89.86

82.21

45.60

LAP

9

 

91.76

90.79

92.97

90.55

83.56

45.93

LDNP

10

 

71.87

71.91

86.47

57.28

44.33

39.26

LDTP

11

 

90.80

90.08

91.70

89.90

81.62

45.44

LFD

12

 

94.18

93.70

94.73

93.62

88.36

47.11

LMP

13

 

95.01

94.47

95.63

94.40

90.04

47.52

LPQ

14

 

91.24

90.34

92.37

90.11

82.50

45.67

LTrP

15

 

91.59

90.62

92.79

90.39

83.21

45.85

LDiP

16

 

86.19

85.44

87.28

85.10

72.41

43.17

WLD

17

 

93.64

93.60

93.71

93.58

87.30

46.83

MBC

18

 

93.72

92.96

94.62

92.82

87.47

46.89

IWBC

19

 

92.72

92.25

93.29

92.16

85.46

46.38

PHOG

20

 

93.41

92.76

94.18

92.64

86.83

46.73

LGIP

21

 

95.87

95.48

96.31

95.44

91.75

47.94

MRELBP

22

 

91.99

91.30

92.84

91.15

84.00

46.03

MBP

23

 

53.88

60.37

81.15

26.61

11.61

34.62

MTP

24

 

88.96

88.85

89.11

88.80

77.92

44.49

LTeP

25

 


 



         Upon extensive evaluation of various feature extractors with diverse classifiers, noteworthy findings have emerged. Notably, the most effective feature extractor across all classifiers, with the exception of the NB classifier, is determined to be MRELBP. However, it is worth highlighting that MRELBP yields no results within the NB classifier due to its lack of features, resulting in a count of zero. Intriguingly, within the context of the NB classifier, the optimal feature extractor is identified as PHOG. These outcomes, outlined in Figure 2, distinctly outline the superiority of specific feature extractors within distinct classifier contexts, thus contributing valuable insights to the overall research findings.


 

 Figure 3: effect of feature extraction on Covid-19 images in different classifier


 


         In our analysis of various works, it has become evident that the features employed in this study surpass several other functions, particularly those highlighted in Table (9). However, the real potential lies in the synergy achieved through combining these features, leading to improved results. This underscores the importance of a holistic approach in leveraging the strengths of individual features for a more comprehensive and effective outcome.


 

 

 

 

 

Table 9: Comparison of proposed classification accuracy with existing state-of-the-art techniques


Previous study

Methods/ classifier

Accuracy (%)

A. I. Khan et al[27]

(CNN) CoroNet

89.60

A. Zargari Khuzani et al[28]

ResNet-18

92.49

L. Wang et al[1]

COVID-Net

93.30

M. Owais et al[29]

Ensemble-Net

94.72

F. Saiz and I. Barandiaran[15]

SDD300

94.92

F. H. Ahmad and S. H. Wady[30]

SVM (CT+GWT+LGIP)

96.18

R. A. Hamaamin et al[31]

HOG +LPQ

97.15

R. A. Hamaamin et al[32]

Resnet 50

98.05

Proposed

MRELBP

95.87


CONCLUSION

         Due to the global prevalence of the COVID-19 pandemic, a multitude of researchers have been diligently striving to devise an intelligent approach for the swift identification of the disease. One prominent technique within this realm involves the utilization of feature extractors to distill pertinent information from images. In this pursuit, a particular study undertook the deployment of a series of feature extractors to scrutinize COVID-19 images, aiming to discern the optimal feature extraction methodology.

        Through a meticulous evaluation of diverse feature extractors, this research endeavors to unearth the most efficacious one. Among the twenty-five feature extractors subjected to assessment, certain features emerged with exceptional efficacy in delineating the disease. Notably, the investigation pinpointed "MRELBP" as the preeminent feature extractor, showcasing a remarkable accuracy rate of 95.87% when employed in conjunction with the SVMP classifier. It is worth mentioning that this study exhibited a comprehensive approach, encompassing an array of classifiers to culminate in its findings. Moreover, a multifaceted set of metrics was employed to gauge the performance of each featured extractor.

The primary issue with the study lies in the time-consuming nature of obtaining results for certain features. Furthermore, some features are sourced from machine learning missions that were not undertaken

REFERENCES

Wei Wang, Yutao Li, Ji Li, Peng Zhang, Xin Wang, "Detecting COVID-19 in Chest X-Ray Images via MCFF-Net", Computational Intelligence and Neuroscience, vol. 2021, Article ID 3604900, 8 pages, 2021. https://doi.org/10.1155/2021/3604900.

Ebrahim Mohammed Senan, Ali Alzahrani, Mohammed Y. Alzahrani, Nizar Alsharif, Theyazn H. H. Aldhyani, "Automated Diagnosis of Chest X-Ray for Early Detection of COVID-19 Disease", Computational and Mathematical Methods in Medicine, vol. 2021, Article ID 6919483, 10 pages, 2021. https://doi.org/10.1155/2021/6919483.

J. Lian, M. Zhang, N. Jiang, W. Bi, and X. Dong, “Feature Extraction of Kidney Tissue Image Based on Ultrasound Image Segmentation,” J. Healthc. Eng., vol. 2021, 2021, doi: 10.1155/2021/9915697.

Nugrahaeni, Ratna Astuti and Kusprasapta Mutijarsa. “Comparative analysis of machine learning KNN, SVM, and random forests algorithm for facial expression classification.” 2016 International Seminar on Application for Technology of Information and Communication (ISemantic) (2016): 163-168.

Ş. Öztürk and B. Akdemir, “ScienceDirect ScienceDirect ScienceDirect Application of Feature Extraction and Classification Methods for Histopathological Image using GLCM , Application of Feature Extraction and Classification Methods for and GLCM , Histopathological Image using a SFTA,” Procedia Comput. Sci., vol. 132, no. Iccids, pp. 40–46, 2018, doi: 10.1016/j.procs.2018.05.057.

Vani Kumari, S., Usha Rani, K. (2020). Analysis on Various Feature Extraction Methods for Medical Image Classification. In: Jyothi, S., Mamatha, D., Satapathy, S., Raju, K., Favorskaya, M. (eds) Advances in Computational and Bio-Engineering. CBE 2019. Learning and Analytics in Intelligent Systems, vol 16. Springer, Cham. https://doi.org/10.1007/978-3-030-46943-6_3.

Keivani, M., Mazloum, J., Sedaghatfar, E., Tavakoli, M.B. (2020). Automated analysis of leaf shape, texture, and color features for plant classification. Traitement du Signal, Vol. 37, No. 1, pp. 17-28. https://doi.org/10.18280/ts.370103.

T. J. Alhindi, S. Kalra, K. H. Ng, A. Afrin, and H. R. Tizhoosh, “Comparing LBP , HOG and Deep Features for Classification of Histopathology Images,” 2018 Int. Jt. Conf. Neural Networks, pp. 1–7, 2018, doi: 10.1109/IJCNN.2018.8489329.

E. Mohammadi, E. Fatemizadeh, H. Sheikhzadeh, and S. Khoubani, “A Textural Approach for Recognizing Architectural Distortion In Mammograms,” pp. 136–140, 2013.

C. Turan and K. Lam, “Histogram-based Local Descriptors for Facial Expression Recognition ( FER ): A comprehensive Study,” J. Vis. Commun. Image Represent., 2018, doi: 10.1016/j.jvcir.2018.05.024.

S. A. Medjahed, “A Comparative Study of Feature Extraction Methods in Images Classification,” no. February, pp. 16–23, 2015, doi: 10.5815/ijigsp.2015.03.03.

S. Bakheet and A. Al-hamadi, “Automatic detection of COVID-19 using pruned GLCM-Based texture features and LDCRF classification,” Comput. Biol. Med., vol. 137, no. August, p. 104781, 2021, doi: 10.1016/j.compbiomed.2021.104781.

Saipullah, K.M., Kim, DH. A robust texture feature extraction using the localized angular phase. Multimed Tools Appl 59, 717–747 (2012). https://doi.org/10.1007/s11042-011-0766-5.

M. Aradhya and D. S. Guru, “Violent Video Event Detection Based on Integrated LBP and GLCM Texture Features Revue d ’ Intelligence Artificielle Violent Video Event Detection Based on Integrated LBP and GLCM Texture Features,” no. May, 2020, doi: 10.18280/ria.340208.

F. Saiz and I. Barandiaran, “COVID-19 Detection in Chest X-ray Images using a Deep Learning Approach,” Int. J. Interact. Multimed. Artif. Intell., vol. 6, no. 2, p. 4, 2020, doi: 10.9781/ijimai.2020.04.003.

A. K. Singh, A. Kumar, M. Mahmud, M. S. Kaiser, and A. Kishore, “COVID-19 Infection Detection from Chest X-Ray Images Using Hybrid Social Group Optimization and Support Vector Classifier,” Cognit. Comput., no. 0123456789, 2021, doi: 10.1007/s12559-021-09848-3.

A. Helwan, M. K. S. Ma’Aitah, H. Hamdan, D. U. Ozsahin, and O. Tuncyurek, “Radiologists versus Deep Convolutional Neural Networks: A Comparative Study for Diagnosing COVID-19,” Comput. Math. Methods Med., vol. 2021, 2021, doi: 10.1155/2021/5527271.

M. Alruwaili, A. Shehab, and S. Abd El-Ghany, “COVID-19 Diagnosis Using an Enhanced Inception-ResNetV2 Deep Learning Model in CXR Images,” J. Healthc. Eng., vol. 2021, no. Dl, 2021, doi: 10.1155/2021/6658058.

D. Ji, Z. Zhang, Y. Zhao, and Q. Zhao, “Research on Classification of COVID-19 Chest X-Ray Image Modal Feature Fusion Based on Deep Learning,” J. Healthc. Eng., vol. 2021, 2021, doi: 10.1155/2021/6799202.

P. Gaur, V. Malaviya, A. Gupta, G. Bhatia, R. B. Pachori, and D. Sharma, “COVID-19 disease identification from chest CT images using empirical wavelet transformation and transfer learning,” Biomed. Signal Process. Control, vol. 71, no. PA, p. 103076, 2022, doi: 10.1016/j.bspc.2021.103076.

Talal S. Qaid, Hussein Mazaar, Mohammad Yahya H. Al-Shamri, Mohammed S. Alqahtani, Abeer A. Raweh, Wafaa Alakwaa, "Hybrid Deep-Learning and Machine-Learning Models for Predicting COVID-19", Computational Intelligence and Neuroscience, vol. 2021, Article ID 9996737, 11 pages, 2021. https://doi.org/10.1155/2021/9996737

Aljabri, Malak, Amal A. Alahmadi, Rami Mustafa A. Mohammad, Menna Aboulnour, Dorieh M. Alomari, and Sultan H. Almotiri. 2022. "Classification of Firewall Log Data Using Multiclass Machine Learning Models" Electronics 11, no. 12: 1851. https://doi.org/10.3390/electronics11121851.

 

Dietterich, T.G. (2000). Ensemble Methods in Machine Learning. In: Multiple Classifier Systems. MCS 2000. Lecture Notes in Computer Science, vol 1857. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45014-9_1

S. A. Jafar Zaidi, I. Chatterjee, and S. Brahim Belhaouari, “COVID-19 Tweets Classification during Lockdown Period Using Machine Learning Classifiers,” Appl. Comput. Intell. Soft Comput., vol. 2022, 2022, doi: 10.1155/2022/1209172.

J. Zeng et al., “Prediction of peak particle velocity caused by blasting through the combinations of boosted-chaid and svm models with various kernels,” Appl. Sci., vol. 11, no. 8, 2021, doi: 10.3390/app11083705.

U. Özkaya, Ş. Öztürk, and M. Barstugan, “Coronavirus (COVID-19) Classification Using Deep Features Fusion and Ranking Technique,” Stud. Big Data, vol. 78, pp. 281–295, 2020, doi: 10.1007/978-3-030-55258-9_17.

A. I. Khan, J. L. Shah, and M. M. Bhat, “CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images,” Comput. Methods Programs Biomed., vol. 196, p. 105581, 2020, doi: 10.1016/j.cmpb.2020.105581.

 

A. Zargari Khuzani, M. Heidari, and S. A. Shariati, “COVID-Classifier: an automated machine learning model to assist in the diagnosis of COVID-19 infection in chest X-ray images,” Sci. Rep., vol. 11, no. 1, pp. 1–6, 2021, doi: 10.1038/s41598-021-88807-2.

M. Owais, H. S. Yoon, T. Mahmood, A. Haider, H. Sultan, and K. R. Park, “Light-weighted ensemble network with multilevel activation visualization for robust diagnosis of COVID19 pneumonia from large-scale chest radiographic database,” Appl. Soft Comput., vol. 108, no. April, p. 107490, 2021, doi: 10.1016/j.asoc.2021.107490.

F. H. Ahmad and S. H. Wady, “COVID-19 Infection Detection from Chest X-Ray Images Using Feature Fusion and Machine Learning,” Sci. J. Cihan Univ. – Sulaimaniya, vol. 5, no. 2, pp. 10–30, 2021.

R. A. Hamaamin, S. H. Wady, and A. W. Kareem, “Classification of COVID-19 on Chest X-Ray Images Through the Fusion of HOG and LPQ Feature Sets,” vol. 4, no. 2, pp. 135–143, 2022, doi: 10.24271/psr.51.

Hamaamin, R. A., Wady, S. H., & Sangawi, A. W. K. (2022). COVID-19 Classification based on Neutrosophic Set Transfer Learning Approach. UHD Journal of Science and Technology, 6(2), 11–18. https://doi.org/10.21928/uhdjst.v6n2y2022.pp11-18.