HANDWRITTEN CHARACTER RECOGNITION IN ASSYRIAN LANGUAGE USING CONVOLUTIONAL NEURAL NETWORK

Revella Eshaya Armya a, Maiwan Bahjat Abdulrazzaq b

a Technical College of Informatics, Akre, Kurdistan Region, Iraq – revella.eshaya@dpu.edu.krd

b Faculty of Science, University of Zakho, Zakho, Kurdistan Region, Iraq – maiwan.abdulrazzaq@uoz.edu.krd

 

Received: 14 Aug., 2023 / Accepted: 1 Nov., 2023 / Published: 28 Mar., 2024.                https://doi.org/10.25271/sjuoz.2024.12.1.1189

ABSTRACT:

Academics and researchers worldwide have paid close attention to biometric handwriting recognition using deep learning as much research has been proposed to enhance biometric recognition in the past and in recent years. Several solutions for character recognition systems in various languages, including Chinese, English, Japanese, Arabic, and Kurdish have been developed. Unfortunately, there has been minimal growth in the Assyrian language. There is still little research on Assyrian handwriting. In this paper, a new Assyrian language dataset was created as part of the procedure by distributing 500 forms consisting of 36 Assyrian characters to people between the ages of 13 and 60 of both genders. The preprocessing operation includes cleaning the noisy data and segmenting each image to 224x224 pixels. This effort resulted in the collection of 18,000 images of these characters to be trained 70% and tested 30% in four CNN models, VGG16, VGG19, MobileNet-V2, and ResNet-50, over 30 epochs to give an accuracy rate of 90.97%, 92.06%, 95.70%, and 94.97%., respectively.

KEYWORDS: Deep Learning, Convolutional Neural Network, Handwritten Character Recognition, Assyrian Language.


1.     INTRODUCTION

        Due to recent advances in deep learning models, handwritten character recognition has virtually addressed the problem for many popular languages. For many other languages, however, the detection of handwritten characters remains challenging due to a lack of sufficiently large labelled datasets required for training deep learning models (Jayasundara et al., 2019).

        Character handwritten recognition systems in a variety of languages, including Chinese, English, Japanese, Arabic, and Kurdish have been the subject of a large number of potential solutions. On the other hand, there was no significant progress in the Assyrian language. As such, the recognition of Assyrian handwritten characters remains a current and relatively unaddressed research problem. Convolutional Neural Networks (CNNs) and other algorithms based on Deep Learning provide the ability to independently acquire the distinctive features of images without relying on human intervention. The Convolutional Neural Network (CNN) architecture is a more sophisticated iteration of the multi-layer perceptron (MLP) foundation. The functionality of the CNN framework has similarities to that of the human brain (Putri, Pratomo, & Azhari, 2023). Humans use their naked eyes to discover and distinguish items by viewing hundreds of objects (Albahli, Nawaz, Javed, & Irtaza). CNN uses the same patterns to see and recognize things. GoogleNet, AlexNet, VGG, and ResNet are some notable CNN examples. CNN networks combine key point detection and classification with little preprocessing and computation (Parikh & Desai, 2022).

        In addition, CNN has accomplished innovative feats due to their abilities to encode in-depth information and spatial awareness. CNNs are adept at comprehending both minute and large details in images, however as layers are pooled, critical data  are lost, and CNNs require a huge number of training samples (usually thousands or tens of thousands per class) to train and classify images successfully. Thus, there is considerable interest in training CNNs with fewer training samples (Jayasundara et al., 2019).

        For all these reasons the present research aimed to Create a new dataset for 500 writers of handwritten Assyrian language that consists of 36 characters, classify the Assyrian handwritten recognition using different Convolutional Neural Network models to achieve the highest accuracy rate, and compare the performance of the Assyrian handwritten recognition model with state-of-the-art models in other languages.

        The original zone of the Assyrian language is Upper Mesopotamia, southeastern Anatolia, northwestern Iran, and the northeastern Levant. This vast region stretches from the plain of Urmia in western Iran to the Nineveh Plains, Kirkuk, Erbil, and Duhok regions in northern Iraq, as well as the northern regions of Syria and southcentral and southern Turkey (Frederick Mario Fales, 2023), (De Ridder, 2018).

        However, Assyrians living in diaspora groups may instead use loanwords borrowed from the languages spoken by those communities (Frederick M Fales, 2021). So, instability in the Middle East over the course of the previous century has resulted in a global diaspora of Assyrian speakers, with the majority currently calling regions like North and South America, Australia, Europe, or Russia home. Those who speak Assyrian are of Assyrian ethnicity and are descended from the first people who inhabited Mesopotamia (Benjamen, 2022).

        The Assyrian language uses the Madnḥāyā Syriac alphabet and is written from right to left. Along with other current Aramaic languages, it is believed that the Assyrian language is endangered because younger Assyrians do not learn the entire language. This is due to the fact that many of them have moved to other nations and adapted to their culture (Muhammad, 2019).

This study aimed to create a new dataset for the Assyrian language characters, that contains 36 characters, train and test the VGG16, VGG19, MobileNet-V2, and ResNet-50 models on the new Assyrian language dataset, and extract the accuracy and loss rate.

Table 1: Assyrian Characters, its Name and the Sound of the Pronunciation (Sada, 2021)

No.

Name of Character

Assyrian Character

Sounds of Character

1

ālap

/a/

2

bēth

/b/

3

gāmal

/g/

4

dālath

/d/

5

/h/

6

wāw

/w/

7

zayin

/z/

8

ḥēth

/x/

9

ṭēth

/ṭ/

10

yodh

/y/

11

First / kāp

/k/

12

Connected End/ kāp

/k/

13

Free End/ kāp

/k/

14

lāmadh

/l/

15

First/ mim

/m/

16

End/ mim

/m/

17

First/ nun

/n/

18

Connected End/ nun

/n/

19

Free End/ nun

/n/

20

semkath

/s/

21

‘ē

/ʕ/

22

/p/

23

ṣādē

/ṣ/

24

qop

/q/

25

rēš

/r/

26

šin

/ʃ/

27

tāu

/t/

28

/w/

29

ghē

/ɣ/

30

/dʒ/

31

dhē

/đ/

32

ḥē

/x/

33

fē - wē

/f - w/

34

thē

/θ/

35

djē

/ʒ/

36

chē

//

2.     CONVOLUTIONAL NEURAL NETWORK (CNN)

        Convolutional neural networks integrate artificial neural networks with contemporary methods of deep learning. They have been utilized for many years in image recognition tasks, such as handwritten character recognition, the topic of this research. CNNs are thought to be the first deep learning approach with successful multilayer hierarchical structure networks robustness. CNNs can help forward propagation network improve their backpropagation algorithm deficiencies by reducing the number of trainable network parameters (Seng, Chiang, Salam, Tan, & Chai, 2021).

        An explanation of the algorithms used to recognize handwriting written in this research:

2.1     VGG16

        Oxford Net is another name for the Visual Geometry Group (VGG). VGG16 is a convolutional neural network that was trained on more than one million images from the ImageNet database. The network is composed of sixteen layers and is capable of categorizing things into one thousand categories. Hence, the network has acquired rich feature representations for an assortment of images. The image input dimensions of the network are 224 by 224 pixels. VGG16 has thirteen convolution layers, which are separated by pooling layers. In deep learning, the final two layers are the fully connected layer and output layer. The layers corresponding to the different algorithms will replace these two levels. Loss 3 classifier and output layer will comprise the final two layers of VGG16. VGG16, which is a well-known baseline algorithm for feature extraction. Each layer is made of filters that extract features. As the number of layers increases, so does the number of filters, allowing for the extraction of additional data. When the number of layers increases, the object's size decreases (Pragathi, Priyadarshini, Saveetha, Banu, & Aarif), (Korichi, Slatnia, Aiadi, Tagougui, & Kherallah, 2020). Figure (1)(a) shows the VGG16 architecture.

Figure 1: (a) VGG16, (b) VGG19 Architecture (Jraba, Elleuch, & Kherallah, 2021)

2.2     VGG19

       VGG19 is a deep learning model for image categorization that has been pre-trained. This network has 19 layers and was trained on one million photos in 1000 categories from the ImageNet database.  These 19 layers consist of 16 convolutional, three fully connected CNN with stride and padding of 1, and 2*2 max pooling layers (Almisreb, Turaev, Saleh, & Al Junid, 2022). This network only contains 3x3 convolutional layers stacked on top of one another to increase depth. On top of that, a max pooling layer is introduced to handle volume size reduction. Max pooling is applied to a 2x2 pixel window. After Max-Pooling, the Model is composed of three Fully-Connected Layers (FC Layers): the first two layers  have 4,096 Nodes, while the third layer is used to accomplish 1000-way ILSVRC classification and hence has 1000 channels (one for each class). After that, a soft-max layer is added. All of the hidden layers in the VGG19 Model have rectified linear activation unit (ReLU) (Khari, Garg, Crespo, & Verdú, 2019). Figure (1)(b) Shows the VGG19 architecture.

2.3     MobileNet-v2


        MobileNet-V2 is a lightweight deep neural network model that utilizes depth-wise separable convolutions to efficiently extract spatial and channel features. This is achieved by decomposing the standard convolution operation into two distinct convolution methods (Ghosh et al., 2020). MobileNet-V2 also provides two global hyperparameters, width and resolution multiplier, to strike a balance between delay and precision (Hamida, El Gannour, Cherradi, Ouajji, & Raihani, 2022).  Figure (2) Shows the MobileNet-v2 architecture.

 


 

Figure 2: MobileNet-v2 Architecture (Jin et al., 2023)

        The depth wise separable convolution is MobileNet-V2's unit. The inverted residual structure and linear bottlenecks are MobileNet-V2, most significant enhancements. The modules that are coupled to residuals constitute the inverted residual structure. The initial step in the process involves employing the projection convolution to augment the dimensionality. Subsequently, the depth convolution is used, and finally, the projection convolution is utilized to reduce the dimensionality (Jin et al., 2023), (Srinivasu et al., 2021).

2.4     ResNet-50

        This model is a ResNet-50 Deep Convolutional Neural Network that has been pre-trained (CNN). The selection of ResNet was motivated by its significant utilization of Batch Normalization and Dropout techniques. These two strategies serve to standardize the model and mitigate the risk of overfitting. In addition, the inclusion of identity mappings or skip connections in Residual Neural Networks (ResNets) addresses the issue of vanishing gradients, so facilitating the training of a more complex model with increased depth, leading to improved performance. Using the ImageNet Large Scale Visual Recognition Competition (ILSVRC) dataset weights, the model was pre-trained. Thus, the initial model accepted photographs with a resolution of 224 by 224 pixels (Chatterjee, Dutta, Ganguly, Chatterjee, & Roy, 2019), (Chatterjee, Dutta, Ganguly, Chatterjee, & Roy, 2020).   Figure (3) Shows the ResNet-50 architecture.

 

Figure 3: ResNet-50 Architecture (Chatterjee & Roy, 2020)

3.     RELATED WORK

        C. Chandankhede and R. Sachdeo, 2023 (Chandankhede & Sachdeo, 2023) created  the handwritten  modi barakhadi datasetwhich was developed by collecting samples from roughly 25 individuals. The 7721-item dataset has just been evaluated. The Otsu binarization technique was utilized to trim and pre-process all individual characters. The performance of pre-processed data on a real-world handwritten character database generated by multiple individuals was evaluated using both methodologies. The testing accuracy of ResNet-50 recognized image is 94.552%, and the model precision is 0.86.

        M. Halder, et. al., 2023 (Halder, Kundu, & Hasan, 2023) created a new convolutional neural network model and evaluated it using the CMATERdb 3.1.2 dataset. The model surpasses previous methodologies discussed in the literature pertaining to the alphabets of the dataset, with an average accuracy for training of 98.78%, an average accuracy for validation of 98.33%, and an average accuracy for testing of 98.21%. The present study has established a comprehensive framework for the identification and classification of individual Bangla characters. This framework serves as a crucial stepping- stone towards further developments and progress in the domain of Bangla handwriting recognition.

       S. D. Pande, et. al., 2022 (Pande et al., 2022) used the most effective techniques for improving recognition rate and configured CNN for effective Devanagari handwritten character recognition using the dataset (DHCD), a strong open dataset with 46 classes of Devanagari characters and 2,000 unique images for each class. After recognition, conflict resolution was crucial for effective recognition. This method enabled the user to resolve disagreements. In terms of precision and training time, this strategy produced favorable outcomes.

        A. A. A. Ali and S. Mallaiah, 2022 (Ali & Mallaiah, 2022) employed in their study, CNN and SVM classifiers were utilized for Arabic handwriting identification using two distinct deep neural network models to obtain accurate features. In addition, they investigated the application of dropout in the suggested model for text recognition in photos of handwritten documents and demonstrated the system's efficiency for handwritten recognition in several Arabic scripts tested on diverse datasets. Simulation findings indicate that the suggested CNN-based-SVM with dropout model outperforms the conventional CNN classifier and the dropout CNN-based-SVM model.

        M. Elleuch, et. al., 2021 (Elleuch, Jraba, & Kherallah, 2021) investigated the application of the transfer learning approach in their suggested models (Inception-v3, ResNet, and VGG16) with Arabic handwritten recognition and demonstrated the effectiveness of the system for Arabic handwritten script recognition utilizing the IFN/ENIT database. Deep CNN-based models that were trained from the ground up were compared to transfer learning techniques. When applied to a set of photos of the handwritten Arabic word IFN/ENIT, ResNet and VGG models with TL yield encouraging results with 98.99% and 98.10% accuracy, respectively.

        H. M. Balaha, et. al., 2021 (Balaha, Ali, Saraya, & Badawy, 2021) introduced a novel deep learning (DL) framework that utilizes two distinct Convolutional Neural Network architectures, namely HMB1 and HMB2. Additionally, they provided several optimization strategies, regularization approaches, and dropout mechanisms. The approach employed by the researchers might potentially serve as a foundation for further investigation into handwritten Arabic text. The performance metrics that were computed were accuracy, recall, precision, and F1. The uniform weight initializer and AdaDelta optimizer achieved the highest levels of accuracy. The implementation of data augmentation techniques resulted in a notable improvement in accuracy. HMB1 reported a testing accuracy of 98.4% by utilizing augmentation on the HMBD dataset, which consisted of 865,840 data points.

        M. M. Yapıcı, et. al., 2021 (Yapıcı, Tekerek, & Topaloğlu, 2021) Presented a Cycle-GAN as a new data augmentation strategy to address the issue of insufficient data in signature verification. In addition, a signature verification system unique to Caps-Net was revealed. Four commonly used convolutional neural network (CNN) approaches are used to test the proposed data augmentation technique: VGG16, VGG19, ResNet-50, and DenseNet-121. The approach has made a substantial contribution to the success of all of the previously stated CNN methods. On the DenseNet-121, and the proposed data augmentation technique provides the most advantages. Using two well-known databases, the authors evaluated the data augmentation technique using the suggested signature verification system, GPDS and MCYT. In comparison to other studies, their verification approach produced the best findings on the MCYT database and the second-best results on the GPDS database.

        M. Shams, et. al., 2020 (Shams, Elsonbaty, & ElSawy, 2020) described an effective deep convolutional neural network architecture for extracting and classifying Arabic handwritten characters dataset, used (AHCD). The researchers employed a dropout support vector machine (SVM) to categorize and identify missing attributes that were not accurately detected by a deep convolutional neural network (DCNN). This was done to enhance the dependability and effectiveness of the proposed framework. In addition, the proposed approach employed K-means clustering as a technique of dividing the multi-stroke Arabic characters into 13 distinct groups that exhibit similarity. In contrast to alternative approaches, the system under consideration exhibits a classification accuracy of 95.07% and a classification error rate of 4.93%.

        S. Jraba, et. al., 2020 (Jraba, Elleuch, & Kherallah, 2020) proposed a novel technique for the identification of the Arabic handwritten words. The study of Arabic handwriting identification is a recent area of focus within the field of computer vision, presenting several promising applications including intelligent systems, video conferencing, and real-time applications. The authors proposed the utilization of Deep Convolutional Neural Networks (DCNN) as a tailored approach for performing the classification task. Prominent architectural models in the field of computer vision encompass ResNet and VGG16. These models have been trained using an enhanced dataset comprising photos sourced from the IFN/ENIT database. This approach has demonstrated commendable performance in conventional pattern recognition tasks. The system that was created had very high rates of identification, as indicated by the data that was obtained.

        M. Elleuch and M. Kherallah, 2020 (Elleuch & Kherallah, 2020) fixed the problem of overfitting by adding regularization techniques to their Convolutional Deep Belief Networks (CDBN) model and used IFN/ENIT datasets with data augmentation to test the proposed model on low and high-level dimensions in Arabic textual (character/ word) images. The experimental results demonstrate that the proposed CDBN architectures outperform convolutional networks for categorizing textual picture data. CDBN were used to automatically learn the most discriminative characteristics from AHS's textual image data. The advantages of deep belief networks and convolutional neural networks can be incorporated into this architecture. In order to solve the issue of overfitting, the authors incorporated regularization techniques into their CDBN model.

4.     METHODOLOGY

4.1     Convolution Neural Network Operations

        Most often, CNNs consist of a number of sequentially linked layers of multi-convolutional processing, then a number of layers of fully connected processing. Inputs from the layer below it is convolved with filters that have been trained for each successive convolutional layer. After the convolution step, the pooling operation is carried out on the output of the current layer in order to lessen the magnitude of the data and cut down on the amount of overfitting that occurs in the network (Safarzadeh & Jafarzadeh, 2020).

        Convolution refers to the mathematical combination of two functions to form a third function. In the context of CNNs, a convolutional layer (called Filter or Kernel) is applied to the input data (image) to produce a feature map. The filter is applied to the input data and its output is formed on the new layer. Figure (4) shows the procedure for performing a Product Dot operation between a 3x3 filter matrix and a 3x3 region of the input image matrix. The resulting matrix elements are summed and the sum is the output value (Destination Pixel) on the feature map. The filter then passes over the input matrix, repeats the dot product with each remaining set of 3x3 regions, and completes the feature map. Multiple filters are used for a single input and the resulting feature maps are linked together to obtain the final result of a single convolutional layer (Dao, 2020).


Figure 4: The process of wrapping the filter with the image in a single layer (Dao, 2020)

4.1.1   Convolutional Layer: Convolutional layers extract picture features first. Because pixels are only related to nearby pixels, convolution maintains the relationship between different parts of a picture (Lamsaf, Ait Kerroum, Boulaknadel, & Fakhri, 2022). Convolution is the first and most crucial stage in filtering a picture with a lower pixel filter to minimize its size while keeping pixel relationships (Boutounte & Ouadid, 2021). A 3x3 filter with a 1x1 stride (1-pixel shift at each step) convolutions the 5x5 picture to a 3x3 output (64% reduction in complexity) (Hossain & Ali, 2019).

        The CNN's Convolution Layer is its most important part. It convolves or multiplies the resulting pixel matrix to create an activation map for the image (Niharmine, Outtaj, & Azouaoui, 2022). CNNs define mathematical convolution differently than mathematics or engineering. NN layers undergo convolution. K-filter convolutional layers (also called kernels). Filters identify corners, edges, and endpoints. Filters are N*N*R grids, where N is the filter's height and width and R is the number of picture channels. Each filter convolves through the input grid, multiplying each pixel by its filter value. The multiplications are then added (Altwaijry & Al-Turaiki, 2021).

        Combining two functions to create a third is called convolution. CNNs create feature maps from image data using a convolutional layer (Filter or Kernel) (Adebayo, Oluwatobi Aworinde, Akinwunmi, Ayandiji, & Olalekan Monsir, 2022). The new layer receives the filter output from the input data. Summating the matrix elements yields the feature map's Destination Pixel. The filter then travels over the input matrix, repeats the dot product with each set of 3 x 3 regions, and finishes the feature map (Truong Quang, Duy, & Nhan, 2020). A single convolutional layer is created by linking feature maps from many filters for a single input (Dao, 2020).

Figure 5: General architecture of CNN (Siddique, Sakib, & Siddique, 2019)

        Mathematically expresses the convolution process if we use a two-dimensional image (I) as input and use a two-dimensional filter (K) of size (m*n). The feature map (S) is obtained according to the mathematical equation (1) (Yao & Zheng, 2023):

        Additionally, the Accuracy for Correctly classified instances divided by the total number of instances, and mathematically is obtained according to the mathematical equation (2) (Vakili, Ghamsari, & Rezaei, 2020):

4.1.2   Activation Function: An activation function is a node (placed at the end or between the layers of neural networks) that helps determine whether a neuron will fire or not (Wang, Li, Song, & Rong, 2020). This study employed the Rectified Linear Unit (ReLU) function, which produces an output of zero when the input value is less than or equal to zero. Alternatively, the resulting output will be equivalent to the initial input value, Mathematical equation (3) explains this function (Kandel & Castelli, 2020):

4.1.3   Pooling or Subsampling: The Pooling process is an essential step in convolution-based systems, as it reduces the dimensions of feature maps and combines a set of values to search for a smaller number of those values, that is, to reduce the dimensions of the feature map (Gholamalinezhad & Khosravi, 2020).

4.1.4   Flattening Layer: Flattening reduces the spatial dimensions of the pooled feature map while preserving the channel dimension. The flattening layer provides an extra dimension even if the inputs are shaped without a channel dimension. Following the flattening operation, the feature matrix is turned into a vector that can be fed into Keras' dense layer, a fully connected neural network (Mishra, Sachan, & Rajpal, 2020).

4.1.5   Fully Connected Layer: This is the final layer that the neural network is fed after the convolution and pooling layers; the classification part consists of a few fully connected layers. These layers accept only one-dimensional data; where it is considered the interface between the individual neurons of one layer and the individual neurons of the next layer. So, a layer is completely connected and serves as a connection point for the neurons of all of the other layers (Dubey & Jain, 2019).

        The activation function (SoftMax) is used as a higher layer (after the fully connected layer), as it begins to deal with the results with a real value that is not appropriately scaled and which may be difficult to deal with, as it converts the number vector into a probability vector between 0 and 1. The SoftMax activation function () can be defined by mathematical equation (4) (Bhatnagar, Gill, & Ghosh, 2020):

x: Values from the neurons of the output layer.

n: The number of neurons.

 : The sum of the exponential values (e) of the output cells.

The fully connected layers finally connect each layer of the max pooling layer to the output neurons (Chauhan, Ghanshala, & Joshi, 2018).

4.2     Dataset

        The research utilizes a dataset created for Handwritten Characters Assyrian Language. The database consists of 36 Assyrian characters with 500 samples in each character. Thus, the dataset consists of 18,000 samples. Initially, the form for the Assyrian language was distributed to people between the ages of 13  to 60 years of both genders to fill it out.

After the researcher selected the sample from 500 people . The form distributed and collected  within 6 months to 4 governorates,  Table (2) shows that:

Table 2: Form Distribution details and Numbers

Governorate

City/District

The number of forms

Duhok

Duhok

256

Zakho

34

Semel

43

Bagere

10

Kori Gavana

5

Erbil

Erbil

22

Ankawa

21

Diana

10

Musel

Musel

7

Hamdaneya

10

Baghdad

Baghdad

82

Total

11

500

 

        Then, the responders were tasked with writing the characters into the empty squares with size 224, 224 pixels provided in the forms table. as shown in Figure (6).

Figure 6: Sample of Assyrian Language full Forms

        All cropping characters images saved with the new label from 1 to1000 preceded by the folder number (class) which is from 0 to34, each character is in a separate folder,  Figure (7) shows a sample of characters after crop.

Figure 7: Sample of Cropping and Labelling Characters

 

         Because we used the VGG16, VGG19, MobileNet-V2, and ResNet-50 algorithms, we were able to reduce the size of each image to 224 by 224. Each image's original dimensions were different. Images in logical form were transformed into images in unsigned integer form throughout the training procedure to obtain relevant features. Eventually, the bmp photos were converted to greyscale. This is accomplished using MATLAB.

Certain perplexing photos in the dataset have an impact on classifier accuracy owing to their personal handwriting peculiarities and characteristics. Another issue that might be found is the scanning noise.  Figure 1 (a) shows examples of chosen characters utilized in the study. All of the preceding stages were completed as part of the preprocessing phase. (b), are modified as some instances of characters that cause confusion during the process.

4.3     Split the Dataset

        The new dataset consists of 18,000 images . The CNN models were trained on 70% of the characters, which means the total number of training dataset is 12,600 images, and tested on 30% of the characters, which means the total number of the testing dataset is 5,400 images. Table (3) describes the Training and Validation partition, and number of character’s images:

 

 

 

 

Table 3: Dataset Partition

Partition

Number of Characters

Training images (70%)

12,600

Validation images (30%)

5,400

Total of images

18,000

4.4     Models Training

        The training step is very important for the model, as this stage is concerned with creating a model from the data given to it. The model is trained on the training dataset to find the correct weights that will be automatically adjusted by the specified algorithm, which helps to reduce performance.

At this stage, the Compile and fit functions were implemented in MATLAB. Within the Compile function, use the network optimizer ((ADAM) Adaptive Moment Estimation).

        And the loss function (categorical_crossentropy) and the scale (accuracy), as for the fit function within this stage, through which the training and investigation data and the number of training cycles are determined (Epoch), where 30 epochs were tested experimentally and it is sufficient to adjust the weights to the best with a learning rate of 0.0001 to update the weights, and Batch Size is set to 100.

4.5     Models Testing

        In this step, the test process was applied to verify whether the model recognized the Assyrian characters correctly on the data allocated for the validation or not.

The confusion matrix was used to test the performance of prediction algorithms based on a set of tests by applying the function of confusion matrix, on the models used in the data assigned to the validation, The prediction results, which are the values of TP, FP, FN, TN were obtained after applying the mentioned equations to calculate the combination measure of precision and recall of the model.

 

4.6     Dimensional Reduction

        Principal Components Analysis (PCA) is a popular approach for both modifying data and reducing dimensionality (Mahmood & Abdulrazzaq, 2022). Calculating the features that explain the majority of the variation in the data is what principal component analysis (PCA) does. It merely locates a subspace that accounts for the vast majority of the variance present in the data, and then it eliminates dimensions that have a low variance (Yang et al., 2020). This is achieved by the modification of the data space by introducing additional qualities that exhibit non-linear relationships with one another. In classification and regression problems, PCA has been used successfully as a way to change and reduce the number of dimensions (Valls, Aler, Galván, & Camacho, 2021).

        In addition to being a statistical tool, principal component analysis makes use of orthogonal transformations. Using the use of principal components analysis (PCA), a set of correlated variables can be changed into a set of uncorrelated variables. PCA is a technique for analysing exploratory data. PCA can also be used to study the relationships between variables. As a result, it has the potential to be utilized to reduce dimensionality (Ma & Yuan, 2019), (Reddy et al., 2020).

5.     RESULTS AND DISCUSSION

        This work was performed using MATLAB R2022b on a machine with an Intel® Core (TM) i7-1165G7 processor, 2.80 GHz, 64-bit where four convolutional neural network models were used . The results of accuracy for the database of Assyrian characters  were obtained from 500 people aged (13-60) who wrote 36 handwritten letters of the Assyrian language . The results of the highest accuracy were as follows: 95.70% in MobileNet-V2, 94.97% in ResNet-50, 92.06% in VGG19, and finally 90.97% in VGG16.

Table 4: Confusion Matrix parameters for Four Models

CNN Models

Precision

Recall

F1_scoure

VGG16

0.9321

0.8910

0.9107

VGG19

0.9402

0.9047

0.9219

MobileNet-v2

0.9667

0.9502

0.9584

ResNet-50

0.9477

0.9178

0.9324

While the extracted results were compared with the results of other researchers who used other languages and datasets, they ranged between better and less due to the different types of datasets used in terms of language, the number of images, and their size.


Table 5:The accuracy Ratio for CNN Methods used in Handwritten Recognition


Reference

CNN Methods

Dataset

No. of Dataset Sample Trian

Feature Selection Dimensionality reduction

Accuracy

(Chandankhede & Sachdeo, 2023)

CNN

Modi barakhadi

7721 characters

ResNet-50

94.55%

(Halder et al., 2023)

CNN New Method

CMATERdb 3.1.2

37,858 images

-

98.33%

(Pande et al., 2022)

CNN including Dropout Layer

DHCD

46 classes

-

99.13%

(Ali & Mallaiah, 2022)

CNN

AHDB

AHCD

IFN/ENIT

HACDB

13311 words

16,800 characters

26.459 words

6.600 shapes

SVM

99%

99.71%

98.58%

99.85%

(Elleuch et al., 2021)

DCNN

IFN/ENIT

10 classes – 450 images

ResNet

VGG16

98.99%

98.10%

(Balaha et al., 2021)

CNN with HMB1 & HMB2

HMBD

Seven-page dataset

-

98.4%

(Yapıcı et al., 2021)

CNN

GPDS MCYT

4000 signatures

VGG16, VGG19

ResNet50, DenseNet121

88.97%

98.06%

(Shams et al., 2020)

DCNN & SVM

AHCD

840 images

K-mean

95.07%

(Jraba et al., 2020)

DCNN

IFN/ENIT

AHS images

ResNet, VGG16

93%

(Elleuch & Kherallah, 2020)

CDBN

IFN/ENIT

Augmented AHS images

-

95.94%

(Ahlawat & Choudhary, 2020)

Hybrid CNN & SVM

MNIST

70000 data Elements

-

99.28%

(Das & Mohanty, 2020)

DCNN

Javanes Script

120 characters

-

99.65%

(Ashiquzzaman, Tushar, Rahman, & Mohsin, 2019)

CNN

CMATER

3000 Digits

-

99.40%

This Research

CNN

HCAL

18000 images for 36 class of characters

VGG16

VGG19

MobileNet-V2

ResNet-50

90.97%

92.06%

95.70%

94.97%


        The majority of the reviewed researchers used convolutional neural networks in different models and on different types of datasets, dividing them into different proportions for training and testing, and their results were very good, which prompted us to use the convolutional neural network and apply it to the new dataset that we created to  get similar results. The results are comparable to those obtained by the researchers, and we benefited from the experience of others who have utilized the CNN models.

        Therefore, while the subject of handwritten character and digit recognition requires major attention, few academic efforts are expended by Assyrian technicians to design a functioning system in this regard. At the same time, logically, the topic has not been addressed sufficiently in academic publications and essays, to the extent that the subjects are concerned with the Assyrian language.

         It can also be noted from the previous table that the researchers who used ResNet and VGG16 using DCNN in the reference (Elleuch et al., 2021) had better results than the results we extracted for the same models using CNN, Despite the reduced number of classes, the training and testing images were similarly less in quantity.

While the research (Chandankhede & Sachdeo, 2023), the result of ResNet-50 was slightly lower than the result of our research, as they used 7,721 images of characters, while we used 18,000 images.

        The majority of the reviewed researchers used convolutional neural networks in different models and on different types of datasets, dividing them into different proportions for training and testing, and their results were very good, which prompted us to use the convolutional neural network and apply it to the new dataset that we created in order to  arrive at similar results. The results are comparable to those obtained by the researchers, and we benefited from the experience of others who have utilized the same models.

        Therefore, while the subject of handwritten character and digit recognition requires major attention, few academic efforts are expended by Assyrian technicians to design a functioning system in this regard. At the same time, logically, the topic has not been addressed sufficiently in academic publications and essays, to the extent that the subjects are concerned with the Assyrian language.


(a)

(b)

Figure 8. Performance of model Training and Validation for VGG16: (a) Accuracy, and (b) Loss


 


(a)

(b)

Figure 9. Performance of model Training and Validation for VGG19: (a) Accuracy, and (b) Loss


(a)

(b)

Figure 10. Performance of model Training and Validation for MobileNet-V2: (a) Accuracy, and (b) Loss


(a)

(b)

Figure 11. Performance of model Training and Validation for ResNet-50: (a) Accuracy, and (b) Loss



CONCLUSION

        Initially, a dataset of 36 Assyrian characters was created. It was collected after distributing 500 forms to people of both  genders, between the ages of 13  to 60. Additionally, the dimensions of the images were modified to 224 * 224 pixels. The maximum accuracy rate in recognizing handwritten Assyrian characters was achieved by extracting accurate data using four models of the convolutional neural network, namely VGG16, VGG19, MobileNet-V2, and ResNet-50. Ultimately, a comparative analysis was conducted to evaluate the efficacy of the Assyrian handwriting recognition model in relation to contemporary models employed in other languages.

References

Adebayo, S., Oluwatobi Aworinde, H., Akinwunmi, A. O., Ayandiji, A., & Olalekan Monsir, A. (2022). Convolutional neural network-based crop disease detection model using transfer learning approach. Indonesian Journal of Electrical Engineering and Computer Science, 29(1), 365. doi:10.11591/ijeecs.v29.i1.pp365-374

Ahlawat, S., & Choudhary, A. (2020). Hybrid CNN-SVM classifier for handwritten digit recognition. Procedia Computer Science, 167, 2554-2560.

Albahli, S., Nawaz, M., Javed, A., & Irtaza, A. (2021). An improved faster-RCNN model for handwritten character recognition. Arabian Journal for Science and Engineering, 46(9), 8509-8523.

Ali, A. A. A., & Mallaiah, S. (2022). Intelligent handwritten recognition using hybrid CNN architectures based-SVM classifier with dropout. Journal of King Saud University-Computer and Information Sciences, 34(6), 3294-3300.

Almisreb, A. A., Turaev, S., Saleh, M. A., & Al Junid, S. A. M. (2022). Arabic Handwriting Classification using Deep Transfer Learning Techniques. Pertanika Journal of Science & Technology, 30(1).

Altwaijry, N., & Al-Turaiki, I. (2021). Arabic handwriting recognition system using convolutional neural network. Neural Computing and Applications, 33(7), 2249-2261.

Ashiquzzaman, A., Tushar, A. K., Rahman, A., & Mohsin, F. (2019). An efficient recognition method for handwritten arabic numerals using CNN with data augmentation and dropout Data management, analytics and innovation (pp. 299-309): Springer.

Balaha, H. M., Ali, H. A., Saraya, M., & Badawy, M. (2021). A new Arabic handwritten character recognition deep learning system (AHCR-DLS). Neural Computing and Applications, 33(11), 6325-6367.

Benjamen, A. (2022). Assyrians in Modern Iraq: Negotiating Political and Cultural Space: Cambridge University Press.

Bhatnagar, S., Gill, L., & Ghosh, B. (2020). Drone image segmentation using machine and deep learning for mapping raised bog vegetation communities. Remote Sensing, 12(16), 2602.

Boutounte, M., & Ouadid, Y. (2021). Characters recognition using keys points and convolutional neural network. Indonesian Journal of Electrical Engineering and Computer Science, 22(3), 1629. doi:10.11591/ijeecs.v22.i3.pp1629-1634

Chandankhede, C., & Sachdeo, R. (2023). Offline MODI script character recognition using deep learning techniques. Multimedia Tools and Applications, 1-12.

Chatterjee, S., Dutta, R. K., Ganguly, D., Chatterjee, K., & Roy, S. (2019). Bengali handwritten character classification using transfer learning on deep convolutional neural network. arXiv preprint arXiv:1902.11133.

Chatterjee, S., Dutta, R. K., Ganguly, D., Chatterjee, K., & Roy, S. (2020). Bengali handwritten character classification using transfer learning on deep convolutional network. Paper presented at the Intelligent Human Computer Interaction: 11th International Conference, IHCI 2019, Allahabad, India, December 12–14, 2019, Proceedings 11.

Chauhan, R., Ghanshala, K. K., & Joshi, R. (2018). Convolutional neural network (CNN) for image detection and recognition. Paper presented at the 2018 first international conference on secure cyber computing and communication (ICSCCC).

Dao, H. (2020). Image classification using convolutional neural networks.

Das, A., & Mohanty, M. N. (2020). Use of deep neural network for optical character recognition Advancements in Computer Vision Applications in Intelligent Systems and Multimedia Technologies (pp. 219-254): IGI Global.

De Ridder, J. J. (2018). Descriptive Grammar of Middle Assyrian: Harrassowitz Verlag.

Dubey, A. K., & Jain, V. (2019). Comparative study of convolution neural network’s relu and leaky-relu activation functions. Paper presented at the Applications of Computing, Automation and Wireless Systems in Electrical Engineering: Proceedings of MARC 2018.

Elleuch, M., Jraba, S., & Kherallah, M. (2021). The Effectiveness of Transfer Learning for Arabic Handwriting Recognition using Deep CNN. Journal of Information Assurance & Security, 16(2).

Elleuch, M., & Kherallah, M. (2020). Off-line Handwritten Arabic text recognition using convolutional DL networks. International Journal of Computer Information Systems and Industrial Management Applications, 12, 104-112.

Fales, F. M. (2021). Neo-Assyrian History of the Akkadian Language (2 vols) (pp. 1347-1395): Brill.

Fales, F. M. (2023). The Assyrian Empire. The Oxford History of the Ancient Near East: Volume IV: the Age of Assyria, 1, 425.

Gholamalinezhad, H., & Khosravi, H. (2020). Pooling methods in deep neural networks, a review. arXiv preprint arXiv:2009.07485.

Ghosh, T., Abedin, M., Chowdhury, S., Tasnim, Z., Karim, T., Reza, S. M. S., . . . Yousuf, M. (2020). Bangla handwritten character recognition using MobileNet V1 architecture. Bulletin of Electrical Engineering and Informatics, 9, 2547-2554. doi:10.11591/eei.v9i6.2234

Halder, M., Kundu, S., & Hasan, M. F. (2023). An Improved Method to Recognize Bengali Handwritten Characters Using CNN, Singapore.

Hamida, S., El Gannour, O., Cherradi, B., Ouajji, H., & Raihani, A. (2022). Handwritten computer science words vocabulary recognition using concatenated convolutional neural networks. Multimedia Tools and Applications, 1-27.

Hossain, M. A., & Ali, M. M. (2019). Recognition of handwritten digit using convolutional neural network (CNN). Global Journal of Computer Science and Technology.

Jayasundara, V., Jayasekara, S., Jayasekara, H., Rajasegaran, J., Seneviratne, S., & Rodrigo, R. (2019). Textcaps: Handwritten character recognition with very small datasets. Paper presented at the 2019 IEEE winter conference on applications of computer vision (WACV).

Jin, G., Liu, Y., Qin, P., Hong, R., Xu, T., & Lu, G. (2023). An End-to-End Steel Surface Classification Approach Based on EDCGAN and MobileNet V2. Sensors, 23(4), 1953.

Jraba, S., Elleuch, M., & Kherallah, M. (2020). Arabic handwritten recognition system using deep convolutional neural networks. Paper presented at the International Conference on Intelligent Systems Design and Applications.

Jraba, S., Elleuch, M., & Kherallah, M. (2021). Arabic handwritten recognition system using deep convolutional neural networks. Paper presented at the Intelligent Systems Design and Applications: 20th International Conference on Intelligent Systems Design and Applications (ISDA 2020) held December 12-15, 2020.

Kandel, I., & Castelli, M. (2020). Transfer learning with convolutional neural networks for diabetic retinopathy image classification. A review. Applied Sciences, 10(6), 2021.

Khari, M., Garg, A. K., Crespo, R. G., & Verdú, E. (2019). Gesture Recognition of RGB and RGB-D Static Images Using Convolutional Neural Networks. Int. J. Interact. Multim. Artif. Intell., 5(7), 22-27.

Korichi, A., Slatnia, S., Aiadi, O., Tagougui, N., & Kherallah, M. (2020). Arabic handwriting recognition: Between handcrafted methods and deep learning techniques. Paper presented at the 2020 21st International Arab Conference on Information Technology (ACIT).

Lamsaf, A., Ait Kerroum, M., Boulaknadel, S., & Fakhri, Y. (2022). Recognition of Arabic handwritten words using convolutional neural network. Indonesian Journal of Electrical Engineering and Computer Science, 26(2), 1148. doi:10.11591/ijeecs.v26.i2.pp1148-1155

Lincy, R. B., & Gayathri, R. (2021). Off-Line Tamil Handwritten Character Recognition Based on Convolutional Neural Network with VGG16 and VGG19 Model. Paper presented at the Advances in Automation, Signal Processing, Instrumentation, and Control: Select Proceedings of i-CASIC 2020.

Ma, J., & Yuan, Y. (2019). Dimension reduction of image deep feature using PCA. Journal of Visual Communication and Image Representation, 63, 102578.

Mahmood, M. R., & Abdulrazzaq, M. B. (2022). Performance evaluation of chi-square and relief-F feature selection for facial expression recognition. Indonesian Journal of Electrical Engineering and Computer Science, 27(3), 1470-1478.

Mishra, S., Sachan, R., & Rajpal, D. (2020). Deep convolutional neural network based detection system for real-time corn plant disease recognition. Procedia Computer Science, 167, 2003-2010.

Muhammad, A. R. (2019). The Assyrian language situation in the Kurdistan region of Iraq. Балтийский гуманитарный журнал, 8(1 (26)), 21-23.

Niharmine, L., Outtaj, B., & Azouaoui, A. (2022). Tifinagh handwritten character recognition using optimized convolutional neural network. International Journal of Electrical and Computer Engineering (IJECE), 12, 4164. doi:10.11591/ijece.v12i4.pp4164-4171

Pande, S. D., Jadhav, P. P., Joshi, R., Sawant, A. D., Muddebihalkar, V., Rathod, S., . . . Das, S. (2022). Digitization of handwritten Devanagari text using CNN transfer learning–A better customer service support. Neuroscience Informatics, 2(3), 100016.

Parikh, M., & Desai, A. (2022). Recognition of Handwritten Gujarati Conjuncts Using the Convolutional Neural Network Architectures: AlexNet, GoogLeNet, Inception V3, and ResNet50. Paper presented at the International Conference on Advances in Computing and Data Sciences.

Pragathi, M., Priyadarshini, K., Saveetha, S., Banu, A. S., & Aarif, K. M. (2019). Handwritten tamil character recognition UsingDeep learning. Paper presented at the 2019 International Conference on Vision Towards Emerging Trends in Communication and Networking (ViTECoN).

Putri, D. U. K., Pratomo, D. N., & Azhari, A. (2023). Hybrid convolutional neural networks-support vector machine classifier with dropout for Javanese character recognition. TELKOMNIKA (Telecommunication Computing Electronics and Control), 21(2), 346-353.

Reddy, G. T., Reddy, M. P. K., Lakshmanna, K., Kaluri, R., Rajput, D. S., Srivastava, G., & Baker, T. (2020). Analysis of dimensionality reduction techniques on big data. IEEE Access, 8, 54776-54788.

Sada, E. (2021). Assyrian-Syriac chants from the liturgy of the Church of the East.

Safarzadeh, V. M., & Jafarzadeh, P. (2020). Offline Persian handwriting recognition with CNN and RNN-CTC. Paper presented at the 2020 25th international computer conference, computer society of Iran (CSICC).

Seng, L. M., Chiang, B. B. C., Salam, Z. A. A., Tan, G. Y., & Chai, H. T. (2021). MNIST Handwritten Digit Recognition with Different CNN Architectures. Journal of Applied Technology and Innovation (e-ISSN: 2600-7304), 5(1), 7.

Shams, M., Elsonbaty, A., & ElSawy, W. (2020). Arabic handwritten character recognition based on convolution neural networks and support vector machine. arXiv preprint arXiv:2009.13450.

Siddique, F., Sakib, S., & Siddique, M. A. B. (2019). Recognition of handwritten digit using convolutional neural network in python with tensorflow and comparison of performance for various hidden layers. Paper presented at the 2019 5th international conference on advances in electrical engineering (ICAEE).

Srinivasu, P. N., SivaSai, J. G., Ijaz, M. F., Bhoi, A. K., Kim, W., & Kang, J. J. (2021). Classification of skin disease using deep learning neural networks with MobileNet V2 and LSTM. Sensors, 21(8), 2852.

Truong Quang, V., Duy, L., & Nhan, N. (2020). Vietnamese handwritten character recognition using convolutional neural network. IAES International Journal of Artificial Intelligence (IJ-AI), 9, 276. doi:10.11591/ijai.v9.i2.pp276-281

Vakili, M., Ghamsari, M., & Rezaei, M. (2020). Performance analysis and comparison of machine and deep learning algorithms for IoT data classification. arXiv preprint arXiv:2001.09636.

Valls, J. M., Aler, R., Galván, I. M., & Camacho, D. (2021). Supervised data transformation and dimensionality reduction with a 3-layer multi-layer perceptron for classification problems. Journal of Ambient Intelligence and Humanized Computing, 12(12), 10515-10527.

Wang, Y., Li, Y., Song, Y., & Rong, X. (2020). The influence of the activation function in a convolution neural network model of facial expression recognition. Applied Sciences, 10(5), 1897.

Yang, W., Zhao, Y., Wang, D., Wu, H., Lin, A., & He, L. (2020). Using principal components analysis and IDW interpolation to determine spatial and temporal changes of surface water quality of Xin’anjiang river in Huangshan, China. International journal of environmental research and public health, 17(8), 2942.

Yao, K., & Zheng, Y. (2023). Fundamentals of Machine Learning Nanophotonics and Machine Learning: Concepts, Fundamentals, and Applications (pp. 77-112): Springer.

Yapıcı, M. M., Tekerek, A., & Topaloğlu, N. (2021). Deep learning-based data augmentation method and signature verification system for offline handwritten signature. Pattern Analysis and Applications, 24(1), 165-179.