Signature Verification Based On Proposed Fast Hyper Deep Neural Network
Signature Verification Based On Proposed Fast Hyper Deep Neural Network
Corresponding Author:
Zainab Hashim
Department of Computer Sciences, University of Technology
Baghdad, Iraq
Email: [email protected]
1. INTRODUCTION
One of the key developments in modern scientific study is the advancement of biometric verification
techniques [1]. Meanwhile, it is well recognized that the present approaches to biometry [2]–[5] have a number
of issues that restrict the range of their applicability. Typically, when high levels of security and authentication
are necessary, biological and behavioral features are utilized. Biological features include the face, fingerprint,
palm, iris, and retina, while behavioral features include signature and voice. In many areas of our lives, such
as banks, educational institutions, attendance monitoring systems, and official document verification, where
the need for authenticity is paramount, handwritten signature verification has become an integral aspect [6].
Based on the technique of acquisition, signatures are divided into offline and online signatures. Online
signatures are obtained using digital devices like tablets or electronic pens that can capture real-time features
as opposed to traditional pen and paper signatures (off-line), which are afterwards scanned as a digital file [7].
The process of deciding whether a signature is real or fake is known as signature verification, which is why it
is described as a two-class problem. Signature verification phases are similar to identification phases, except
in the classification phase, where its legitimacy to that class will be verified and the tested signature class will
be known [8]. The two basic approaches employed in the verification process are model-based verification and
distance-based verification. In the model-based approach, models such as the convolutional neural network
(CNN), hidden Markov model (HMM), and support vector machine (SVM) are created to represent the
distribution of data. While in the distance-based method, dynamic time wrapping (DTW) compares the test
signature with the reference signature using distance measures [9]. The major aim of this work is the
development of an accurate writer-independent off-line signature verification model that protects from
unexpected forgery with a small error rate. Finding more effective and relevant hybrid features to represent the
handwritten signature as well as building a robust classifier to train these features can help overcome this issue.
The following succinct summary of contributions that try to address the research's primary issue and improve
classification accuracy:
− The proposed work extracts meaningful and perfect signature features based on fusion of appearance-based
features, texture-based features, and frequency-based features.
− Using a new fast hyper deep neural networks (FHDNN) architecture to classify the extracted features. The
proposed deep model achieves improvement in results compared with the previous approaches based on
accuracy and equal error rate (EER) measures.
− The system achieved great performance on it by using two types of datasets from the more popular
challenge handwritten signature datasets without any data augmentation techniques.
This research has the following framework: section 2 introduces recent research on signature verification
systems. Section 3, explains the proposed model stages in details. In section 4, the findings and comments are
covered. Section 5 contains the conclusions and recommendations for future work.
2. RELATED WORKS
Many researchers have investigated the use of handwritten signatures to confirm identity and have
developed processing techniques. One of the first research on signature verification was carried out in 1977.
The research involved features that were taken from horizontal and vertical regions of each signature [10]. The
researches then carried on.
Banerjee et al. [11] created a language-invariant off-line signature verification approach. The
modified signal of the signature image was used to extract 4 various kinds of features, including statistical,
shape-based, similarity-based, and frequency-based features. A unique wrapper feature selection method built
on the red deer algorithm has also been developed to minimize the feature dimension. The authentication and
verification procedures were carried out with the Nave Bayes classifier.
Ruiz et al. [12] used the Siamese neural networks to help solve the off-line handwritten signature
verification. To enhance the number of samples and the diversity required for training deep neural networks, 3
different kinds of synthetic data have been examined: the global public dataset of synthetic (GPDS-synthetic)
dataset, compositional synthetic signature generation from shape primitives, and augmented data samples from
the GAVAB dataset. Combining real and fake signatures for training led to the best verification outcomes.
Tayeb et al. [13] suggested a method for validating written signatures on bank checks. In order to
identify anomalies, this model used a convolutional neural network (CNN) to analyze the pixels in a signature
image. The ability of a CNN to extract features helps speed up feature analysis and processing for signature
verification.
Hadjadj et al. [14] offered a method for determining whether a signature is authentic by using the
textural information in the image of the signature. The local ternary patterns (LTP) and the orientated basic
image features (oBIFs) are two textural descriptors that were utilized to describe the signature images. The
distances between pairs of real and fake signatures were utilized to train SVM classifiers. Signature images
were projected in feature space (a unique SVM for every single descriptor).
Navid et al. [15] aimed to utilize convolutional neural networks (CNN) to automate the process of
signature verification. This model was built on top of the VGG-19, a pre-trained convolutional neural network.
Bhunia et al. [16] proposed a one-class support vector machines (SVMs) in order to create a writer-dependent
signature verification method based on two distinct texture feature types, namely discrete wavelet and local
quantized patterns (LQP) features, in order to generate two distinct authenticity scores for a given signature.
Arisoy [6] proposed a writer-independent signature verification system based on one-shot learning.
Siamese neural network was performed in order to recognize authentic signatures from fake ones. All these
works above are based mostly on deep learning neural networks to implement signature verification processes.
In this paper, we present a new fast hyper deep neural network model based on hybrid features in order to
verify handwritten signatures.
Signature verification based on proposed fast hyper deep neural network (Zainab Hashim)
964 ISSN: 2252-8938
Table 1. Samples of handwritten signature images from SigComp2011 and CEDAR datasets
Sample 1 Sample 2 Sample 3
SigComp2011 Chinese and Dutch
CEDAR
Because each pixel requires less data to be delivered in a grayscale image, utilizing this format will
simplify the signature extraction procedure and speed up processing. As indicated in Table 2, the 256 possible
shades of gray, that vary from black to white, are stored as an 8-bit integer representing the grayscale level.
Table 2. Gray-scale conversion of signature images from SigComp2011 and CEDAR datasets
Sample 1 Sample 2 Sample 3
SigComp2011 Chinese and Dutch
CEDAR
same size, whereas our raw collected images can be of varying sizes. We use bicubic interpolation as a resizing
method, the bicubic interpolation estimates the pixels in the (i, j) positions using a sampling distance of 16
adjacent pixels (4×4) [21] as shown in (3).
Where:
3 2
W − 1(SX ) = −S +2S
2
−S
−3S3 +4S2 +2
W1(SX ) = 2
3 2
W2(SX ) = S −S
2
Table 3. Histogram equalized signature images from SigComp2011 and CEDAR datasets
Sample 1 Sample 2 Sample 3
SigComp2011 Chinese and Dutch
CEDAR
CEDAR
CEDAR
paper, we will use PCA as feature extractor to extract appearance-based signature features from each image
based on Eigensignature method producing (50*1) feature vector which later will be combined with FFT and
GLCM feature vectores. The signature image matrices in two dimensions will be converted into one dimension
row vector. The Eigenvector, a signature code set, is created using the Eigensignature method, which extracts
important information from a signature image. The signature database, which contains signature codes from
earlier training, is then used to compare this signature code. The principal component is calculated as
follows: First, the resized signature images are read and a training set of total M images is created in order to
use them in computing the average mean. The input image is then subtracted from the average mean, as
indicated in (4) [24].
∅𝑖 = 𝑋̅𝑖 − 𝑋 (4)
C = 1/𝑀 ∑𝑚
𝑛−1 ∅𝑛 ∅𝑛 (5)
Most effective Eigen values are sorted and selected after computing the covariance matrix's Eigenvalues and
Eigenvectors. A collection of Eigenvectors' highest Eigenvalues are picked. The training samples are projected
onto Eigenimages to obtain feature space.
where 𝑃̃ is the estimated probability of the groups of pairs of surrounding gray-levels in the image and N
is the overall number of gray-levels employed (the GLCM dimensions) [28].
− Energy: is a metric of similarity or angular second momentum (ASM), which evaluates the consistency of
the textural representation, or the repeating of pixel pairs, as illustrated in (7). It is in charge of identifying
texture disorders. Energy can reach a maximum value of 1 [29].
𝑁−1
̃
Energy = ∑𝑖,𝑗=0(𝑝𝑖𝑗 )
2
(7)
− Entropy: this is typically categorized as an initial measure of the level of chaos in an image, is another
crucial GLCM property to distinguish an image texture. In (8) can be used to quickly calculate the GLCM
derived entropy from the GLCM elements [26], which is inversely proportional to GLCM energy.
𝑁−1
Entropy = − ∑𝑖,𝑗=0 𝑃𝑖 𝑗 log 𝑃𝑖 𝑗 (8)
− Homogeneity: is also called inverse difference moment (IDM). The GLCM's diagonal elements with high
values indicate that the visual texture is highly homogeneous. The homogeneity is greatest when the image
pixel values are same [29]. Due to the GLCM's large yet adverse relationship between contrast and
homogeneity, homogeneity decreases as contrast increases while being constant in energy [29]. The IDM
is shown in (9).
𝑁−1 ̃
𝑃 𝑖𝑗
𝐼𝐷𝑀 = ∑ 1+(𝑖−𝑗)2
(9)
𝑖,𝑗=0
− Mean: compared to other GLCM textural features, it seems to be the best way to measure GLCM texture.
The GLCM Mean is not simply the total of all the original values of pixels in the image window; instead,
it is numerically equal to the GLCM dissimilarity, in which each pixel is valued by its frequency of
appearance and a specific neighboring pixels [30] as shown in (10).
𝑁−1 𝑁−1
𝑢𝑖 = ∑𝑖,𝑗=0 𝑖 𝑃̃𝑖 𝑗 𝑢𝑗 = ∑𝑖,𝑗=0 𝑗 𝑃̃𝑖 𝑗 (10)
𝑚 𝑛
1 −(𝑖×𝑥×𝜋(𝑥 + ))
f(x, y) = ∑𝑀−1 𝑁−1
𝑀=0 ∑𝑁=0 𝑓(𝑚, 𝑛)𝑒
𝑀 𝑁 (13)
𝑀.𝑁
The pixel at location (m, n) is represented by f(m, n), while F(x, y) is the function to represent the
image in the frequency domain with respect to position x and y, M × N indicates the image's dimension, and i
is sqrt (-1). Then, we apply vector quantization on spectral signature images to convert it to feature vector.
Finally, the PCA (50*1) feature vector, GLCM (6*1) feature vector and FFT (50*1) feature vector are
combined producing (106*1) signature feature vector.
Signature verification based on proposed fast hyper deep neural network (Zainab Hashim)
968 ISSN: 2252-8938
The proposed model is trained with 64 batches, 100 epochs, and a learning rate of 0.001. Our model design
uses the adaptive momentum (Adam) optimizer, which has a learning rate of 0.001.
used as classifier of the resulted vectors. The dense layer's stated number of neurons will have an impact on
the output shape. Dense layer carries out the action: activation (dot (input, kernel) +bias) is equal to output,
Table 7 summarized all the layers of the proposed hyper deep model.
Table 7. Layers Summary of the proposed fast hyper deep model (FHDNN)
Layer Filters Parameters
Conv1D 16 Stride=1; Kernal-size=3
Max-pooling 16 Stride=1; Pool size=1
Conv1D 32 Stride=1; Kernal-size=3
Max-pooling 32 Stride=1; Pool size=1
Conv1D 32 Stride=1; Kernal-size=3
Max-pooling 32 Stride=1; Pool size=1
Dense 32 Linear Activation Function
Conv1D 64 Stride=1; Kernal-size=3
Max-pooling 64 Stride=1; Pool size=1
Conv1D 64 Stride=1; Kernal-size=3
Max-pooling 64 Stride=1; Pool size=1
Conv1D 128 Stride=1; Kernal-size=3
Max-pooling 128 Stride=1; Pool size=1
Dense 128 Linear activation function
Conv1D 256 Stride=1; Kernal-size=3
Max-pooling 256 Stride=1; Pool size=1
Dense 512 Linear activation function
Conv1D 512 Stride=1; Kernal-size=3
Max-pooling 512 Stride=1; Pool size=1
Conv1D 35 Stride=1; Kernal-size=3
Flatten None
Dense SoftMax activation function
Where TP, TN, FP, and FN stand for the corresponding true positive, true negative, false positive, and false
negative results. In (15) defines F1-score as the harmonic mean of precision and recall. In (16) and (17)
illustrate precision and recall.
2∗𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛∗𝑅𝑒𝑐𝑎𝑙𝑙
𝐹1𝑆𝑐𝑜𝑟𝑒 = (15)
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛+𝑅𝑒𝑐𝑎𝑙𝑙
𝑇𝑃
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 = (16)
𝑇𝑃+𝐹𝑃
𝑇𝑃
𝑅𝑒𝑐𝑎𝑙𝑙 = (17)
𝑇𝑃+𝐹𝑁
In case of the SigComp2011 dataset, when our method is compared to other state-of-the-art methods
utilizing accuracy and EER on the Dutch and Chinese SigComp2011 dataset, it is evident that our method
outperforms them, with accuracy value of 100% and EER of 0.0 as presented in Table 8. The suggested strategy
has improved accuracy metrics with a growing range of (24.02%) compared to (Ismail Hadjadj (2019) Chinese
[13]) that achives an accuracy of 75.98%.
Signature verification based on proposed fast hyper deep neural network (Zainab Hashim)
970 ISSN: 2252-8938
Table 8. Compares the outcomes of our model using the SigComp2011 dataset with those of other systems
Method Accuracy EER
Arisoy (2021) [6] 90.11% -
Banerjee et al. (2021) Dutch [11] 99.28% 0.03
Banerjee et al. (2021) Chinese [11] 99.12% 0.06
Ruiz et al. (2020) [12] 99.44% 3.93
Tayeb et al. (2017) [13] 98.8% -
Hadjadj et al. (2019) Dutch [14] 97.74% -
Hadjadj et al. (2019) Chinese [14] 75.98% -
Cozzens et al. (2018) Dutch [32] 84.74% -
Solar et al. (2020) Dutch [33] 99.44% 0.03
Alvarez et al. (2016) Dutch [34] 97.00% -
Alvarez et al. (2016) Chinese [34] 95.00% -
Kao and Wen (2020) Dutch [35] 98.96% -
Proposed method 100% 0.0
In addition to the two public datasets that we used, our model was also tested with a private dataset
we collected from 31 signers, with 12 signature images for each signer as shown in Table 9. Every image has
a signature made with a red, blue, or black pen. The results show an accuracy of 99.33% on this dataset.
Gray-scale signatures
Histogram equilization
signatures
Gaussian blured
signatures
Resized signatures
Table 10 compares the outcomes from our model's use of the CEDAR dataset with several novel
techniques. Our model gets better results and exceeded the pre-trained models when it comes to both accuracy
and EER as shown in Figure 3. The small EER value indicates better performance. However, in our case, EER
is equal to 0.0, which means that the proportion of misclassified genuine signatures and the proportion of
misclassified fake signatures are almost the same as shown in (18), and this is due to the datasets we used. In
practical applications EER should be near to zero.
𝐹𝐴𝑅+𝐹𝑅𝑅
𝐸𝐸𝑅 = 2
(18)
Where FAR is false accept rate and FRR is false reject rate.
The proposed model is used with features extracted from PCA, GLCM, and FFT which lead to finding
the best results of (precision, recall, F-measure), Table 11 shows the proposed system results on SigComp2011
and CEDAR datasets. In term of speed, our model achieves great results in training each dataset on hp laptop
with processor (11th Gen Intel(R) Core (TM) i7-1195G7 @ 2.90GHz 2.92 GHz), RAM size 16.0 GB, the
screen card is NVIDIA GeForce MX350 and the operating system is windows10. For SigComp2011, each
epoch in the proposed model executed within 6 sec that make the total training time is approximately 10
minutes, while for CEADAR dataset each epoch is executed in about 18 sec and that make the total training
time 30 minute, and when we compare these results with the VGG16 and VGG19 models that were applied to
the same datasets, it appeared that the suggested model outperformed the previous models in relation to speed
as shown in Table 12. We also compared the accuracy and loss of our model with VGG16 model as shown in
Table13. Our model overcomes the pre-trained model.
Table 10. Compares the outcomes of our model using the CEDAR dataset with those of other systems
Method Accuracy EER
Banerjee et al. (2021) [11] 99.36% 0.01
Hadjadj et al. (2017) [13] 97.99% -
Navid et al. (2019) [15] 88% -
Bhunia et al. (2019) [16] - 7.59
Solar et al. (2020) [33] 99.15% 4.91
Nurullah Çalik et al. (2019) [36] 98.49% -
Jampour and Naserasadi (2019) [37] 98.76% -
Li Liu et al. (2021) [38] - 6.74
Maruyama et al. (2021) [39] - 0.82
Alsuhimat and Mohamad [40] 87.7% 11.40
Kadhm et al. [41] 99.7% -
Proposed Method 100% 0.0
Figure 3. The accuracy obtained from our model on SigComp2011 and CEDAR dataset compared to
other systems
Table 11. The proposed model results of SigComp2011 and CEDAR datasets
Dataset Precision Recall F-score Loss
SigComp2011 1.00 0.487 0.655 0.00001011
CEDAR 1.00 0.507 0.672 0.00000060849
Table 12. Comparision between the proposed model and VGG16 and VGG19 in term of speed
Method SigComp2011 CEDAR
(speed for each epoch) (speed for each epoch)
VGG16 53 sec 179 sec
VGG19 51 sec 91 sec
The proposed model 6 sec 18 sec
Table 13. The proposed model results of SigComp2011 and CEDAR compared with pre-trained model
Method Accuracy Loss Dataset
VGG16 49.85 8.17 SigComp2011
50.71 7.94 CEDAR
The proposed model 100.00 0.00001011 SigComp2011
100.00 0.00000060849 CEDAR
5. CONCLUSION
This study provides a hybrid feature-based method for handwritten signature verification and a
proposed fast hyper deep neural network (FHDNN) that is applicable for writer-independent scenario. So as to
assess the efficiency and performance of our suggested model, we employ two well-known datasets;
SigComp2011 and CEDAR. As initial stage we perform four pre-processing stages on them. Then, we use
PCA, GLCM, and FFT as feature extraction methods to build hybrid feature vector for each image. After that,
these features are inputted into a proposed model, the proposed model's primary goal is to categorize the hybrid
features that were derived from the earlier stage to identify a false signature from an authentic one. This model
was built with 31 layers; nine of them were convolutional layers, eight Max-pooling layers, nine Leaky-ReLU
layers, four Dense layers and one Flatten layer.
The suggested technique enhances the verification accuracy and outperforms the other previous
modern techniques, with an accuracy value of 100% for both datasets and a speeding up the training time to
about 10 minutes for the SigComp2011 dataset and 30 minutes for the CEDAR dataset. Additionally, it has a
Signature verification based on proposed fast hyper deep neural network (Zainab Hashim)
972 ISSN: 2252-8938
high precision rate, which can be attributed to the model's architecture and the choice of effective signature
features.
In future work, we will collect handwritten signatures in Arabic language to evaluate the verification
performance of the proposed model on different languages. Also, we will build a model that depends on both
writer-dependent and writer-independent approaches. Meanwhile, we'll keep investigating signature features
to have the best verification effect with the least amount of training data.
REFERENCES
[1] Z. Hashim, H. M. Ahmed, and A. H. Alkhayyat, “A comparative study among handwritten signature verification methods using
machine learning techniques,” Scientific Programming, vol. 2022, pp. 1–17, Oct. 2022, doi: 10.1155/2022/8170424.
[2] M. A. Taha and H. M. Ahmed, “Iris features extraction and recognition based on the local binary pattern technique,” in 2021 International
Conference on Advanced Computer Applications (ACA), IEEE, Jul. 2021, pp. 16–21. doi: 10.1109/ACA52198.2021.9626827.
[3] M. A. Taha and H. M. Ahmed, “A fuzzy vault development based on iris images,” EUREKA: Physics and Engineering, no. 5, pp.
3–12, Sep. 2021, doi: 10.21303/2461-4262.2021.001997.
[4] H. M. A. Salman and S. Hameed, “Eye detection using Helmholtz principle,” Baghdad Science Journal, vol. 16, no. 4(Suppl.), p.
1087, Dec. 2019, doi: 10.21123/bsj.2019.16.4(Suppl.).1087.
[5] H. M. Ahmad and S. R. Hameed, “Eye diseases classification using hierarchical multilabel artificial neural network,” in 2020 1st.
Information Technology To Enhance e-learning and Other Application (IT-ELA, IEEE, Jul. 2020, pp. 93–98. doi: 10.1109/IT-
ELA50150.2020.9253120.
[6] M. V. Arisoy, “Signature verification using Siamese neural network one-shot learning,” International Journal of Engineering and
Innovative Research, vol. 3, no. 3, pp. 248–260, Sep. 2021, doi: 10.47933/ijeir.972796.
[7] M. Saleem and B. Kovari, “Online signature verification based on signer dependent sampling frequency and dynamic time warping,”
in 2020 7th International Conference on Soft Computing & Machine Intelligence (ISCMI), IEEE, Nov. 2020, pp. 182–186. doi:
10.1109/ISCMI51676.2020.9311604.
[8] N. H. Al-banhawy, H. Mohsen, and N. Ghali, “Signature identification and verification systems: a comparative study on the online
and offline techniques,” Future Computing and Informatics Journal, vol. 5, no. 1, pp. 28–45, Dec. 2020, doi: 10.54623/fue.fcij.5.1.3.
[9] Y. Zhou, J. Zheng, H. Hu, and Y. Wang, “Handwritten signature verification method based on improved combined features,”
Applied Sciences, vol. 11, no. 13, p. 5867, Jun. 2021, doi: 10.3390/app11135867.
[10] H. Nemmour and Y. Chibani, “Off-line signature verification using artificial immune recognition system,” in 2013 International
Conference on Electronics, Computer and Computation (ICECCO), IEEE, Nov. 2013, pp. 164–167. doi:
10.1109/ICECCO.2013.6718254.
[11] D. Banerjee, B. Chatterjee, P. Bhowal, T. Bhattacharyya, S. Malakar, and R. Sarkar, “A new wrapper feature selection method for
language-invariant offline signature verification,” Expert Systems with Applications, vol. 186, p. 115756, Dec. 2021, doi:
10.1016/j.eswa.2021.115756.
[12] V. Ruiz, I. Linares, A. Sanchez, and J. F. Velez, “Off-line handwritten signature verification using compositional synthetic
generation of signatures and Siamese neural networks,” Neurocomputing, vol. 374, pp. 30–41, Jan. 2020, doi:
10.1016/j.neucom.2019.09.041.
[13] S. Tayeb et al., “Toward data quality analytics in signature verification using a convolutional neural network,” in 2017 IEEE
International Conference on Big Data (Big Data), IEEE, Dec. 2017, pp. 2644–2651. doi: 10.1109/BigData.2017.8258225.
[14] I. Hadjadj, A. Gattal, C. Djeddi, M. Ayad, I. Siddiqi, and F. Abass, “Offline signature verification using textural descriptors,” 2019,
pp. 177–188. doi: 10.1007/978-3-030-31321-0_16.
[15] S. M. A. Navid, S. H. Priya, N. H. Khandakar, Z. Ferdous, and A. B. Haque, “Signature verification using convolutional neural
network,” in 2019 IEEE International Conference on Robotics, Automation, Artificial-intelligence and Internet-of-Things
(RAAICON), IEEE, Nov. 2019, pp. 35–39. doi: 10.1109/RAAICON48939.2019.19.
[16] A. K. Bhunia, A. Alaei, and P. P. Roy, “Signature verification approach using fusion of hybrid texture features,” Neural Computing
and Applications, vol. 31, no. 12, pp. 8737–8748, Dec. 2019, doi: 10.1007/s00521-019-04220-x.
[17] M. Liwicki et al., “Signature verification competition for online and offline skilled forgeries (SigComp2011),” in 2011 International
Conference on Document Analysis and Recognition, IEEE, Sep. 2011, pp. 1480–1484. doi: 10.1109/ICDAR.2011.294.
[18] M. A. Shaikh, T. Duan, M. Chauhan, and S. N. Srihari, “Attention based writer independent verification,” in 2020 17th International
Conference on Frontiers in Handwriting Recognition (ICFHR), IEEE, Sep. 2020, pp. 373–379. doi:
10.1109/ICFHR2020.2020.00074.
[19] S. Rajendran, R. Dorothy, R. M. Joany, R. J. Rathish, S. Santhana Prabha, and S. Rajendran, “Image enhancement by Histogram
equalization,” Int. J. Nano. Corr. Sci. Engg, vol. 2, no. 4, pp. 21–30, 2015, [Online]. Available:
https://www.researchgate.net/publication/283727396
[20] P. Singhal, A. Verma, and A. Garg, “A study in finding effectiveness of Gaussian blur filter over bilateral filter in natural scenes
for graph based image segmentation,” in 2017 4th International Conference on Advanced Computing and Communication Systems
(ICACCS), IEEE, Jan. 2017, pp. 1–6. doi: 10.1109/ICACCS.2017.8014612.
[21] B. K. Triwijoyo and A. Adil, “Analysis of medical image resizing using bicubic interpolation algorithm,” Jurnal Ilmu Komputer,
vol. 14, no. 1, p. 20, Apr. 2021, doi: 10.24843/JIK.2021.v14.i01.p03.
[22] M. Malaisamy, “Principal component analysis based feature vector extraction,” Indian Journal of Science and Technology, vol. 8,
no. 35, Dec. 2015, doi: 10.17485/ijst/2015/v8i35/77760.
[23] A. Wirdiani, T. Lattifia, I. K. Supadma, B. J. K. Mahar, D. A. N. Taradhita, and A. Fahmi, “Real-time face recognition with
eigenface method,” International Journal of Image, Graphics and Signal Processing, vol. 11, no. 11, pp. 1–9, Nov. 2019, doi:
10.5815/ijigsp.2019.11.01.
[24] E. A. Khorsheed and Z. A. Nayef, “Face recognition algorithms: a review,” Academic Journal of Nawroz University, vol. 11, no. 3,
pp. 202–207, Aug. 2022, doi: 10.25007/ajnu.v11n3a1432.
[25] B. Sebastian, A. Unnikrishnan, and K. Balakrishnan, “Gray level co-occurrence matrices: generalisation and some new features,”
May 2012, [Online]. Available: http://arxiv.org/abs/1205.4831
[26] S. Bakheet and A. Al-Hamadi, “Automatic detection of COVID-19 using pruned GLCM-Based texture features and LDCRF
classification,” Computers in Biology and Medicine, vol. 137, p. 104781, Oct. 2021, doi: 10.1016/j.compbiomed.2021.104781.
[27] K. Harrar, K. Messaoudene, and M. Ammar, “Combining GLCM with LBP features for knee osteoarthritis prediction: data from the
Osteoarthritis initiative,” ICST Transactions on Scalable Information Systems, p. 171550, Jul. 2018, doi: 10.4108/eai.20-10-2021.171550.
[28] H. A. Dwaich and H. A. Abdulbaqi, “Signature texture features extraction using GLCM approach in Android studio,” Journal of
Physics: Conference Series, vol. 1804, no. 1, p. 012043, Feb. 2021, doi: 10.1088/1742-6596/1804/1/012043.
[29] N. Iqbal, R. Mumtaz, U. Shafi, and S. M. H. Zaidi, “Gray level co-occurrence matrix (GLCM) texture based crop classification
using low altitude remote sensing platforms,” PeerJ Computer Science, vol. 7, p. e536, May 2021, doi: 10.7717/peerj-cs.536.
[30] A. K. Aggarwal, “Learning texture features from GLCM for classification of brain tumor MRI images using random forest
classifier,” WSEAS Transactions on Signal Processing, vol. 18, pp. 60–63, Apr. 2022, doi: 10.37394/232014.2022.18.8.
[31] D. Yanqing, Y. Guoqing, and Z. Yanjie, “Remote sensing image content retrieval based on frequency spectral energy,” Procedia
Computer Science, vol. 107, pp. 448–453, 2017, doi: 10.1016/j.procs.2017.03.088.
[32] B. Cozzens et al., “Signature verification using a convolutional neural network,” 2018.
[33] J. Ruiz-del-Solar, C. Devia, P. Loncomilla, and F. Concha, “Offline signature verification using local interest points and
descriptors,” 2008, pp. 22–29. doi: 10.1007/978-3-540-85920-8_3.
[34] G. Alvarez, M. Bryant, B. Sheffer, and M. Bryant, “Offline signature verification with convolutional neural networks,” Technical
report, Stanford University, p. 8, 2016.
[35] H.-H. Kao and C.-Y. Wen, “An offline signature verification and forgery detection method based on a single known sample and an
explainable deep learning approach,” Applied Sciences, vol. 10, no. 11, p. 3716, May 2020, doi: 10.3390/app10113716.
[36] N. Çalik, O. C. Kurban, A. R. Yilmaz, T. Yildirim, and L. Durak Ata, “Large-scale offline signature recognition via deep neural
networks and feature embedding,” Neurocomputing, vol. 359, pp. 1–14, Sep. 2019, doi: 10.1016/j.neucom.2019.03.027.
[37] M. Jampour and A. Naserasadi, “Chaos game theory and its application for offline signature identification,” IET Biometrics, vol. 8,
no. 5, pp. 316–324, Sep. 2019, doi: 10.1049/iet-bmt.2018.5188.
[38] L. Liu, L. Huang, F. Yin, and Y. Chen, “Offline signature verification using a region based deep metric learning network,” Pattern
Recognition, vol. 118, p. 108009, Oct. 2021, doi: 10.1016/j.patcog.2021.108009.
[39] T. M. Maruyama, L. S. Oliveira, A. S. Britto, and R. Sabourin, “Intrapersonal parameter optimization for offline handwritten
signature augmentation,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 1335–1350, 2021, doi:
10.1109/TIFS.2020.3033442.
[40] F. M. Alsuhimat and F. S. Mohamad, “Offline signature verification using long short-term memory and histogram orientation
gradient,” Bulletin of Electrical Engineering and Informatics, vol. 12, no. 1, pp. 283–292, Feb. 2023, doi: 10.11591/eei.v12i1.4024.
[41] M. S. Kadhm, M. J. Mohammed, and H. Ayad, “An accurate signature verification system based on proposed HSC approach and
ANN architecture,” Indonesian Journal of Electrical Engineering and Computer Science, vol. 21, no. 1, p. 215, Jan. 2021, doi:
10.11591/ijeecs.v21.i1.pp215-223.
BIOGRAPHIES OF AUTHORS
Zainab Hashim received the B.S.c degree in computer science from University
of Technology, Baghdad, Iraq, in 2016, the M.Sc. degree in computer/artificial intelligence
from University of Technology, Baghdad, Iraq, in 2019. She is currently a PhD student in
computer science from University of Technology. Her research interests include artificial
intelligence, machine learning, and biometrics. She can be contacted at email:
[email protected].
Hanaa Mohsin Assistant Professor Dr. Hanaa Moshin Ahmed Salman. Obtained
her M.Sc. and Ph.D. from the University of Technology, Baghdad, Iraq, in 2002 and 2006,
respectively. Currently she is a lecturer in Computer Science and a member of the Scientific
Committee and Promotion Committee in the Department of Computer Science. Dr. Hanaa has
more than 23 years of experience and she has supervised graduate students. Her primary
research interests include cryptography, computer security, biometrics, image processing, and
computer graphics. She can be contacted at email: [email protected].
Signature verification based on proposed fast hyper deep neural network (Zainab Hashim)