Skip to main content

ORIGINAL RESEARCH article

Front. Robot. AI, 07 May 2021
Sec. Robot Vision and Artificial Perception
This article is part of the Research Topic Current Trends in Image Processing and Pattern Recognition View all 8 articles

Behavioral Biometric Data Analysis for Gender Classification Using Feature Fusion and Machine Learning

Shivanand S. GornaleShivanand S. Gornale1Sathish Kumar
Sathish Kumar1*Abhijit PatilAbhijit Patil1Prakash S. HiremathPrakash S. Hiremath2
  • 1Department of Computer Science, Rani Channamma University, Belagavi, India
  • 2Department of MCA, KLE Technological University, Hubballi, India

Biometric security applications have been employed for providing a higher security in several access control systems during the past few years. The handwritten signature is the most widely accepted behavioral biometric trait for authenticating the documents like letters, contracts, wills, MOU’s, etc. for validation in day to day life. In this paper, a novel algorithm to detect gender of individuals based on the image of their handwritten signatures is proposed. The proposed work is based on the fusion of textural and statistical features extracted from the signature images. The LBP and HOG features represent the texture. The writer’s gender classification is carried out using machine learning techniques. The proposed technique is evaluated on own dataset of 4,790 signatures and realized an encouraging accuracy of 96.17, 98.72 and 100% for k-NN, decision tree and Support Vector Machine classifiers, respectively. The proposed method is expected to be useful in design of efficient computer vision tools for authentication and forensic investigation of documents with handwritten signatures.

Introduction

Now days, Biometrics is a widely established measure in security and authentication processes in several system applications. According to International Standard Organization (ISO), biometric is defined as a means of biological process for recognizing and analyzing an individual based on their physiological (fingerprint, face, iris, palm prints) and behavioral characteristics (Signature, Gaits, keystroke, voice/speech) (Fairhurst et al., 2017). Among these traits, handwritten signature is the most preferred age-old behavioral trait, which is used to authenticate the documents due to its individuality and consistence features. Signatures were first knowingly recorded in the Talmud (Steinsalz, 1992) in the IV century. Since then the nature of handwriting signature has been continuously evolving, in the past decades, from traditional way of handwritten signature with pen or pencil on a piece of paper to digitally recording signature, i.e. on a tablet by digitizing pen. Nowadays, with the advent of the electronic scanning devices, handwritten signatures are digitized and then stored for use in automatic document verification and authentication processes. Signatures basically comprise horizontal, vertical and curving strokes with multilingual texts which mostly denote name and last name, as the description made by its handler/person, representing his agreement or presence on that document. Handwritten signature is significant as a handy tool for identifying and verifying the document as well as for authorizing the user of a document. Even today, signatures are still considered as the most reliable and adequate evidences in the court of law. Signature, being a behavioral biometric trait, measures the identity of an individual uniquely. The handwritten-signature-trait contains many dynamic and inherent behavioral features, which are useful in the determination of a person’s soft characteristics like gender, handedness, personality and age information, etc. (Jain et al., 2004) (Singh and Verma, 2019).

In this paper, the objective of the proposed work is to design a machine learning framework for analyzing the gender classification of writers based on their handwritten signatures using fusion of textural and statistical features. The LBP and HOG features represent the texture. The k-NN, decision tree and SVM classifiers are used to perform classification. Further, the paper is organized as follows. Related Work contains the literature on the work related to the signatures based gender classification. Proposed Methodology focuses on the proposed methodology. In Experimental Results and Analysis, experimental results are analyzed and compared with the existing results. Finally, in Conclusion, the conclusions are given.

Related Work

In the literature on the recognition of handwritten signature as a biometric trait, a considerable number of studies have been reported on specific applications such as signature verification, handwriting recognition and writer identification. However, work related to writer’s gender classification domain is very scanty. In this section, a review of related studies on gender classification is presented.

A. A. M. Abushariah et al. (Abushariah et al., 2012) have performed signature based gender classification using the global features, namely, height, width and area, on own image dataset of 3,000 signatures of 50 males and 50 females. By using artificial neural network (ANN), an average accuracy of 76.20 and 74.20% for male and female, respectively, has been obtained.

Tasmina Islam et al. (Islam and Fairhurst, 2019) have investigated the effect of writer style, age, and gender on Natural Revocability Analysis in Handwritten Signature Biometrics. A dataset of a total of 4,190 signatures were collected from different gender and age groups, and sixty features were extracted. Signatures were analyzed using the Analysis of Variance (ANOVA) technique and it is found that adopting different style signatures will not affect the new naturally revoked signature. There is no much difference in the static and dynamic features of original and new signatures. Natural revocability can be the feasible option in demographic factors such as age and gender of the handwritten signatures.

Prasentjit Maji et al. (Maji et al., 2015) have performed gender classification based on offline handwritten signature by using basic statistical features over a dataset of 500 signatures. By using back propagation neural network, an accuracy of 84%has been achieved.

Cavalcante Bandeira, et al. (Cavalcante Bandeira et al., 2019) have investigated the impact of the combination of a handwritten signature and keyboard keystroke dynamics for gender prediction. From 100 participants, keystroke characteristics and handwritten signature images were collected. By extracting basic static statistical and dynamic features, an average accuracy of 68.03% is obtained with multilayer perceptron.

Moumita Pal, et al. (pal et al., 2018) have performed gender classification using Euler number based feature extraction on 500 Hindi handwritten signatures images. Using back propagation neural network classifier, an average accuracy of 88.80% is obtained.

Many researchers have worked on writer identification of handwritten signatures and handwritings (Al-ma'adeed and Hassaïne, 2014; Ibrahim et al., 2014; Siddiqi et al., 2014; SRIVASTAVA, 2014; Mirza et al., 2016; Bouadjenek et al., 2017; Topaloglu and Ekmekci, 2017; Navya et al., 2018; Baboria et al., 2019; Faundez-Zanuy et al., 2020). However, it is observed that very less effort is done on gender identification based on offline handwritten signatures. These signatures may contain multilingual texts. Therefore, to bridge this research gap, a frame work is proposed to identify writer’s gender from handwritten signatures of individuals with varying ages.

Proposed Methodology

The general steps involved in the proposed writer’s gender classification based on handwritten signature biometric are shown in Figure 1.

FIGURE 1
www.frontiersin.org

FIGURE 1. General methodology for the gender classification using handwritten signature.

Dataset

From the literature review, it has been found that the publicly available standard image datasets of handwritten signatures do not have gender annotation. With this motivation, own dataset is created to carry out the experiments with the proposed method. The nature of this own offline handwritten signature database is described below.

• The signature samples are collected from persons with varying ages. These samples consist of multilingual scripts of Kannada, Hindi, Marathi, and English.

• Each individual has done 10 signatures on a white A4 paper sheet using blue or black colored ball pen.

• To avoid geometrical variations further, the papers with sample signatures have been scanned using the EPSON DS-1630 color scanner with a resolution of 300 DPI.

• The database consists of 4,790 signatures, which are acquired from 479 subjects, out of whom 250 are male volunteers and 229 are female volunteers.

• All the individuals, who had knowledge of English and other languages, have been educated about the purpose of collection of signature samples.

Some samples of these signatures are shown in Figure 2.

FIGURE 2
www.frontiersin.org

FIGURE 2. Handwritten signature samples from the own dataset.

Preprocessing

The main aim of pre-processing is to improve the data and remove unwanted distortion or noise and also to enhance important features of handwritten signatures. The various techniques are taken into consideration such as converting the raw handwritten signature input image to a standard Grayscale, than normalizing images by cropping into a standard size of 150 × 150. As shown in the Figure 3.

FIGURE 3
www.frontiersin.org

FIGURE 3. Preprocessing of signatures.

Feature Extraction

Feature extraction for object recognition in the image is the most important part of the gender classification approach (Gornale et al., 2020). In this proposed work, three different feature extraction approaches (the first one is the HOG approach and the second one is the LBP approach and the third one is the Statistical approach) and Textural features are used. The details of these approaches are given below.

Local Binary Patterns

LBP was first introduced by Ojala et al., in 1996, as an efficient and powerful texture descriptor, which is widely used in image processing and computer vision areas for feature and histogram representation (Ojala et al., 1996). The original LBP operator considers eight neighborhood of a pixel, with Center pixel being used as threshold (Shivanand, 2017). It generates an 8-bit binary pattern for each pixel depending on its 3 × 3 neighborhood, which is converted to the corresponding decimal value, as illustrated in Figure 4. The distribution of these binary patterns, represented by the equivalent decimal, serve as the local texture descriptor. In general, for a pixel at ic = (xc, yc), the LBP is calculated by using the following formula:

LBP (xc,yc)=n=0N1s(inic)2n,(1)

where in is a neighbor pixel of the pixel at ic and s = 1 if in ≥ ic otherwise s = 0. From each signature image, 58 LBP features are extracted (Shivanand, 2017).

FIGURE 4
www.frontiersin.org

FIGURE 4. LBP computation.

Histogram of Oriented Gradients

The HOG, first introduced by Dalal and Triggs (Dalal and Triggs, 2005), is one of the efficient local texture descriptor that is often utilized in the computer vision and image processing for capturing the distribution of the edges of an object or its local intensity gradients value (R et al., 2019). The HOG feature vector is calculated from the image gradient detectors, each convolved with a simple convolution window as follows:

Gx=I(x+1,y)I(x1,y),Gy=I(x,y+1)I(x,y1),(2)

where, I (x, y) is the pixel intensity at location (x, y) in the input image; Gx and Gy are the horizontal and vertical components of the gradients, respectively.

Statistical Features

Statistical features are used to characterize the location and variability of the pixel distribution in a signature image. In this work, the statistical features and shape features of handwritten signatures that are computed are as following:

Area is the number of pixels in an image.

Skewness=E(xμ)2σ3.

• Euler Number = (number of objects)−(number of holes).

Euler(n)=2n(n,12).

• Perimeter is the number of boundary pixels.

Perimeter = bwperim(BW)

• Eccentricity is ratio of the distance between the foci of the ellipse and major axis length. Eccentricitye=(1b2a2)12.

• EquivDiameter is the diameter of the circle with same area as region. Equiv-diameterd=Areaπ2.

• Standard Deviation is one of the quantitative measurement. StandardDeviation=1N1i=1N|Aiμ|2.

Textural Features

Entropyfilter:M=Entropy(H)=i=1npi(log2pi),

Standardfilter= stdfilt(f),

Rangefilter= rangefilt(I).

All the features determine the center element of the neighborhood (Gornale et al., 2016).

Feature Level Fusion

The feature level fusion is sparingly utilized due to its complexity of larger dimensions in fused feature space (Gornale et al., 2016). However, the handwritten signature have inherently rich sets of characteristics and thus, intuitively, fusion of features is expected to be more effective. Thus, in this work, the features of handwritten signatures obtained from HOG, statistical and LBP techniques have been fused together by using concatenation rule (Gornale et al., 2019).

Classification

K-Nearest Neighbor (k-NN) Classifier: k-NN classifier classifies the data based upon the different kinds of distances selected for the purpose (Gornale et al., 2016). It classifies the object feature vectors into k classes based on a distance metric as similarity measure. Herein, the city block distance is used and it is given as following:

DCity-block(M,N)=J=0n|MJNJ|.(13)

Decision Tree

Decision tree is a supervised machine learning technique for inducing a decision tree from training data. The decision tree predictive model will map the observations against an item to deduce the conclusion about its targeted value. Decision trees are grown using training data. Starting at the root node, the data is recursively split into subsets. In each step the best split is determined based on a criterion (Reinders et al., 2019). Commonly used criteria are G index and entropy index:

GIndex:G(E)=1j=1cpj2,(14)
EntropyindexH(E)=j=1cpjlogpj.(15)

The decision tree works on randomly chosen n training samples from the handwritten signature dataset. It creates a root and assigns the sample data to each node, this is repeated until all nodes comprise a single sample from same class, randomly selects m variables from M possible variables. Picks the best split feature and threshold the samples using the G index and entropy index and splits the node into two child nodes and pass the corresponding subsets.

Support Vector Machine

SVM is a supervised binary classifier. It works on a decision boundary in multidimensional feature vector space that is able to separate the classes of training set of vectors with known class labels. Geometrically, SVM are those training patterns that are closest to the decision boundary. SVM algorithm merely seeks out the separating hyperplane with the highest margin (Gornale. et al., 2016).

The proposed method is represented in the form of Algorithm as given below:

Input: Handwritten signature image.

Output: Writer’s gender classification.

Step 1: Input a handwritten signature image.

Step 2: Perform pre-processing of input image that includes noise removal and resizing by normalizing to a size of 150 × 150.

Step 3: Compute features

1) Local Binary Patterns, Histogram of Oriented Gradients and Statistical features individually.

2) Fusion of Local Binary Patterns, Histogram of Oriented Gradients and Statistical features.

Step 4: Classify the fused features using k-NN, decision tree and SVM classifiers and output the writer’s gender class.

End.

Experimental Results and Analysis

The experimentation of the proposed method is carried out using own dataset which is described in Dataset. The dataset contains a total 4,790 signature images which are collected from 250 male and 229 female volunteers belonging to different age groups. Firstly, the performance of LBP, HOG and statistical features is investigated separately. With local binary patterns (LBP), 58 features are computed and then classified. By using HOG, 10,404 features are extracted and then classified. Next, 8 statistical features are extracted and then classified. In each case, k-NN, decision tree and SVM binary classifiers are employed.

Then, feature level fusion of Local Binary Patterns, Histogram of Oriented Gradients and Statistical and Textural features has been performed by concatenation of feature vectors. The fused features are stored as final feature vector. Finally, these features are classified by applying k-NN, decision tree and SVM binary classifiers.

To initiate the experiment, independent performance of LBP features of the handwritten signatures images is examined. It is observed that the highest accuracy of 96.11% is achieved by Decision Tree, and the lowest accuracy of 92.31% obtained by K-NN classifier. Similarly, in case of the independent analysis of HOG, it is found that the highest accuracy of 97.78% is achieved by Decision Tree, and the lowest accuracy of 94.55% obtained by K-NN classifier. Also, in case of the independent analysis of Statistical features, the highest accuracy of 92.65% is achieved by Decision Tree, and the lowest accuracy of 82.12% obtained by K-NN classifier.

Further, the fusion of different features were tested over the signature database and it was noted that, by fusing HOG and LBP features, the highest accuracy of 97.87% is achieved by Decision Tree, and the lowest accuracy of 94.53% obtained by K-NN classifier. Further, with the fusion of LBP and Statistical features, the highest accuracy of 96.20% is achieved by Decision Tree, and the lowest accuracy of 82.25% obtained by K-NN classifier. Also, in case of the fusion of HOG and the statistical features, the highest accuracy of 98.37% is achieved by Decision Tree, and the lowest accuracy of 81.67% obtained by K-NN classifier. Finally, the fusion of LBP, HOG, Statistical features yielded enhanced accuracy of 98.72% by Decision Tree, and 96.17% obtained by K-NN classifier. Table 1 contains the experimental results in terms of classification accuracy and confusion matrix, for each of the features and fusion of features, obtained by the binary classifiers, namely, k-NN, decision tree and SVM.

TABLE 1
www.frontiersin.org

TABLE 1. Classification accuracy of the proposed method and confusion matrix, for each of the features and fusion of features, obtained by the binary classifiers, namely, k-NN, decision tree and SVM.

The performance of the proposed algorithm is analyzed in terms of the metrics, namely, precision, F_Score, True Positive Rate (Recall) and False Positive Rate (Specificity), which are defined by Eqs. 1619.

Precision= TPTP+FP,(16)
F_Score= 2*Precision*True Positive RatePrecsion+True Positive Rate,(17)
True Positive Rate= TPTP+FN,(18)
False Positive Rate= FPFP+TN,(19)

where, TN denotes True Negative, TP denotes True Positive, FP denotes False Positive and FN denotes False Negative. The performance comparison of the individual features with the fused feature is presented in Table 2 for each of the classifiers used in the classification experiments and also in Figure 5 graphically.

TABLE 2
www.frontiersin.org

TABLE 2. The performance comparison of the individual features with the fused features is for each of the classifiers, namely, k-NN, Decision Tree and SVM.

FIGURE 5
www.frontiersin.org

FIGURE 5. Graphical representation of features and classifiers.

Comparative Analysis

To realize the effectiveness of the proposed method, it is compared with the similar works present in literature and are depicted in Table 3. A. A. M. Abushariah et al. (Abushariah et al., 2012) have performed signature based gender classification using the global features, namely, Height, width and Area on own database of 3,000 signatures of 50 males and 50 females. By using Artificial Neural Network (ANN), an average accuracy of 76.20 and 74.20% for male and female was obtained. Prasentjit Maji et al. (Maji et al., 2015) have performed gender classification based on offline handwritten signature by using basic statistical features over a dataset of 500 signatures. By using Back Propagation Neural Network, an accuracy of 84% was achieved. Danilo R.C. Bandeira et al. (Cavalcante Bandeira et al., 2019) have investigated the impact of the combination of a handwritten signature and keyboard keystroke dynamics for gender prediction. From 100 participants keystroke characteristics and handwritten images were collected. By extracting basic Static statistical and dynamic features obtained an average accuracy of 68.03% with Multi-Layer perceptron. Moumita Pal et al. (pal et al., 2018) have performed gender classification using Euler number based feature extraction on 500 Hindi handwritten signatures images. Based on Back Propagation Neural Network classification, an average accuracy of 88.80% is obtained. The drawback of the reported works is that the database of limited sizes are used for experiments. Whereas, the proposed method outperformed all these methods by using the fusion of LBP, HOG, Statistical and Texture features with decision tree classifier, wherein a relatively larger database consisting of 4,790 signature images is used. The proposed method based on feature fusion has yielded an accuracy of 98.72%, which is an encouraging result. Table 3 shows a summary of the related methods in the literature, features, classifiers and datasets used to achieve the reported classification accuracy. It is observed the proposed method based on feature fusion has yielded higher accuracy of gender classification than other methods.

TABLE 3
www.frontiersin.org

TABLE 3. Summary of the related methods in the literature, features, classifiers and datasets used to achieve the reported classification accuracy.

Conclusion

In the present study, an effective method for gender identification based on handwritten signature is proposed using feature fusion and machine learning. Although from the related work, it is observed that researchers who have worked on gender classification using different methodologies and obtained some promising results with own datasets or standard datasets, there is still a large scope for developing a robust algorithm using effective discriminatory features. In the proposed method, feature-level fusion of LBP, HOG, Statistical and Textural features is experimented on own dataset comprising 4,790 signatures images (250 males and 229 females) of good quality. It is found that the proposed algorithm produces accuracies of 96.17% for k-NN, 98.72% for Decision Tree and 100% for Support Vector Machine, respectively. In future, the classification problem of gender determination from handwritten signatures and other biometrics using deep learning will be studied.

Data Availability Statement

The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.

Ethics Statement

The sample signatures of persons are collected from persons who voluntarily participated after due notification of the purpose of this research work. These samples are used only for the research work.

Author Contributions

SG: Paper Writing, SK: Data collection and Experimental analysis, AP: Testing and Validation, PH: Scientific editing and Proof Reading.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The authors are grateful to the reviewers for their helpful comments and suggestions which improved the quality of this paper considerably.

References

Abushariah, A. A. M., Gunawan, T. S., and Chebil, J. (2012). Automatic Person Identification System Using Handwritten Signatures, 3–5.

Al-ma'adeed, Somaya, and Hassaïne, Abdelâali. (2014). Automatic Prediction of Age, Gender, and Nationality in Offline Handwriting. EURASIP J. Image Video Process. 10. doi:10.1186/1687-5281-2014-10

CrossRef Full Text | Google Scholar

Baboria, B., Kaur, P., and Gupta, B. (2019). “Comparative Analysis of Statistical-GA Writer Identification Abstract. Int. J. Scientific Res. Eng. Dev. 2 (4), 574–581.2581-7175

Google Scholar

Bouadjenek, N., Nemmour, H., and Chibani, Y. (2017). “Writer's Gender Classification Using HOG and LBP Features,” in Recent Advances in Electrical Engineering and Control Applications, 317–325. doi:10.1007/978-3-319-48929-2_24

CrossRef Full Text | Google Scholar

Cavalcante Bandeira, D. R., de Paula Canuto, A. M., Da Costa-Abreu, M., Fairhurst, M., Li, C., and Costa do Nascimento, D. S. (2019). Investigating the Impact of Combining Handwritten Signature and Keyboard Keystroke Dynamics for Gender Prediction." 8th Brazilian Conference on Intelligent Systems (BRACIS). doi:10.1109/bracis.2019.00031

CrossRef Full Text | Google Scholar

Dalal, N., and Triggs, B. (2005). Histograms of Oriented Gradients for Human Detection: Histograms of Oriented Gradients for Human Detection. IEEE Computing. Society. Conference. Computer. Vision. And Pattern Recognition, 886–893. [Online]. Available: http://lear.inrialpes.fr.ss.

Fairhurst, M., Li, C., and Da Costa‐Abreu, M. (2017). Predictive Biometrics: a Review and Analysis of Predicting Personal Characteristics from Biometric Data. IET biom. 6 (6), 369–378. doi:10.1049/iet-bmt.2016.0169

CrossRef Full Text | Google Scholar

Faundez-Zanuy, M., Fierrez, J., Ferrer, M. A., Diaz, M., Tolosana, R., and Plamondon, R. (2020). Handwriting Biometrics: Applications and Future Trends in E-Security and E-Health. Cogn. Comput. 12, 940–953. doi:10.1007/s12559-020-09755-z

CrossRef Full Text | Google Scholar

Gornale, Shivanand. S., Patil, Abhijit., and Kruthi, R. (2019). Multimodal Biometrics Data Based Gender Classification Using Machine Vision Int. J. Innovative Tech. Exploring Eng. (Ijitee) ISSN 8 (11).2278-3075

Google Scholar

Gornale, Shivanand. S., Patravali, Pooja. U., and Manza, Ramesh. R. (2016). Detection of Osteoarthritis Using Knee X-Ray Image Analyses: A Machine Vision Based Approach. Int. J. Comp. Appl. 145 (1). doi:10.5120/ijca20169105440975 – 8887

CrossRef Full Text | Google Scholar

Gornale, S., Patil, A., and C., V. (2016). Fingerprint Based Gender Identification Using Discrete Wavelet Transform and Gabor Filters. Ijca 152, 34–37. doi:10.5120/ijca2016911794

CrossRef Full Text | Google Scholar

Gornale., Shivanand. S., Patil, A., and Prabha, (2016). Statistical Features Based Gender Identification Using SVM. International Journal for Scientific Research and Development, 241–244.

Ibrahim, A. S., Youssef, A. E., and Abbott, A. L. (2014). “Global vs. Local Features for Gender Identification Using Arabic and English Handwriting,” in IEEE International Symposium on Signal Processing and Information Technology (ISSPIT). Noida, India, 000155–000160. doi:10.1109/ISSPIT.2014.7300580

CrossRef Full Text | Google Scholar

Islam, T., and Fairhurst, M. (2019). Investigating the Effect of Writer Style, Age and Gender on Natural Revocability Analysis in Handwritten Signature Biometric. 2019 Eighth International Conference on Emerging Security Technologies. (EST). Essex Business School, University of Essex: Colchester, UK. doi:10.1109/est.2019.8806234

CrossRef Full Text | Google Scholar

Jain, A. K., Dass, S. C., and Nandakumar, K. (2004). Can Soft Biometric Traits Assist User Recognition? Proc. SPIE - Int. Soc. Opt. Eng. 5404. doi:10.1117/12.542890

CrossRef Full Text | Google Scholar

Maji, P., Chatterjee, S., Chakraborty, S. N. K., Samanta, S., and Dey, N. (2015). Effect of Euler Number as a Feature in Gender Recognition System from Offline Handwritten Signature Using Neural Networks. 2nd International Conference on Computing for Sustainable Global Development. (New Delhi, India: INDIACom), 1869–1873.

Google Scholar

Mirza, A., Moetesum, M., Siddiqi, I., and Djeddi, C. (2016). Gender Classification from Offline Handwriting Images Using Textural Features, 395–398. doi:10.1109/ICFHR.2016.0080

CrossRef Full Text

Navya, B. J., Shivakumara, P., Shwetha, G. C., Roy, S., Guru, D. S., Pal, U., et al. (2018). Adaptive Multi-Gradient Kernels for Handwritting Based Gender Identification. 16th International Conference on Frontiers in Handwriting Recognition (ICFHR). NY, USA: Niagara Falls, 392–397. doi:10.1109/ICFHR-2018.2018.00075

CrossRef Full Text | Google Scholar

Ojala, T., Pietikäinen, M., and Harwood, D. (1996). A Comparative Study of Texture Measures with Classification Based on Featured Distributions. Pattern Recognition 29 (1), 51–59. doi:10.1016/0031-3203(95)00067-4

CrossRef Full Text | Google Scholar

pal, M., Bhattacharyya, S., and Sarkar, T. (2018). “Euler Number Based Feature Extraction Technique for Gender Discrimination from Offline Hindi Signature Using SVM & BPNN Classifier,” in Emerging Trends in Electronic Devices and Computational Techniques (Kolkata: EDCT)), 1–6. doi:10.1109/EDCT.2018.8405084

CrossRef Full Text | Google Scholar

R, K., Patil, A., and Gornale, S. (2019). Fusion of Local Binary Pattern and Local Phase Quantization Features Set for Gender Classification Using Fingerprints. ijcse 7, 22–29. doi:10.26438/ijcse/v7i1.2229

CrossRef Full Text | Google Scholar

Reinders, C., Ackermann, H., Yang, M. Y., and Rosenhahn, B. (2019). Learning Convolutional Neural Networks for Object Detection with Very Little Training Data. Multimodal Scene Understanding 65, 65–100. doi:10.1016/b978-0-12-817358-9.00010-x

CrossRef Full Text | Google Scholar

Shivanand, S. (2017). Gornale, Basavanna M, and Kruthi R “Fingerprint Based Gender Classification Using Local Binary Pattern”. Int. J. Comput. Intelligence Res. 13 (2), 261–271. ISSN 0973-1873.

Google Scholar

Gornale, Shivanand. S., Patil, Abhijit., and Ramchandra, Kruti., “Multimodal Biometrics Data Analysis for Gender Estimation Using Deep Learning”, Int. J. Data Sci. Anal. (2020). 6(2): 64–68. doi:10.11648/j.ijdsa.20200602.11http://www.sciencepublishinggroup.com/j/(Print); ISSN: 2575-1891 (Online).

CrossRef Full Text | Google Scholar

Siddiqi, I., Djeddi, C., Raza, A., and Souici-meslati, L. (2014). Automatic Analysis of Handwriting for Gender Classification. Pattern Anal. Applic 18 (4), 887–899. doi:10.1007/s10044-014-0371-0

CrossRef Full Text | Google Scholar

Singh, Harshdeep., and Verma, Priyanka. (2019). Anita and Navjot Kaur “Analysis of Signature Features in Different Age Groups”. Peer Reviewed J. Forensic Genet. Sci. 14 (2). doi:10.32474/PRJFGS.2019.03.000160

CrossRef Full Text | Google Scholar

Srivastava, Swati. (2014). Analysis of Male and Female Handwriting. Int. J. Adv. Res. Comp. Sci. [S.l.] 5 (6), 162–164. doi:10.26483/ijarcs.v5i5.2209

CrossRef Full Text | Google Scholar

Steinsalz, Rabbi. Adin. (1992). Ketubot (With Commentary), the Talmud, Tractate Ketubot. New York, NY: Random House. (Bibin_To Compare characteristics of Electronic signatures and Handwritten Signatures.pdf (aditya.ac.in)).

Topaloglu, M., and Ekmekci, S. (2017). Gender Detection and Identifying One’s Handwriting with Handwriting Analysis. Expert Syst. Appl. 79, 236–243. doi:10.1016/j.eswa.2017.03.001

CrossRef Full Text | Google Scholar

Keywords: biometrics, gender classification, offline handwritten signature, LBP, HOG, K-NN, decision tree, support vector machine

Citation: Gornale SS, Kumar S, Patil A and Hiremath PS (2021) Behavioral Biometric Data Analysis for Gender Classification Using Feature Fusion and Machine Learning. Front. Robot. AI 8:685966. doi: 10.3389/frobt.2021.685966

Received: 26 March 2021; Accepted: 20 April 2021;
Published: 07 May 2021.

Edited by:

Manza Raybhan Ramesh, Dr Babasaheb Ambedkar Marathwada University, India

Reviewed by:

Arjun Mane, Government Institute of Forensic Science Nagpur, India
V. S. Malemath, KLE Dr. M.S. Sheshgiri College of Engineering and Technology, India

Copyright © 2021 Gornale, Kumar, Patil and Hiremath. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sathish Kumar, c2F0aGlzaGt1bWFyc3QyNUBnbWFpbC5jb20=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.