Skip to main content

ORIGINAL RESEARCH article

Front. Big Data, 22 July 2022
Sec. Cybersecurity and Privacy

Boosting Face Presentation Attack Detection in Multi-Spectral Videos Through Score Fusion of Wavelet Partition Images

  • 1Department of Computer Science and Engineering, Indraprastha Institute of Information Technology Delhi (IIIT-Delhi), New Delhi, India
  • 2Department of Computer Science and Engineering, Indian Institute of Technology (IIT) Jodhpur, Jodhpur, India
  • 3Department of Computer Science and Electrical Engineering, Texas A&M University, KIngsville, TX, United States

Presentation attack detection (PAD) algorithms have become an integral requirement for the secure usage of face recognition systems. As face recognition algorithms and applications increase from constrained to unconstrained environments and in multispectral scenarios, presentation attack detection algorithms must also increase their scope and effectiveness. It is important to realize that the PAD algorithms are not only effective for one environment or condition but rather be generalizable to a multitude of variabilities that are presented to a face recognition algorithm. With this motivation, as the first contribution, the article presents a unified PAD algorithm for different kinds of attacks such as printed photos, a replay of video, 3D masks, silicone masks, and wax faces. The proposed algorithm utilizes a combination of wavelet decomposed raw input images from sensor and face region data to detect whether the input image is bonafide or attacked. The second contribution of the article is the collection of a large presentation attack database in the NIR spectrum, containing images from individuals of two ethnicities. The database contains 500 print attack videos which comprise approximately 1,00,000 frames collectively in the NIR spectrum. Extensive evaluation of the algorithm on NIR images as well as visible spectrum images obtained from existing benchmark databases shows that the proposed algorithm yields state-of-the-art results and surpassed several complex and state-of-the-art algorithms. For instance, on benchmark datasets, namely CASIA-FASD, Replay-Attack, and MSU-MFSD, the proposed algorithm achieves a maximum error of 0.92% which is significantly lower than state-of-the-art attack detection algorithms.

1. Introduction

Face recognition algorithms have been gaining more interest than ever, both for their increasing usage (Guo and Zhang, 2019; Sepas-Moghaddam et al., 2019) and their limitations (Agarwal et al., 2016, 2019c; Siddiqui et al., 2016; Cavazos et al., 2019; Marcel et al., 2019; Mukudi and Hills, 2019; Wu et al., 2019; Ghosh et al., 2020; Singh et al., 2020). While researchers are attempting to make the face recognition algorithms generalizable to unseen scenarios, it is important that the face presentation attack detection (PAD) algorithms are also generalizable and inclusive (Agarwal et al., 2019a,b; Sanghvi et al., 2021). Existing research in face presentation attack detection has primarily focused on detecting different kinds of attacks captured in the visible spectrum. While the algorithms have received nearly perfect classification performance on individual attacks, the focus is on designing unified algorithms for different kinds of presentation attacks in visible images (Liu et al., 2019b; Mehta et al., 2019).

The usage of face recognition algorithms in the near infrared spectrum is also increasing. With surveillance cameras operating in both visible (VIS) and near infrared (NIR) images, the PAD algorithms must be developed for NIR spectrum face images as well. It is well observed that the illustration of attacks in both spectra is quite different, and hence, attack detection in images of different spectrums require specialized algorithms. Realizing this, recently researchers have started working toward PAD in NIR spectrum images as well (Agarwal et al., 2017; Zhang et al., 2020). This research focuses on extending the usability and efficiency of presentation attack detection algorithms in multiple spectrums and ethnicity variations.

1.1. Related Study

In this section, we present a brief overview of the existing algorithms in multi-spectrum (visible + near infrared) presentation attack detection. Face presentation attack detection in other than the visible spectrum is still in nascent stages. Pavlidis and Symosek (Pavlidis and Symosek, 2000) presented an algorithm to detect disguised faces system using images captured in the NIR spectrum. Yi et al. (2014) performed experiments with a print attack by printing the photo of 100 clients on coarse paper in both VIS and NIR spectrum. However, the database was not made publicly available. Later, Chingovska et al. (2016) and Raghavendra et al. (2017) also analyzed the vulnerability of face recognition systems in the NIR spectrum. They demonstrated that face recognition systems working in the NIR spectrum are also susceptible to presentation attacks such as replay and print. Therefore, effective presentation attack detection algorithms are required to protect the surveillance system operating in the NIR spectrum. Chingovska et al. (2016) also released the first publicly available VIS and NIR presentation attack database (MSSPOOF). The database comprises images from 21 subjects. Three photos of each client in the VIS and NIR spectrum are selected and printed using the black and white printer. The database contains 70 real access images and 144 attack images for each client. They demonstrated more than 88% Spoof False Accept Rate (SFAR) using the NIR attack on the NIR face recognition system. However, no counter-presentation attack algorithm is presented in this article.

Later, Raghavendra et al. (2016) proposed a presentation attack detection algorithm on the MSSPOOF (multispectral spoof) database using the combination of Laplacian pyramid and Fourier transform. They reported an Average Classification Error Rate (ACER) of 2.1% and 0.74% on the VIS and NIR spectrum, respectively. Raghavendra et al. (2017) prepared the extended multispectral presentation attack database (EMSPAD) from 50 subjects and attack samples were prepared by printing the images from two different printers. The database is captured in seven different spectral bands ranging from 425 to 930 nm. In total, the database contains 7, 000 print attack images and 3, 500 real access images from 50 subjects. A vulnerability of the face recognition system using printed attack samples is shown with perfect SFAR in 680 nm spectra and more than 98% SFAR in 930 nm spectra. Agarwal et al. (2017) have proposed the first large-scale multi-spectrum face presentation attack database consisting of both 2D and 3D attack instruments. The real and attack videos are captured in several constrained and unconstrained environments and in three imaging spectrums namely visible (VIS), near infrared (NIR), and thermal. Liu and Kumar (2018) proposed different CNN architectures to detect the presentation attack using multi-spectral face images. They prepared the presentation attack database from 13 masked subjects and 9 real subjects. The limitation of the study is the unavailability of the database. Sun et al. (2018) proposed consistent measure based presentation attack detection in multi-spectral imaging. Bhattacharjee et al. (2018) have shown the vulnerability of VIS, NIR, and Thermal face recognition using deep CNN models under custom silicone mask attacks. George et al. (2019) presented a new database along with multi-channel CNN for face PAD.

Heusch et al. (2020) have presented a database in multiple spectrums to effectively present a study on the effect of SWIR imaging on presentation attack detection. Zhang et al. (2020) have a multi-spectral database comprising 2D print attacks. Along with that the squeeze and excitation network using ResNet-18 as a backbone network is also proposed to counter presentation attacks. Li et al. (2020) extended the database by incorporating multiple attacks including a 3D print and silica gel mask. The other popular face presentation attack databases are SiW (Liu et al., 2018), SiW-M (Liu et al., 2019b), and OULU-NPU (Boulkenafet et al., 2017b); however, all are captured in the visible spectrum. The details of the existing face presentation studies can also be found in the comprehensive evaluation (Jia et al., 2020b) and handbook (Bhattacharjee et al., 2019; Marcel et al., 2019).

1.2. Research Contributions

The literature review shows that there are a few publicly available databases in the NIR spectrum and a significant lack of research on automated algorithms for presentation attack detection in the NIR spectrum images and algorithms that can detect PAD in multiple spectrums. Inspired by these, there are two primary contributions of this research. The first contribution is that we have prepared a large video-based attack database in the NIR spectrum from more than 340 subjects of two different ethnicities: Indian and Chinese. The second contribution of this article is a state-of-the-art face presentation attack detection algorithm for different kinds of attacks in the NIR + VIS spectrum. The efficacy of the proposed algorithm is demonstrated through extensive experiments. The proposed algorithm not only yields effective performance with spectrum variations but is also generalized against attack instruments, ethnicities, and databases. Experimental evaluation of multiple databases shows that the proposed algorithm surpasses several deep learning, non-deep learning, and state-of-the-art algorithms by a significant margin.

The article is organized as follows. In the next section, the proposed large-scale face presentation attack dataset is described along with the experimental protocol used to perform the face PAD task. The section also describes the impact of the proposed dataset in evaluating the sensitivity of the face recognition algorithms. Section 3 describes the proposed unified face PAD algorithm effective in handling not only individual spectrums such as VIS and NIR but also a combination of them. In Section 4, the effectiveness of the proposed algorithm to handle multiple spectrums is evaluated using several existing multi-spectrum datasets including the proposed dataset. Furthermore, to benchmark, our study is similar to existing studies, in Section 5 extensive experimental evaluation has been performed on the standard visible (VIS) spectrum datasets. The extensive comparison on each dataset with state-of-the-art algorithms are presented and showcase that the proposed algorithm outperforms them with a significant margin. Finally, the conclusion and future research directions are briefly described.

2. Proposed Spoof-in-NIR Database

As discussed in the previous section, there are few databases in the NIR spectrum. Therefore, to promote the research in this emerging field, we first prepared a large video PAD database in the NIR spectrum, termed the Spoof-in-NIR database. In this section, we present the details of the proposed NIR presentation attack database. The NIR database contains face images from two completely different ethnicities: Indian and Chinese. The database contains the attack videos captured using the printed photograph of 400 genuine users. The database will be made publicly available to the research community1.

2.1. Camera Setup

To collect the database in the NIR spectrum, a camera is mounted on the tripod, and subjects are asked to stand at a distance of approximately 1 m and look into the camera. To ensure that the videos are captured in a relatively uncontrolled scenario, no other special instructions are given to the subjects. The database is captured in two different environments and background conditions: one inside and the other outside the building at night time. To ensure that the videos are only captured in the NIR spectrum, a visible cut filter is placed in front of the camera. GO-5000-USB camera2 is used to capture the videos at the frame resolution of 2, 048 × 2, 560. Frames are captured at the rate of 20 fps and stored as raw pixels so that the quality of images is not degraded because of compression.

2.2. Indian Face Presentation Attack Database

Each ethnicity subset comprises two sets: bonafide and attacked. The bonafide/real videos of the Indian database are captured from 152 subjects at night time using a NIR camera. To provide variability in the data, videos are captured at two different locations and comprise variations in background and illumination conditions. The subjects also perform natural motions such as eye blinking and head movement.

To capture the attacked videos, frontal images of all the subjects are first captured using a DSLR camera in day time constrained environment. A black and white printout of frontal images of 150 subjects is taken using an HP color printer and videos of these images are captured to prepare the attacked video subset. Attack videos of the Indian subset are captured in two different sessions with illumination variations. Attack and real videos are captured for a duration of 12 s to exhibit a real world surveillance scenario.

In total, 300 attacks and 152 real videos are collected as part of this Indian database. The resolution of the real and attack frame is 2, 048 × 2, 560. Face detection is performed using the Viola-Jones face detector (Viola and Jones, 2004). Characteristics of the database are given in Table 1 and samples are shown in Figure 1A.

TABLE 1
www.frontiersin.org

Table 1. Characteristics of the proposed NIR face presentation attack database.

FIGURE 1
www.frontiersin.org

Figure 1. Sample face images from the proposed Spoof-in-NIR database. Images are shown from (A) Indian ethnicity and (B) Chinese ethnicity.

2.3. Chinese Face Presentation Attack Database

To prepare the attack database with Chinese ethnicities, real visible spectrum frontal images of 98 subjects are randomly selected from the NIR-VIS 2.0 database (Li et al., 2013). Similar to the Indian database, these images are also printed on A4 paper using HP color printer. These prints are then placed on a fixed medium to capture the attack videos. A total of 98 videos comprising 7, 840 images form the NIR attacked subset while real NIR samples from 725 individuals acquired from the CASIA NIR-VIS2.0 database (Li et al., 2007) comprise the bonafide Chinese subset. In total, Chinese database contains 12, 469 real face images and 7, 799 attack face images. Characteristics of the database are given in Table 1 and sample images are shown in Figure 1B.

2.4. Experimental Protocol

To facilitate benchmarking the performance of different algorithms on this database, we propose a protocol as summarized in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Protocol for the proposed Spoof-in-NIR database experiments.

Indian NIR dataset is divided into subject independent 15 folds, where at a time one fold is used for training the linear SVM classifier for attack detection, and the remaining folds are used for evaluating the classifier. In the Indian subset, attack videos are captured in two different sessions; therefore, the results for both the sessions are calculated separately. To report the results of a particular session, attack videos captured in that session are used. Similarly, the Chinese NIR database is divided into five random folds. Due to the unavailability of the videos in the real subset of the Chinese NIR dataset, only frame based results are calculated. For the Indian NIR database, both video based and frame based protocols are designed. In video-based protocol, every video is classified as real or attack, and in the frame-based protocol, an individual entity of a video (frame) is classified as real or attack. The protocol will help evaluate the algorithm trained in limited and constrained data settings while testing large-scale data coming from multiple folds. In place of single fold train-test evaluation, multi-folds ensure the extensive assessment and generalizability of the algorithm.

2.5. Vulnerability of Face Recognition Against Attack

Researchers have demonstrated that several publicly available commercial face recognition systems are prone to presentation attacks (Wen et al., 2015; Agarwal et al., 2017). To show the effect of the proposed NIR face database on the Commercial-Off-the-Shelf (COTS) systems, a face identification experiment is performed. The gallery images are single frontal images of each subject. For instance, the gallery for the Indian NIR attack database comprises 150 images, one for each subject. Each attack video is used as a probe to perform the identification using FaceVACS3. It is important to observe that with all the attacked images as a probe, COTS yields the rank-1 identification accuracy of 100% for the Chinese NIR subset while Indian NIR attack videos show more than 98% identification accuracy at rank-1. The face identification experiment thus shows the vulnerability of the existing face recognition system against the attack in the NIR spectrum.

3. Proposed Presentation Attack Detection Algorithm

Face presentation attack detection algorithms proposed in the literature are primarily based on encoding texture measures and image artifacts such as moiŕe pattern (Patel et al., 2016), local binary patterns (Määttä et al., 2011; Boulkenafet et al., 2016; Ramachandra and Busch, 2017), and more recently, deep learning algorithms have been proposed (Chen et al., 2019; Tu et al., 2019; Ma et al., 2020). A recent study (Jia et al., 2020b) shows that hand-crafted image feature-based algorithms have the potential of handling multiple challenging presentation attacks while being computationally feasible. Therefore, in this article, we present a presentation attack detection algorithm based on combining the image features extracted from global as well as local facial regions. Raw images provide a global view of the real and attack data. This global information can help in extracting discriminative information such as foreground-background inconsistency and attacking medium boundary, while local facial regions may help in extracting textural information/artifact of the face. We hypothesize that since the attack images are mostly recaptured from the camera, they can be different in high-frequency content from its counterpart i.e., real face. The proposed presentation attack detection algorithm encodes these assertions to differentiate between real and attacked images. Figure 2 illustrates the steps involved in the proposed algorithm. To highlight the low and high frequency of discrimination, the first step of the algorithm involves applying the wavelet decomposition to the input image/frame. Wavelet provides the multi-resolution decomposition of the frame and highlights the edges in multiple directions: horizontal, vertical, and diagonal (Fowler, 2005). Discrete wavelet transform (DWT) downsamples the subbands to M/2 × N/2, where M and N are the height and width of the image respectively. Since downsampling leads to loss of information, we apply redundant discrete wavelet transform (RDWT). RDWT preserves the image size by creating subbands of the same size as the input image, thereby generating overcomplete representations.

FIGURE 2
www.frontiersin.org

Figure 2. Illustrating the proposed presentation attack detection pipeline.

As mentioned earlier, the effect of presentation attack may be visible in either facial regions or even as abnormalities concerning foreground and background regions. Therefore, wavelet decomposition is applied in two steps. In the first step, RDWT decomposition is applied on the complete input image without face detection, while the second step involves first applying face detection followed by tessellating the facial region into nine non-overlapping facial patches.

Global information, which is directly computed from the raw images, is computed without tessellating the frames into multiple blocks. To compute the information from the local texture, the face region is first divided into multiple blocks, and then each block is decomposed using a wavelet filter. The patches generated are of size 32 × 32. The textural features are then extracted by applying Haralick features (Haralick and Shanmugam, 1973) on the RDWT decomposed subbands, which can encode the image distortion-based information present in the recaptured images such as intensity distribution and homogeneity.

The basic building block of Haralick features is the Co-occurrence Matrix, which contains the information of counts of how many times a certain pixel value is in the neighborhood of other possible pixel values. The Haralick features used in this research are listed below:

li,j=Ci,j/S

where C is the Co-occurrence Matrix, and S is the sum of all elements of C. li, j is the likelihood of occurrence of the pixel value i in the neighborhood of pixel value j. Ng is the number of gray levels in matrix C.

• Global Homogeneity:

h1=r=1Ngc=1Ngl(r,c)2    (1)

• Local Homogeneity:

h2=r=1Ngc=1Ng11+(r-c)2l(r,c)    (2)

• Correlation Measures:

h3=ERC-ERC1max{ER,EC}    (3)
h4=(1-exp{-2(ERC2-ERC)})12    (4)

where,

ERC=-r=1Ngc=1Ngl(r,c)log[l(r,c)]
ERC1=-r=1Ngc=1Ngl(r,c)log{lr(r)lc(c)}
ERC2=-r=1Ngc=1Nglr(r)lc(c)log{lr(r)lc(c)}

• Average:

h5=r=22Ngrlr+c(r)    (5)

• Intensity Variation:

h6=n=0Ng-1n2{r=1Ngc=1Ngl(r,c)},|r-c|=n    (6)

• Pixel Dependency:

h7=r=1Ngc=1Ng(rc)l(r,c)-μrμcσrσc    (7)

• Variance

h8=r=1Ngc=1Ng(r-μ)2l(r,c)    (8)

• Total Variance:

h9=r=22Ng(r-h10)2lr+c(r)    (9)

• Total Entropy:

h10=-r=22Nglr+c(r)log[lr+c(r)]    (10)

• Entropy:

h11=-r=1Ngc=1Ngl(r,c)log[l(r,c)]    (11)

• Variance Difference:

h12=r=0Ng-1r2lr-c(r)    (12)

• Entropy Difference:

h13=-r=0Ng-1lr-c(r)log[lr-c(r)]    (13)

Here, μr, μc, σr, and σc are the mean and SDs of lr and lc, respectively. r and c represent the row and column index in the co-occurrence matrix. lr+c(r) is the likelihood of co-occurrence matrix coordinates summing to r+c, and ER, and EC are the entropies of lr and lc.

The process described above is repeated for each channel in the RGB image. While the Haralick features are used to encode the textural information, RDWT provides the high-frequency information presented in multiple orientations. For classification, a linear Support Vector Machine (SVM) (Vapnik, 2013) classifier is used based on its popularity in the PAD research (Ramachandra and Busch, 2017; Jia et al., 2020a). Two different SVM classifiers are trained: one for the features extracted from the raw sensor images and the second for the features extracted from the face region. The final score of a testing frame is calculated as the weighted sum of the scores obtained from two different trained SVM. To summarize, the steps involved in the proposed algorithm are discussed below:

1. The input image is decomposed into individual RGB channels

2. Each channel is then decomposed into four wavelet sub-bands using Redundant Discrete Wavelet Transform (RDWT) (Fowler, 2005),

3. Haralick texture features are computed over each wavelet sub-bands and the original image without decomposition,

4. Haralick features obtained from each subband are concatenated and input to a linear Support Vector Machine (SVM) classifier for classification,

5. The face region is cropped using the eye coordinates obtained using the Viola-Jones face detector,

6. Face region is divided into nine non-overlapping blocks,

7. Each face block is decomposed into four wavelet sub-bands using Redundant Discrete Wavelet Transform,

8. Haralick texture features are computed over each wavelet sub-bands and the original face block without decomposition,

9. Haralick features obtained from each subband and original face are concatenated and input to a linear Support Vector Machine (SVM) classifier for classification,

10. The scores obtained from SVM in Steps 4 and 9 are fused using weighted sum rule fusion and then thresholded for classification.

4. Results on Multispectral PAD Databases

The performance of the proposed algorithm is computed on the proposed Spoof-in-NIR database along with the publicly available multispectral MSSPOOF database (Chingovska et al., 2016) and the CASIA-SURF database (Zhang et al., 2020). The results are reported in terms of Bonafide (real) Presentation Classification Error Rate (BPCER), Attack Presentation Classification Error Rate (APCER), and Average Classification Error Rate (ACER) also known as Half Total Error Rate (HTER). BPCER is defined as the rate of the bonafide (real) samples classified as attack samples. APCER is the proportion of the attack samples classified as bonafide (real) samples. The results are also reported using Equal Error Rate (EER).

APCER=1-i=1Aξ|A|
BPCER=i=1Bς|B|
ACER=(APCER+BPCER)/2

where, |A| and |B| is the total number of presentation attack and real images, respectively. ξ and ς is 0 if an attack image is classified as real (bonafide), else 1.

FAR=FalsePositive|A|
FRR=FalseNegative|B|
HTER=(FAR+FRR)/2

HTER is an average of FRR and FAR. EER is a specific value of HTER at which FAR is equal to FRR.

First, the results of the proposed algorithm on existing MSSPOOF and CASIA-SURF databases along with comparisons to state-of-the-art (SOTA) algorithms are reported. Later, the experiments performed on the proposed Spoof-in-NIR database are described. The comparison of the proposed algorithm with SOTA algorithms highlights the efficacy of PAD in multiple spectrums.

4.1. Results on MSSPOOF Database

Multispectral spoof database contains images with print attack performed in four different ways: 1) original image is captured in the visible spectrum, while the visible image is recaptured using visible and near infrared spectrum camera, and 2) original image is captured in near-infrared spectrum while NIR image is recaptured back using visible and near infrared filter camera. The database contains a total of 630 real images in visible and 624 images in near infrared spectrums pertaining to 21 subjects, in 7 different environmental conditions. The database is divided into three disjoint sets: train, dev, and test. Raghavendra et al. (2016) proposed two different protocols on the MSSPOOF database:

1. Individual spectrum: NIR and VIS, individual spectrum data in training and testing.

2. Combined spectrum: both spectrum data in training and testing.

We have reported the results using these two protocols along with the experiments defined in this paper to fully utilize the characteristics of the database. In the individual spectrum experiments: the presentation attack detection SVM model is trained using individual spectrum data such as for visible spectrum experiments. The partition belonging to the visible spectrum is used for training and evaluating the classifier. In the combined spectrum experiment, the train set belonging to both the spectra is used to learn the classifier, and similarly, for evaluation, test sets belonging to both spectra are combined.

4.1.1. Results of Individual Spectrum Experiments

The results for individual spectrum experiments are summarized in Table 3. The proposed algorithm achieves 0% ACER and BPCER in both the spectrums which shows the consistency of the algorithm across spectrums. The second best performing algorithm is by Raghavendra et al. (2016) (LaMTiF) and it achieves 0.74% and 2.08% ACER in NIR and VIS spectrum, respectively. LaMTiF uses a combination of Laplacian pyramid based decomposition to extract the high-frequency information and Short Term Fourier Transform (STFT) for time and frequency features. The limitation of the LaMTiF algorithm is the high error rate for attack detection. Table 3 also shows the results of several other texture based algorithms and it can be observed that the algorithms are generally ineffective in detecting either the bonafide presentation or attack presentation. The proposed algorithm is robust to both kinds of data, which is desired and required for real-world presentation attack detection algorithms integrated with face recognition systems. The misclassification of bonafide data as attack data can frustrate the genuine user because of the need to give face data again and again for recognition. At the same time allowing attack data as bonafide data can cause serious harm to the face recognition system. The proposed algorithm yields the best EER (i.e., 0%) on the MSSPOOF database.

TABLE 3
www.frontiersin.org

Table 3. Results (%) on individual and combined spectrum set of MSSPOOF database.

4.1.2. Results on Combined Spectrum Experiments

In the combined spectrum experiment, the data belonging to both spectrums are utilized to make a joint decision. To learn the presentation attack detection model, the training set given in the database for both the spectrums is used and evaluation is performed on the test set of both the spectrums collectively. Table 3 shows the results of the proposed and existing algorithms on the combined set.

The proposed algorithm achieves 0.0% ACER on the combined spectrum database which is significantly better than existing algorithms. The results show that the proposed algorithm is robust towards different kinds of data. Similar to individual spectrum results, the combined spectrum shows the ineffectiveness of the existing algorithms in detecting attack or bonafide data. The GLCM features yield the perfect error rate in detecting the presentation attack but rejecting the bonafide almost all the time. Similarly, the second best algorithm yields 3% APCER. The proposed algorithm yields 1.39% EER on the joint spectrum dataset.

4.2. Experiments on CASIA-SURF Database

The recently proposed CASIA-SURF database (Zhang et al., 2020) is one of the most extensive databases for face presentation attack detection problems both in terms of modalities and subjects. The database consists of 21, 000 videos of 1,000 subjects in RGB (color), IR, and depth modalities. For each subject, one real video is captured while six fake videos are captured using eye, mouth, and nose regions cut from the flat and curved printed face. For example, in attacks 1 and 2, the eye region is cut from the flat and curved face photo. Similarly, in other attacks, either the eyes and nose or eyes, nose, and mouth, all portion is cut from flat and curved face photo. The color, depth, and IR videos are acquired using the Intel RealSense SR300 camera. The real faces are first printed out using an A4 color printer and later used by the attackers while exhibiting real life motions such as turning left or right, moving up or down, and walk-in or away from the camera. The database is divided into training, validation, and testing set and contains 148,089, 48,789, and 295,644 cropped images, respectively.

In this research, we have used the pre-defined protocol for evaluation, and the results are reported on fused modalities. The baseline algorithm (Zhang et al., 2019) consists of the ResNet-18 model as a backbone model where the first three ResNet blocks are used for feature extraction. The features from each modality, i.e., color, depth, and IR, are then fused using a squeeze and excitation module. In the end, two blocks of ResNet are used for discriminative features learning, followed by global average pooling. The whole pipeline is trained using a softmax classifier. The comparison of the proposed algorithm with the baseline algorithm in terms of error rates is given in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Error rates (%) of the proposed and baseline algorithms on the CASIA-SURF database (Zhang et al., 2019).

The scores computed over different modalities are fused using a weighted sum. The weight parameter across different modalities is learned, which yields the lowest EER on the validation set. The fusion of all three modalities, i.e., color, depth, and IR, outperforms the baseline algorithm by 37.5% (i.e., reduces from 2.4% to 1.5%) in terms of ACER. The APCER of the baseline algorithm (i.e., 3.8%) is more than two times that of the proposed algorithm (i.e., 1.6%). The proposed fusion of color and IR with depth modality surpasses the baseline algorithm in identifying bonafide (i.e., real) images by at least 66%. As shown in Figure 3, the APCER of the proposed algorithm is significantly better than the recently proposed face PAD algorithm by Zhang et al. (2020). For example, when the fusion of color and IR data is performed, the APCER of the proposed and existing algorithm (Zhang et al., 2020) is 10.1% and 36.5%, respectively. The fusion of all modalities in the proposed algorithm yields 1.6% APCER, whereas, the existing algorithm yields 1.9% APCER. These superior performances of the proposed algorithm on one of the largest multi-modal presentation attack databases establish the efficacy of the proposed algorithm in identifying physical fake face data.

FIGURE 3
www.frontiersin.org

Figure 3. Comparing Attack Presentation Classification Error Rate (APCER) (%) of the proposed algorithm with Zhang et al. (2020).

4.3. Experiments on Proposed Spoof-in-NIR Database

The proposed Spoof-in-NIR database contains images and videos of subjects from two different ethnicities: Indian and Chinese. Indian NIR database contains 152 real videos and 300 attack videos collected in two sessions. Chinese NIR database contains 12, 469 real frames taken from CASIA VIS-NIR 2.0 (Li et al., 2013) and 7, 799 attack frames. According to the protocol described in Section 2.4, the experiments are performed on the proposed Spoof-in-NIR database. To make our approach in-line with the recent literature (Zhang et al., 2019, 2020; Jia et al., 2020b) which has utilized the ResNet as a backbone or to make the comparison, we have also performed a comparison with the ResNet-18 model (He et al., 2016) pre-trained on ImageNet (Deng et al., 2009). The model is fine-tuned for PAD for 50 epochs using the Adam optimizer and adaptive learning rate, where the initial value is set to 0.0001.

The results of the proposed algorithm on the Spoof-in-NIR database are summarized in Table 5. The proposed algorithm achieves 4.3% and 0.8% EER on sessions 1 and 2 of the Indian NIR spectrum dataset, respectively, for video-based experiments. Similarly, for the frame-based evaluation, an EER of 19.7% and 23.2% is achieved for sessions 1 and 2, respectively. On the Chinese NIR dataset, the proposed algorithm yields 20.7% EER using a frame-based experiment. The reported results show that the classification of an individual frame is difficult compared to video, where the information from multiple frames helps improve the results. Apart from that, the proposed algorithm surpasses the deep CNN architecture, i.e., ResNet, by a significant margin. Due to the imbalanced nature of images/videos of real and spoof classes in the proposed dataset especially in the Chinese subset, we have also used another evaluation metric namely the weighted F-1 score. The weighted F-1 score on the Chinese PAD subset of the proposed and ResNet-18 architecture is 0.8230 and 0.6907, respectively. On the Indian subset, the average weighted F-1 values for video-based classification are 0.9687 and 0.8675 which are obtained from the proposed and ResNet algorithm, respectively. Comparing the performance with MSSPOOF (Chingovska et al., 2016) and CASIA-SURF (Zhang et al., 2020) databases further show the proposed database is more challenging. The availability of this database to the research community will further improve the state-of-the-art presentation attack detection in multiple spectrums.

TABLE 5
www.frontiersin.org

Table 5. Video and frame based EER (μ±σ)% on the proposed NIR print attack database.

5. Experiments on Existing VIS Spectrum Databases

To further demonstrate the effectiveness of the second contribution of this article i.e., the development of a robust algorithm across different attacks, additional experiments are performed on the existing benchmark databases in the VIS spectrum. The performance of the proposed algorithm is evaluated on the CASIA-Face Anti-Spoofing Database (FASD) (Zhang et al., 2012), Replay-Attack (Chingovska et al., 2012), MSU-MFSD (Wen et al., 2015), 3D Mask Attack Database (3DMAD) (Erdogmus and Marcel, 2013), MSU USSA database (Patel et al., 2016), WMCA (George et al., 2019), SiW-M (Liu et al., 2019b), Silicone Mask Attack Database (SMAD) (Manjani et al., 2017), and 3D wax figure face database (WFFD) (Jia et al., 2019). These databases cover a wide spectrum of attacks such as print, photo, a replay of video, 3D hard resin masks, and the most challenging silicone mask. Table 6 summarizes the characteristics of these databases and a brief description of each is provided below.

TABLE 6
www.frontiersin.org

Table 6. Characteristics of the existing VIS spectrum attack database used in this research.

CASIA-FASD database (Zhang et al., 2012) contains three different kinds of attacks: cut photo (eye portions are cut to perform the eye blink), warped photo (to make it cylindrical as a real face), and replay of a video. It contains the videos in three different image qualities: low, normal, and high. Replay-Attack database (Chingovska et al., 2012) is captured in controlled and adverse environments. In the controlled environment, the background was kept fixed and the fluorescent lamp was used for illumination. In the adverse environment, the background is random and natural light is the source of illumination. MSU-MFSD database (Wen et al., 2015) is captured from 35 subjects and is one of the mobile face attack databases. Real videos are captured from two different devices: a built-in camera of a MacBook Air 13-inch laptop and a front facing camera of a Google Nexus 5 Android phone. To capture the attack, two different high-resolution cameras are used: a Canon 550D single-lens reflex camera and an iPhone 5S back facing camera.

CASIA-FASD, Replay-Attack, and MSU-MFSD databases are challenging but contain 2D attacks only. To assess the effectiveness of the algorithm on the 3D attack, the 3DMAD database (Erdogmus and Marcel, 2013) is also used in this article. Advancement in the 3D reconstruction and 3D printer makes the availability of 3D masks easier. These masks can be worn and can effectively hide the identity of the person in day-to-day life. These masks are hard to detect in comparison to wearing photo paper masks. 3DMAD database is captured from 17 subjects where each subject is wearing a different 3D mask. It is captured in three sessions, where the real access videos are captured in the first two sessions and the third session covers the 3D mask attack. For each subject ten real and five 3D attack videos are captured, with a total of 255 videos in the database. While the 3DMAD database is the challenging 3D mask attack database but it has some limitations. The masks used to prepare the database are hard resin masks that do not allow movements similar to the natural face. In the real world, some challenging cases are found where the robbers have used silicone masks to hide their identity from the surveillance cameras. These silicone masks are soft masks that can properly fit the face and can move with the face. Manjani et al. (2017) have prepared the Silicone Mask Attack Database (SMAD) which is currently the most challenging kind of attack to detect.

To tackle the limitations such as diversity in terms of background, illumination, and image quality, Patel et al. (2016) prepared one of the largest databases. Such a database is essential to obtain generalizable and robust anti-spoofing methods, particularly in face unlock scenarios on smartphones. To create such a database we selected 1,000 live subject images of celebrities from the Weakly Labeled Face Database4. The public set of the MSU USSA database for face anti-spoofing consists of 9, 360 images (out of which 1, 040 are real images and 8, 320 spoof attack images) of 1, 040 subjects. To perform the experiments standard database protocol of 5-fold cross-validation is performed.

On the grandtest protocol of the WMCA database, the proposed algorithm achieves 2.4% ACER; whereas, the existing algorithms such as FASNet (Lucena et al., 2017), DeepPixBis (George and Marcel, 2019), and MC-ResNetPAD (Parkin and Grinchuk, 2019) achieve 11.44%, 6.0%, and 2.6%, respectively. On the unseen protocols of SiW-M, the proposed algorithm achieves the average EER (%) and ACER (%) of 12.8% and 14.6%, respectively. The ACER (%) of Auxiliary (Liu et al., 2018), Deep Tree Network (Liu et al., 2019b), DeepPixBis (George and Marcel, 2019), and MCCNN (George and Marcel, 2020) is 23.6%, 16.8%, 19.6%, and 18.6%, respectively. The extensive results on these challenging databases further prove the effectiveness of the proposed algorithm.

Recently, a new modality of 3D attack is highlighted where wax figure faces are used as a possible adversary on face recognition systems (Jia et al., 2019). The authors have shown that state-of-the-art face recognition algorithms such as OpenFace (Amos et al., 2016) and Face++5 are vulnerable to wax figure faces. These spoof faces have achieved at least 92% Impostor Attack Presentation Match Rate (IAMPR) across multiple protocols. Therefore, the identification of wax faces from real faces is important and challenging because of properties similar to real faces. In this research, we have used two working conditions (protocols) provided by the authors. In the first protocol (Prot. 1), images captured under different recording devices and environments are used, whereas, in another protocol (Prot. 3), images captured in different and same recording devices and environments are combined. Protocol 1 consists of 600 trains, 200 development, and 440 test images; while, protocol 3 consists of 1,320 trains, 440 development, and 440 test images.

To compare the results with existing state-of-the-art algorithms, the original protocol of each database is followed and the results are reported both in terms of intra-database and cross database scenarios. Table 7 shows the comparison of the proposed algorithm with existing algorithms in video based attack detection. On one of the most challenging presentation attack i.e., silicone mask, the proposed algorithm outperforms the state-of-the-art performances (Manjani et al., 2017; Shao et al., 2019). EER and HTER of the proposed algorithm on video-based detection are 7.7% and 6.9% which is more than 37% and 47% lower, respectively. Similarly, on the CASIA-FASD database, the proposed algorithm gives the lowest EER value of 0.92%. Perfect EER on Replay-Attack, MSU-MFSD, and 3DMAD shows the robustness of the algorithm across different attacks and acquisition/attack devices. The proposed algorithm outperforms various deep learning algorithms (Manjani et al., 2017; Tu and Fang, 2017; Ma et al., 2020; Pinto et al., 2020; Song et al., 2020), multi cue fusion algorithms (Patel et al., 2016; Zhang et al., 2018), motion algorithms (Edmunds and Caplier, 2018), and texture algorithms (Boulkenafet et al., 2016; Peng et al., 2020). The ultra deep neural network proposed by Tu and Fang (Tu and Fang, 2017) combines a pre-trained deep residual network with Long Short Term Memory (LSTM) and yields an EER of 1.22% and 1.03% on CASIA-FASD and Replay-Attack databases, respectively. The EER of the proposed algorithm is at least 24% better than (Tu and Fang, 2017) on CASIA-FASD while 0% EER is achieved on the Replay-Attack database. As shown in Figure 4, the proposed algorithm outperforms several state-of-the-art presentation attack detection algorithms including the recent deep forest (Cai and Chen, 2019) algorithm on one of the largest MSU-USSA databases. The EER and standard deviation of the proposed, Deep Forest (Cai and Chen, 2019) and LBP + Color moment (Patel et al., 2016) algorithm are 1.1±0.3%, 1.6±0.6%, and 3.9±0.8%, respectively.

TABLE 7
www.frontiersin.org

Table 7. Comparison with existing results on the video based presentation attack detection.

FIGURE 4
www.frontiersin.org

Figure 4. Comparison with existing results including deep forest (Cai and Chen, 2019) on the MSU USSA database for presentation attack detection.

Table 8 shows that the proposed algorithm either achieves state-of-the-art or competitive results even for frame-based classification with all the challenging face spoofing databases. The EER value of 0.8%, 4.95%, and 0.0% is achieved on Replay-Attack, CASIA-FASD, and MSU-MFSD database respectively in frame based detection. HTER of 2.1% is achieved on the Replay-Attack database in the grand test attack scenario which is lower than various texture based algorithms (Boulkenafet et al., 2016; Peng et al., 2017). On the SMAD database, the proposed algorithm shows an improvement of more than 25% to 47% from the baseline performance (Manjani et al., 2017). The detection error rate using the proposed and existing algorithms on wax figure faces is reported in Table 9. Similar to other challenging attacks, the proposed algorithm outperforms several existing algorithms for wax face detection including hand-crafted and deep learning algorithms. The ACER of the proposed algorithm is 17.02% better than VGG-16 based wax face detection.

TABLE 8
www.frontiersin.org

Table 8. Comparison with SOTA results on the frame based presentation attack detection in terms of EEE (%) and HTER (%).

TABLE 9
www.frontiersin.org

Table 9. Wax figure face detection error rates (%) on the unconstrained (protocol 1) and real-world protocol (protocol 3) of WFFD database (Jia et al., 2019).

The proposed algorithm utilizes the weighted score fusion of the classifier trained using the features computed by tessellating the face region and full image input. Therefore, to study the importance of the individual region and segregate them into small regions, we have performed experiments on multiple databases. Interestingly, it is found that the features from the global region without input tessellation are useful in handling presentation attacks. For example, on MSU-MFSD and Replay-Attack, the raw input images show effectiveness in detecting the presentation attacks by achieving at most 0.4% EER. The fusion of face and full image further improves the detection performance. On CASIA-FASD, the fusion reduces the EER from 1.1% (face region) to 0.92% (face + raw image).

To further show the generalizability of the proposed PAD algorithm, cross database experiments are also performed, and the results are reported in Table 10. For video based scenarios, when the anti-spoofing algorithm is trained on the CASIA-FASD database, average HTER values of 26.7% and 35.3% are reported on the MSU-MFSD and Replay-Attack databases, respectively. The average HTER on CASIA-FASD and Replay-Attack is 23.7% and 32.2% respectively when the anti-spoofing model is trained using the MSU-MFSD database. When the model is learned using the Replay-Attack database and tested on each subset of CASIA-FASD and MSU-MFSD the average HTER value reported is 33.3% and 23.9%, respectively. The anti-spoof model trained using the Replay-Attack database shows better generalizability and may be attributed to the fact that it is captured in different illumination, devices, and background. The comparison of the proposed countermeasure with the existing anti-spoofing algorithms is shown in Table 10.

TABLE 10
www.frontiersin.org

Table 10. Comparison with existing results on the video based presentation attack detection under cross dataset setting.

The proposed algorithm outperforms the state-of-the-results when the countermeasure is trained using Replay-Attack and MSU-MFSD database. The proposed algorithm improves the HTER of the 2nd best performing algorithm (Peng et al., 2020) from 39.6% to 23.7% on the CASIA-FASD database when the model is trained on MSU-MFSD. Similarly, the proposed algorithm improves the performance on MSU-MFSD by 1.1% when the classifier is trained using Replay-Attack. Recently proposed algorithms based on deep CNN by Chen et al. (2019) and Sun et al. (2020b) yield an HTER value of 34.7% and 37.3% on the CASIA database when the model is trained on Replay-Attack, whereas the HTER of the proposed algorithm is at least 1.4% lower. The patch and depth CNN-v2 (Liu et al., 2019a) outperform the proposed algorithm in a cross-database setting; however, the significant drawbacks of the algorithm are the processing speed and high memory requirement. As claimed by the authors, their algorithm can process 1–2 frames per second, hence challenging for large scale implementation or deployment on mobile devices. Simultaneously, the proposed algorithm is highly computationally inexpensive both in terms of speed and memory requirement. Another strength of the proposed algorithm concerning patch and depth CNN-v2 (Liu et al., 2019a) is that it outperforms under the same database but unseen subjects train-test setting (reported in Table 7).

We have also performed the presentation attack detection using the individual color channel of RGB. To perform these experiments, random 30 frames from each video are first selected for feature extraction. The performance is reported on three challenging benchmark databases: CASIA-FASD, Replay-Attack, and MSU-MFSD. Results with 30 frames of a train and test video on the CASIA-FASD, Replay-Attack, and MSU-MFSD databases are reported in Figure 5A. R channel yields the lowest EER value of 3.3% on CASIA-FASD database which is lower than (Feng et al., 2016; Patel et al., 2016). On the Replay-Attack database, the lowest EER value of 0.1% is given by the B channel which is equal to the EER reported by Boulkenafet et al. (2017a). Boulkenafet et al. (2017a) have used all frames of a video while we have used only 30 frames for such comparable performance.

FIGURE 5
www.frontiersin.org

Figure 5. Ablation study of the proposed PAD algorithm in terms of its performance on the individual color channel of the images. Apart from that the practicality of the proposed algorithm to be deployed in resource constraint devices using computational speed. (A) Classification results with individual color channel and RGB; (B) Computational complexity of the PAD algorithms.

Efficacy of the proposed algorithm: In summary, the strengths of the proposed generalized PAD algorithm are listed below:

• The proposed algorithm can be implemented in real time. The feature extraction time on core i7@ 3.4GHz CPU machine with a Matlab environment is 0.1 frames per second (FPS) (Figure 5B). The huge deployment of face unlocking on mobile devices6 needs protection from presentation attacks. The proposed algorithm with such low computational time and memory requirement can also be implemented on mobile devices;

• The proposed algorithm outperforms various state-of-the-algorithms including 3D CNN (Li et al., 2018), SfSNet (Pinto et al., 2020), Multi-Regional CNN (Ma et al., 2020), deep dictionary (Manjani et al., 2017), and SE-ResNet18 (Wang et al., 2020), for a variety of presentation attacks including silicone mask attack (Manjani et al., 2017) and 2D attacks;

• The proposed algorithm is also able to handle new modality of 3D attacks i.e., wax faces (Jia et al., 2019). The average EER of the proposed algorithm is at least 37.4% and 12.6% lower than VGG-16 deep learning and multi-scale (M-scale) LBP texture features;

• The proposed algorithm is generalizable across imaging spectrum (VIS/NIR), attacks (2D/3D), acquisition devices (Mobile/High-def), and quality of images (Low/High).

6. Conclusion

Similar to the visible spectrum, face recognition in the near-infrared (NIR) spectrum is also vulnerable to presentation attacks. In the literature, there is very limited research on developing efficient and inclusive countermeasures for the attack in the NIR spectrum and designing a unified algorithm to design and evaluate the performance of PAD algorithms toward continuously evolving presentation attacks in multiple spectra. In this research, we contribute to this space by creating a large NIR PAD face database that comprises videos with different kinds of attacks on Indian and Chinese ethnicities. We next present a presentation attack detection algorithm for efficiently differentiating between bonafide and attacked images in the NIR spectrum. The generalizability of the proposed algorithm is demonstrated by evaluating the performance of 11 existing databases and comparing it with state-of-the-art results reported in the literature. It is observed that the proposed algorithm yields the best results on almost all the databases using all three metrics of APCER, BPCER, and EER. In cross-database evaluations, while the proposed algorithm yields the best results, the error rates are comparatively higher. In future study, we plan to improve the effectiveness of the algorithm so that the error rates can be further reduced without increasing the computational complexity.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

AA conducted the experiment(s) and analyzed the results. All authors conceived the proposed algorithm and experiment(s), reviewed and updated the article, and performed the proofreading of the paper.

Funding

AA was partially supported by the Visvesvaraya Ph.D. Fellowship. MV is partially supported through the Swarnajayanti Fellowship by the Government of India. This research is partially supported through a research grant from the Ministry of Electronics and Information Technology, Government of India.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

Agarwal, A., Sehwag, A., Singh, R., and Vatsa, M. (2019a). “Deceiving face presentation attack detection via image transforms,” in IEEE BigMM (Singapore: IEEE), 373–382.

Google Scholar

Agarwal, A., Sehwag, A., Vatsa, M., and Singh, R. (2019b). “Deceiving the protector: Fooling face presentation attack detection algorithms,” in IEEE ICB (Crete: IEEE), 1–6.

Google Scholar

Agarwal, A., Singh, R., and Vatsa, M. (2016). “Face anti-spoofing using haralick features,” in IEEE BTAS (Niagara Falls, NY: IEEE), 1–6.

Google Scholar

Agarwal, A., Vatsa, M., and Singh, R. (2019c). “CHIF: convoluted histogram image features for detecting silicone mask based face presentation attack,” in IEEE BTAS (Tampa, FL: IEEE), 1–5.

Google Scholar

Agarwal, A., Yadav, D., Kohli, N., Singh, R., Vatsa, M., and Noore, A. (2017). “Face presentation attack with latex masks in multispectral videos,” in IEEE CVPRW (Honolulu, HI: IEEE), 81–89.

Google Scholar

Amos, B., Ludwiczuk, B., Satyanarayanan, M., et al. (2016). Openface: A General-Purpose Face Recognition Library With Mobile Applications. Pittsburgh, PA: CMU School of Computer Science.

Google Scholar

Anjos, A., and Marcel, S. (2011). “Counter-measures to photo attacks in face recognition: a public database and a baseline,” in IEEE IJCB (Washington, DC: IEEE), 1–7.

Google Scholar

Atoum, Y., Liu, Y., Jourabloo, A., and Liu, X. (2017). “Face anti-spoofing using patch and depth-based cnns,” in IEEE IJCB (Denver, CO: IEEE), 319–328.

Google Scholar

Azeddine, B., Kamal, E. A., Abdelkrim, O., Djamel, S., and Mourad, O. (2017). Face antispoofing based on frame difference and multilevel representation. J. Electr. Imaging 26, 26-26–14. doi: 10.1117/1.JEI.26.4.043007

CrossRef Full Text | Google Scholar

Bhattacharjee, S., Mohammadi, A., Anjos, A., and Marcel, S. (2019). “Recent advances in face presentation attack detection,” in Handbook of Biometric Anti-Spoofing (Springer), 207–228.

Google Scholar

Bhattacharjee, S., Mohammadi, A., and Sébastien, M. (2018). “Spoofing deep face recognition with custom silicone masks,” in IEEE BTAS (Redondo Beach, CA: IEEE).

PubMed Abstract | Google Scholar

Boulkenafet, Z., Komulainen, J., and Hadid, A. (2015). “Face anti-spoofing based on color texture analysis,” in IEEE ICIP (Quebec City, QC: IEEE), 2636–2640.

Google Scholar

Boulkenafet, Z., Komulainen, J., and Hadid, A. (2016). Face spoofing detection using colour texture analysis. IEEE TIFS 11, 1818–1830. doi: 10.1109/TIFS.2016.2555286

CrossRef Full Text | Google Scholar

Boulkenafet, Z., Komulainen, J., and Hadid, A. (2017a). Face antispoofing using speeded-up robust features and fisher vector encoding. IEEE SPL 24, 141–145. doi: 10.1109/LSP.2016.2630740

CrossRef Full Text | Google Scholar

Boulkenafet, Z., Komulainen, J., Li, L., Feng, X., and Hadid, A. (2017b). “Oulu-npu: a mobile face presentation attack database with real-world variations,” in IEEE F&G (Washington, DC: IEEE), 612–618.

Google Scholar

Bresan, R., Pinto, A., Rocha, A., Beluzo, C., and Carvalho, T. (2019). Facespoof buster: a presentation attack detector based on intrinsic image properties and deep learning. arXiv preprint arXiv:1902.02845. doi: 10.48550/arXiv.1902.02845

CrossRef Full Text | Google Scholar

Cai, R., and Chen, C. (2019). Learning deep forest with multi-scale local binary pattern features for face anti-spoofing. arXiv preprint:1910.03850. doi: 10.48550/arXiv.1910.03850

CrossRef Full Text | Google Scholar

Cavazos, J. G., Phillips, P. J., Castillo, C. D., and O'Toole, A. J. (2019). Accuracy comparison across face recognition algorithms: where are we on measuring race bias? arXiv preprint arXiv:1912.07398. doi: 10.48550/arXiv.1912.07398

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, H., Hu, G., Lei, Z., Chen, Y., Robertson, N. M., and Z., et al. (2019). Attention-based two-stream convolutional networks for face spoofing detection. IEEE Trans. Inf. For. Security 15, 578–593. doi: 10.1109/TIFS.2019.2922241

CrossRef Full Text | Google Scholar

Chingovska, I., Anjos, A., and Marcel, S. (2012). “On the effectiveness of local binary patterns in face anti-spoofing,” in BIOSIG (Darmstadt: IEEE).

Google Scholar

Chingovska, I., Erdogmus, N., Anjos, A., and Marcel, S. (2016). “Face recognition systems under spoofing attacks,” in Face Recognition Across the Imaging Spectrum (Springer), 165–194.

Google Scholar

de Freitas Pereira, T., Anjos, A., Martino, J. M. D., and Marcel, S. (2013). “Can face anti-spoofing countermeasures work in a real world scenario?” in ICB (Madrid: IEEE), 1–8.

Google Scholar

de Freitas Pereira, T., Komulainen, J., Anjos, A., De Martino, J. M., Hadid, A., Pietikäinen, M., et al. (2014). Face liveness detection using dynamic texture. EURASIP JIVP 2014, 1–15. doi: 10.1186/1687-5281-2014-2

CrossRef Full Text | Google Scholar

Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. (2009). “Imagenet: a large-scale hierarchical image database,” in IEEE CVPR (Miami, FL: IEEE), 248–255.

Google Scholar

Edmunds, T., and Caplier, A. (2018). Motion-based countermeasure against photo and video spoofing attacks in face recognition. JVCIR 50, 314–332. doi: 10.1016/j.jvcir.2017.12.004

CrossRef Full Text | Google Scholar

Erdogmus, N., and Marcel, S. (2013). “Spoofing in 2d face recognition with 3d masks and anti-spoofing with kinect,” in IEEE BTAS (Arlington, VA: IEEE), 1–6.

Feng, L., Po, L.-M., Li, Y., Xu, X., Yuan, F., Cheung, T. C.-H., et al. (2016). Integration of image quality and motion cues for face anti-spoofing: a neural network approach. JVCIR 38, 451–460. doi: 10.1016/j.jvcir.2016.03.019

CrossRef Full Text | Google Scholar

Fowler, J. E. (2005). The redundant discrete wavelet transform and additive noise. IEEE SPL 12, 629–632. doi: 10.1109/LSP.2005.853048

CrossRef Full Text | Google Scholar

Galbally, J., Marcel, S., and Fierrez, J. (2014). Image quality assessment for fake biometric detection: application to iris, fingerprint, and face recognition. IEEE TIP 23, 710–724. doi: 10.1109/TIP.2013.2292332

PubMed Abstract | CrossRef Full Text | Google Scholar

George, A., and Marcel, S. (2019). “Deep pixel-wise binary supervision for face presentation attack detection,” in IEEE ICB (Crete: IEEE), 1–8.

Google Scholar

George, A., and Marcel, S. (2020). Learning one class representations for face presentation attack detection using multi-channel convolutional neural networks. IEEE TIFS 16, 361–375. doi: 10.1109/TIFS.2020.3013214

CrossRef Full Text | Google Scholar

George, A., Mostaani, Z., Geissenbuhler, D., Nikisins, O., Anjos, A., and Marcel, S. (2019). Biometric face presentation attack detection with multi-channel convolutional neural network. IEEE TIFS 15, 42–55. doi: 10.1109/TIFS.2019.2916652

CrossRef Full Text | Google Scholar

Ghosh, S., Singh, R., and Vatsa, M. (2020). Subclass heterogeneity aware loss for cross-spectral cross-resolution face recognition. IEEE TBIOM 2, 245–256. doi: 10.1109/TBIOM.2020.2984324

CrossRef Full Text | Google Scholar

Guo, G., and Zhang, N. (2019). A survey on deep learning based face recognition. Comput. Vision Image Understand. 189, 102805. doi: 10.1016/j.cviu.2019.102805

CrossRef Full Text | Google Scholar

Gustavo, B. d. S., João, P. P., and Aparecido, N. M. (2018). On the learning of deep local features for robust face spoofing detection. arXiv preprint arXiv:1806.07492v07491. doi: 10.1109/SIBGRAPI.2018.00040

CrossRef Full Text | Google Scholar

Haralick, R. M., and Shanmugam, K. (1973). Textural features for image classification. IEEE TSMC 3, 610–621. doi: 10.1109/TSMC.1973.4309314

CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in IEEE CVPR (Las Vegas, NV: IEEE), 770–778.

PubMed Abstract | Google Scholar

Heusch, G., George, A., Geissbühler, D., Mostaani, Z., and Marcel, S. (2020). Deep models and shortwave infrared information to detect face presentation attacks. IEEE TBIOM 2, 399–409. doi: 10.1109/TBIOM.2020.3010312

CrossRef Full Text | Google Scholar

Jia, S., Guo, G., and Xu, Z. (2020a). A survey on 3d mask presentation attack detection and countermeasures. Pattern Recog. 98, 107032. doi: 10.1016/j.patcog.2019.107032

CrossRef Full Text | Google Scholar

Jia, S., Guo, G., Xu, Z., and Wang, Q. (2020b). Face presentation attack detection in mobile scenarios: a comprehensive evaluation. Image Vis Comput. 93, 103826. doi: 10.1016/j.imavis.2019.11.004

CrossRef Full Text | Google Scholar

Jia, S., Hu, C., Guo, G., and Xu, Z. (2019). “A database for face presentation attack using wax figure faces,” in ICIAP (Trento: Springer), 39–47.

Google Scholar

Jourabloo, A., Liu, Y., and Liu, X. (2018). “Face de-spoofing: Anti-spoofing via noise modeling,” in ECCV (Munich), 290–306.

Google Scholar

Komulainen, J., Hadid, A., Pietikäinen, M., Anjos, A., and Marcel, S. (2013). “Complementary countermeasures for detecting scenic face spoofing attacks,” in ICB (Madrid: IEEE), 1–7.

Google Scholar

Li, A., Tan, Z., Li, X., Wan, J., Escalera, S., Guo, G., et al. (2020). Casia-surf cefa: a benchmark for multi-modal cross-ethnicity face anti-spoofing. arXiv preprint arXiv:2003.05136. doi: 10.48550/arXiv.2003.05136

CrossRef Full Text | Google Scholar

Li, H., He, P., Wang, S., Rocha, A., Jiang, X., and Kot, A. C. (2018). Learning generalized deep feature representation for face anti-spoofing. IEEE TIFS 13, 2639–2652. doi: 10.1109/TIFS.2018.2825949

CrossRef Full Text | Google Scholar

Li, S., Yi, D., Lei, Z., and Liao, S. (2013). “The casia nir-vis 2.0 face database,” in IEEE CVPRW (Portland, OR: IEEE), 348–353.

Google Scholar

Li, S. Z., Chu, R., Liao, S., and Zhang, L. (2007). Illumination invariant face recognition using near-infrared images. IEEE TPAMI 29, 627–639. doi: 10.1109/TPAMI.2007.1014

PubMed Abstract | CrossRef Full Text | Google Scholar

Lin, C., Liao, Z., Zhou, P., Hu, J., and Ni, B. (2018). “Live face verification with multiple instantialized local homographic parameterization,” in IJCAI (Stockholm), 814–820.

Google Scholar

Liu, J., and Kumar, A. (2018). “Detecting presentation attacks from 3d face masks under multispectral imaging,” in IEEE CVPRW (Salt Lake City, UT: IEEE), 47–52.

Google Scholar

Liu, Y., Jourabloo, A., and Liu, X. (2018). “Learning deep models for face anti-spoofing: Binary or auxiliary supervision,” in IEEE CVPR (Salt Lake City, UT: IEEE), 389–398.

Google Scholar

Liu, Y., Stehouwer, J., Jourabloo, A., Atoum, Y., and Liu, X. (2019a). “Presentation attack detection for face in mobile phones,” in Selfie Biometrics (Springer), 171–196.

PubMed Abstract | Google Scholar

Liu, Y., Stehouwer, J., Jourabloo, A., and Liu, X. (2019b). “Deep tree learning for zero-shot face anti-spoofing,” in IEEE CVPR (Long Beach, CA: IEEE), 4680–4689.

Google Scholar

Lucena, O., Junior, A., Moia, V., Souza, R., Valle, E., and Lotufo, R. (2017). “Transfer learning using convolutional neural networks for face anti-spoofing,” in ICIAR (Montreal, QC), 27–34.

Google Scholar

Ma, Y., Wu, L., Li, Z., and liu, F. (2020). A novel face presentation attack detection scheme based on multi-regional convolutional neural networks. Pattern Recogn. Lett. 131, 261–267. doi: 10.1016/j.patrec.2020.01.002

CrossRef Full Text | Google Scholar

Määttä, J., Hadid, A., and Pietikäinen, M. (2011). “Face spoofing detection from single images using micro-texture analysis,” in IEEE IJCB (Washington, DC: IEEE), 1–7.

Google Scholar

Manjani, I., Tariyal, S., Vatsa, M., Singh, R., and Majumdar, A. (2017). Detecting silicone mask-based presentation attack via deep dictionary learning. IEEE TIFS 12, 1713–1723. doi: 10.1109/TIFS.2017.2676720

CrossRef Full Text | Google Scholar

Marcel, S., Nixon, M. S., Fierrez, J., and Evans, N. (2019). Handbook of Biometric Anti-Spoofing: Presentation Attack Detection. Springer.

Google Scholar

Mehta, S., Uberoi, A., Agarwal, A., Vatsa, M., and Singh, R. (2019). “Crafting a panoptic face presentation attack detector,” in ICB (Crete), 1–6.

Google Scholar

Mukudi, P. B., and Hills, P. J. (2019). The combined influence of the own-age,-gender, and-ethnicity biases on face recognition. Acta Psychol. 194, 1–6. doi: 10.1016/j.actpsy.2019.01.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Pan, S., and Deravi, F. (2018). “Facial biometrie presentation attack detection using temporal texture co-occurrence,” in IEEE ISBA (Singapore: IEEE), 1–7.

Google Scholar

Parkin, A., and Grinchuk, O. (2019). “Recognizing multi-modal face spoofing with face recognition networks,” in IEEE/CVF CVPRW (Long Beach, CA: IEEE).

Google Scholar

Patel, K., Han, H., and Jain, A. K. (2016). Secure face unlock: spoof detection on smartphones. IEEE TIFS 11, 2268–2283. doi: 10.1109/TIFS.2016.2578288

CrossRef Full Text | Google Scholar

Pavlidis, I., and Symosek, P. (2000). “The imaging issue in an automatic face/disguise detection system,” in IEEE CVPRW (Hilton Head, SC: IEEE), 15–24.

Google Scholar

Peng, F., Qin, L., and Long, M. (2017). Face presentation attack detection using guided scale texture. Multimed Tools Appl. 77, 8883–8909. doi: 10.1007/s11042-017-4780-0

CrossRef Full Text | Google Scholar

Peng, F., Qin, L., and Long, M. (2020). Face presentation attack detection based on chromatic co-occurrence of local binary pattern and ensemble learning. JVCIR 66, 102746. doi: 10.1016/j.jvcir.2019.102746

CrossRef Full Text | Google Scholar

Pinto, A., Goldenstein, S., Ferreira, A., Carvalho, T., Pedrini, H., and Rocha, A. (2020). Leveraging shape, reflectance and albedo from shading for face presentation attack detection. IEEE TIFS 15, 3347–3358. doi: 10.1109/TIFS.2020.2988168

CrossRef Full Text | Google Scholar

Pinto, A., Pedrini, H., Schwartz, W. R., and Rocha, A. (2015a). Face spoofing detection through visual codebooks of spectral temporal cubes. IEEE TIP 24, 4726–4740. doi: 10.1109/TIP.2015.2466088

PubMed Abstract | CrossRef Full Text | Google Scholar

Pinto, A., Schwartz, W. R., Pedrini, H., and de Rezende Rocha, A. (2015b). Using visual rhythms for detecting video-based facial spoof attacks. IEEE TIFS 10, 1025–1038. doi: 10.1109/TIFS.2015.2395139

CrossRef Full Text | Google Scholar

Qin, L., Zhang, L.-B., Peng, F., and Long, M. (2017). “Content-independent face presentation attack detection with directional local binary pattern,” in Biometric Recognition (Springer International Publishing), 118–126.

Google Scholar

Raghavendra, R., Kiran, R., Sushma, V., Faouzi, C., and Busch, C. (2017). “On the vulnerability of extended multispectral face recognition systems towards presentation attacks,” in IEEE ISBA (New Delhi: IEEE).

Google Scholar

Raghavendra, R., Raja, K. B., Marcel, S., and Busch, C. (2016). “Face presentation attack detection across spectrum using time-frequency descriptors of maximal response in laplacian scale-space,” in IPTA (Oulu), 1–6.

Google Scholar

Ramachandra, R., and Busch, C. (2017). Presentation attack detection methods for face recognition systems: a comprehensive survey. ACM Comput. Surveys 50, 8. doi: 10.1145/3038924

PubMed Abstract | CrossRef Full Text | Google Scholar

Sanghvi, N., Singh, S. K., Agarwal, A., Vatsa, M., and Singh, R. (2021). “Mixnet for generalized face presentation attack detection,” in ICPR (Milan: IEEE), 5511–5518.

Google Scholar

Sepas-Moghaddam, A., Pereira, F. M., and Correia, P. L. (2019). Face recognition: a novel multi-level taxonomy based survey. IET Biometr. 9, 58–67. doi: 10.1049/iet-bmt.2019.0001

CrossRef Full Text | Google Scholar

Shao, R., Lan, X., and Yuen, P. C. (2019). Joint discriminative learning of deep dynamic textures for 3d mask face anti-spoofing. IEEE TIFS 14, 923–938. doi: 10.1109/TIFS.2018.2868230

CrossRef Full Text | Google Scholar

Siddiqui, T. A., Bharadwaj, S., Dhamecha, T. I., Agarwal, A., Vatsa, M., Singh, R., et al. (2016). “Face anti-spoofing with multifeature videolet aggregation,” in IEEE ICPR (Cancun: IEEE), 1035–1040.

Google Scholar

Singh, R., Agarwal, A., Singh, M., Nagpal, S., and Vatsa, M. (2020). On the robustness of face recognition algorithms against attacks and bias. AAAI 34, 13583–13589. doi: 10.1609/aaai.v34i09.7085

CrossRef Full Text | Google Scholar

Song, X., Wu, Q., Yu, D., Hu, G., and Wu, X. (2020). Face anti-spoofing detection using least square weight fusion of channel-based feature classifiers. Technical report, EasyChair.

Google Scholar

Sun, W., Song, Y., Chen, C., Huang, J., and Kot, A. C. (2020a). Face spoofing detection based on local ternary label supervision in fully convolutional networks. IEEE TIFS 15, 3181–3196. doi: 10.1109/TIFS.2020.2985530

CrossRef Full Text | Google Scholar

Sun, W., Song, Y., Zhao, H., and Jin, Z. (2020b). A face spoofing detection method based on domain adaptation and lossless size adaptation. IEEE Access 8, 66553–66563. doi: 10.1109/ACCESS.2020.2985453

CrossRef Full Text | Google Scholar

Sun, X., Huang, L., and Liu, C. (2018). Multispectral face spoofing detection using vis-nir imaging correlation. IJWMIP 16, 1840003. doi: 10.1142/S0219691318400039

CrossRef Full Text | Google Scholar

Tirunagari, S., Poh, N., Windridge, D., Iorliam, A., Suki, N., and Ho, A. T. (2015). Detection of face spoofing using visual dynamics. IEEE TIFS 10, 762–777. doi: 10.1109/TIFS.2015.2406533

CrossRef Full Text | Google Scholar

Tu, X., and Fang, Y. (2017). “Ultra-deep neural network for face anti-spoofing,” in ICONIP (Guangzhou: Springer International Publishing), 686–695.

Google Scholar

Tu, X., Zhao, J., Xie, M., Du, G., Zhang, H., Li, J., et al. (2019). Learning generalizable and identity-discriminative representations for face anti-spoofing. arXiv preprint arXiv:1901.05602. doi: 10.1145/3402446

CrossRef Full Text | Google Scholar

Vapnik, V. (2013). The Nature of Statistical Learning Theory. Springer Science & Business Media.

Google Scholar

Viola, P., and Jones, M. J. (2004). Robust real-time face detection. IJCV 57, 137–154. doi: 10.1023/B:VISI.0000013087.49260.fb

CrossRef Full Text | Google Scholar

Wang, G., Han, H., Shan, S., and Chen, X. (2020). Unsupervised adversarial domain adaptation for cross-domain face presentation attack detection. IEEE TIFS.16, 56–69. doi: 10.1109/TIFS.2020.3002390

CrossRef Full Text | Google Scholar

Wang, S.-Y., Yang, S.-H., Chen, Y.-P., and Huang, J.-W. (2017). Face liveness detection based on skin blood flow analysis. Symmetry 9, 305. doi: 10.3390/sym9120305

CrossRef Full Text | Google Scholar

Wen, D., Han, H., and Jain, A. K. (2015). Face spoof detection with image distortion analysis. IEEE TIFS 10, 746–761. doi: 10.1109/TIFS.2015.2400395

CrossRef Full Text | Google Scholar

Wu, X., Huang, H., Patel, V. M., He, R., and Sun, Z. (2019). Disentangled variational representation for heterogeneous face recognition. AAAI 33, 9005–9012. doi: 10.1609/aaai.v33i01.33019005

CrossRef Full Text | Google Scholar

Yang, J., Lei, Z., and Li, S. Z. (2014). Learn convolutional neural network for face anti-spoofing. arXiv preprint arXiv:1408.5601. doi: 10.48550/arXiv.1408.5601

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, J., Lei, Z., Liao, S., and Li, S. Z. (2013). “Face liveness detection with component dependent descriptor,” in ICB (Madrid), 1–6.

Google Scholar

Yeh, C., and Chang, H. (2018). “Face liveness detection based on perceptual image quality assessment features with multi-scale analysis,” in IEEE WACV (Lake Tahoe, NV: IEEE), 49–56.

Google Scholar

Yi, D., Lei, Z., Zhang, Z., and Li, S. Z. (2014). “Face anti-spoofing: multi-spectral approach,” in Handbook of Biometric Anti-Spoofing (Springer), 83–102.

Google Scholar

Zhang, L.-B., Peng, F., Qin, L., and Long, M. (2018). Face spoofing detection based on color texture markov feature and support vector machine recursive feature elimination. JVCIR 51, 56–69. doi: 10.1016/j.jvcir.2018.01.001

CrossRef Full Text | Google Scholar

Zhang, S., Liu, A., Wan, J., Liang, Y., Guo, G., Escalera, S., et al. (2020). Casia-surf: A large-scale multi-modal benchmark for face anti-spoofing. IEEE TBIOM 2, 182–193. doi: 10.1109/TBIOM.2020.2973001

CrossRef Full Text | Google Scholar

Zhang, S., Wang, X., Liu, A., Zhao, C., Wan, J., Escalera, S., et al. (2019). “A dataset and benchmark for large-scale multi-modal face anti-spoofing,” in IEEE CVPR (Long Beach, CA: IEEE), 919–928.

Google Scholar

Zhang, Z., Yan, J., Liu, S., Lei, Z., Yi, D., and Li, S. Z. (2012). “A face antispoofing database with diverse attacks,” in IAPR ICB (New Delhi: IEEE), 26–31.

Google Scholar

Keywords: face recognition (FR), presentation attack detection (PAD), multi-spectral, security, generalized PAD

Citation: Agarwal A, Singh R, Vatsa M and Noore A (2022) Boosting Face Presentation Attack Detection in Multi-Spectral Videos Through Score Fusion of Wavelet Partition Images. Front. Big Data 5:836749. doi: 10.3389/fdata.2022.836749

Received: 15 December 2021; Accepted: 16 May 2022;
Published: 22 July 2022.

Edited by:

Huan Zhang, University of California, Los Angeles, United States

Reviewed by:

Chittaranjan Pradhan, KIIT University, India
Xiaoyun Wang, Nvidia, United States

Copyright © 2022 Agarwal, Singh, Vatsa and Noore. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Akshay Agarwal, akshaya@iiitd.ac.in

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.