- 1Industry-Academy Cooperation Team, Hanyang University, Seoul, Republic of Korea
- 2Bionics Research Center, Korea Institute of Science and Technology, Seoul, Republic of Korea
- 3Department of HY-KIST Bio-Convergence, Hanyang University, Seoul, Republic of Korea
This study aimed at developing a noncontact authentication system using event-related pupillary response (ErPR) epochs in an augmented reality (AR) environment. Thirty participants were shown in a rapid serial visual presentation consisting of familiar and unknown human photographs. ErPR was compared with event-related potential (ERP). ERP and ErPR amplitudes for familiar faces were significantly larger compared with those for stranger faces. The ERP-based authentication system exhibited perfect accuracy using a linear support vector machine classifier. A quadratic discriminant analysis classifier trained using ErPR features achieved high accuracy (97%) and low false acceptance (0.03) and false rejection (0.03) rates. The correlation coefficients between ERP and ErPR amplitudes were 0.452–0.829, and the corresponding Bland–Altman plots showed a fairly good agreement between them. The ErPR-based authentication system allows noncontact authentication of persons without the burden of sensor attachment via low-cost, noninvasive, and easily implemented technology in an AR environment.
1 Introduction
A biometric authentication system protects significant information from people or organizations and plays an important role in the identification and authentication infrastructure (Kaongoen et al., 2017). Biometric systems are used to verify the identity of people based on their unique physiological and/or behavioral personal characteristics (Rahman et al., 2021), such as fingerprints, palm prints, face scans, iris scans, ear shapes, and vocal tract systems (Sabeti et al., 2020). Identification of people has been widely used to prevent the leakage of private information and unauthorized access of security systems in various fields such as banking, online transactions, border control, military, retail, healthcare, law enforcement, and enterprises (Rahman et al., 2022). However, many biometric traits are prone to stealing and forgery due to the advancement of related scientific techniques (Galbally et al., 2014; Ashenaei et al., 2022). Thus, exploring unique biological traits is important for biometric purposes (Rahman and Nakanishi, 2022). To enhance safety and security, many researchers have attempted to secure alternative biometric traits, such as electroencephalography (EEG) (Ashenaei et al., 2022).
In terms of robustness against hacking and forgery, EEG signals are a superior biometric approach because they have unique attributes not possessed by past biometric methods (Wang M. et al., 2021). In terms of security, EEG-based biometrics have the following advantages: 1) difficulty of duplication and theft because they are uncapturable, 2) alive biometric, and 3) unlimited replacement (Ashenaei et al., 2022). Additionally, biometric traits from a deceased (and warm) body can be used to allow authentication of security systems-; however, a dead brain cannot generate the EEG oscillations required for authentication (Norton et al., 2017). Finally, EEG signals are robust to the demand for coercive password entry because they can be easily affected by external pressures (Wu et al., 2018a). EEG-based biometric systems are primarily categorized into spontaneous EEG and evoked potentials (EPs) using signal acquisition protocols (Wang M. et al., 2021; Ashenaei et al., 2022). Resting-state EEG is a spontaneous signal that is naturally generated by the brain without any stimuli. It has been utilized in previous studies as a biometric trait (Thomas and Vinod, 2018; Kim and Kim, 2019; Maiorana, 2021) because of its flexible acquisition of signals and suitability for continuous monitoring compared with stimuli-based EPs (Thomas and Vinod, 2018). However, the stability of resting-state-based biometrics is relatively poor (Wang M. et al., 2019). In contrast, to elicit unique signals based on strict protocols, event-related potential (ERP)-based biometrics require users to pay attention to repetitive sensory stimuli (i.e., visually EPs) (Wang M. et al., 2021). ERP is a significant biometric trait that can reflect high-level neural resources such as attention and memory, which can perceive only people with knowledge of their intrinsic information (Sabeti et al., 2020). Additionally, the performance of detecting ERP components (i.e., P3) has been enhanced by improving the feature extraction and classification algorithms using deep learning methods (Wang H. et al., 2019; Wang H. et al., 2021; Wang et al., 2023a). Generally, strict protocols involving cognitive tasks are more distinctive between individuals and reproducible compared with those without tasks, but they are time-consuming (Wang et al., 2023a). Thus, an ERP-based biometric system can be a high-performance and safe authentication system, as proven in previous studies (Kaongoen et al., 2017; Chan et al., 2018; Wu et al., 2018b; Chen et al., 2020; Kasim and Tosun, 2021; Zhao et al., 2021). Although EEG signals have been studied as a unique biological trait of biometric authentication because of their advantages for safety and security (Shams et al., 2022), the attachment of a sensor on the head to acquire EEG signals is a major obstacle (Chang S. et al., 2020; Park et al., 2022). Thus, EEG biometrics, having the disadvantages of sensor attachment, inconvenience, complexity, onerous processes, and susceptibility to muscle noise, have lower usability than other biometrics (Rhee et al., 2022). While most ERP-based systems have been studied based on monitor screens, virtual or augmented reality (AR)-based smart glasses have the advantage of providing more flexibility such as freeing both hands and allowing the use of multiple devices (Uhlmann et al., 2019).
Pupil images can be easily measured using an eye-tracker device as an add-on to virtual reality or AR glasses. Pupillary rhythms, i.e., pupil size change (PSC), are reliably modulated by functional brain processes, such as cognition (Papetti et al., 2020), perception (Bradley et al., 2017), attention (Unsworth et al., 2018), memory (Stolte et al., 2020), and emotion (Cherng et al., 2020) via neural connectivity. Previous studies have demonstrated that PSC is strongly associated with neural activity in brain regions or networks that involve the locus coeruleus–norepinephrine system, dorsal attention network, posterior and anterior cingulate cortex, insular cortex, basal ganglia, and lingual gyrus (Joshi et al., 2016; Ceh et al., 2021; Groot et al., 2021; Mäki-Marttunen, 2021). Thus, PSC signals have a great potential for use in biometrics as an alternative to ERP analysis. The pupil-based biometric as a biometric trait reflecting neural activity is robust to stealing and forgery and is simple and convenient compared with ERP in terms of usability. Previous studies have reported a significant correlation between ERP components and PSC (Widmann et al., 2018; Schütte et al., 2021; Selezneva and Wetzel, 2022). Several studies have also attempted to directly compare ERP and PSC epochs and have reported mutual similarities and replaceability (Park et al., 2019; Dahl et al., 2020; Park et al., 2022). The PSC epoch caused by a target stimulus is called event-related pupillary response (ErPR).
The aim of this study is to develop an ErPR-based biometric authentication system using a rapid serial visual presentation (RSVP) paradigm with human photograph stimuli in AR glasses. The RSVP paradigm, which can present a large number of stimuli in a short time, can elicit strong EPs caused by a stimulus paradigm consisting of targets and nontargets (Acqualagna and Blankertz, 2013), and has been utilized in many visual-based brain-computer interface studies (Zhang H. et al., 2022; Wang et al., 2023b; Wang et al., 2024). Additionally, numerous ERP-based studies have reported that ERP components (i.e., N2, P3, N4, and P6) are significant features for distinguishing familiar from stranger human faces (Hanso et al., 2010; Huang et al., 2017; Chang W. et al., 2020). We elicited stronger subject-specific ErPR from pupil images and ERP from EEG signals using our RSVP paradigm, which included familiar (target) or stranger (nontarget) human photographs. Two biometric traits were compared using accuracy, area under the receiver operating characteristics curve (AUC), false rejection rate (FRR), and false acceptance rate (FAR). Detailed information on the proposed biometric method and results are described in the following sections.
2 Materials and methods
2.1 Subjects
Thirty healthy volunteers (15 men and 15 women), aged between 22 and 33 years (mean age, 27.20 ± 3.34 years) participated in this experiment. All the participants had normal or corrected-to-normal vision (i.e., over 0.8) and were right-handed. Each participant participated voluntarily and was paid 30,000 KRW. None of them had any history of serious medical or psychological illnesses. Written informed consent was obtained from all participants, and they were notified of the restrictions and requirements. All the experimental protocols were approved by the Ethics Committee of the Korea Institute of Science and Technology, Seoul, South Korea (approval number: 2021-012).
2.2 Experimental procedure and stimuli
The participants were required to provide photographs of familiar faces of their family or friends of the same gender. In total, 300 photographs of familiar faces were collected from 30 participants, with 10 photographs per participant. For each subject, ten photographs were randomly presented throughout the entire experiment as the target stimuli. In all the trials, a total of 900 photographs of random Korean individuals (450 men and 450 women) were collected and used as nontarget stimuli. Repeated exposure to a stranger’s face (i.e., familiarization) may induce ERP patterns similar to those of a familiar face (Campbell and Tanaka, 2021). Thus, all nontarget stimuli were presented only once without duplication, and the order was counterbalanced. To minimize the effects of gender and race on ERP (Ito and Urland, 2005), all photographs used in this experiment consisted of a Korean person, and photographs of the same gender as the subjects were presented as both the target and nontarget stimuli. All the photographs were set up to be the same orientation and size.
Each participant wore a Microsoft HoloLens 2 (Microsoft Corp., Redmond, WA, United States) in an electrically shielded room and sat in a comfortable armchair. An electrically shielded room was used to minimize the risk of external interference during the measurement of EEG signals and to increase the concentration of the subjects. Participants were required to perform an ERP task, and EEG signals and pupil images were measured during the task. The overall process of the experiment was recorded using a monitoring camera, as shown in Figure 1.
Figure 1. Overview of the experimental setup. (A) Experimental management and monitoring. (B) Experimental environment. (C) Example of an AR stimulus. (D) Pupil Labs’ AR binocular eye-tracker add-on AR device. (E) AR device (HoloLens 2).
As Figure 2 shows, the participants executed the ERP task for approximately 5 min. The ERP task began by displaying a cross on the AR screen’s center for 2 s followed by ten face photographs, each shown for 2 s. The ten face photographs were arranged randomly, with only one featuring as the target familiar face selected randomly from a database of the subject’s photographs. The other nine photographs presented random strangers’ faces as non-targets. Each photographic stimulus was presented center screen for 100 ms, and a block screen was shown for 100 ms between photographs (i.e., the distance between two subsequent stimuli is 200 ms). One trial comprised ten face photograph stimuli and lasted 2 seconds. One block comprised five trials, each separated by a 2-s interval, totaling 20 s. The entire experiment consisted of 50 trials (ten blocks) with an inter-block interval of 5 s resulting in a total duration of 245 s. Before and after the experiment, there were periods of preparation and relaxation, each lasting for 30 s. The stimuli displayed at the center of the screen and the size of the photographs, including the black screen were 400 × 500 pixels. The stimuli were presented on a transparent display on an AR headset (MS Microsoft HoloLens 2™ Microsoft Corp., Redmond, WA, United States). The event triggers for the stimuli were synchronized with the EEG and ErPR data through User Datagram Protocol communication. When the stimuli began on the AR device, the recording software for EEG and ErPR on the laptop received UDP packets from the AR device. The data was then saved with the specific time corresponding to when the UDP packet was received at the start of the stimuli.
2.3 Data acquisition and signal processing
A 64-channel BioSemi ActiveTwo system (BioSemi BV, WG-Plein, Amsterdam, Netherlands) was used to acquire EEG signals from the participants at a sampling rate of 2048 Hz with an EEG cap of active Ag/AgCl electrodes through a conductive water-based gel (CG04 Saline Base Signa Gel, Parker Laboratories Inc., Fairfield, NJ, United States) arranged in an international 10–20 system (ground: common mode sense; reference: driven right-leg electrodes). The electrode impedance between the measurement and ground electrodes was maintained below 10 kΩ. To avoid contaminating meaningful patterns of ERP, muscle artifacts of oculomotor were removed from the raw EEG signals using independent component analysis based on visual inspection (Katus et al., 2020; Li et al., 2020). Pupil images were taken on Pupil Labs’ AR binocular eye-tracker (Pupil Labs, Berlin, BB, Germany) add-ons for the HoloLens 2 at a sampling rate of 200 fps with a resolution of 192 × 192 pixels using Pupil Core software (Pupil Labs, Berlin, BB, Germany). This software provides data related to eye movement, including gaze position and pupil diameter. A previous study confirmed that pupil size increases as illumination decreases. Under five different lighting conditions, the average pupil diameter measured 3.5 mm at 550 lx, 4.2 mm at 350 lx, 5.2 mm at 150 lx, 5.03 mm at 40 lx, and 5.4 mm at 2 lx. The pupil size significantly increased when the illumination changed from 550 to 150 lx; however, lighting conditions of 150, 40, and 2 lx minimally impacted the changes in pupil size (Maqsood and Schumacher, 2017). To minimize the effect on pupil size caused by a significant change in ambient light, ambient light in the electrically shielded room was controlled at 150 lx or less (Park et al., 2021; Park et al., 2022). During the experiment, the ambient light was measured using a Visible Light SD Card Logger (Sper Scientific Meters Ltd., Scottsdale, AZ, United States) at a 1 Hz sampling rate in both the experiment room (105.47 ± 2.22 lx) and AR device (30.70 ± 7.75 lx).
2.4 Data processing
The procedure for signal processing in the ERP was as follows: 1) The EEG signals were down sampled from 2048 to 200 Hz, and then filtered using a fourth-order Butterworth band-pass filter (0.1–50 Hz). 2) The filtered EEG data were segmented into EEG epochs with lengths of 800 ms from 200 ms before the onset of each stimulus to represent the stimulus. 3) All EEG epochs were corrected at baseline by averaging EEG epochs lasting 800 ms using 200 ms of data before target onset (Mitchell et al., 2016). 4) EEG epochs in all trials were averaged with lengths of 800 ms. 5) The amplitude and latency were defined by calculating difference values in amplitude between the lowest and highest points and the time value of the highest point in the ERP epoch, respectively, within a time window of 200–750 ms (Kaongoen et al., 2017), as shown in Figure 3A. The time windows were divided into P3a (200–350 ms), P3b (400–490 ms), and LPP (530–750 ms) (Takeda et al., 2014; Causse et al., 2016). Amplitude and latency were extracted from three time domains and four brain regions (Fz, Cz, Pz, and Oz electrodes) respectively. All signal processing and data analyses were performed using EEGlab, which is a MATLAB toolbox (R2020b; MathWorks Inc., Natick, MA, United States).
Figure 3. Overview of signal processing with definition of features in (A) ERP and (B) ErPR. (a) EEG raw signals. (b), (h) Time log of target onset. (c) Epochs of separated EEG signals based on target onset. (d) Grand averages for all separated EEG epochs, and definition of amplitude and latency in ERP including time windows of P3a, P3b, and LPP. (e) Procedure for detecting pupil area. (f) Signals of pupil diameter (raw data). (g) Signals of PSC calculated from frame difference. (i) Epochs of separated PSC data based on target onset. (j) Grand averages for all separated PSC epochs, and definition of amplitude and latency in ErPR including time windows of P3a, P3b, and LPP.
The procedure for signal processing in ErPR was as follows: 1) The pupil diameter was obtained from data offered by the Pupil Core software at a sampling rate 200 fps, and data of the dominant eye for each subject were used for analysis. The dominant eye of each individual was identified using the hole-in-the-card test (Li et al., 2010). 2) PSCs were calculated using the difference between the frames for each pupil diameter. 3) All PSC epochs were corrected to a baseline by the averaged PSC epochs lasting 800 ms using 200 ms of data before the target onset. 4) The PSC epochs in all the trials were averaged with lengths of 800 ms, and the average PSC epoch was defined as ErPR. 5) The amplitude and latency of the ErPR epoch were defined by calculating the difference between the lowest and highest PSC points and the time value of the highest PSC point, respectively, within the time windows of P3a (200–350 ms), P3b (400–490 ms), and the LPP (530–750 ms), consistent with the ERP epochs, as shown in Figure 3B. All signal processing was performed using the MATLAB signal processing toolbox (R2020b, MathWorks Inc., Natick, MA, United States).
2.5 Statistical analysis and classification
This study has a design within subject, wherein two stimuli such as target (i.e., familiar face) and nontarget photograph (i.e., stranger face) are tested on each test subject. Thus, for the statistical analysis, a paired-samples t-test was used to compare the responses of individual participants between target and nontarget stimuli based on the Shapiro–Wilk normality test (p > 0.05). The recommended total sample size, calculated using G*Power software (ver. 3.1.9.7; Heinrich-Heine-Universität Düsseldorf, Düsseldorf, Germany), was 54 samples (α = 0.01, 1 – β = 0.95, effect size = 0.5), and this study (i.e., 60 samples size) satisfied the recommended sample size from G*Power (Faul et al., 2007). The significant level to test hypotheses was controlled by the number of individual hypotheses (i.e., α = 0.05∕n) to resolve the problem of type-I errors by multiple comparisons (Jafari and Ansari-Pour, 2019) as follows: the statistically significant level was set to 0.0017 (i.e., α = 0.05/30; 24 ERP and six ErPR features). Moreover, this study confirmed the practical significance of using an effect size of Cohen’s d, with the standard values of 0.20, 0.50, and 0.80 regarded as small, medium, and large, respectively (Huck et al., 1974). All statistical analyses were conducted using IBM SPSS Statistics for Windows (SPSS Corp., Armonk, NY, United States).
To determine the best classification algorithm for the two conditions (ERP and ErPR features), two machine-learning algorithms were used: 1) linear support vector machine (LSVM), 2) quadratic discriminant analysis (QDA), 3) Naïve Bayes (NB), 4) logistic regression (LR), and 5) radial basis function support vector machine (RBF-SVM). Optimization results for each classification method were obtained through five-fold cross-validation using “scikit-learn” (ver. 0.24.2) of Python (ver. 3.6.9). To assess practical classification performance, we reduced the number of trials while maintaining a 1:9 ratio between familiar and unfamiliar stimuli. To extract the features, we computed the averaged ERP and ErPR epochs over the familiar and unfamiliar stimuli trials. From the averaged ERP epochs, we extracted 12 features, which comprised four channels (Fz, Cz, Pz, and Oz) and three indicators (P3a, P3b, and LPP) for the ERP features. From the averaged ErPR epochs, we also extracted three features including three indicators (P3a, P3b, and LPP) for ErPR features. The structures of the features for each condition were: 1) 60 samples (30 subjects and two conditions) × 12 ERP features and 2) 60 samples (30 subjects and two conditions) × three ErPR features. Accuracy refers to the average accuracy of the five-fold cross-validation. Moreover, the FAR, FRR, and AUC were evaluated.
The accuracy was calculated using the proportion of the total number of correct predictions, as shown in Equation 1.
A true positive (TP) is a correctly classified target. A false negative (FN) is an incorrectly classified target. A true negative (TN) is a correctly classified nontarget. A false positive (FP) is an incorrectly classified nontarget. FAR is the proportion of identification instances in which unauthorized persons are incorrectly accepted, and FRR is that in which authorized persons are incorrectly rejected. These values were calculated using Equation 2.
3 Results
3.1 Averaged plot of ERP and ErPR epochs from all subjects
Figure 4 shows the average ERP plot of each channel (Fz, Cz, Pz, and Oz) and the ErRPs for the target and non-target stimuli. Each average plot includes 50 target epochs and 450 nontarget epochs. The solid and dotted lines indicate the ERP or ErPR epochs of the participants after being presented with the target and nontarget stimuli, respectively. The evoked positive potentials within P3a (200–350 ms), P3b (400–490 ms), and LPP (530–750 ms) were clearly observed in both the target and nontarget ERP epochs in the Fz, Cz, Pz, and Oz regions. The increasing ERP amplitude trend was observed for the target stimuli from P3a, P3b, and LPP compared with the nontarget stimuli in all electrode channels. No clear difference was observed in EEG latency. These trends of amplitude and latency of ERP are similar to those observed in the ErPR epoch.
Figure 4. Averaged plot of ERP at Fz (A), Cz (B), Pz (C), and Oz (D) and ErPR (E) from all subjects for target and nontarget stimuli.
3.2 Amplitude and latency of ERP epoch
Figure 5 shows the results of a paired-sample t-test between the target and nontarget stimuli in terms of the amplitude and latency of ERP. The amplitude of ERP of target stimuli in P3a (200–350 ms) component was significantly larger than those of nontarget stimuli in Fz [t (58) = 8.445, p < 0.001, Cohen’s d = 2.527], Cz [t (58) = 8.637, p < 0.001, Cohen’s d = 2.599], Pz [t (58) = 9.595, p < 0.001, Cohen’s d = 2.667], and Oz [t (58) = 3.789, p < 0.001, Cohen’s d = 1.036]. The amplitude of ERP of target stimuli in P3b (400–490 ms) component was significantly larger than those of nontarget stimuli in Fz [t (58) = 10.027, p < 0.001, Cohen’s d = 2.509], Cz [t (58) = 9.243, p < 0.001, Cohen’s d = 2.306], Pz [t (58) = 6.497, p < 0.001, Cohen’s d = 1.925], and Oz [t (58) = 4.786, p < 0.001, Cohen’s d = 1.353]. The amplitude of ERP of target stimuli in LPP (530–750 ms) component was significantly larger than those found in nontarget stimuli in Fz [t (58) = 12.661, p < 0.001, Cohen’s d = 2.967], Cz [t (58) = 10.470, p < 0.001, Cohen’s d = 2.298], Pz [t (58) = 4.367, p < 0.001, Cohen’s d = 1.290], and Oz [t (58) = 6.830, p < 0.001, Cohen’s d = 1.809].
Figure 5. Comparisons of ERP amplitude and latency for target and nontarget stimuli in (A) P3a, (B) P3b, and (C) LPP with a paired-samples t-test. The error bars show the standard error in each condition (*, p < 0.05, ***, p < 0.001).
The latency of ERP in P3a, P3b, and LPP revealed that there was no significant difference between target and nontarget stimuli in all electrode sites as follows: P3a in Fz [t (58) = 1.752, p = 0.0903], Cz [t (58) = 2.341, p = 0.0263, adjusted by the Bonferroni correction], Pz [t (58) = 3.201, p = 0.0033, adjusted by the Bonferroni correction], and Oz [t (58) = 6.830, p = 0.0678]; P3b in Fz [t (58) = 2.905, p = 0.0070, adjusted by the Bonferroni correction], Cz [t (58) = 2.419, p = 0.0220, adjusted by the Bonferroni correction], Pz [t (58) = 1.990, p = 0.0561], and Oz [t (58) = 2.494, p = 0.0186, adjusted by the Bonferroni correction]; LPP in Fz [t (58) = 2.731, p = 0.0106, adjusted by the Bonferroni correction], Cz [t (58) = 2.819, p = 0.0086, adjusted by the Bonferroni correction], Pz [t (58) = 3.141, p = 0.0039, adjusted by the Bonferroni correction], and Oz [t (58) = 2.875, p = 0.0075, adjusted by the Bonferroni correction], as shown in Figure 5.
3.3 Amplitude and latency of ErPR epoch
Figure 6 shows the results of a paired-sample t-test for target and nontarget stimuli in amplitude and latency of ErPR. The amplitude of the ErPR of the target stimuli was significantly larger than those of the nontarget stimuli in the P3a [t (58) = 7.275, p < 0.001, Cohen’s d = 2.242], P3b [t (58) = 8.165, p < 0.001, Cohen’s d = 2.235], and LPP [t (58) = 5.472, p < 0.001, Cohen’s d = 1.610]. The latency of the ErPR of target stimuli was significantly delayed compared to that of nontarget stimuli at P3a [t (58) = 4.298, p < 0.001, Cohen’s d = 1.140]. No significant differences were found between the ErPR latencies for the target and nontarget stimuli in the P3b [t (58) = 1.861, p = 0.0729] and LPP [t (58) = 1.685, p = 0.1028].
Figure 6. Comparisons of ErPR amplitude and latency for target and nontarget stimuli in each P3a, P3b, and LPP with a paired-samples t-test. The error bars show the standard error in each condition (***, p < 0.001).
3.4 Correlation and Bland–Altman plot among ERP and ErPR
The Pearson correlation coefficient between the ERP and ErPR amplitudes was statistically significant. (1) P3a at Fz (r = 0.685, p < 0.001), Cz (r = 0.722, p < 0.001), Pz (r = 0.733, p < 0.001), and Oz (r = 0.466, p < 0.001). (2) P3b at Fz (r = 0.829, p < 0.001), Cz (r = 0.745, p < 0.001), Pz (r = 0.514, p < 0.001), and Oz (r = 0.452, p < 0.001). (3) LPP at Fz (r = 0.628, p < 0.001), Cz (r = 0.643, p < 0.001), Pz (r = 0.558, p < 0.001), and Oz (r = 0.652, p < 0.001). The Pearson correlation coefficient between the ERP and ErPR latency was found to be statistically significant. (1) P3a at Fz (r = 0.639, p < 0.001), Cz (r = 0.706, p < 0.001), Pz (r = 0.660, p < 0.001), and Oz (r = 0.702, p < 0.001). (2) P3b at Fz (r = 0.356, p < 0.01), Cz (r = 0.403, p < 0.01), Pz (r = 0.338, p < 0.01), and Oz (r = 0.435, p < 0.001). (3) LPP at Fz (r = 0.469, p < 0.001), Cz (r = 0.482, p < 0.001), Pz (r = 0.533, p < 0.001), and Oz (r = 0.476, p < 0.001), as shown in Figure 7.
Figure 7. Results of correlation analysis between ERP and ErPR amplitude and latency for target and nontarget stimuli in P3a, P3b, and LPP. (A) depicts the correlation of amplitude between ERP and ErPR, with each row representing the amplitude of P3a, P3b, and LPP as indicated by the subtitles in the boxes. Each column shows the results for different channels: Fz, Cz, Pz, and Oz. (B) depicts the correlation for latency in ERP and ErPR, with each plot showing the results for P3a latency, P3b latency, and LPP latency. Each column represents the results for the channels Fz, Cz, Pz, and Oz. Significant findings were plotted as linear regression lines (p < 0.01, p < 0.001).
Figure 8 shows the results of the Bland–Altman plot between the ERP and ErPR features. It was used to visualize the differences in measurements between the two different variables. The x- and y-axes of the plot display the mean values of two variables and difference between the two variables, respectively, and involved the following three lines: the mean difference between two variables (i.e.,
Figure 8. Representative Bland–Altman plots for (A) amplitude and (B) latency from ERP and ErPR epochs. The solid central line in each plot represents the mean difference between the two variables, and the upper and lower dotted lines represent the limit of the 95% confidence interval (
3.5 Classification
To distinguish between the target and nontarget stimuli, the ERP and ErPR classification performances were compared. We used the twelve features for the ERP condition and the three features for the ErPR condition, which were the amplitudes of the P3a, P3b, and LPP components based on statistical significance. LSVM and RBF-SVM were the superior classifier with 100% accuracy (versus 98% with QDA) when using ERP features, while QDA with 97% accuracy outperformed LSVM (83% accuracy) on ErPR features. Overall, the classification accuracy using the ERP features was 3% greater than that with the ErPR features. Details of the classification results are listed in Table 1. Additionally, a permutation test was conducted to determine the accuracy and generalization ability of five classifiers (repeated 10,000 times), and all classifiers for both the ERP and ErPR datasets were significant (p < 0.0001) in the permutation test.
Table 1. Results of classification using LSVM, QDA, NB, LR, and RBF-SVM (five-fold cross-validation) among target and nontarget epochs for ERP and ErPR (N = 50).
Figure 9 shows the FRR, FAR, and accuracy of the proposed authentication system as functions of the number of trials based on the QDA classifier. The accuracy declines rapidly as the number of trials decreases. The proposed ERP- and ErPR-based authentication systems exhibited accuracies of 95% when the number of trials was less than 34 and 44, respectively. Although the accuracy increases with more tasks, the time required for authentication also increases.
Figure 9. Trends of (A) false rejection rate (FRR), (B) false acceptance rate (FAR), and (C) accuracy by the number of RSVP trials in an ErPR-based authentication system (QDA classifier).
3.6 Real-time system for individual identification in AR environment
The real-time system proposed in this study consists of a Microsoft HoloLens 2 (Microsoft Corp., Redmond, WA, United States), Pupil Labs’ AR binocular eye-tracker (Pupil Labs, Berlin, BB, Germany) add-ons for the HoloLens 2, and a personal computer for analysis. As shown in Figure 10, the target system could be accessed using two-factor individual identification in AR environment. The procedure of individual identification to access the target system is as follows: 1) The user wears the Microsoft HoloLens 2 headset and operates the target system (Figure 10A). 2) Then, the user uses the “Sign in” button to attempt authentication to access the target system (Figure 10B). 3) The proposed system conducts primary authentication by analyzing the user’s iris pattern (Figure 10C). Iris recognition is developed using publicly available open-source code from GitHub (https://github.com/thuyngch/Iris-Recognition). 4) The database comprises photographs voluntarily registered by users who intend to utilize the authentication system, alongside randomly collected photographs of individuals. A unique identifier resulting from iris recognition is assigned to the user, and a random sequence of photographs is generated using target photographs stored in the database (Figure 10D). 5) The authentication tasks are performed, and the pupil image is measured (Figure 10E). 6) The proposed system conducts authentication by analyzing the user’s ErPR pattern. If the user is the real client, ErPR response occurs in reaction to the target photographs, and the authentication system allows access to the target system (Figure 10F). Figure 10G shows the real-time system architecture for individual identification using ErPR response. The ErPR-based authentication system was developed using the MATLAB App designer (2020b, Mathworks Inc., Natick, MA, United States) and Unity 2018.1 (Unity Technologies, San Francisco, CA, United States). In the two-factor-based real-time identity recognition system, the iris recognition accuracy was 100%, and the results of the ErPR-based authentication system are consistent with those reported in Section 3.5 Classification.
Figure 10. Overview of a real-time system for individual identification in an AR environment. (A) Overview of the authentication system in AR environment. (B) “Sign in” screen. (C) Authentication using iris recognition. (D) Database of human photographs. (E) ErPR-based authentication tasks. (F) Final decision of the proposed authentication system for accessing the target system. (G) A real-time system architecture for ErPR-based authentication system: (a) Input infrared image; (b) Detecting pupil area and measuring pupil size; (c) ErPR epoch; (d) Controlling trials of authentication task; (e) Setting IP connection between the AR device and analysis PC; (f) Final decision of individual identification.
4 Discussion
We proposed a novel ErPR-based identity authentication system that uses familiar and stranger human faces for RSVP. We demonstrated that the RSVP stimulus based on facial familiarity elicited distinct ErPR and ERP traits in each user and that the ErPR trait can be utilized as an alternative to the ERP-based authentication system. This study assessed the classification performance, which involves accuracy, AUC, FAR, and FRR, between a target (i.e., familiar face) and nontarget (i.e., stranger face) in our RSVP paradigm for two biometric traits. Additionally, to compare the similarities between them, correlation coefficients and Bland–Altman plots were analyzed.
The average ERP epoch of each subject revealed EP components such as P3a (200–350 ms), P3b (400–490 ms), and LPP (530–750 ms) with significant differences in amplitude, although the latency did not show significant differences between targets and nontargets. P3 and LPP components are well-known representative indicators of cognitive processes. The P3 component can be elicited if the user’s brain fully perceives the stimulus. The LPP component is related to post-processing of consciousness, thus indicating advanced cognition and regulation of information (Zhang R. et al., 2022). These components are elicited by strict protocols, in which low-probability “target stimuli” and high-probability “nontarget stimuli” (i.e., a target rate of 10% or 20%) are mixed (Jijomon and Vinod, 2021). Kaongoen et al. (2017) proposed a two-factor authentication system using P300 ERP responses from photographic stimuli; furthermore, the P300 epochs (i.e., 200–750 ms) in the target condition showed a higher amplitude compared to the nontarget. Zeng et al. (2018) developed an identity authentication system using the RSVP paradigm, including the self-face and non-self-face. They reported significant differences in P3a and P3b amplitudes induced by familiar and unfamiliar photographs. Rathi et al. (2021) proposed an authentication system using P300 speller, consisting of pictures of different object pictures with a 2 × 2 matrix. They found that P300 amplitude in the target condition was significantly larger than that in the nontarget condition. Other studies have demonstrated that the amplitude in P3 and LPP components of ERP in target are significantly larger than those in the nontarget, and our findings are consistent with these studies (Lee et al., 2017; Kim et al., 2018; Sabeti et al., 2020; Rathi et al., 2022). However, this study differs from previous studies in that it uses AR glasses in the authentication systems. AR-glass-based authentication systems can provide users with more flexibility than monitor screens, such as freeing both hands and enabling the use of multiple devices (Uhlmann et al., 2019). The proposed ERP-based authentication system achieved perfect accuracy in terms of FAR and FRR using the LSVM and RBF-SVM classifier (5-fold cross-validation).
The pupillary rhythm (i.e., ErPR)-based authentication system proposed in this study exhibited lower performance than ERP, but achieved high performance in accuracy (97%), FAR (0.03), and FRR (0.03) using the QDA classifier (5-fold cross-validation). Similar to the ERP epoch, the amplitude of the average ErPR epoch of each subject in the target stimulus was significantly larger than that of the nontarget stimuli. The PSC is significantly associated with the brain regions related to cognitive processing, involving locus coeruleus–norepinephrine, posterior and anterior cingulate cortex, paracingulate cortex, orbitofrontal cortex, right anterior insular cortex, dorsal anterior cingulate, basal ganglia, lingual gyrus, and thalamus (Joshi et al., 2016; Larsen and Waters, 2018; DiNuzzo et al., 2019; Ceh et al., 2021; Groot et al., 2021; Mäki-Marttunen, 2021). The neural resource caused by cognition for stimuli is reflected in pupil size via a top-down executive control network in the following steps: 1) Alert, an early component (Pa), 2) acceleration of Pa, and 3) executive control by a prominent late component (Pe) (Geva et al., 2013). Many previous studies have reported a strong correlation between PSC and ERP components in cognitive processing. (1) The pupil dilation response is associated with the amplitude of the P3a component in the top-down control of involuntary orienting of attention (Selezneva and Wetzel, 2022). (2) Pupil dilation is related to the amplitude of LPP during cognitive reappraisal (Strauss et al., 2016). (3) The pupil dilation (i.e., ErPR reinstatement data) caused by arousal-related norepinephrine release related to attention is correlated with stronger EEG α-β desynchronization (i.e., event-related desynchronization) and ERP signals (Dahl et al., 2020). (4) Increasing pupil size has been correlated with the amplitude of the P300 and N400 components in cognitive load (Kuipers and Thierry, 2011; Tao et al., 2019) and cognitive flexibility (Kuipers and Thierry, 2013). In our previous studies, we found that the amplitude of the P3 and LPP components in both ERP and ErPR epochs significantly decreased with increased mental load and showed a strong positive correlation between them (Park and Whang, 2018; Park et al., 2019; Park et al., 2022). In this study, the amplitudes of ERP and ErPR epochs were directly related to each other, based on the results for correlation coefficients (i.e., in the range of 0.452–0.829) and the Bland–Altman plot (i.e., fairly good agreement) between them. We identified sufficient evidence that the ErPR of the pupil rhythm could be utilized as an alternative to ERP in authentication systems. The ErPR-based authentication system, especially in an AR environment (i.e., eye-tracker add-on AR glasses), can provide good usability in a simple, economical, and contactless manner.
5 Conclusion
This study aimed to develop an infrared camera-based noncontact authentication system using ErPR epochs obtained from pupillary rhythms in an AR environment. The proposed ErPR-based authentication system achieved high performance but showed lower performance than previous EEG signal-based authentication systems (Kaongoen et al., 2017; Chan et al., 2018; Wu et al., 2018b; Chen et al., 2020; Kasim and Tosun, 2021; Zhao et al., 2021). However, the approach presented in this paper allows noncontact authentication for people without the burden of sensor attachment via low-cost, noninvasive, and easily implemented technologies in an AR environment. Although the time required for authentication and effect of variations in ambient light levels must be improved, the proposed method has considerable potential for use in person-authentication systems. Future studies will attempt to overcome the disadvantages of this study.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by Ethics Committee of the Korea Institute of Science and Technology. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
SP: Conceptualization, Investigation, Methodology, Validation, Writing–original draft, Writing–review and editing. JH: Data curation, Formal Analysis, Software, Visualization, Writing–original draft. LK: Conceptualization, Funding acquisition, Supervision, Writing–review and editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Challengeable Future Defense Technology Research and Development Program through the Agency for Defense Development (ADD) funded by the Defense Acquisition Program Administration (DAPA) in 2023 (No. 912911601).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Acqualagna L., Blankertz B. (2013). Gaze-independent BCI-spelling using rapid serial visual presentation (RSVP). Clin. Neurophysiol. 124, 901–908. doi:10.1016/j.clinph.2012.12.050
Ashenaei R., Asghar Beheshti A. A., Yousefi Rezaii T. (2022). Stable EEG-Based biometric system using functional connectivity based on time-frequency features with optimal channels. Biomed. Signal Process. Control. 77, 103790. doi:10.1016/j.bspc.2022.103790
Bradley M. M., Sapigao R. G., Lang P. J. (2017). Sympathetic ANS modulation of pupil diameter in emotional scene perception: effects of hedonic content, brightness, and contrast. Psychophysiology 54, 1419–1435. doi:10.1111/psyp.12890
Campbell A., Tanaka J. W. (2021). When a stranger becomes a friend: measuring the neural correlates of real-world face familiarisation. Vis. Cogn. 29, 689–707. doi:10.1080/13506285.2021.2002993
Causse M., Peysakhovich V., Fabre E. F. (2016). High working memory load impairs language processing during a simulated piloting task: an ERP and pupillometry study. Front. Hum. Neurosci. 10, 240. doi:10.3389/fnhum.2016.00240
Ceh S. M., Annerer-Walcher S., Koschutnig K., Körner C., Fink A., Benedek M. (2021). Neurophysiological indicators of internal attention: an fMRI-eye-tracking coregistration study. Cortex 143, 29–46. doi:10.1016/j.cortex.2021.07.005
Chan H. L., Kuo P. C., Cheng C. Y., Chen Y. S. (2018). Challenges and future perspectives on electroencephalogram-based biometrics in person recognition. Front. Neuroinform. 12, 66. doi:10.3389/fninf.2018.00066
Chang S., Dong W., Jun H. (2020). Use of electroencephalogram and long short-term memory networks to recognize design preferences of users toward architectural design alternatives. J. Comput. Des. Eng. 7, 551–562. doi:10.1093/jcde/qwaa045
Chang W., Wang H., Yan G., Liu C. (2020). An EEG based familiar and unfamiliar person identification and classification system using feature extraction and directed functional brain network. Expert Syst. Appl. 158, 113448. doi:10.1016/j.eswa.2020.113448
Chen J. X., Mao Z. J., Yao W. X., Huang Y. F. (2020). EEG-based biometric identification with convolutional neural network. Multimed. Tools Appl. 79, 10655–10675. doi:10.1007/s11042-019-7258-4
Cherng Y. G., Baird T., Chen J. T., Wang C. A. (2020). Background luminance effects on pupil size associated with emotion and saccade preparation. Sci. Rep. 10, 15718. doi:10.1038/s41598-020-72954-z
Dahl M. J., Mather M., Sander M. C., Werkle-Bergner M. (2020). Noradrenergic responsiveness supports selective attention across the adult lifespan. J. Neurosci. 40, 4372–4390. doi:10.1523/JNEUROSCI.0398-19.2020
DiNuzzo M., Mascali D., Moraschi M., Bussu G., Maugeri L., Mangini F., et al. (2019). Brain networks underlying eye’s pupil dynamics. Front. Neurosci. 13, 965. doi:10.3389/fnins.2019.00965
Faul F., Erdfelder E., Lang A. G., Buchner A. (2007). G*Power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav. Res. Methods. 39, 175–191. doi:10.3758/bf03193146
Galbally J., Satta R., Gemo M., Beslay L. (2014). Biometric spoofing: a JRC case study in 3D face recognition. Luxembourg: Publications Office of the European Union.
Gardener S. L., Lyons-Wall P., Martins R. N., Rainey-Smith S. R. (2020). Validation and reliability of the Alzheimer’s disease-Commonwealth Scientific and Industrial Research Organisation food frequency questionnaire. Nutrients 12, 3605. doi:10.3390/nu12123605
Geva R., Zivan M., Warsha A., Olchik D. (2013). Alerting, orienting or executive attention networks: differential patters of pupil dilations. Front. Behav. Neurosci. 7, 145. doi:10.3389/fnbeh.2013.00145
Groot J. M., Boayue N. M., Csifcsák G., Boekel W., Huster R., Forstmann B. U., et al. (2021). Probing the neural signature of mind wandering with simultaneous fMRI-EEG and pupillometry. NeuroImage. 224, 117412. doi:10.1016/j.neuroimage.2020.117412
Hanso L., Bachmann T., Murd C. (2010). Tolerance of the ERP signatures of unfamiliar versus familiar face perception to spatial quantization of facial images. Psychology 01, 199–208. doi:10.4236/psych.2010.13027
Huang W., Wu X., Hu L., Wang L., Ding Y., Qu Z. (2017). Revisiting the earliest electrophysiological correlate of familiar face recognition. Int. J. Psychophysiol. 120, 42–53. doi:10.1016/j.ijpsycho.2017.07.001
Huck S. W., Cormier W. H., Bounds W. G. (1974). Reading Statistics and research. New York, NY: Harper and Row.
Ito T. A., Urland G. R. (2005). The influence of processing objectives on the perception of faces: an ERP study of race and gender perception. Cogn. Affect. Behav. Neurosci. 5, 21–36. doi:10.3758/cabn.5.1.21
Jafari M., Ansari-Pour N. (2019). Why, when and how to adjust your p values? Cell J. 20, 604–607. doi:10.22074/cellj.2019.5992
Jijomon C. M., Vinod A. P. (2021). Detection and classification of long-latency own-name auditory evoked potential from electroencephalogram. Biomed. Signal Process. Control. 68 (102724), 102724. doi:10.1016/j.bspc.2021.102724
Joshi S., Li Y., Kalwani R. M., Gold J. I. (2016). Relationships between pupil diameter and neuronal activity in the locus coeruleus, colliculi, and cingulate cortex. Neuron 89, 221–234. doi:10.1016/j.neuron.2015.11.028
Kaongoen N., Yu M., Jo S. (2017). Two-factor authentication system using P300 response to a sequence of human photographs. IEEE Trans. Syst. Man. Cybern. Syst. 50, 1178–1185. doi:10.1109/TSMC.2017.2756673
Kasim Ö., Tosun M. (2021). Biometric authentication from photic stimulated EEG records. Appl. Artif. Intell. 35, 1407–1419. doi:10.1080/08839514.2021.1981660
Katus L., Mason L., Milosavljevic B., McCann S., Rozhko M., Moore S. E., et al. (2020). ERP markers are associated with neurodevelopmental outcomes in 1–5 month old infants in rural Africa and the UK. NeuroImage 210, 116591. doi:10.1016/j.neuroimage.2020.116591
Kim D., Kim K. (2019). Resting state EEG-based biometric system using concatenation of quadrantal functional networks. IEEE Access 7, 65745–65756. doi:10.1109/ACCESS.2019.2917918
Kim S. P., Kang J. H., Jo Y. C., Oakley I. (2018). “Development of a multi-modal personal authentication,” in 2017 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Kuala Lumpur, Malaysia, December 12–15, 2017 (IEEE), 712–715.
Kuipers J. R., Thierry G. (2011). N400 amplitude reduction correlates with an increase in pupil size. Front. Hum. Neurosci. 5, 61. doi:10.3389/fnhum.2011.00061
Kuipers J. R., Thierry G. (2013). ERP-pupil size correlations reveal how bilingualism enhances cognitive flexibility. Cortex 49, 2853–2860. doi:10.1016/j.cortex.2013.01.012
Larsen R. S., Waters J. (2018). Neuromodulatory correlates of pupil dilation. Front. Neural Circuits. 12, 21. doi:10.3389/fncir.2018.00021
Lee S., Kang J. H., Oakley I., Kim S. P. (2017). Late positive event-related potentials in electroencephalography can distinguish acquaintances from strangers when both groups recognize faces. Washington DC: Society for Neuroscience.
Li F., Tao Q., Peng W., Zhang T., Si Y., Zhang Y., et al. (2020). Inter-subject P300 variability relates to the efficiency of brain networks reconfigured from resting-to task-state: evidence from a simultaneous event-related EEG-fMRI study. NeuroImage 205, 116285. doi:10.1016/j.neuroimage.2019.116285
Li J., Lam C. S., Yu M., Hess R. F., Chan L. Y., Maehara G., et al. (2010). Quantifying sensory eye dominance in the normal visual system: a new technique and insights into variation across traditional tests. Invest. Ophthalmol. Vis. Sci. 51, 6875–6881. doi:10.1167/iovs.10-5549
Maiorana E. (2021). Learning deep features for task-independent EEG-based biometric verification. Pattern Recognit. Lett. 143, 122–129. doi:10.1016/j.patrec.2021.01.004
Mäki-Marttunen V. (2021). Pupil-based states of brain integration across cognitive states. Neuroscience 471, 61–71. doi:10.1016/j.neuroscience.2021.07.016
Maqsood F., Schumacher U. (2017). Effects of varying light conditions and refractive error on pupil size. Cogent Med. 4, 1338824. doi:10.1080/2331205X.2017.1338824
Martin Bland J. M. M., Altman D. G. (1986). Statistical methods for assessing agreement between two methods of clinical measurement. Lancet 327, 307–310. doi:10.1016/S0140-6736(86)90837-8
Mitchell M. B., Shirk S. D., McLaren D. G., Dodd J. S., Ezzati A., Ally B. A., et al. (2016). Recognition of faces and names: multimodal physiological correlates of memory and executive function. Brain Imaging Behav. 10, 408–423. doi:10.1007/s11682-015-9420-6
Norton L., Gibson R. M., Gofton T., Benson C., Dhanani S., Shemie S. D., et al. (2017). Electroencephalographic recordings during withdrawal of life-sustaining therapy until 30 minutes after declaration of death. Can. J. Neurol. Sci. 44, 139–145. doi:10.1017/cjn.2016.309
Papetti A., Gregori F., Pandolfi M., Peruzzini M., Germani M. (2020). A method to improve workers’ well-being toward human-centered connected factories. J. Comput. Des. Eng. 7, 630–643. doi:10.1093/jcde/qwaa047
Park S., Kim L., Ha J., Mun S. (2022). Infrared webcam-based non-contact measurement of event-related potentials from event-related pupillary responses: an approach focused on mental workload. J. Comput. Des. Eng. 9, 1327–1347. doi:10.1093/jcde/qwac059
Park S., Mun S., Ha J., Kim L. (2021). Non-contact measurement of motion sickness using pupillary rhythms from an infrared camera. Sensors (Basel) 21, 4642. doi:10.3390/s21144642
Park S., Mun S., Lee D. W., Whang M. (2019). IR-camera-based measurements of 2D/3D cognitive fatigue in 2D/3D display system using task-evoked pupillary response. Appl. Opt. 58, 3467–3480. doi:10.1364/AO.58.003467
Park S., Whang M. (2018). Infrared camera-based non-contact measurement of brain activity from pupillary rhythms. Front. Physiol. 9, 1400. doi:10.3389/fphys.2018.01400
Rahman A., Chowdhury M. E. H., Khandakar A., Kiranyaz S., Zaman K. S., Reaz M. B. I., et al. (2021). Multimodal EEG and keystroke dynamics based biometric system using machine learning algorithms. IEEE Access 9, 94625–94643. doi:10.1109/ACCESS.2021.3092840
Rahman A., Chowdhury M. E. H., Khandakar A., Tahir A. M., Ibtehaz N., Hossain M. S., et al. (2022). Robust biometric system using session invariant multimodal EEG and keystroke dynamics by the ensemble of self-ONNs. Comput. Biol. Med. 142, 105238. doi:10.1016/j.compbiomed.2022.105238
Rahman M. A., Nakanishi I. (2022). “Person authentication using brain waves evoked by individual-related and imperceptible visual stimuli,” in 2022 International Conference of the Biometrics Special Interest Group (BIOSIG), Darmstadt, Germany, September 4–16, 2022 (IEEE), 1–5.
Rathi N., Singla R., Tiwari S. (2021). A novel approach for designing authentication system using a picture based P300 speller. Cogn. Neurodyn. 15, 805–824. doi:10.1007/s11571-021-09664-3
Rathi N., Singla R., Tiwari S. (2022). A comparative study of classification methods for designing a pictorial P300-based authentication system. Med. Biol. Eng. Comput. 60, 2899–2916. doi:10.1007/s11517-022-02626-9
Rhee J. H., Ma J. H., Seo J., Cha S. H. (2022). Review of applications and user perceptions of smart home technology for health and environmental monitoring. J. Comput. Des. Eng. 9, 857–889. doi:10.1093/jcde/qwac030
Sabeti M., Boostani R., Moradi E. (2020). Event related potential (ERP) as a reliable biometric indicator: a comparative approach. Array 6, 100026. doi:10.1016/j.array.2020.100026
Schütte F., Wartenburger I., Weymar M., Rabovsky M. (2021). Is the P600 linked to phasic noradrenaline release? Relating ERPs to pupil size in a sentence processing paradigm. Proc. Archit. Mech. Lang. Process.
Selezneva E., Wetzel N. (2022). The impact of probabilistic cues on sound-related pupil dilation and ERP responses in 7–9-year-old children. Cogn. 5, 86–106. doi:10.1080/25742442.2022.2048592
Shams T. B., Hossain M. S., Mahmud M. F., Tehjib M. S., Hossain Z., Pramanik M. I. (2022). EEG-based biometric authentication using machine learning: a comprehensive survey. ECTI-EEC. 20, 225–241. doi:10.37936/ecti-eec.2022202.246906
Stolte M., Gollan B., Ansorge U. (2020). Tracking visual search demands and memory load through pupil dilation. J. Vis. 20, 21. doi:10.1167/jov.20.6.21
Strauss G. P., Ossenfort K. L., Whearty K. M. (2016). Reappraisal and distraction emotion regulation strategies are associated with distinct patterns of visual attention and differing levels of cognitive demand. PLOS ONE 11, e0162290. doi:10.1371/journal.pone.0162290
Takeda Y., Okuma T., Kimura M., Kurata T., Takenaka T., Iwaki S. (2014). Electrophysiological measurement of interest during walking in a simulated environment. Int. J. Psychophysiol. 93, 363–370. doi:10.1016/j.ijpsycho.2014.05.012
Tao D., Tan H., Wang H., Zhang X., Qu X., Zhang T. (2019). A systematic review of physiological measures of mental workload. Int. J. Environ. Res. Public Health. 16, 2716. doi:10.3390/ijerph16152716
Thomas K. P., Vinod A. P. (2018). EEG-based biometric authentication using gamma band power during rest state. Circuits Syst. Signal Process. 37, 277–289. doi:10.1007/s00034-017-0551-4
Uhlmann E., Franke D., Hohwieler E. (2019). Smart Maintenance – dynamic model-based instructions for service operations. Procedia CIRP 81, 1417–1422. doi:10.1016/j.procir.2019.04.327
Unsworth N., Robison M. K., Miller A. L. (2018). Pupillary correlates of fluctuations in sustained attention. J. Cogn. Neurosci. 30, 1241–1253. doi:10.1162/jocn_a_01251
Wang H., Huang H., Liu Y., Xu H., Li T. (2019). An event related potential electroencephalogram signal analysis method based on denoising auto-encoder neural network. Contr. Theory. Appl. 36, 589–595. doi:10.7641/CTA.2018.70910
Wang H., Pei Z., Xu L., Xu T., Bezerianos A., Sun Y., et al. (2021). Performance enhancement of P300 detection by multiscale-CNN. IEEE Trans. Instrum. Meas. 70, 1–12. doi:10.1109/TIM.2021.3067943
Wang H., Wang Z., Sun Y., Yuan Z., Xu T., Li J. (2024). A cascade xDAWN EEGNet structure for unified visual-evoked related potential detection. IEEE Trans. Neural. Syst. Rehabil. Eng. 32, 2270–2280. doi:10.1109/TNSRE.2024.3415474
Wang M., El-Fiqi H., Hu J., Abbass H. A. (2019). Convolutional neural networks using dynamic functional connectivity for EEG-based person identification in diverse human states. IEEE Trans. Inf. Forensic Secur., Trans. IEEE 14, 3259–3272. doi:10.1109/TIFS.2019.2916403
Wang M. M., Kasmarik K., Bezerianos A., Tan K. C., Abbass H. (2021). On the channel density of EEG signals for reliable biometric recognition. Pattern Recognit. Lett. 147, 134–141. doi:10.1016/j.patrec.2021.04.003
Wang Z., Chen C., Li J., Wan F., Sun Y., Wang H. (2023a). ST-CapsNet: linking spatial and temporal attention with capsule network for P300 detection improvement. IEEE Trans. Neural. Syst. Rehabil. Eng. 31, 991–1000. doi:10.1109/TNSRE.2023.3237319
Wang Z., Zhang H., Ji Z., Yang Y., Wang H. (2023b). A review of deep learning methods for cross-subject rapid serial visual presentation detection in World Robot Contest 2022. Brain Sci. Adv. 9, 195–209. doi:10.26599/bsa.2023.9050013
Widmann A., Schröger E., Wetzel N. (2018). Emotion lies in the eye of the listener: emotional arousal to novel sounds is reflected in the sympathetic contribution to the pupil dilation response and the P3. Biol. Psychol. 133, 10–17. doi:10.1016/j.biopsycho.2018.01.010
Wu Q., Yan B., Zeng Y., Zhang C., Tong L. (2018b). Anti-deception: reliable EEG-based biometrics with real-time capability from the neural response of face rapid serial visual presentation. Biomed. Eng. OnLine. 17, 55. doi:10.1186/s12938-018-0483-7
Wu Q., Zeng Y., Zhang C., Tong L., Yan B. (2018a). An EEG-based person authentication system with open-set capability combining eye blinking signals. Sensors (Basel) 18, 335. doi:10.3390/s18020335
Zeng Y., Wu Q., Yang K., Tong L., Yan B., Shu J., et al. (2018). EEG-based identity authentication framework using face rapid serial visual presentation with optimized channels. Sensors (Basel) 19 (6), 6. doi:10.3390/s19010006
Zhang H., Wang Z., Yu Y., Yin H., Chen C., Wang H. (2022). An improved EEGNet for single-trial EEG classification in rapid serial visual presentation task. Brain Sci. Adv. 8, 111–126. doi:10.26599/BSA.2022.9050007
Zhang R., Zeng Y., Tong L., Shu J., Lu R., Li Z., et al. (2022). EEG identity authentication in multi-domain features: a multi-scale 3D-CNN approach. Front. Neurorobot. 16, 901765. doi:10.3389/fnbot.2022.901765
Keywords: augmented reality, authentication, biometrics, event-related potential, event-related pupillary response
Citation: Park S, Ha J and Kim L (2024) Event-related pupillary response-based authentication system using eye-tracker add-on augmented reality glasses for individual identification. Front. Physiol. 15:1325784. doi: 10.3389/fphys.2024.1325784
Received: 22 October 2023; Accepted: 05 August 2024;
Published: 13 August 2024.
Edited by:
Recep Avci, The University of Auckland, New ZealandCopyright © 2024 Park, Ha and Kim. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Laehyun Kim, bGFlaHl1bmtAa2lzdC5yZS5rcg==