Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry, 06 February 2024
Sec. Autism

Self-administered questionnaires enhance emotion estimation of individuals with autism spectrum disorders in a robotic interview setting

Shunta KonishiShunta Konishi1Masaki KuwataMasaki Kuwata1Yoshio MatsumotoYoshio Matsumoto2Yuichiro YoshikawaYuichiro Yoshikawa3Keiji TakataKeiji Takata4Hideyuki HaraguchiHideyuki Haraguchi4Azusa KudoAzusa Kudo5Hiroshi IshiguroHiroshi Ishiguro3Hirokazu Kumazaki,*Hirokazu Kumazaki5,6*
  • 1Department of Human and Engineered Environmental Studies, Graduate School of Frontier Sciences, The University of Tokyo, Chiba, Japan
  • 2Department of Medical and Robotic Engineering Design, Faculty of Advanced Engineering, Tokyo University of Science, Tokyo, Japan
  • 3Department of Systems Innovation, Graduate School of Engineering Science, Osaka University, Osaka, Japan
  • 4National Center of Neurology and Psychiatry, Department of Preventive Intervention for Psychiatric Disorders, National Institute of Mental Health, Tokyo, Japan
  • 5Department of Neuropsychiatry, Graduate School of Biomedical Sciences, Nagasaki University, Nagasaki, Japan
  • 6College of Science and Engineering, Kanazawa University, Kanazawa, Japan

Background: Robots offer many unique opportunities for helping individuals with autism spectrum disorders (ASD). Determining the optimal motion of robots when interacting with individuals with ASD is important for achieving more natural human-robot interactions and for exploiting the full potential of robotic interventions. Most prior studies have used supervised machine learning (ML) of user behavioral data to enable robot perception of affective states (i.e., arousal and valence) and engagement. It has previously been suggested that including personal demographic information in the identification of individuals with ASD is important for developing an automated system to perceive individual affective states and engagement. In this study, we hypothesized that assessing self-administered questionnaire data would contribute to the development of an automated estimation of the affective state and engagement when individuals with ASD are interviewed by an Android robot, which will be linked to implementing long-term interventions and maintaining the motivation of participants.

Methods: Participants sat across a table from an android robot that played the role of the interviewer. Each participant underwent a mock job interview. Twenty-five participants with ASD (males 22, females 3, average chronological age = 22.8, average IQ = 94.04) completed the experiment. We collected multimodal data (i.e., audio, motion, gaze, and self-administered questionnaire data) to train a model to correctly classify the state of individuals with ASD when interviewed by an android robot. We demonstrated the technical feasibility of using ML to enable robot perception of affect and engagement of individuals with ASD based on multimodal data.

Results: For arousal and engagement, the area under the curve (AUC) values of the model estimates and expert coding were relatively high. Overall, the AUC values of arousal, valence, and engagement were improved by including self-administered questionnaire data in the classification.

Discussion: These findings support the hypothesis that assessing self-administered questionnaire data contributes to the development of an automated estimation of an individual’s affective state and engagement. Given the efficacy of including self-administered questionnaire data, future studies should confirm the effectiveness of such long-term intervention with a robot to maintain participants’ motivation based on the proposed method of emotion estimation.

1 Introduction

Autism spectrum disorders (ASD) are a set of diverse neurodevelopment disorders characterized by difficulty with social interactions and behavioral difficulties. Individuals with ASD may behave, communicate, interact, and learn in ways that are different from most other people. The estimated 5-year lifetime cumulative incidence of ASD in children born between 2009 and 2014 in Japan is 2.75% (1). Regarding the economic impact, the cost of supporting an individual with ASD is expensive (e.g., his or her lifetime is estimated over $3.6 million in USA) (2). There are a variety of interventions for ASD. Many individuals with ASD cannot easily sustain high motivation and concentration in human interventions (3). Indeed, the dynamic facial features and expressions of humans may induce sensory and emotional overload and distraction (4). This overload can hamper interactions, as these individuals tend to actively avoid sensory stimulation and instead focus on more predictable elementary features. In addition, individuals with ASD struggle to generalize skills learned in intervention to everyday use, which is one of the greatest barriers to intervention success (57).

There is increasing anecdotal and case-based evidence that robots can offer many unique opportunities for individuals with ASD to learn social skills (812). It is also known that Android robots, whose appearances and movements resemble those of humans, can greatly promote social skill learning (1318). Additionally, android robots exhibit various facial expressions (e.g., smiling, nodding, and brow movements) during speech and can provide subtle nonverbal cues. Therefore, generating intelligent three-dimensional learning environments using android robots may represent a powerful avenue for enhancing skills with generalization to real-world settings. Previous studies (13, 14) using android robot in a mock job interview setting showed skill enhancement with generalization to the real-world settings.

If a patient does not develop a positive attitude toward the robot, the intervention becomes challenging. As each individual with ASD has strong likes and dislikes (19), the optimal motion of android robots for facilitating communication differs among these individuals (20). However, in previous studies using android robots, the android parameters were uniform rather than personalized (1318). In addition, the emotions and attention of the individuals with ASD may easily change during the intervention (21).

Moreover, many individuals with ASD have atypical and diverse ways of expressing their affective-cognitive states (22, 23). To address their heterogeneity, elucidating the optimal motion of robots for a given individual considering their traits and state is important for achieving more natural human-robot interactions and for exploiting the full potential of robotic interventions.

To achieve optimal robot motion for individuals with ASD, it is important to develop autonomous robots that can learn and recognize behavioral cues and respond smoothly to an individual’s real-time state (24). Most prior studies have focused on applying supervised machine learning (ML) to enable robot perception of user engagement directly from user behavioral data (e.g., child vocalizations, facial and body expressions, and physiological data such as the heart rate) (2527). However, these studies have struggled to develop computational models to estimate an individual’s state from behavioral data, owing to the lack of consideration of individuals’ ASD symptoms.

A previous study demonstrated the technical feasibility of considering an individual’s symptoms when using ML to enable robot perception of affect and engagement in individuals with ASD (28). In this study, the Childhood Autism Rating Scale (CARS) (29) was used to assess the presence and severity of symptoms of ASD. The study revealed that the expert assessment provided by CARS data (29) improved estimation of the state of individuals with ASD when interacting with robots. Their idea of including personal demographic information in the estimation of the state of individuals with ASD is important for developing an automatic system that can perceive individual affective states and engagement. On the other hand, the CARS takes a long time to administer, and specialists (and special training) are needed to conduct the CARS.

Self-administered questionnaires are known to have several limitations, including insufficient objectivity; however, they do not require the involvement of trained clinicians and can be completed rapidly. In clinical situations, self-administered questionnaires, such as the Autism Spectrum Quotient-Japanese version (AQ-J (30);), Adolescent/Adult Sensory Profile (AASP;31), and LSAS (32), can provide therapists with valuable information on individuals with ASD. Ease of obtaining personal demographic information is important when developing computational models for individuals with ASD. We hypothesized that assessing self-administered questionnaire data contributes to the development of an automated estimation of affective state and engagement when individuals with ASD are interviewed by an Android robot, which will be linked to implementing long-term interventions and maintaining participant motivation.

Automated estimation of an individual’s affective states and engagement is considered important for long-term intervention, as the states and engagement of participants can change daily and/or dynamically during the interaction. This study therefore aimed to assess the contribution of self-administered questionnaire data to develop an automated estimation of an individual’s affective states and engagement when individuals with ASD were interviewed by an android robot. The goal of the current study was to obtain an accurate emotion estimator for individuals during interactions by maximizing the area under the curve (AUC) values, a conventional measure used in the field of machine learning.

2 Materials and methods

2.1 Samples

2.1.1 Population

The inclusion criteria for individuals with ASD were as follows: 1) diagnosed with ASD based on the Diagnostic and Statistical Manual of Mental Disorders, fifth edition (DSM-5), by the supervising study psychiatrist, and 2) not currently taking medication. Thirty-five individuals with ASD participated in this study. Instead of adopting a standard statistical test paradigm with power analysis using G*Power, we adopted a five-fold cross-validation test paradigm to evaluate the goodness of fit (AUC) of the model, which is commonly used in machine learning (33). To determine the number of samples used for model fitting, we collected a similar number of samples [35] to that reported by Rudovic (28). Three participants lost concentration during the experiment and were unable to complete the experiment. Seven participants were unable to correctly perform gaze calibration. Finally, 25 samples (males 22, females 3, average chronological age = 22.8, average IQ = 94.04) completed the experiment without any technical challenges or distress that would have led to the termination of the session. All samples had no regular jobs. Details are presented in Table 1. At the time of enrollment, the diagnoses of all participants were confirmed by a psychiatrist with more than fifteen years of experience in ASD using standardized criteria taken from the Diagnostic Interview for Social and Communication Disorders (DISCO) (34). The DISCO has good psychometric properties (35). To exclude other psychiatric diagnoses, the Mini-International Neuropsychiatric Interview (MINI) (36) was administered.

Table 1
www.frontiersin.org

Table 1 Descriptive statistics of participants.

2.1.2 Ethical procedure

The present study was approved by the Ethics Committee of Kanazawa University and was conducted in accordance with the Declaration of Helsinki. The authors had no conflict of interest. Participants were recruited by flyers that explained the content of the experiment. After receiving a complete explanation of the study, all participants and their guardians agreed to participate in the study. Written informed consent was obtained from the individuals and/or the legal guardian (of minors) for the publication of any potentially identifiable images or data included in this article.

2.2 Self-Administered questionnaire

All participants completed the Autism Spectrum Quotient-Japanese version (AQ-J) (30), a self-administered questionnaire used to measure autistic traits and evaluate ASD-specific behaviors and symptoms. The AQ-J is a self-administered questionnaire with five subscales (social skills, attention switching, attention to detail, imagination, and communication). Previous work with the AQ-J has been replicated across cultures (37) and ages (38). The AQ is sensitive to the broader autism phenotype (39). In this study, we did not use the AQ-J score as a cutoff for ASD and used only the DSM-5 and DISCO to diagnose ASD and to determine whether to include participants in our study. It takes approximately 10 minutes to complete the AQ-J.

Full-scale IQ scores were obtained with the Japanese Adult Reading Test (JART), a standardized cognitive function test used to estimate the premorbid intelligence quotients (IQ) of examinees with cognitive impairments (40). The JART has good validity for measuring IQ. The JART results are comparable to those of the WAIS-III (40). It takes approximately 10 minutes to complete the JART. Usually, the JART is conducted in a face-to-face interview setting. The participant is instructed to read the characters. In this study, we prepared a self-administered version of the JART in which participants described the reading.

The severity of social anxiety symptoms was measured using the Liebowitz Social Anxiety Scale (LSAS) (32), a 24-item self-rated scale that measures the role of social phobia in life across various situations. This self-administered questionnaire included 13 items that relate to performance anxiety and 11 concern social interaction situations. Each item was separately rated in terms of “fear” and “avoidance” using a 4-point categorical scale. Therefore, there are 48-items in total. According to receiver-operating curve analyses, an LSAS score of 30 is correlated with minimal symptoms and is the best cutoff value for distinguishing individuals with and without social anxiety disorder (41). It takes approximately 10 minutes to complete the LSAS.

The Adolescent/Adult Sensory Profile (AASP) is a self-administered questionnaire that measures sensory processing in individuals aged 11 years and older (31). The internal consistency coefficients of the AASP range from 0.64 to 0.78 for the quadrant scores. In this study, before the experiment, the participants reported how often they exhibited certain behaviors related to sensory experiences on a scale of one (almost never) to five (almost always). The AASP examines four different “quadrants” of sensory processing: low registration, sensation seeking, sensory sensitivity, and sensation avoiding. As the AASP does not categorize responses according to perceptual domains (e.g., auditory, visual, tactile), a perceptual domain analysis was not performed in this study. This scale takes approximately 10 minutes to complete. Please see the details of feature set (i.e., self-administered questionnaire) in Table 2.

Table 2
www.frontiersin.org

Table 2 Description of the behavioral data and self-administered questionnaire data.

2.3 Interviewer robot system

The robot used in this study was an Android ST by A-Lab Co., Ltd. (Figure 1), which is a female humanoid robot with an appearance similar to that of a real human. Its artificial body has the same proportions, facial features, hair color, and hairstyle as a human. The synthesized voice of the android is also similar to that of a human. To elicit the belief that the robot behaved and responded autonomously without any failures, we adopted a Wizard-of-Oz (WOz) method, similar to that conventionally used in robotics studies (42). Facial expressions (i.e., smiling, nodding, and brow movements) can be generated in addition to utterances during conversation.

Figure 1
www.frontiersin.org

Figure 1 Android ST.

2.4 Procedures

In the experiment, a participant entered the room and sat across the table from the android, as shown in Figure 2. The gaze tracking device on the table was then calibrated to measure the participant’s gaze. In the beginning of the interaction with the android, the participant is asked to adjust the volume and the speed of the synthesized voice to comfortable levels by changing the parameters. There were 5 volume levels and 5 speed levels.

Figure 2
www.frontiersin.org

Figure 2 The experimental room.

The android played the role of the interviewer, sitting in front of the participant, and the WOZ operator sat behind a partition, as shown in Figure 2. Each participant underwent four sessions of mock job interviews as the interviewee. Each session corresponded to one of the four conditions regarding android behavior, namely, with and without idle body motions and eye-blinking motions. The order of the four conditions was randomly assigned and counterbalanced among participants to reduce the order effect. The reason for including four conditions is that individuals with ASD vary regarding preferred behaviors (20).

In each interview session, question-and-answer conversations were conducted 9 or 10 times, initiated by the android interviewer. The android asked questions based on predefined sentence lists (see the Supplementary Material) and waited until the participant answers. This simple form of conversation was adopted to reduce variation in the dialog structure.

2.5 Behavioral measurement

In each session, the behavior of the participant was recorded using a standard webcam (Logicool C980GR) and a gaze tracker (Tobii X2-30) installed on the desk as shown in Figure 3C. In each interview session, the participant answered the android interviewer 9 or 10 times. A set of behavioral data (audio, motion, and gaze) was collected that corresponded to each speech segment of the participant. Details of the measurement and feature extraction methods are described below. Please see the details of feature set (i.e., audio, motion, and gaze) in Table 2.

25.1 Audio data

The audio data of the participant while interacting with the android robot were collected and stored in a computer and then analyzed afterward. For feature extraction, open-source speech and music interpretation by large-space extraction (OpenSMILE) (43) was adopted. It is a software toolkit for audio analysis, processing and classification, especially for speech and music applications. For the output of OpenSMILE, 88-dimensional feature values in the GeMAPS v01b feature set were utilized as previously described by Eyben et al. (44), including ZCR, HNR, F0, and MFCC. The zero crossing rate (ZCR) is the rate at which the sound waveform data cross zero in a given period. It is a key feature for classifying percussive sounds and tends to be higher for the speech part of the signal. The harmonics to noise ratio (HNR) is the ratio of harmonic and noise components that provides an indication of the general frequency of the speech signal by quantifying the relationship between the periodic component (harmonics) and the aperiodic component (noise). F0 is the fundamental frequency, and MFCC is the Mel frequency cepstral coefficient of the sound.

2.5.2 Motion data

Video recordings of the upper body of subjects were collected with a webcam during the session. These data were converted into 16-dimensional motion feature data by image processing using OpenPose (45). It is an open-source library that uses part confidence maps to estimate the joint positions of the human body at high speed. In this study, eight key points on the upper body (nose, right/left eyes, right/left ears, neck, and right/left shoulders) were measured in 2-dimension (2D), as shown in Figure 3A. Subsequently, standard deviations of the 2D positions of the eight key points in a time series were calculated and used as 16-dimensional feature values, which indicated the degree of the participants’ movement/steadiness during each speech segment.

Figure 3
www.frontiersin.org

Figure 3 (A) Example gaze distributions of two-dimensional gaze information. (B) Example upper body posture for the calculation of motion data. (C) Overview of the experimental scene.

2.5.3 Gaze data

The gaze data of the participant in the interaction were collected by a gaze tracker to determine how much the participant looked at the android. The system requires calibration before measurements can be collected. Figure 3B shows examples of the acquired 2D gaze distribution. Based on the gaze coordinates, the ratio of time elapsed while looking at the android’s face and body in each speech segment was calculated, yielding 2D feature values.

2.6 Annotation of emotional state

In this study, the emotional state of participants was represented based on Russell’s circumplex model (46). This model has the following three axes: 1) arousal, which is an indicator of concentration; 2) valence, which is an indicator of pleasure or displeasure; and 3) engagement, which is an indicator of interest. To apply machine learning techniques to estimate the emotional state of participants from the measured multimodal data, the ground truth emotion data were generated by manual annotation of the recorded video sequences. The methods and criteria of emotion annotation were established by an experienced psychologist who later instructed two research assistants in coding. The research assistants performed independent annotations and communication regarding the criteria until an intraclass correlation coefficient (ICC) of 0.8 or higher was obtained. The emotion annotations occurred only in the speech segments when the participant answered the robot interviewer. Therefore, the number of annotations was equal to the number of questions in the interview.

2.7 Data analysis

The collected behavioral data (audio, motion, and gaze) and the self-administered questionnaire data were used to train a model to estimate the emotional state of subjects. The training model was LightGBM (47). It is an ensemble machine learning algorithm that uses a gradient boosting framework based on decision trees (GBDT) and has faster training speed, lower memory usage, and better accuracy than other boosting algorithms. The binary classification model in Microsoft LightGBM v4.0.0, an open-source library for Python, was used in this study.

We tested seven combinations of behavioral information (1) audio only, 2) motion only, 3) gaze only, 4) audio+motion, 5) audio+gaze, 6) motion+gaze, and 7) audio+motion+gaze) to investigate the importance of each type of behavioral data. We also tested the estimation with and without self-administered questionnaire data to investigate the importance of the self-administered questionnaire data (see Figure 4). The performance of the estimation model was evaluated based on the AUC of a fivefold cross-validation. The AUC, which is the area under the receiver operating characteristic (ROC) curve, is the measure of the ability of a binary classifier in machine learning to distinguish between two classes. The ROC curve was used to indicate the connection/trade-off between clinical sensitivity and specificity of the cutoff. The AUC has a value between 0 and 1, and higher AUC indicates better performance in distinguishing between classes.

Figure 4
www.frontiersin.org

Figure 4 Overview of the automatic estimation system for determining individual emotional state based on machine learning. The collected behavioral data (audio, motion, and gaze) were used to train models with and without self-administered questionnaire data. LightGBM was utilized as the machine learning method.

3 Results

Participant performance was carefully monitored to ensure that all participants, except for three, were focused during the trial and remained highly motivated from the beginning to the end of the experiment. In total, 893 sets of feature vectors of behavioral measurements and annotated emotional states were obtained for 25 samples, which were then used to build the model for estimation.

To evaluate the accuracy of the model based on the variables used, we adopted a five-fold cross-validation paradigm that is conventionally used in machine learning (36). Table 3 shows the performance of emotion estimation models based on behavioral data. Each value indicates the AUC (mean and standard deviation) in the five-fold cross-validation test with and without self-administered questionnaire data. The graphs in Figure 5 show the ROC curves for the five validations and the average curve for the classifier with the highest AUC to classify arousal, valence, and engagement with and without SAQ. Figure 6 shows the contribution of each feature of the self-administered questionnaire data in terms of the Gini importance in the LightGBM.

Table 3
www.frontiersin.org

Table 3 Estimation of emotions from behavioral data in terms of AUC.

Figure 5
www.frontiersin.org

Figure 5 ROC curves showing the five validations and their average curve for the classifier with the highest AUC to classify arousal (left), valence (middle), and engagement (right) without (top) and with SAQ (bottom).

Figure 6
www.frontiersin.org

Figure 6 The contribution of self-administered questionnaire data to estimations of arousal, valence, and engagement (in terms of Gini importance) calculated with LightGBM, as shown in the “audio+motion+gaze with self-administered questionnaire” condition.

For arousal, in accordance with our hypothesis, the performance of the estimation model with self-administered questionnaire data was higher than that without self-administered questionnaire data for all combinations of behavioral data. The model with the highest AUC was audio+motion+gaze with self-administered questionnaire data (as shown in Table 3). Of the self-administered questionnaire variables, the low registration score had the highest contribution to estimations of arousal.

For valence, the AUC was higher when the self-administered questionnaire data were included for all combinations of the behavioral data except for the “audio only” case. The estimators with the highest AUC values were motion with self-administered questionnaire data, audio+motion with self-administered questionnaire data. Of the self-administered questionnaire variables, the sensory avoidance score had the highest contribution to estimations of valence.

For engagement, in accordance with our hypothesis, the AUC of the estimator was higher when the self-administered questionnaire data were included for all combinations of behavioral data. The estimators with the highest AUC values were audio+motion+gaze with self-administered questionnaire data and audio+motion with self-administered questionnaire data. Of the self-administered questionnaire variables, the sensory seeking score had the highest contribution to estimations of engagement.

4 Discussion

In this study, we assessed the contributions of self-administered questionnaire data to the development of an automatic system for detecting individual affective states and engagement during interviews with an android robot among individuals with ASD. Our experiments showed that leveraging the self-administered questionnaire data enhanced the emotional (i.e., arousal, valence, and engagement) estimation of individuals with ASD approximately by 0.02 on average in AUC. A previous study (28) that used the CARS also suggested that these data could facilitate the robot’s perception of affect and engagement in individuals with ASD. Previous work that focused on prediction, a different performance criterion, namely ICC, was used to verify accuracy. Therefore, it is not directly comparable with our study because it focused on the improvement of classification using the performance criterion of the AUC. However, the CARS requires more time (generally two to three hours, including the time needed for observation and filling out the score) than self-questionnaires (generally 40 minutes). In addition, the CARS requires the involvement of a specialist, whereas self-questionnaires do not need require such intervention. Therefore, the use of self-administered questionnaires may be more feasible for future implementation of robotic interventions.

In this study, among the self-administered questionnaire variables, the sensory profile would have a contribution to robot perception of affect and engagement. Kumazaki et al. (20) revealed that in an interview setting with an android robot, the sensory profile is an important factor for estimating the attitude of individuals with ASD toward android robots, which is in line with the results of this study.

Kim et al. (48) suggested that improving audio-based emotion estimation for individuals with ASD could allow the robotic system to properly assess the engagement of individuals, which is consistent with the results of this study. Rudovic et al. (28) previously suggested that audio data are insufficient to estimate a subject’s state, and that it is important to reduce background noise to utilize audio data. Unlike the previous study (28), in the present study, audio data would outperform the other single modalities in terms of the assessment of arousal and engagement, followed by a combination of motion and gaze data. This is considered to be caused by our experimental setup, where the environment was carefully prepared to be quiet, with reduced background noise.

Regarding the assessment of valence, motion data would outperform the other single modalities, followed by the combination of audio and gaze data. A previous study in the general population (49) suggested that motion is important for estimating valence, which is in line with the results of this study. Individuals with ASD exhibit abnormalities in posture (50) and coordination of balance (51). Given these factors, it is logical that motion data had important contributions to the estimation of the affective state of individuals with ASD in this study.

We showed that adding data from self-administered questionnaires would enable us to address heterogeneity in the representations of affective states and engagement in individuals with ASD (52). These results are linked to achieving more personalized and natural human-robot interactions and exploiting the full potential of robotic interventions.

This study had several limitations that should be addressed in future research. Firstly, the sample size was relatively small; therefore, future studies with larger sample sizes are warranted to validate our results. In this study, we only measured audio, motion, and gaze data in real time. However, monitoring other data, such as physiological data, may be useful to improve the AUC value. The number of self-administered questionnaires utilized in the study was also limited, meaning we were unable to fully capture the complete range of individual characteristics that influence affective states and engagement in individuals with ASD. Future studies including additional items are required to overcome these limitations. We discuss the data in light of ASD engagement behavior. Future studies should deliberate on ASD behaviors, such as empathy and nervousness. In addition, we conducted only semi-structured interviews. To create programs using Android robots that can be applied to a variety of situations, future studies with a variety of interview settings are required. Furthermore, the focus of this study on the interaction between individuals with ASD and an Android robot in a simulated job-interview setting may not mirror real-life social interactions entirely. Future studies investigating a variety of real-life social interactions are required. Our data showed that assessing self-administered questionnaire data contributes to the development of an automated estimation of an affective state and engagement when individuals with ASD are interviewed by android robots. Considering that the behavior of individuals with ASD toward robots is superior to humans (812), it is not clear whether the developed classifier works even for behavioral data obtained in the interaction with a human interviewer. In addition, we did not investigate whether the automated estimation of an individual’s affective state and engagement could be used to implement long-term interventions and maintain the motivation of participants. Future studies are required to ascertain the efficacy of data collection using self-administered questionnaires.

In the field of robotic interventions for individuals with ASD, few studies have demonstrated skill acquisition that is considered clinically meaningful and generalized beyond the specific robot encounter (53). To implement interventions with long-term effects and maintain the motivation of participants, it is important to develop technology to accurately and automatically perceive participant affect and engagement. The importance of adding data from self-administered questionnaires for emotion estimation reported in this paper could serve as a reference for the development of android robots that detect the effects and engagement of individuals, which would be the next step in establishing interventions involving android robots.

In this study, we assessed the contributions of self-administered questionnaire data to the development of an automatic system for detecting the affective states and engagement of individuals with ASD during interviews with an android robot. We leveraged self-administered questionnaire data and found that the estimation performance was enhanced for arousal, valence, and engagement. Our results also would support our hypothesis that the assessment of self-administered questionnaire data can contribute to the development of an automated estimation of an individual’s affective state and engagement when individuals with ASD are interviewed by an android robot. In the field of robotic interventions for individuals with ASD, few studies have demonstrated skill acquisition that is clinically meaningful and generalizable beyond the specific robot encounter. To implement long-term interventions and maintain the motivation of participants, technology that can automatically detect the affect and engagement of individuals in terms of personal traits is needed. Future studies should confirm the effectiveness of long-term intervention with a robot that can maintain the participants’ motivation based on the proposed method of emotion estimation.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the Ethics Committee of Kanazawa University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individuals and/or the legal guardian (of minors) for the publication of any potentially identifiable images or data included in this article.

Author contributions

MK and HK designed the study, conducted the experiments, conducted the statistical analyses, analyzed and interpreted the data, and drafted the manuscript. MK, YM, YY, KT, HH, HI and HK, MK, and AK conceptualized the study, participated in its design, assisted with data collection and scoring of behavioral measures, analyzed and interpreted the data, drafted the manuscript, and critically revised the manuscript for important intellectual content. HK approved the final version to be published. All the authors have read and approved the final version of the manuscript.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by Grants-in-Aid for Scientific Research from the Japan Society for the Promotion of Science (22H04874) and Moonshot R&D (grant number JPMJMS2011).

Acknowledgments

We would like to thank all participants and their families for their cooperation in this study. We also thank M. Miyao and H. Ito for their assistance with our experiments.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2024.1249000/full#supplementary-material

References

1. Sasayama D, Kuge R, Toibana Y, Honda H. Trends in Autism Spectrum Disorder Diagnoses in Japan, 2009 to 2019. JAMA Netw Open (2021) 4:e219234. doi: 10.1001/jamanetworkopen.2021.9234

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Cakir J, Frye RE, Walker SJ. The lifetime social cost of autism: 1990–2029. Res Autism Spectr Disord (2020) 72:101502. doi: 10.1016/j.rasd.2019.101502

CrossRef Full Text | Google Scholar

3. Warren ZE, Zheng Z, Swanson AR, Bekele E, Zhang L, Crittendon JA, et al. Can robotic interaction improve joint attention skills? J Autism Dev Disord (2015) 45:3726–34. doi: 10.1007/s10803-013-1918-4

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Johnson CP, Myers SM. Identification and evaluation of children with autism spectrum disorders. Pediatrics (2007) 120:1183–215. doi: 10.1542/peds.2007-2361

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Karkhaneh M, Clark B, Ospina MB, Seida JC, Smith V, Hartling L. Social Stories ™ to improve social skills in children with autism spectrum disorder: a systematic review. Autism (2010) 14:641–62. doi: 10.1177/1362361310373057

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Vismara LA, Rogers SJ. Behavioral treatments in autism spectrum disorder: what do we know? Annu Rev Clin Psychol (2010) 6:447–68. doi: 10.1146/annurev.clinpsy.121208.131151

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Wass SV, Porayska-Pomsta K. The uses of cognitive training technologies in the treatment of autism spectrum disorders. Autism (2014) 18:851–71. doi: 10.1177/1362361313499827

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Pennisi P, Tonacci A, Tartarisco G, Billeci L, Ruta L, Gangemi S, et al. Autism and social robotics: a systematic review. Autism Res (2016) 9:165–83. doi: 10.1002/aur.1527

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Kumazaki H, Muramatsu T, Yoshikawa Y, Matsumoto Y, Ishiguro H, Kikuchi M, et al. Optimal robot for intervention for individuals with autism spectrum disorders. Psychiatry Clin Neurosci (2020) 74:581–6. doi: 10.1111/pcn.13132

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Diehl JJ, Schmitt LM, Villano M, Crowell CR. The clinical use of robots for individuals with autism spectrum disorders: a critical review. Res Autism Spectr Disord (2012) 6:249–62. doi: 10.1016/j.rasd.2011.05.006

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Ricks DJ, Colton MB. Trends and considerations in robot-assisted autism therapy. In: 2010 IEEE International Conference on Robotics and Automation. Anchorage, USA: IEEE (2010). p. 4354–9.

Google Scholar

12. Scassellati B, Admoni H, Matarić M. Robots for use in autism research. Annu Rev BioMed Eng (2012) 14:275–94. doi: 10.1146/annurev-bioeng-071811-150036

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Kumazaki H, Muramatsu T, Yoshikawa Y, Matsumoto Y, Ishiguro H, Mimura M, et al. Role-play-based guidance for job interviews using an android robot for individuals with autism spectrum disorders. Front Psychiatry (2019) 10:239. doi: 10.3389/fpsyt.2019.00239

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Kumazaki H, Muramatsu T, Yoshikawa Y, Corbett BA, Matsumoto Y, Higashida H, et al. Job interview training targeting nonverbal communication using an android robot for individuals with autism spectrum disorder. Autism (2019) 23:1586–95. doi: 10.1177/1362361319827134

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Yoshikawa Y, Kumazaki H, Matsumoto Y, Miyao M, Kikuchi M, Ishiguro H. Relaxing gaze aversion of adolescents with autism spectrum disorder in consecutive conversations with human and android robot-a preliminary study. Front Psychiatry (2019) 10:370. doi: 10.3389/fpsyt.2019.00370

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Takata K, Yoshikawa Y, Muramatsu T, Matsumoto Y, Ishiguro H, Mimura M, et al. Social skills training using multiple humanoid robots for individuals with autism spectrum conditions. Front Psychiatry (2023) 14:1168837. doi: 10.3389/fpsyt.2023.1168837

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Kumazaki H, Muramatsu T, Yoshikawa Y, Matsumoto Y, Ishiguro H, Mimura M. Android robot was beneficial for communication rehabilitation in a patient with schizophrenia comorbid with autism spectrum disorders. Schizophr Res (2023) 254:116–7. doi: 10.1016/j.schres.2023.02.009

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Kumazaki H, Muramatsu T, Yoshikawa Y, Matsumoto Y, Takata K, Ishiguro H, et al. Android robot promotes disclosure of negative narratives by individuals with autism spectrum disorders. Front Psychiatry (2022) 13:899664. doi: 10.3389/fpsyt.2022.899664

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Trevarthen C, Delafield-Butt JT. Autism as a developmental disorder in intentional movement and affective engagement. Front Integr Neurosci (2013) 7:49. doi: 10.3389/fnint.2013.00049

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Kumazaki H, Muramatsu T, Yoshikawa Y, Matsumoto Y, Kuwata M, Takata K, et al. Differences in the optimal motion of android robots for the ease of communications among individuals with autism spectrum disorders. Front Psychiatry (2022) 13:883371. doi: 10.3389/fpsyt.2022.883371

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Egger HL, Dawson G, Hashemi J, Carpenter KLH, Espinosa S, Campbell K, et al. Automatic emotion and attention analysis of young children at home: a ResearchKit autism feasibility study. NPJ Digit Med (2018) 1:20. doi: 10.1038/s41746-018-0024-6

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Lyons V, Fitzgerald M. atypical sense of self in autism spectrum disorders: a neuro- cognitive perspective. In: Michael F, editor. Recent Advances in Autism Spectrum Disorders. Rijeka: IntechOpen (2013). p. Ch. 31.

Google Scholar

23. Chasson G, Jarosiewicz SR. Social competence impairments in autism spectrum disorders. In: Patel VB, Preedy VR, Martin CR, editors. Comprehensive Guide to Autism. New York, NY: Springer New York (2014). p. 1099–118.

Google Scholar

24. Kim E, Paul R, Shic F, Scassellati B. Bridging the research gap: making HRI useful to individuals with autism. J Hum Robot Interact (2012) 1:26–54. doi: 10.5898/JHRI.1.1.Kim

CrossRef Full Text | Google Scholar

25. Rudovic O, Lee J, Mascarell-Maricic L, Schuller B, Picard R. Measuring engagement in robot-assisted autism therapy: a cross-cultural study. Front Robot AI (2017) 4:36. doi: 10.3389/frobt.2017.00036

CrossRef Full Text | Google Scholar

26. Jain S, Thiagarajan B, Shi Z, Clabaugh C, Matarić MJ. Modeling engagement in long-term, in-home socially assistive robot interventions for children with autism spectrum disorders. Sci Robot (2020) 5:eaaz3791. doi: 10.1126/scirobotics.aaz3791

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Rudovic O, Utsumi Y, Lee J, Hernandez J, Ferrer EC, Schuller B, et al. CultureNet: a deep learning approach for engagement intensity estimation from face images of children with autism. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (2018). Madrid, Spain:IEEE. p. 339–46.

Google Scholar

28. Rudovic O, Lee J, Dai M, Schuller B, Picard RW. Personalized machine learning for robot perception of affect and engagement in autism therapy. Sci Robot (2018) 3:eaao6760. doi: 10.1126/scirobotics.aao6760

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Schopler E, Reichler RJ, DeVellis RF, Daly K. Toward objective classification of childhood autism: childhood autism rating scale (CARS). J Autism Dev Disord (1980) 10:91–103. doi: 10.1007/bf02408436

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Wakabayashi A, Tojo Y, Baron-Cohen S, Wheelwright S. The autism-spectrum quotient (AQ) Japanese version: evidence from high-functioning clinical group and normal adults. Shinrigaku Kenkyu (2004) 75:78–84. doi: 10.4992/jjpsy.75.78

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Brown CE, Dunn W. Adolescent/Adult Sensory Profile. San Antonio, TX: Therapy Skill Builders, The Psychological Corporation (2002).

Google Scholar

32. Liebowitz MR. Social phobia. Mod Probl Pharmacopsychiatry (1987) 22:141–73. doi: 10.1159/000414022

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Hanley JA, McNeil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology (1982) 143(1):29–36. doi: 10.1148/radiology.143.1.7063747

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Leekam SR, Libby SJ, Wing L, Gould J, Taylor C. The diagnostic interview for social and communication disorders: algorithms for ICD-10 childhood autism and wing and gould autistic spectrum disorder. J Child Psychol Psychiatry (2002) 43:327–42. doi: 10.1111/1469-7610.00024

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Wing L, Leekam SR, Libby SJ, Gould J, Larcombe M. The diagnostic interview for social and communication disorders: background, inter-rater reliability and clinical use. J Child Psychol Psychiatry (2002) 43:307–25. doi: 10.1111/1469-7610.00023

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Lecrubier Y, Sheehan DV, Weiller E, Amorim P, Bonora I, Sheehan KH, et al. The mini international neuropsychiatric interview (MINI): a short diagnostic structured interview: reliability and validity according to the CIDI. Eur Psychiatry (1997) 12:224–31. doi: 10.1016/S0924-9338(97)83296-8

CrossRef Full Text | Google Scholar

37. Wakabayashi A, Baron-Cohen S, Uchiyama T, Yoshida Y, Tojo Y, Kuroda M, et al. The autism-spectrum quotient (AQ) children’s version in Japan: a cross-cultural comparison. J Autism Dev Disord (2007) 37:491–500. doi: 10.1007/s10803-006-0181-3

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Auyeung B, Baron-Cohen S, Wheelwright S, Allison C. The autism spectrum quotient: children’s version (AQ-Child). J Autism Dev Disord (2008) 38:1230–40. doi: 10.1007/s10803-007-0504-z

PubMed Abstract | CrossRef Full Text | Google Scholar

39. Wheelwright S, Auyeung B, Allison C, Baron-Cohen S. Defining the broader, medium and narrow autism phenotype among parents using the Autism Spectrum Quotient (AQ). Mol Autism (2010) 1:10. doi: 10.1186/2040-2392-1-10

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Hirata-Mogi S, Koike S, Toriyama R, Matsuoka K, Kim Y, Kasai K. Reliability of a paper-and-pencil version of the Japanese adult reading test short version. Psychiatry Clin Neurosci (2016) 70:362. doi: 10.1111/pcn.12400

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Mennin DS, Fresco DM, Heimberg RG, Schneier FR, Davies SO, Liebowitz MR. Screening for social anxiety disorder in the clinical setting: using the liebowitz social anxiety scale. J Anxiety Disord (2002) 16:661–73. doi: 10.1016/s0887-6185(02)00134-2

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Nishio S, Taura K, Sumioka H, Ishiguro H. Teleoperated android robot as emotion regulation media. Int J Soc Robot (2013) 5:563–73. doi: 10.1007/s12369-013-0201-3

CrossRef Full Text | Google Scholar

43. Eyben F, Wöllmer M, Schuller B. Opensmile: the munich versatile and fast open-source audio feature extractor. In: Proceedings of the 18th ACM International Conference on Multimedia. Firenze, Italy: Association for Computing Machinery (2010). p. 1459–62.

Google Scholar

44. Eyben F, Scherer KR, Schuller BW, Sundberg J, André E, Busso C, et al. The geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing. IEEE Trans Affect Comput (2016) 7:190–202. doi: 10.1109/TAFFC.2015.2457417

CrossRef Full Text | Google Scholar

45. Cao Z, Simon T, Wei S-E, Sheikh Y. Realtime multi-person 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE (2017). p. 7291–9.

Google Scholar

46. Russell JA. A circumplex model of affect. J Pers Soc Psychol (1980) 39:1161–78. doi: 10.1037/h0077714

CrossRef Full Text | Google Scholar

47. Sheng D, Cheng X. Analysis of the Q-T dispersion and T wave alternans in patients with epilepsy. Yangtze Med (2017) 1:109–16. doi: 10.4236/ym.2017.12011

CrossRef Full Text | Google Scholar

48. Kim JC, Azzi P, Jeon M, Howard AM, Park CH. Audio-based emotion estimation for interactive robotic therapy for children with autism spectrum disorder. In: 2017 14th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI). Jeju, South Korea: IEEE (2017). p. 39–44.

Google Scholar

49. Li W, Alemi O, Fan J, Pasquier P. Ranking-based affect estimation of motion capture data in the valence-arousal space. In: Proceedings of the 5th International Conference on Movement and Computing. Genoa, Italy: Association for Computing Machinery (2018). p. Article 18.

Google Scholar

50. Bojanek EK, Wang Z, White SP, Mosconi MW. Postural control processes during standing and step initiation in autism spectrum disorder. J Neurodev Disord (2020) 12:1. doi: 10.1186/s11689-019-9305-x

PubMed Abstract | CrossRef Full Text | Google Scholar

51. Stins JF, Emck C. Balance performance in autism: a brief overview. Front Psychol (2018) 9:901. doi: 10.3389/fpsyg.2018.00901

PubMed Abstract | CrossRef Full Text | Google Scholar

52. Stewart ME, Russo N, Banks J, Miller L, Burack JA. Sensory characteristics in ASD. Mcgill J Med (2009) 12:108. doi: 10.26443/mjm.v12i2.280

PubMed Abstract | CrossRef Full Text | Google Scholar

53. Scassellati B, Boccanfuso L, Huang CM, Mademtzi M, Qin M, Salomons N, et al. Improving social skills in children with ASD using a long-term, in-home social robot. Sci Robot (2018) 3:eaat7544. doi: 10.1126/scirobotics.aat7544

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: autism spectrum disorders, machine learning, self-administered questionnaire, affective state, automated estimation

Citation: Konishi S, Kuwata M, Matsumoto Y, Yoshikawa Y, Takata K, Haraguchi H, Kudo A, Ishiguro H and Kumazaki H (2024) Self-administered questionnaires enhance emotion estimation of individuals with autism spectrum disorders in a robotic interview setting. Front. Psychiatry 15:1249000. doi: 10.3389/fpsyt.2024.1249000

Received: 28 June 2023; Accepted: 23 January 2024;
Published: 06 February 2024.

Edited by:

Antonio Narzisi, Stella Maris Foundation (IRCCS), Italy

Reviewed by:

Irini Giannopulu, Creative Robotics Lab, Australia
Adham Atyabi, University of Colorado Colorado Springs, United States
George Fragulis, University of Western Macedonia, Greece

Copyright © 2024 Konishi, Kuwata, Matsumoto, Yoshikawa, Takata, Haraguchi, Kudo, Ishiguro and Kumazaki. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hirokazu Kumazaki, a3VtYXpha2lAdGlhcmEub2NuLm5lLmpw

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.